π€ GPT-2 Bulgarian LoRA Adapter (Merged)
I will be training a much larger sample in the coming days (1k is small - but my computers bandwidth is smaller)
This model is a fine-tuned and merged version of openai-community/gpt2-medium
, adapted to Bulgarian using the LoRA technique. Training was performed on a filtered sample of the Bulgarian subset of the CC100 dataset using PEFT.
π§ Model Details
- Base Model:
openai-community/gpt2-medium
- LoRA Rank: 8
- Target Modules:
c_attn
- Dataset:
cc100.bg
(1000 filtered samples) - Max Seq Length: 512 tokens
- Batch Size: 2 (with gradient accumulation)
- Steps: 1000
- Merged Model: Yes (LoRA weights fused into base model)
π¬ Example Usage
This model is a fine-tuned and merged version of openai-community/gpt2-medium
, adapted to Bulgarian using the LoRA technique. Training was performed on a filtered sample of the Bulgarian subset of the CC100 dataset using PEFT.
π§ Model Details
- Base Model:
openai-community/gpt2-medium
- LoRA Rank: 8
- Target Modules:
c_attn
- Dataset:
cc100.bg
(1000 filtered samples) - Max Seq Length: 512 tokens
- Batch Size: 2 (with gradient accumulation)
- Steps: 1000
- Merged Model: Yes (LoRA weights fused into base model)
π¬ Example Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("your-username/gpt2-bulgarian-merged")
tokenizer = AutoTokenizer.from_pretrained("your-username/gpt2-bulgarian-merged")
inputs = tokenizer("ΠΡΠ»Π³Π°ΡΠΈΡ Π΅ ΠΈΠ·Π²Π΅ΡΡΠ½Π° ΡΡΡ ΡΠ²ΠΎΠΈΡΠ΅", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
π Intended Use
For educational purposes, experimentation, and research on low-resource language modeling in Bulgarian.
β οΈ Limitations
- Trained on a small 1k sample.
- No toxic content filtering or safety tuning.
- Should not be used in production without further validation.
π€ Author
Developed by Vanessa Beck on Google Colab using π€ Transformers + PEFT.
- Downloads last month
- 5
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support