π¦ llama38bbatch1model-merged
This is a fine-tuned LLaMA-3 8B model using QLoRA + PEFT + SFTTrainer, trained on a custom dataset (final_batch1
, 2
, and 3.csv
) by @hemanthvamsikrishna.
π Training Details
- Base Model:
meta-llama/Meta-Llama-3-8B
- Fine-tuning Method: QLoRA (4-bit), PEFT (LoRA)
- Trainer:
trl.SFTTrainer
- Epochs: 3
- Batch Size: 4 (with gradient accumulation = 4)
- Max Sequence Length: 512
- LoRA Targets:
['q_proj', 'k_proj', 'v_proj', 'o_proj', 'gate_proj', 'up_proj', 'down_proj']
π¦ Files Included
- β
Sharded model weights (
pytorch_model-*.bin
) - β Tokenizer & Config files
- β LoRA merged with base weights
π§ͺ Example Usage
from transformers import pipeline
pipe = pipeline("text-generation", model="hemanthvamsikrishna/llama38bbatch1model-merged")
output = pipe("What is the future of AI?", max_new_tokens=100)
print(output[0]['generated_text'])
- Downloads last month
- 19
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for hemanthvamsikrishna/llama38bbatch1model-merged
Base model
meta-llama/Meta-Llama-3-8B