spankevich/llm-course-hw3-lora
Text Generation
•
Updated
•
2
Collection of models from the third LLM course homework. It containes three LLMs fine-tuned using LoRA, QLoRA, and DoRA.
Note it was used to fine-tune OuteAI/Lite-Oute-1-300M-Instruct for tweet tone classification problem. Default model achieved 0.08 f1-score, while fine-tuned version achieved 0.53 f1-score in less than 8 minutes of fine-tuning on a single A100
Note It was used to fine-tune the same model as LoRA. Fine-tuned version achieved 0.51 f1-score in less than 15 minutes of fine-tuning on a single A100
Note It was used to fine-tune a bigger model TinyLlama/TinyLlama-1.1B-Chat-v1.0. By default it achieved around 0.20f1-score. Fine-tuned version achieved 0.54 f1-score in less than 4 minutes of fine-tuning on a single A100