|
--- |
|
library_name: transformers |
|
license: apache-2.0 |
|
language: |
|
- fr |
|
|
|
tags: |
|
- phi-3 |
|
- french |
|
- phi-3-3B |
|
--- |
|
|
|
## Model Card for Model ID |
|
|
|
Chocolatine v1.0 |
|
3.82B params. |
|
Window context = 4k tokens |
|
|
|
This is a DPO fine-tune of Microsoft's Phi-3-mini-4k-instruct, trained to master the French language perfectly. |
|
|
|
 |
|
|
|
### Model Description |
|
|
|
Fine-tuned with the 12k DPO Intel/orca_dpo_pairs translated in French : AIffl/french_orca_dpo_pairs. |
|
Chocolatine is a general model and can itself be finetuned to be specialized for specific use cases. |
|
More infos & Benchmarks very soon ^^ |
|
|
|
### Limitations |
|
|
|
Chocolatine is a quick demonstration that a base 3B model can be easily fine-tuned to specialize in a particular language. |
|
It does not have any moderation mechanisms. |
|
|
|
- **Developed by:** Jonathan Pacifico, 2024 |
|
- **Model type:** LLM |
|
- **Language(s) (NLP):** French |
|
- **License:** MIT |