jpacifico's picture
Update README.md
98d049b verified
|
raw
history blame
990 Bytes
---
library_name: transformers
license: apache-2.0
language:
- fr
tags:
- phi-3
- french
- phi-3-3B
---
## Model Card for Model ID
Chocolatine v1.0
3.82B params.
Window context = 4k tokens
This is a DPO fine-tune of Microsoft's Phi-3-mini-4k-instruct, trained to master the French language perfectly.
![image/jpeg](https://github.com/jpacifico/French-Alpaca/blob/main/Assets/chocolatine.png?raw=true)
### Model Description
Fine-tuned with the 12k DPO Intel/orca_dpo_pairs translated in French : AIffl/french_orca_dpo_pairs.
Chocolatine is a general model and can itself be finetuned to be specialized for specific use cases.
More infos & Benchmarks very soon ^^
### Limitations
Chocolatine is a quick demonstration that a base 3B model can be easily fine-tuned to specialize in a particular language.
It does not have any moderation mechanisms.
- **Developed by:** Jonathan Pacifico, 2024
- **Model type:** LLM
- **Language(s) (NLP):** French
- **License:** MIT