File size: 990 Bytes
4b53ac5
 
63aa8ef
 
 
 
 
 
 
 
4b53ac5
 
63aa8ef
4b53ac5
63aa8ef
 
 
4b53ac5
63aa8ef
4b53ac5
63aa8ef
4b53ac5
 
 
63aa8ef
 
 
4b53ac5
63aa8ef
4b53ac5
98d049b
63aa8ef
4b53ac5
63aa8ef
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
library_name: transformers
license: apache-2.0
language:
- fr

tags:
- phi-3
- french
- phi-3-3B
---

## Model Card for Model ID

Chocolatine v1.0  
3.82B params.  
Window context = 4k tokens  

This is a DPO fine-tune of Microsoft's Phi-3-mini-4k-instruct, trained to master the French language perfectly.  

![image/jpeg](https://github.com/jpacifico/French-Alpaca/blob/main/Assets/chocolatine.png?raw=true)  

### Model Description

Fine-tuned with the 12k DPO Intel/orca_dpo_pairs translated in French : AIffl/french_orca_dpo_pairs.  
Chocolatine is a general model and can itself be finetuned to be specialized for specific use cases.  
More infos & Benchmarks very soon ^^  

### Limitations

Chocolatine is a quick demonstration that a base 3B model can be easily fine-tuned to specialize in a particular language.  
It does not have any moderation mechanisms.

- **Developed by:** Jonathan Pacifico, 2024
- **Model type:** LLM 
- **Language(s) (NLP):** French
- **License:** MIT