Update README.md
Browse files
README.md
CHANGED
@@ -14,14 +14,13 @@ pipeline_tag: text-generation
|
|
14 |
|
15 |
### Chocolatine-3B-Instruct-DPO-v1.2
|
16 |
|
|
|
|
|
|
|
17 |
DPO fine-tuned of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct) (3.82B params)
|
18 |
using the [jpacifico/french-orca-dpo-pairs-revised](https://huggingface.co/datasets/jpacifico/french-orca-dpo-pairs-revised) rlhf dataset.
|
19 |
Training in French also improves the model in English, surpassing the performances of its base model.
|
20 |
-
*The model supports 128K context length*.
|
21 |
-
|
22 |
-
### OpenLLM Leaderboard
|
23 |
|
24 |
-
TBD.
|
25 |
|
26 |
### MT-Bench-French
|
27 |
|
|
|
14 |
|
15 |
### Chocolatine-3B-Instruct-DPO-v1.2
|
16 |
|
17 |
+
Best version of Chocolatine-3B for French.
|
18 |
+
*The model supports 128K context length*.
|
19 |
+
|
20 |
DPO fine-tuned of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct) (3.82B params)
|
21 |
using the [jpacifico/french-orca-dpo-pairs-revised](https://huggingface.co/datasets/jpacifico/french-orca-dpo-pairs-revised) rlhf dataset.
|
22 |
Training in French also improves the model in English, surpassing the performances of its base model.
|
|
|
|
|
|
|
23 |
|
|
|
24 |
|
25 |
### MT-Bench-French
|
26 |
|