Nexesenex commited on
Commit
6057827
·
verified ·
1 Parent(s): 44c5cab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -19,9 +19,9 @@ Also known as : Llama_3.x_70b_Dolmen_v1.0 (1.1 will come soon)
19
 
20
  This model is essentially a Llama 3.1 smart brick based on by a 3.0->3.3 "port", to be used in second level merges.
21
 
22
- I might abandon the 3 stages "smart merges" (like Smarteaz) because they are dilluting too much the source models used with the merge-stock technique once I add more models on the top of them, even if the benches and PPL were good, and it was dilluted furthermore into the level 4/5 merges I was doing afterwards. It was too much of a soup.
23
 
24
- This time, for the base, I used a Llama 3.0 Dolphin 2.9.1/Llama 3.3 instruct abliterated merge, in order to get both the capabilities of each model, and notably Dolphin, not ported on Llama 70b 3.1 or 3.3 by CognitiveComputations.
25
 
26
  Then, I added the best 'instructions oriented' finetunes I know, simple as that.
27
 
 
19
 
20
  This model is essentially a Llama 3.1 smart brick based on by a 3.0->3.3 "port", to be used in second level merges.
21
 
22
+ I might abandon the 3 stages "smart merges" (like Smarteaz) because they are dilluting too much the source models used with the merge-stock technique once I add more models on the top of them. Even if the benches and PPL were good, and the prose as well, it ends up being too dilluted furthermore into the level 4/5 merges I'm doing afterwards.
23
 
24
+ So, this time, for the base, I used a Llama 3.0 Dolphin 2.9.1/Llama 3.3 instruct abliterated merge, in order to get both the capabilities of each model, and notably Dolphin, not ported on Llama 70b 3.1 or 3.3 by CognitiveComputations.
25
 
26
  Then, I added the best 'instructions oriented' finetunes I know, simple as that.
27