aashish1904 commited on
Commit
0aa01eb
·
verified ·
1 Parent(s): 2640a71

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +98 -0
README.md ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - wzhouad/gemma-2-9b-it-WPO-HB
6
+ - UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
7
+ - google/gemma-2-9b-it
8
+ - princeton-nlp/gemma-2-9b-it-SimPO
9
+ library_name: transformers
10
+ tags:
11
+ - mergekit
12
+ - merge
13
+ - merge
14
+
15
+
16
+ ---
17
+
18
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
19
+
20
+ # QuantFactory/gemma-2-9B-it-advanced-v2.1-GGUF
21
+ This is quantized version of [jsgreenawalt/gemma-2-9B-it-advanced-v2.1](https://huggingface.co/jsgreenawalt/gemma-2-9B-it-advanced-v2.1) created using llama.cpp
22
+
23
+ # Original Model Card
24
+
25
+ # Gemma Advanced V2.1
26
+
27
+ This is a merge of the 'smartest' advanced fine-tunes available for Gemma-2-9b-it. It includes WPO, SimPO, and SPPO. The merge was performed via the SOTA 'della' merge method. Merge parameters have been hand-tuned for best results. The Q8_0 quant is highly recommended until better quants come along.
28
+
29
+ ## Notes and observations:
30
+ * The extreme temperature sensitivity from V1 has been fixed, no longer needs to be run at lower temperatures
31
+ * Has a somewhat different writing style than any of the parent models
32
+ * Great instruction following
33
+ * Tracks plot details well and has good situational understanding
34
+ * Seems to have a good understanding of psychology, emotions and creative writing
35
+ * More 'sane' than base gemma-it, SPPO, or SimPO - not as prone to 'Cruella De Vil' or 'Evil Sorceress' like SPPO or SimPO, when portraying characters
36
+ * Would likely serve as a good base for further merges
37
+ * I'm looking for a job, if you're hiring. I'm a skilled Python developer who brings strong devops skills along with an ever-growing knowledge of machine learning pipelines and models. Message me if you want to talk about what I can bring to your team.
38
+ * Overall, this feels like a very useful and successful merge.
39
+
40
+ Quantized GGUFs can be found here: https://huggingface.co/jsgreenawalt/gemma-2-9B-it-advanced-v2.1-GGUF
41
+ I'll link to other quants as they appear.
42
+
43
+ # sample ollama Modelfile
44
+ ```yaml
45
+ FROM /path/to/file/gemma-2-9B-it-advanced-v2.1-Q8_0.gguf
46
+ PARAMETER stop "<start_of_turn>"
47
+ PARAMETER stop "<end_of_turn>"
48
+ PARAMETER num_ctx 8192
49
+ TEMPLATE """<start_of_turn>user
50
+ {{ if .System }}{{ .System }} {{ end }}{{ .Prompt }}<end_of_turn>
51
+ <start_of_turn>model
52
+ {{ .Response }}<end_of_turn>"""
53
+ ```
54
+
55
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
56
+
57
+ ## Merge Details
58
+ ### Merge Method
59
+
60
+ This model was merged using the della merge method using [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it) as a base.
61
+
62
+ ### Models Merged
63
+
64
+ The following models were included in the merge:
65
+ * [wzhouad/gemma-2-9b-it-WPO-HB](https://huggingface.co/wzhouad/gemma-2-9b-it-WPO-HB)
66
+ * [princeton-nlp/gemma-2-9b-it-SimPO](https://huggingface.co/princeton-nlp/gemma-2-9b-it-SimPO)
67
+ * [UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3](https://huggingface.co/UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3)
68
+
69
+ ### Configuration
70
+
71
+ The following YAML configuration was used to produce this model:
72
+
73
+ ```yaml
74
+ models:
75
+ - model: google/gemma-2-9b-it
76
+ - model: wzhouad/gemma-2-9b-it-WPO-HB
77
+ parameters:
78
+ density: 0.55
79
+ weight: 0.6
80
+ - model: princeton-nlp/gemma-2-9b-it-SimPO
81
+ parameters:
82
+ density: 0.35
83
+ weight: 0.6
84
+ - model: UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
85
+ parameters:
86
+ density: 0.25
87
+ weight: 0.4
88
+ merge_method: della
89
+ base_model: google/gemma-2-9b-it
90
+ parameters:
91
+ normalize: true
92
+ int8_mask: true
93
+ lambda: 1.0
94
+ epsilon: 0.1
95
+ dtype: float16
96
+
97
+ ```
98
+