dahara1 commited on
Commit
47e0829
·
verified ·
1 Parent(s): 878b995

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -17
README.md CHANGED
@@ -15,20 +15,7 @@ language:
15
  This gguf version aims to create a gguf with [improved quantization parameters](https://huggingface.co/unsloth/QwQ-32B) and [improved Japanese language capabilities](https://huggingface.co/dahara1/imatrix-jpn-test).
16
  Details are under verification.
17
 
18
- ### currnet sample parameters.
19
-
20
- ```
21
- temperature = 0.6
22
- top-k = 40 (20 to 40 suggested)
23
- min-p = 0.00 (optional, but 0.01 works well, llama.cpp default is 0.1)
24
- top-p = 0.95
25
- repetition-penalty = 1.0
26
- dry-multiplier 0.5
27
- Chat template: <|im_start|>user\nCreate a Flappy Bird game in Python.<|im_end|>\n<|im_start|>assistant\n<think>\n
28
- ```
29
- Reference information
30
- [Tutorial: How to Run QwQ-32B effectively](https://docs.unsloth.ai/basics/tutorial-how-to-run-qwq-32b-effectively#tutorial-how-to-run-qwq-32b)
31
-
32
 
33
  おそらくllama-serverの仕様が変わったためか、私はllama-server経由で正常に動かす事がまだできていません(繰り返しが発生してしまう)
34
  Probably because the llama-server specifications have changed, I have not yet been able to get it to work properly via llama-server (repetition occurs).
@@ -166,8 +153,17 @@ llama_perf_context_print: total time = 864163.89 ms / 1681 tokens
166
  ```
167
 
168
 
 
169
 
170
-
171
-
172
-
 
 
 
 
 
 
 
 
173
 
 
15
  This gguf version aims to create a gguf with [improved quantization parameters](https://huggingface.co/unsloth/QwQ-32B) and [improved Japanese language capabilities](https://huggingface.co/dahara1/imatrix-jpn-test).
16
  Details are under verification.
17
 
18
+ ## example
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  おそらくllama-serverの仕様が変わったためか、私はllama-server経由で正常に動かす事がまだできていません(繰り返しが発生してしまう)
21
  Probably because the llama-server specifications have changed, I have not yet been able to get it to work properly via llama-server (repetition occurs).
 
153
  ```
154
 
155
 
156
+ ## currnet sample parameters.
157
 
158
+ ```
159
+ temperature = 0.6
160
+ top-k = 40 (20 to 40 suggested)
161
+ min-p = 0.00 (optional, but 0.01 works well, llama.cpp default is 0.1)
162
+ top-p = 0.95
163
+ repetition-penalty = 1.0
164
+ dry-multiplier 0.5
165
+ Chat template: <|im_start|>user\nCreate a Flappy Bird game in Python.<|im_end|>\n<|im_start|>assistant\n<think>\n
166
+ ```
167
+ Reference information
168
+ [Tutorial: How to Run QwQ-32B effectively](https://docs.unsloth.ai/basics/tutorial-how-to-run-qwq-32b-effectively#tutorial-how-to-run-qwq-32b)
169