feihu.hf commited on
Commit
bac7b87
·
1 Parent(s): 505404f

update README

Browse files
Files changed (4) hide show
  1. README.md +1 -0
  2. config.json +4 -6
  3. generation_config.json +12 -6
  4. tokenizer_config.json +0 -1
README.md CHANGED
@@ -3,6 +3,7 @@ library_name: transformers
3
  license: apache-2.0
4
  license_link: https://huggingface.co/Qwen/Qwen3-14B/blob/main/LICENSE
5
  pipeline_tag: text-generation
 
6
  ---
7
 
8
  # Qwen3-14B-AWQ
 
3
  license: apache-2.0
4
  license_link: https://huggingface.co/Qwen/Qwen3-14B/blob/main/LICENSE
5
  pipeline_tag: text-generation
6
+ base_model: Qwen/Qwen3-14B
7
  ---
8
 
9
  # Qwen3-14B-AWQ
config.json CHANGED
@@ -4,20 +4,19 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
7
  "eos_token_id": 151645,
8
  "head_dim": 128,
9
  "hidden_act": "silu",
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 17408,
13
- "max_position_embeddings": 65536,
14
- "max_window_layers": 28,
15
  "model_type": "qwen3",
16
  "num_attention_heads": 40,
17
  "num_hidden_layers": 40,
18
  "num_key_value_heads": 8,
19
- "pad_token_id": 151643,
20
- "qkv_bias": false,
21
  "quantization_config": {
22
  "bits": 4,
23
  "group_size": 128,
@@ -33,8 +32,7 @@
33
  "tie_word_embeddings": false,
34
  "torch_dtype": "float16",
35
  "transformers_version": "4.51.3",
36
- "use_cache": false,
37
- "use_qk_norm": true,
38
  "use_sliding_window": false,
39
  "vocab_size": 151936
40
  }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
  "eos_token_id": 151645,
9
  "head_dim": 128,
10
  "hidden_act": "silu",
11
  "hidden_size": 5120,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 17408,
14
+ "max_position_embeddings": 40960,
15
+ "max_window_layers": 40,
16
  "model_type": "qwen3",
17
  "num_attention_heads": 40,
18
  "num_hidden_layers": 40,
19
  "num_key_value_heads": 8,
 
 
20
  "quantization_config": {
21
  "bits": 4,
22
  "group_size": 128,
 
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "float16",
34
  "transformers_version": "4.51.3",
35
+ "use_cache": true,
 
36
  "use_sliding_window": false,
37
  "vocab_size": 151936
38
  }
generation_config.json CHANGED
@@ -1,7 +1,13 @@
1
  {
2
- "_from_model_config": true,
3
- "do_sample": true,
4
- "eos_token_id": 151645,
5
- "pad_token_id": 151643,
6
- "transformers_version": "4.51.3"
7
- }
 
 
 
 
 
 
 
1
  {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "temperature": 0.6,
10
+ "top_k": 20,
11
+ "top_p": 0.95,
12
+ "transformers_version": "4.51.0"
13
+ }
tokenizer_config.json CHANGED
@@ -231,7 +231,6 @@
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
234
- "extra_special_tokens": {},
235
  "model_max_length": 131072,
236
  "pad_token": "<|endoftext|>",
237
  "split_special_tokens": false,
 
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
 
234
  "model_max_length": 131072,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,