danielhanchen commited on
Commit
21b18ed
·
verified ·
1 Parent(s): 7ae2b51

Add files using upload-large-folder tool

Browse files
Qwen3-235B-A22B-UD-IQ1_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243bf999121284e4bb1866451694c52cb4ca4fc66ac2d21063ce21af63a3b367
3
+ size 1153521504
Qwen3-235B-A22B-UD-IQ2_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8878e64324d5e31b5bd650dd6167a0f10f347e14e03906a21d4d0e0d6732b61d
3
+ size 1212766048
Qwen3-235B-A22B-UD-Q4_K_XL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8053b7be334a98a20d62d77bd446647106ec76b43e6fe3fb5b2ad3b53b8556a0
3
+ size 30752254816
config.json CHANGED
@@ -11,7 +11,7 @@
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 8192,
15
  "max_position_embeddings": 40960,
16
  "max_window_layers": 94,
17
  "mlp_only_layers": [],
 
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 12288,
15
  "max_position_embeddings": 40960,
16
  "max_window_layers": 94,
17
  "mlp_only_layers": [],