danielhanchen commited on
Commit
7548fff
·
verified ·
1 Parent(s): f29a54e

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "rope_theta": 1000000.0,
22
- "sliding_window": null,
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
- "transformers_version": "4.49.0.dev0",
26
  "unsloth_fixed": true,
27
  "use_cache": true,
28
  "use_sliding_window": false,
 
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "rope_theta": 1000000.0,
22
+ "sliding_window": 32768,
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.49.0",
26
  "unsloth_fixed": true,
27
  "use_cache": true,
28
  "use_sliding_window": false,
generation_config.json CHANGED
@@ -11,5 +11,5 @@
11
  "temperature": 0.7,
12
  "top_k": 20,
13
  "top_p": 0.8,
14
- "transformers_version": "4.49.0.dev0"
15
  }
 
11
  "temperature": 0.7,
12
  "top_k": 20,
13
  "top_p": 0.8,
14
+ "transformers_version": "4.49.0"
15
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c309ac1c08153adb2846748be27a184646297975607ac809a5bff7aa36e2822
3
- size 4957560304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67347b23fb4165b652eb6611f5e1f2a06dfcddba8e909df1b2b0b1857bee06c2
3
+ size 3968658944
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a22e9460cf702833fbe02be74b40815d19f1347b5093c67fac48e87ecbbe02d7
3
- size 1214366696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a40d941d0e7e0b966ad8b62bb6d6b7c88cce1299197b599d9d0a4ce59aabfc1d
3
+ size 2203268048
model.safetensors.index.json CHANGED
@@ -173,7 +173,7 @@
173
  "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
174
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
175
  "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
176
- "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
177
  "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
178
  "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
179
  "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
@@ -184,90 +184,90 @@
184
  "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
185
  "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
186
  "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
187
- "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
188
- "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
189
- "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
190
- "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
191
- "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
192
- "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
193
- "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
194
- "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
195
- "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
196
- "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
197
- "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
198
- "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
199
- "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
200
- "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
201
- "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
202
- "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
203
- "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
204
- "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
205
- "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
206
- "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
207
- "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
208
- "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
209
- "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
210
- "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
211
- "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
212
- "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
213
- "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
214
- "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
215
- "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
216
- "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
217
- "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
218
- "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
219
- "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
220
- "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
221
- "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
222
- "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
223
- "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
224
- "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
225
- "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
226
- "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
227
- "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
228
- "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
229
- "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
230
- "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
231
- "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
232
- "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
233
- "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
234
- "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
235
- "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
236
- "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
237
- "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
238
- "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
239
- "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
240
- "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
241
- "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
242
- "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
243
- "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
244
- "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
245
- "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
246
- "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
247
- "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
248
- "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
249
- "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
250
- "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
251
- "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
252
- "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
253
- "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
254
- "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
255
- "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
256
- "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
257
- "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
258
- "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
259
  "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
260
  "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
261
  "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
262
  "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
263
  "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
264
- "model.layers.28.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
265
- "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
266
- "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
267
- "model.layers.28.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
268
- "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
269
- "model.layers.28.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
270
- "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
271
  "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
272
  "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
273
  "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
 
173
  "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
174
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
175
  "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
176
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
177
  "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
178
  "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
179
  "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
184
  "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
185
  "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
186
  "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
187
+ "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
188
+ "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
189
+ "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
190
+ "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
191
+ "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
192
+ "model.layers.22.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
193
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
194
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
195
+ "model.layers.22.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
196
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
197
+ "model.layers.22.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
198
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
199
+ "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
200
+ "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
201
+ "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
202
+ "model.layers.23.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
203
+ "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
204
+ "model.layers.23.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
205
+ "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
206
+ "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
207
+ "model.layers.23.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
208
+ "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
209
+ "model.layers.23.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
210
+ "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
211
+ "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
212
+ "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
213
+ "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
214
+ "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
215
+ "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
216
+ "model.layers.24.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
217
+ "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
218
+ "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
219
+ "model.layers.24.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
220
+ "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
221
+ "model.layers.24.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
222
+ "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
223
+ "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors",
224
+ "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
225
+ "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
226
+ "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
227
+ "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
228
+ "model.layers.25.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
229
+ "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
230
+ "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
231
+ "model.layers.25.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
232
+ "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
233
+ "model.layers.25.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
234
+ "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
235
+ "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors",
236
+ "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
237
+ "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
238
+ "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
239
+ "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
240
+ "model.layers.26.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
241
+ "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
242
+ "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
243
+ "model.layers.26.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
244
+ "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
245
+ "model.layers.26.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
246
+ "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
247
+ "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors",
248
+ "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
249
+ "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
250
+ "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
251
+ "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
252
+ "model.layers.27.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
253
+ "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
254
+ "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
255
+ "model.layers.27.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
256
+ "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
257
+ "model.layers.27.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
258
+ "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
259
  "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
260
  "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
261
  "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
262
  "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
263
  "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.28.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
265
+ "model.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
266
+ "model.layers.28.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
267
+ "model.layers.28.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
268
+ "model.layers.28.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
269
+ "model.layers.28.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
270
+ "model.layers.28.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
271
  "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
272
  "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
273
  "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",