danielhanchen commited on
Commit
4eca56a
·
verified ·
1 Parent(s): 9ff3dd8

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -4,7 +4,6 @@
4
  "Qwen2VLForConditionalGeneration"
5
  ],
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
10
  "hidden_size": 3584,
@@ -32,34 +31,68 @@
32
  "multi_modal_projector",
33
  "merger",
34
  "modality_projection",
35
- "visual.blocks.0.attn",
36
- "visual.blocks.0.mlp",
37
- "visual.blocks.1.attn",
38
- "visual.blocks.1.mlp",
39
- "visual.blocks.2.attn",
40
- "visual.blocks.2.mlp",
41
- "visual.blocks.3.attn",
42
- "visual.blocks.3.mlp",
43
- "visual.blocks.4.attn",
44
- "visual.blocks.4.mlp",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  "visual.blocks.5.attn",
46
- "visual.blocks.5.mlp",
47
  "visual.blocks.6.attn",
48
- "visual.blocks.6.mlp",
 
49
  "visual.blocks.7.attn",
 
50
  "visual.blocks.7.mlp",
 
 
 
 
51
  "visual.blocks.8.attn",
52
- "visual.blocks.8.mlp",
53
- "visual.blocks.9.attn",
54
- "visual.blocks.9.mlp",
55
- "visual.blocks.10.attn",
56
- "visual.blocks.10.mlp",
57
- "visual.blocks.11.attn",
58
- "visual.blocks.11.mlp",
59
- "visual.blocks.12.attn",
60
- "visual.blocks.12.mlp",
61
- "visual.blocks.31.attn.proj",
62
- "visual.blocks.31.mlp.fc2"
63
  ],
64
  "llm_int8_threshold": 6.0,
65
  "load_in_4bit": true,
@@ -80,7 +113,7 @@
80
  "sliding_window": 32768,
81
  "tie_word_embeddings": false,
82
  "torch_dtype": "bfloat16",
83
- "transformers_version": "4.46.3",
84
  "unsloth_fixed": true,
85
  "use_cache": true,
86
  "use_sliding_window": false,
@@ -88,7 +121,8 @@
88
  "vision_config": {
89
  "in_chans": 3,
90
  "model_type": "qwen2_vl",
91
- "spatial_patch_size": 14
 
92
  },
93
  "vision_end_token_id": 151653,
94
  "vision_start_token_id": 151652,
 
4
  "Qwen2VLForConditionalGeneration"
5
  ],
6
  "attention_dropout": 0.0,
 
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 3584,
 
31
  "multi_modal_projector",
32
  "merger",
33
  "modality_projection",
34
+ "visual.blocks.22.mlp",
35
+ "visual.blocks.30.mlp",
36
+ "visual.blocks.26.mlp",
37
+ "visual.blocks.23.mlp",
38
+ "visual.blocks.24.attn",
39
+ "visual.blocks.21.attn",
40
+ "visual.blocks.29.attn",
41
+ "visual.merger.mlp",
42
+ "visual.blocks.20.attn",
43
+ "visual.blocks.24.mlp",
44
+ "visual.blocks.23.attn",
45
+ "visual.blocks.20.mlp",
46
+ "visual.blocks.28.mlp",
47
+ "visual.blocks.27.mlp",
48
+ "visual.blocks.30.attn",
49
+ "visual.blocks.18.mlp",
50
+ "visual.blocks.25.attn",
51
+ "visual.blocks.26.attn",
52
+ "visual.blocks.21.mlp",
53
+ "visual.blocks.19.attn",
54
+ "visual.blocks.16.mlp",
55
+ "visual.blocks.25.mlp",
56
+ "visual.blocks.17.mlp",
57
+ "visual.blocks.13.mlp",
58
+ "visual.blocks.27.attn",
59
+ "visual.blocks.15.attn",
60
+ "visual.blocks.15.mlp",
61
+ "visual.blocks.17.attn",
62
+ "visual.blocks.16.attn",
63
+ "visual.blocks.13.attn",
64
+ "visual.blocks.28.attn",
65
+ "visual.blocks.14.mlp",
66
+ "visual.blocks.18.attn",
67
+ "visual.blocks.14.attn",
68
+ "visual.blocks.12.attn",
69
+ "visual.blocks.12.mlp",
70
+ "visual.blocks.9.attn",
71
+ "visual.blocks.11.mlp",
72
+ "visual.blocks.9.mlp",
73
+ "visual.blocks.10.mlp",
74
+ "visual.blocks.11.attn",
75
+ "visual.blocks.10.attn",
76
  "visual.blocks.5.attn",
 
77
  "visual.blocks.6.attn",
78
+ "visual.blocks.3.mlp",
79
+ "visual.blocks.5.mlp",
80
  "visual.blocks.7.attn",
81
+ "visual.blocks.8.mlp",
82
  "visual.blocks.7.mlp",
83
+ "visual.blocks.3.attn",
84
+ "visual.blocks.1.attn",
85
+ "visual.blocks.4.mlp",
86
+ "visual.blocks.6.mlp",
87
  "visual.blocks.8.attn",
88
+ "visual.blocks.4.attn",
89
+ "visual.blocks.1.mlp",
90
+ "visual.blocks.2.mlp",
91
+ "visual.blocks.2.attn",
92
+ "visual.blocks.0.attn",
93
+ "visual.blocks.0.mlp",
94
+ "visual.blocks.31.attn",
95
+ "visual.blocks.31.mlp"
 
 
 
96
  ],
97
  "llm_int8_threshold": 6.0,
98
  "load_in_4bit": true,
 
113
  "sliding_window": 32768,
114
  "tie_word_embeddings": false,
115
  "torch_dtype": "bfloat16",
116
+ "transformers_version": "4.49.0",
117
  "unsloth_fixed": true,
118
  "use_cache": true,
119
  "use_sliding_window": false,
 
121
  "vision_config": {
122
  "in_chans": 3,
123
  "model_type": "qwen2_vl",
124
+ "spatial_patch_size": 14,
125
+ "torch_dtype": "bfloat16"
126
  },
127
  "vision_end_token_id": 151653,
128
  "vision_start_token_id": 151652,
generation_config.json CHANGED
@@ -10,5 +10,5 @@
10
  "temperature": 0.01,
11
  "top_k": 1,
12
  "top_p": 0.001,
13
- "transformers_version": "4.46.3"
14
  }
 
10
  "temperature": 0.01,
11
  "top_k": 1,
12
  "top_p": 0.001,
13
+ "transformers_version": "4.49.0"
14
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e245d65133e1766982241347c33092b07cdcf8745943291b641e3a8aa6bbeec3
3
- size 6356703012
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2f687bb8c5437cb08c8e04ea321ed316ff6cda33b88ad754cf655364029e5be
3
+ size 6850192973
preprocessor_config.json CHANGED
@@ -22,8 +22,8 @@
22
  "resample": 3,
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
- "max_pixels": 12845056,
26
- "min_pixels": 3136
27
  },
28
  "temporal_patch_size": 2
29
  }
 
22
  "resample": 3,
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
  },
28
  "temporal_patch_size": 2
29
  }
tokenizer_config.json CHANGED
@@ -134,6 +134,7 @@
134
  "clean_up_tokenization_spaces": false,
135
  "eos_token": "<|im_end|>",
136
  "errors": "replace",
 
137
  "model_max_length": 32768,
138
  "pad_token": "<|vision_pad|>",
139
  "padding_side": "left",
 
134
  "clean_up_tokenization_spaces": false,
135
  "eos_token": "<|im_end|>",
136
  "errors": "replace",
137
+ "extra_special_tokens": {},
138
  "model_max_length": 32768,
139
  "pad_token": "<|vision_pad|>",
140
  "padding_side": "left",