htlou commited on
Commit
87de4e8
·
verified ·
1 Parent(s): 5d5cd66

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +83 -0
  2. added_tokens.json +4 -0
  3. all_results.json +20 -0
  4. checkpoint-120/added_tokens.json +4 -0
  5. checkpoint-120/config.json +68 -0
  6. checkpoint-120/generation_config.json +6 -0
  7. checkpoint-120/model-00001-of-00004.safetensors +3 -0
  8. checkpoint-120/model-00002-of-00004.safetensors +3 -0
  9. checkpoint-120/model-00003-of-00004.safetensors +3 -0
  10. checkpoint-120/model-00004-of-00004.safetensors +3 -0
  11. checkpoint-120/model.safetensors.index.json +694 -0
  12. checkpoint-120/preprocessor_config.json +52 -0
  13. checkpoint-120/special_tokens_map.json +30 -0
  14. checkpoint-120/tokenizer.json +0 -0
  15. checkpoint-120/tokenizer.model +3 -0
  16. checkpoint-120/tokenizer_config.json +70 -0
  17. checkpoint-120/trainer_state.json +441 -0
  18. checkpoint-120/training_args.bin +3 -0
  19. checkpoint-160/added_tokens.json +4 -0
  20. checkpoint-160/config.json +68 -0
  21. checkpoint-160/generation_config.json +6 -0
  22. checkpoint-160/model-00001-of-00004.safetensors +3 -0
  23. checkpoint-160/model-00002-of-00004.safetensors +3 -0
  24. checkpoint-160/model-00003-of-00004.safetensors +3 -0
  25. checkpoint-160/model-00004-of-00004.safetensors +3 -0
  26. checkpoint-160/model.safetensors.index.json +694 -0
  27. checkpoint-160/preprocessor_config.json +52 -0
  28. checkpoint-160/special_tokens_map.json +30 -0
  29. checkpoint-160/tokenizer.json +0 -0
  30. checkpoint-160/tokenizer.model +3 -0
  31. checkpoint-160/tokenizer_config.json +70 -0
  32. checkpoint-160/trainer_state.json +577 -0
  33. checkpoint-160/training_args.bin +3 -0
  34. checkpoint-200/added_tokens.json +4 -0
  35. checkpoint-200/config.json +68 -0
  36. checkpoint-200/generation_config.json +6 -0
  37. checkpoint-200/model-00001-of-00004.safetensors +3 -0
  38. checkpoint-200/model-00002-of-00004.safetensors +3 -0
  39. checkpoint-200/model-00003-of-00004.safetensors +3 -0
  40. checkpoint-200/model-00004-of-00004.safetensors +3 -0
  41. checkpoint-200/model.safetensors.index.json +694 -0
  42. checkpoint-200/preprocessor_config.json +52 -0
  43. checkpoint-200/special_tokens_map.json +30 -0
  44. checkpoint-200/tokenizer.json +0 -0
  45. checkpoint-200/tokenizer.model +3 -0
  46. checkpoint-200/tokenizer_config.json +70 -0
  47. checkpoint-200/trainer_state.json +713 -0
  48. checkpoint-200/training_args.bin +3 -0
  49. checkpoint-240/added_tokens.json +4 -0
  50. checkpoint-240/config.json +68 -0
README.md ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: other
4
+ base_model: llava-hf/llava-v1.6-mistral-7b-hf
5
+ tags:
6
+ - llama-factory
7
+ - full
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: AA_text_image_to_text
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # AA_text_image_to_text
18
+
19
+ This model is a fine-tuned version of [llava-hf/llava-v1.6-mistral-7b-hf](https://huggingface.co/llava-hf/llava-v1.6-mistral-7b-hf) on the AA_text_image_to_text dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.4527
22
+ - Rewards/chosen: -0.6857
23
+ - Rewards/rejected: -4.3940
24
+ - Rewards/accuracies: 0.8165
25
+ - Rewards/margins: 3.7083
26
+ - Logps/rejected: -242.1480
27
+ - Logps/chosen: -207.1762
28
+ - Logits/rejected: -2.3240
29
+ - Logits/chosen: -2.3485
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+
43
+ ## Training procedure
44
+
45
+ ### Training hyperparameters
46
+
47
+ The following hyperparameters were used during training:
48
+ - learning_rate: 1e-06
49
+ - train_batch_size: 8
50
+ - eval_batch_size: 8
51
+ - seed: 42
52
+ - distributed_type: multi-GPU
53
+ - num_devices: 8
54
+ - gradient_accumulation_steps: 4
55
+ - total_train_batch_size: 256
56
+ - total_eval_batch_size: 64
57
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
58
+ - lr_scheduler_type: cosine
59
+ - lr_scheduler_warmup_steps: 10
60
+ - num_epochs: 3.0
61
+
62
+ ### Training results
63
+
64
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
65
+ |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
66
+ | 0.4889 | 0.2899 | 40 | 0.4642 | 1.1544 | -0.1887 | 0.7944 | 1.3431 | -200.0950 | -188.7752 | -1.9876 | -2.0351 |
67
+ | 0.3941 | 0.5797 | 80 | 0.4218 | -0.2275 | -2.2919 | 0.8044 | 2.0644 | -221.1273 | -202.5944 | -1.9449 | -1.9901 |
68
+ | 0.3717 | 0.8696 | 120 | 0.4387 | -0.2101 | -2.4885 | 0.8286 | 2.2784 | -223.0936 | -202.4208 | -2.0902 | -2.1229 |
69
+ | 0.1459 | 1.1594 | 160 | 0.4288 | -0.4029 | -3.3928 | 0.8286 | 2.9899 | -232.1363 | -204.3488 | -2.2733 | -2.3007 |
70
+ | 0.1455 | 1.4493 | 200 | 0.4255 | -0.5338 | -3.6331 | 0.8165 | 3.0992 | -234.5387 | -205.6577 | -2.2466 | -2.2697 |
71
+ | 0.1358 | 1.7391 | 240 | 0.4247 | -0.2714 | -3.6715 | 0.8327 | 3.4001 | -234.9227 | -203.0333 | -2.3605 | -2.3806 |
72
+ | 0.0938 | 2.0290 | 280 | 0.4128 | -0.3136 | -3.7007 | 0.8266 | 3.3870 | -235.2147 | -203.4556 | -2.3725 | -2.3933 |
73
+ | 0.0592 | 2.3188 | 320 | 0.4438 | -0.5767 | -4.1235 | 0.8165 | 3.5467 | -239.4429 | -206.0869 | -2.3109 | -2.3358 |
74
+ | 0.0673 | 2.6087 | 360 | 0.4553 | -0.6264 | -4.3005 | 0.8206 | 3.6740 | -241.2126 | -206.5837 | -2.3254 | -2.3497 |
75
+ | 0.0728 | 2.8986 | 400 | 0.4520 | -0.6855 | -4.3942 | 0.8185 | 3.7087 | -242.1503 | -207.1744 | -2.3247 | -2.3492 |
76
+
77
+
78
+ ### Framework versions
79
+
80
+ - Transformers 4.45.2
81
+ - Pytorch 2.4.0+cu121
82
+ - Datasets 2.21.0
83
+ - Tokenizers 0.20.3
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
all_results.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_logits/chosen": -2.3485426902770996,
4
+ "eval_logits/rejected": -2.32399582862854,
5
+ "eval_logps/chosen": -207.17623901367188,
6
+ "eval_logps/rejected": -242.1480255126953,
7
+ "eval_loss": 0.45268043875694275,
8
+ "eval_rewards/accuracies": 0.8165322542190552,
9
+ "eval_rewards/chosen": -0.6856781244277954,
10
+ "eval_rewards/margins": 3.7083182334899902,
11
+ "eval_rewards/rejected": -4.393996715545654,
12
+ "eval_runtime": 246.3688,
13
+ "eval_samples_per_second": 15.919,
14
+ "eval_steps_per_second": 0.252,
15
+ "total_flos": 4881795388538880.0,
16
+ "train_loss": 0.21833302825689316,
17
+ "train_runtime": 16364.2441,
18
+ "train_samples_per_second": 6.47,
19
+ "train_steps_per_second": 0.025
20
+ }
checkpoint-120/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
checkpoint-120/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/data/align-anything/hantao/models/llava-v1.6-mistral-7b-hf",
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "hidden_size": 4096,
7
+ "ignore_index": -100,
8
+ "image_grid_pinpoints": [
9
+ [
10
+ 336,
11
+ 672
12
+ ],
13
+ [
14
+ 672,
15
+ 336
16
+ ],
17
+ [
18
+ 672,
19
+ 672
20
+ ],
21
+ [
22
+ 1008,
23
+ 336
24
+ ],
25
+ [
26
+ 336,
27
+ 1008
28
+ ]
29
+ ],
30
+ "image_seq_length": 576,
31
+ "image_token_index": 32000,
32
+ "model_type": "llava_next",
33
+ "projector_hidden_act": "gelu",
34
+ "text_config": {
35
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
36
+ "architectures": [
37
+ "MistralForCausalLM"
38
+ ],
39
+ "intermediate_size": 14336,
40
+ "max_position_embeddings": 32768,
41
+ "model_type": "mistral",
42
+ "num_key_value_heads": 8,
43
+ "rms_norm_eps": 1e-05,
44
+ "rope_theta": 1000000.0,
45
+ "sliding_window": null,
46
+ "torch_dtype": "bfloat16",
47
+ "vocab_size": 32064
48
+ },
49
+ "tie_word_embeddings": false,
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.45.2",
52
+ "use_cache": false,
53
+ "use_image_newline_parameter": true,
54
+ "vision_config": {
55
+ "hidden_size": 1024,
56
+ "image_size": 336,
57
+ "intermediate_size": 4096,
58
+ "model_type": "clip_vision_model",
59
+ "num_attention_heads": 16,
60
+ "num_hidden_layers": 24,
61
+ "patch_size": 14,
62
+ "projection_dim": 768,
63
+ "vocab_size": 32000
64
+ },
65
+ "vision_feature_layer": -2,
66
+ "vision_feature_select_strategy": "default",
67
+ "vocab_size": 32064
68
+ }
checkpoint-120/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.45.2"
6
+ }
checkpoint-120/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e40b873bd71b600c08220e434771c07090facd4443ba0c6b07e93f0af4be79
3
+ size 4921618624
checkpoint-120/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8c6699524f4655083b3649f18b14a5ae434f40f28991d2c4c25b21c4066893b
3
+ size 4915917672
checkpoint-120/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c84c6f427c02e2982e1462d89cea773d96ffd94d3ff71fc5f48c7ead78d72cb6
3
+ size 4915917680
checkpoint-120/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:563c89f0c94059695c61aea783f7a4610659f158fdb25e7db0e335105da16f50
3
+ size 380134008
checkpoint-120/model.safetensors.index.json ADDED
@@ -0,0 +1,694 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15133495296
4
+ },
5
+ "weight_map": {
6
+ "image_newline": "model-00001-of-00004.safetensors",
7
+ "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
8
+ "language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
9
+ "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
10
+ "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
11
+ "language_model.model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
12
+ "language_model.model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
13
+ "language_model.model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
14
+ "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
+ "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
17
+ "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
18
+ "language_model.model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
19
+ "language_model.model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
20
+ "language_model.model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
21
+ "language_model.model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
22
+ "language_model.model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
23
+ "language_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
24
+ "language_model.model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
25
+ "language_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
26
+ "language_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
27
+ "language_model.model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
28
+ "language_model.model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
29
+ "language_model.model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
30
+ "language_model.model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
31
+ "language_model.model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
32
+ "language_model.model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
33
+ "language_model.model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
34
+ "language_model.model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
35
+ "language_model.model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
36
+ "language_model.model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "language_model.model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
38
+ "language_model.model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
39
+ "language_model.model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
40
+ "language_model.model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
41
+ "language_model.model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
42
+ "language_model.model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
43
+ "language_model.model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
44
+ "language_model.model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
45
+ "language_model.model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
46
+ "language_model.model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
47
+ "language_model.model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
48
+ "language_model.model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
49
+ "language_model.model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
50
+ "language_model.model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "language_model.model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
+ "language_model.model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
53
+ "language_model.model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
54
+ "language_model.model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
55
+ "language_model.model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
56
+ "language_model.model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
57
+ "language_model.model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
58
+ "language_model.model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
59
+ "language_model.model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
60
+ "language_model.model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
61
+ "language_model.model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
62
+ "language_model.model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
63
+ "language_model.model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
64
+ "language_model.model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
65
+ "language_model.model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
66
+ "language_model.model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
67
+ "language_model.model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
68
+ "language_model.model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
69
+ "language_model.model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
70
+ "language_model.model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
71
+ "language_model.model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
72
+ "language_model.model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
73
+ "language_model.model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
74
+ "language_model.model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
75
+ "language_model.model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
76
+ "language_model.model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
77
+ "language_model.model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
78
+ "language_model.model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
79
+ "language_model.model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
80
+ "language_model.model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
81
+ "language_model.model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
82
+ "language_model.model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
83
+ "language_model.model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
84
+ "language_model.model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
85
+ "language_model.model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
86
+ "language_model.model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
+ "language_model.model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "language_model.model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
89
+ "language_model.model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
90
+ "language_model.model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
91
+ "language_model.model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
92
+ "language_model.model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
93
+ "language_model.model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
94
+ "language_model.model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
95
+ "language_model.model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
96
+ "language_model.model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
97
+ "language_model.model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
98
+ "language_model.model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
99
+ "language_model.model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors",
100
+ "language_model.model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
101
+ "language_model.model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
102
+ "language_model.model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
103
+ "language_model.model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
104
+ "language_model.model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
105
+ "language_model.model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
106
+ "language_model.model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
107
+ "language_model.model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
108
+ "language_model.model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors",
109
+ "language_model.model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
110
+ "language_model.model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
111
+ "language_model.model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
112
+ "language_model.model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
113
+ "language_model.model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
114
+ "language_model.model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
115
+ "language_model.model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
116
+ "language_model.model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
117
+ "language_model.model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
118
+ "language_model.model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
119
+ "language_model.model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
120
+ "language_model.model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
121
+ "language_model.model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
122
+ "language_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
123
+ "language_model.model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
124
+ "language_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
125
+ "language_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
126
+ "language_model.model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
127
+ "language_model.model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
128
+ "language_model.model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
129
+ "language_model.model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
130
+ "language_model.model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
131
+ "language_model.model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
132
+ "language_model.model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
133
+ "language_model.model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
134
+ "language_model.model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
135
+ "language_model.model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
136
+ "language_model.model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
137
+ "language_model.model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
138
+ "language_model.model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
139
+ "language_model.model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
140
+ "language_model.model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
141
+ "language_model.model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
142
+ "language_model.model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
143
+ "language_model.model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
144
+ "language_model.model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
145
+ "language_model.model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
146
+ "language_model.model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
147
+ "language_model.model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
148
+ "language_model.model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
149
+ "language_model.model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
150
+ "language_model.model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
151
+ "language_model.model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
152
+ "language_model.model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
153
+ "language_model.model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
154
+ "language_model.model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
155
+ "language_model.model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
156
+ "language_model.model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
157
+ "language_model.model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
158
+ "language_model.model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
159
+ "language_model.model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
160
+ "language_model.model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
161
+ "language_model.model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
162
+ "language_model.model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
163
+ "language_model.model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
164
+ "language_model.model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
165
+ "language_model.model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
166
+ "language_model.model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
167
+ "language_model.model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
168
+ "language_model.model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
169
+ "language_model.model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
170
+ "language_model.model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
171
+ "language_model.model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
172
+ "language_model.model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
173
+ "language_model.model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
174
+ "language_model.model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
175
+ "language_model.model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
176
+ "language_model.model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
177
+ "language_model.model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
178
+ "language_model.model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
179
+ "language_model.model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
180
+ "language_model.model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "language_model.model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
182
+ "language_model.model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
183
+ "language_model.model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
184
+ "language_model.model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
185
+ "language_model.model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
186
+ "language_model.model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
187
+ "language_model.model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
188
+ "language_model.model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
189
+ "language_model.model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
190
+ "language_model.model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
191
+ "language_model.model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
192
+ "language_model.model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
193
+ "language_model.model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
194
+ "language_model.model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
+ "language_model.model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
+ "language_model.model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
197
+ "language_model.model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
198
+ "language_model.model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors",
199
+ "language_model.model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
200
+ "language_model.model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
201
+ "language_model.model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
202
+ "language_model.model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
203
+ "language_model.model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
204
+ "language_model.model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
205
+ "language_model.model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
206
+ "language_model.model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
207
+ "language_model.model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors",
208
+ "language_model.model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
209
+ "language_model.model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
210
+ "language_model.model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
211
+ "language_model.model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
212
+ "language_model.model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
213
+ "language_model.model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
214
+ "language_model.model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
215
+ "language_model.model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
216
+ "language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
217
+ "language_model.model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
218
+ "language_model.model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
219
+ "language_model.model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
220
+ "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
221
+ "language_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
222
+ "language_model.model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
223
+ "language_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
224
+ "language_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
225
+ "language_model.model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors",
226
+ "language_model.model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
227
+ "language_model.model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
228
+ "language_model.model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
229
+ "language_model.model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
230
+ "language_model.model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
+ "language_model.model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
+ "language_model.model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
233
+ "language_model.model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
234
+ "language_model.model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors",
235
+ "language_model.model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
236
+ "language_model.model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
237
+ "language_model.model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
238
+ "language_model.model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
239
+ "language_model.model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
240
+ "language_model.model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
241
+ "language_model.model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
242
+ "language_model.model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
243
+ "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
244
+ "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
245
+ "language_model.model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
246
+ "language_model.model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
247
+ "language_model.model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
248
+ "language_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
249
+ "language_model.model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
250
+ "language_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
251
+ "language_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
252
+ "language_model.model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
253
+ "language_model.model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
254
+ "language_model.model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
255
+ "language_model.model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
256
+ "language_model.model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
257
+ "language_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
258
+ "language_model.model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
259
+ "language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
260
+ "language_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
261
+ "language_model.model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
262
+ "language_model.model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
263
+ "language_model.model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
264
+ "language_model.model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
265
+ "language_model.model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
266
+ "language_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
+ "language_model.model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
+ "language_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
269
+ "language_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
270
+ "language_model.model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
271
+ "language_model.model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
272
+ "language_model.model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
273
+ "language_model.model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
274
+ "language_model.model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
275
+ "language_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
276
+ "language_model.model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
277
+ "language_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
278
+ "language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
279
+ "language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors",
280
+ "language_model.model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
281
+ "language_model.model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
282
+ "language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
283
+ "language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
284
+ "language_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
285
+ "language_model.model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
286
+ "language_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
287
+ "language_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
288
+ "language_model.model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
289
+ "language_model.model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
290
+ "language_model.model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
291
+ "language_model.model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
292
+ "language_model.model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
293
+ "language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
294
+ "language_model.model.layers.9.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
295
+ "language_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
296
+ "language_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
297
+ "language_model.model.norm.weight": "model-00004-of-00004.safetensors",
298
+ "multi_modal_projector.linear_1.bias": "model-00001-of-00004.safetensors",
299
+ "multi_modal_projector.linear_1.weight": "model-00001-of-00004.safetensors",
300
+ "multi_modal_projector.linear_2.bias": "model-00001-of-00004.safetensors",
301
+ "multi_modal_projector.linear_2.weight": "model-00001-of-00004.safetensors",
302
+ "vision_tower.vision_model.embeddings.class_embedding": "model-00001-of-00004.safetensors",
303
+ "vision_tower.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00004.safetensors",
304
+ "vision_tower.vision_model.embeddings.position_embedding.weight": "model-00001-of-00004.safetensors",
305
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00004.safetensors",
306
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00004.safetensors",
307
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00004.safetensors",
308
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00004.safetensors",
309
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00004.safetensors",
310
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00004.safetensors",
311
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00004.safetensors",
312
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00004.safetensors",
313
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
315
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
316
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
317
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
318
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
319
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
320
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
321
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00004.safetensors",
322
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00004.safetensors",
323
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00004.safetensors",
324
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00004.safetensors",
325
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00004.safetensors",
326
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00004.safetensors",
327
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00004.safetensors",
328
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00004.safetensors",
329
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
330
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
331
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
332
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
333
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
334
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
335
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
336
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
337
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00004.safetensors",
338
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00004.safetensors",
339
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00004.safetensors",
340
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00004.safetensors",
341
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00004.safetensors",
342
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00004.safetensors",
343
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00004.safetensors",
344
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00004.safetensors",
345
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
346
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
347
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
348
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
349
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
350
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
351
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
352
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
353
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00004.safetensors",
354
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00004.safetensors",
355
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00004.safetensors",
356
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00004.safetensors",
357
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00004.safetensors",
358
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00004.safetensors",
359
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00004.safetensors",
360
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00004.safetensors",
361
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
362
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
363
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
364
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
365
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
366
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
367
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
368
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
369
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00004.safetensors",
370
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00004.safetensors",
371
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00004.safetensors",
372
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00004.safetensors",
373
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00004.safetensors",
374
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00004.safetensors",
375
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00004.safetensors",
376
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00004.safetensors",
377
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
378
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
379
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
380
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
381
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
382
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
383
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
384
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
385
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00004.safetensors",
386
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00004.safetensors",
387
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00004.safetensors",
388
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00004.safetensors",
389
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00004.safetensors",
390
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00004.safetensors",
391
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00004.safetensors",
392
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00004.safetensors",
393
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
394
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
395
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
396
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
397
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
398
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
399
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
400
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
401
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00004.safetensors",
402
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00004.safetensors",
403
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00004.safetensors",
404
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00004.safetensors",
405
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00004.safetensors",
406
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00004.safetensors",
407
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00004.safetensors",
408
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00004.safetensors",
409
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
410
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
411
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
412
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
413
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
414
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
415
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
416
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
417
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00004.safetensors",
418
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00004.safetensors",
419
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00004.safetensors",
420
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00004.safetensors",
421
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00004.safetensors",
422
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00004.safetensors",
423
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00004.safetensors",
424
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00004.safetensors",
425
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
426
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
427
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
428
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
429
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
430
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
431
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
432
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
433
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00004.safetensors",
434
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00004.safetensors",
435
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00004.safetensors",
436
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00004.safetensors",
437
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00004.safetensors",
438
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00004.safetensors",
439
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00004.safetensors",
440
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00004.safetensors",
441
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
442
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
443
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
444
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
445
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
446
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
447
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
448
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
449
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00004.safetensors",
450
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00004.safetensors",
451
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00004.safetensors",
452
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00004.safetensors",
453
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00004.safetensors",
454
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00004.safetensors",
455
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00004.safetensors",
456
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00004.safetensors",
457
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
458
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
459
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
460
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
461
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
462
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
463
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
464
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
465
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00004.safetensors",
466
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00004.safetensors",
467
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00004.safetensors",
468
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00004.safetensors",
469
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00004.safetensors",
470
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00004.safetensors",
471
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00004.safetensors",
472
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00004.safetensors",
473
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
474
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
475
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
476
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
477
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
478
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
479
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
480
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
481
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00004.safetensors",
482
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00004.safetensors",
483
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00004.safetensors",
484
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00004.safetensors",
485
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00004.safetensors",
486
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00004.safetensors",
487
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00004.safetensors",
488
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00004.safetensors",
489
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
490
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
491
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
492
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
493
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
494
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
495
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
496
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
497
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00004.safetensors",
498
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00004.safetensors",
499
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00004.safetensors",
500
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00004.safetensors",
501
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00004.safetensors",
502
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00004.safetensors",
503
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00004.safetensors",
504
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00004.safetensors",
505
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
506
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
507
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
508
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
509
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
510
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
511
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
512
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
513
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00004.safetensors",
514
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00004.safetensors",
515
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00004.safetensors",
516
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00004.safetensors",
517
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00004.safetensors",
518
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00004.safetensors",
519
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00004.safetensors",
520
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00004.safetensors",
521
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
522
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
523
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
524
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
525
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
526
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
527
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
528
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
529
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00004.safetensors",
530
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00004.safetensors",
531
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00004.safetensors",
532
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00004.safetensors",
533
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00004.safetensors",
534
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00004.safetensors",
535
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00004.safetensors",
536
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00004.safetensors",
537
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
538
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
539
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
540
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
541
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
542
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
543
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
544
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
545
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00004.safetensors",
546
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00004.safetensors",
547
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00004.safetensors",
548
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00004.safetensors",
549
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00004.safetensors",
550
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00004.safetensors",
551
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00004.safetensors",
552
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00004.safetensors",
553
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
554
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
555
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
556
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
557
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
558
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
559
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
560
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
561
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00004.safetensors",
562
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00004.safetensors",
563
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00004.safetensors",
564
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00004.safetensors",
565
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00004.safetensors",
566
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00004.safetensors",
567
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00004.safetensors",
568
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00004.safetensors",
569
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
570
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
571
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
572
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
573
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
574
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
575
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
576
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
577
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00004.safetensors",
578
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00004.safetensors",
579
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00004.safetensors",
580
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00004.safetensors",
581
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00004.safetensors",
582
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00004.safetensors",
583
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00004.safetensors",
584
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00004.safetensors",
585
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
586
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
587
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
588
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
589
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
590
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
591
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
592
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
593
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00004.safetensors",
594
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00004.safetensors",
595
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00004.safetensors",
596
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00004.safetensors",
597
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00004.safetensors",
598
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00004.safetensors",
599
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00004.safetensors",
600
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00004.safetensors",
601
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
602
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
603
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
604
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
605
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
606
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
607
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
608
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
609
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00004.safetensors",
610
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00004.safetensors",
611
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00004.safetensors",
612
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00004.safetensors",
613
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00004.safetensors",
614
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00004.safetensors",
615
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00004.safetensors",
616
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00004.safetensors",
617
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
618
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
619
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
620
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
621
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
622
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
623
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
624
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
625
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00004.safetensors",
626
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00004.safetensors",
627
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00004.safetensors",
628
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00004.safetensors",
629
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00004.safetensors",
630
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00004.safetensors",
631
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00004.safetensors",
632
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00004.safetensors",
633
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
634
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
635
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
636
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
637
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
638
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
639
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
640
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
641
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00004.safetensors",
642
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00004.safetensors",
643
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00004.safetensors",
644
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00004.safetensors",
645
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00004.safetensors",
646
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00004.safetensors",
647
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00004.safetensors",
648
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00004.safetensors",
649
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
650
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
651
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
652
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
653
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
654
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
655
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
656
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
657
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00004.safetensors",
658
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00004.safetensors",
659
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00004.safetensors",
660
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00004.safetensors",
661
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00004.safetensors",
662
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00004.safetensors",
663
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00004.safetensors",
664
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00004.safetensors",
665
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
666
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
667
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
668
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
669
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
670
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
671
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
672
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
673
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00004.safetensors",
674
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00004.safetensors",
675
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00004.safetensors",
676
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00004.safetensors",
677
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00004.safetensors",
678
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00004.safetensors",
679
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00004.safetensors",
680
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00004.safetensors",
681
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
682
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
683
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
684
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
685
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
686
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
687
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
688
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
689
+ "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
690
+ "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors",
691
+ "vision_tower.vision_model.pre_layrnorm.bias": "model-00001-of-00004.safetensors",
692
+ "vision_tower.vision_model.pre_layrnorm.weight": "model-00001-of-00004.safetensors"
693
+ }
694
+ }
checkpoint-120/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
checkpoint-120/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-120/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-120/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
checkpoint-120/tokenizer_config.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ '<s>' }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '`[INST] `' + content + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 1000000000000000019884624838656,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "split_special_tokens": false,
67
+ "tokenizer_class": "LlamaTokenizer",
68
+ "unk_token": "<unk>",
69
+ "use_default_system_prompt": false
70
+ }
checkpoint-120/trainer_state.json ADDED
@@ -0,0 +1,441 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.8695652173913043,
5
+ "eval_steps": 40,
6
+ "global_step": 120,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.036231884057971016,
13
+ "grad_norm": 61.66789827681496,
14
+ "learning_rate": 5e-07,
15
+ "logits/chosen": -2.7321553230285645,
16
+ "logits/rejected": -2.7100937366485596,
17
+ "logps/chosen": -182.5845489501953,
18
+ "logps/rejected": -189.55001831054688,
19
+ "loss": 0.689,
20
+ "rewards/accuracies": 0.3187499940395355,
21
+ "rewards/chosen": -0.0021577859297394753,
22
+ "rewards/margins": 0.005646524019539356,
23
+ "rewards/rejected": -0.007804309483617544,
24
+ "step": 5
25
+ },
26
+ {
27
+ "epoch": 0.07246376811594203,
28
+ "grad_norm": 44.96703657621111,
29
+ "learning_rate": 1e-06,
30
+ "logits/chosen": -2.753889799118042,
31
+ "logits/rejected": -2.7519516944885254,
32
+ "logps/chosen": -197.34320068359375,
33
+ "logps/rejected": -184.00961303710938,
34
+ "loss": 0.6275,
35
+ "rewards/accuracies": 0.675000011920929,
36
+ "rewards/chosen": 0.030874451622366905,
37
+ "rewards/margins": 0.18904080986976624,
38
+ "rewards/rejected": -0.15816636383533478,
39
+ "step": 10
40
+ },
41
+ {
42
+ "epoch": 0.10869565217391304,
43
+ "grad_norm": 51.54402636298773,
44
+ "learning_rate": 9.996221126793764e-07,
45
+ "logits/chosen": -2.694437265396118,
46
+ "logits/rejected": -2.691904067993164,
47
+ "logps/chosen": -203.14883422851562,
48
+ "logps/rejected": -204.52386474609375,
49
+ "loss": 0.5851,
50
+ "rewards/accuracies": 0.75,
51
+ "rewards/chosen": 0.6205412745475769,
52
+ "rewards/margins": 0.9350436925888062,
53
+ "rewards/rejected": -0.314502477645874,
54
+ "step": 15
55
+ },
56
+ {
57
+ "epoch": 0.14492753623188406,
58
+ "grad_norm": 35.11282014809116,
59
+ "learning_rate": 9.984890219128145e-07,
60
+ "logits/chosen": -2.609405517578125,
61
+ "logits/rejected": -2.5795176029205322,
62
+ "logps/chosen": -188.33395385742188,
63
+ "logps/rejected": -192.52633666992188,
64
+ "loss": 0.5144,
65
+ "rewards/accuracies": 0.699999988079071,
66
+ "rewards/chosen": 0.8838651776313782,
67
+ "rewards/margins": 1.275496244430542,
68
+ "rewards/rejected": -0.39163118600845337,
69
+ "step": 20
70
+ },
71
+ {
72
+ "epoch": 0.18115942028985507,
73
+ "grad_norm": 37.05591291167376,
74
+ "learning_rate": 9.966024404228493e-07,
75
+ "logits/chosen": -2.4429798126220703,
76
+ "logits/rejected": -2.4225034713745117,
77
+ "logps/chosen": -179.79977416992188,
78
+ "logps/rejected": -179.25279235839844,
79
+ "loss": 0.5036,
80
+ "rewards/accuracies": 0.7562500238418579,
81
+ "rewards/chosen": 0.40532931685447693,
82
+ "rewards/margins": 0.8947975039482117,
83
+ "rewards/rejected": -0.4894680976867676,
84
+ "step": 25
85
+ },
86
+ {
87
+ "epoch": 0.21739130434782608,
88
+ "grad_norm": 33.27618771325657,
89
+ "learning_rate": 9.939652198703783e-07,
90
+ "logits/chosen": -2.319044589996338,
91
+ "logits/rejected": -2.320253372192383,
92
+ "logps/chosen": -188.62039184570312,
93
+ "logps/rejected": -193.8806915283203,
94
+ "loss": 0.4987,
95
+ "rewards/accuracies": 0.762499988079071,
96
+ "rewards/chosen": 0.6481167078018188,
97
+ "rewards/margins": 1.2183058261871338,
98
+ "rewards/rejected": -0.5701891183853149,
99
+ "step": 30
100
+ },
101
+ {
102
+ "epoch": 0.2536231884057971,
103
+ "grad_norm": 38.645073192304736,
104
+ "learning_rate": 9.905813465442354e-07,
105
+ "logits/chosen": -2.242053747177124,
106
+ "logits/rejected": -2.2155580520629883,
107
+ "logps/chosen": -203.94546508789062,
108
+ "logps/rejected": -194.8705291748047,
109
+ "loss": 0.5091,
110
+ "rewards/accuracies": 0.7250000238418579,
111
+ "rewards/chosen": 0.8871867060661316,
112
+ "rewards/margins": 1.2737131118774414,
113
+ "rewards/rejected": -0.3865264356136322,
114
+ "step": 35
115
+ },
116
+ {
117
+ "epoch": 0.2898550724637681,
118
+ "grad_norm": 30.937642063399068,
119
+ "learning_rate": 9.864559353357187e-07,
120
+ "logits/chosen": -2.0998620986938477,
121
+ "logits/rejected": -2.094836473464966,
122
+ "logps/chosen": -182.6898956298828,
123
+ "logps/rejected": -185.73983764648438,
124
+ "loss": 0.4889,
125
+ "rewards/accuracies": 0.7749999761581421,
126
+ "rewards/chosen": 1.0410809516906738,
127
+ "rewards/margins": 1.0222995281219482,
128
+ "rewards/rejected": 0.018781563267111778,
129
+ "step": 40
130
+ },
131
+ {
132
+ "epoch": 0.2898550724637681,
133
+ "eval_logits/chosen": -2.035081624984741,
134
+ "eval_logits/rejected": -1.9875677824020386,
135
+ "eval_logps/chosen": -188.77523803710938,
136
+ "eval_logps/rejected": -200.09498596191406,
137
+ "eval_loss": 0.4642000198364258,
138
+ "eval_rewards/accuracies": 0.7943548560142517,
139
+ "eval_rewards/chosen": 1.1544201374053955,
140
+ "eval_rewards/margins": 1.3431099653244019,
141
+ "eval_rewards/rejected": -0.18868987262248993,
142
+ "eval_runtime": 248.6586,
143
+ "eval_samples_per_second": 15.773,
144
+ "eval_steps_per_second": 0.249,
145
+ "step": 40
146
+ },
147
+ {
148
+ "epoch": 0.32608695652173914,
149
+ "grad_norm": 29.205255063362404,
150
+ "learning_rate": 9.815952220071804e-07,
151
+ "logits/chosen": -1.9733244180679321,
152
+ "logits/rejected": -1.9242098331451416,
153
+ "logps/chosen": -195.65652465820312,
154
+ "logps/rejected": -221.744384765625,
155
+ "loss": 0.4587,
156
+ "rewards/accuracies": 0.762499988079071,
157
+ "rewards/chosen": 1.3803393840789795,
158
+ "rewards/margins": 1.8610279560089111,
159
+ "rewards/rejected": -0.4806883931159973,
160
+ "step": 45
161
+ },
162
+ {
163
+ "epoch": 0.36231884057971014,
164
+ "grad_norm": 29.678687938529745,
165
+ "learning_rate": 9.76006553766365e-07,
166
+ "logits/chosen": -1.8085625171661377,
167
+ "logits/rejected": -1.772071123123169,
168
+ "logps/chosen": -198.9405517578125,
169
+ "logps/rejected": -203.29978942871094,
170
+ "loss": 0.4531,
171
+ "rewards/accuracies": 0.768750011920929,
172
+ "rewards/chosen": 0.8211376070976257,
173
+ "rewards/margins": 1.646414041519165,
174
+ "rewards/rejected": -0.8252763748168945,
175
+ "step": 50
176
+ },
177
+ {
178
+ "epoch": 0.39855072463768115,
179
+ "grad_norm": 30.73732601293307,
180
+ "learning_rate": 9.696983781607415e-07,
181
+ "logits/chosen": -1.7908868789672852,
182
+ "logits/rejected": -1.7599233388900757,
183
+ "logps/chosen": -183.01809692382812,
184
+ "logps/rejected": -172.02801513671875,
185
+ "loss": 0.4595,
186
+ "rewards/accuracies": 0.8374999761581421,
187
+ "rewards/chosen": 0.607209324836731,
188
+ "rewards/margins": 1.649515151977539,
189
+ "rewards/rejected": -1.042305827140808,
190
+ "step": 55
191
+ },
192
+ {
193
+ "epoch": 0.43478260869565216,
194
+ "grad_norm": 41.49881496846909,
195
+ "learning_rate": 9.626802303086209e-07,
196
+ "logits/chosen": -1.9050066471099854,
197
+ "logits/rejected": -1.8625679016113281,
198
+ "logps/chosen": -186.8179931640625,
199
+ "logps/rejected": -194.26707458496094,
200
+ "loss": 0.4276,
201
+ "rewards/accuracies": 0.8187500238418579,
202
+ "rewards/chosen": 0.4140642285346985,
203
+ "rewards/margins": 1.712418794631958,
204
+ "rewards/rejected": -1.2983543872833252,
205
+ "step": 60
206
+ },
207
+ {
208
+ "epoch": 0.47101449275362317,
209
+ "grad_norm": 36.480881476794956,
210
+ "learning_rate": 9.549627184863528e-07,
211
+ "logits/chosen": -2.1638290882110596,
212
+ "logits/rejected": -2.0793471336364746,
213
+ "logps/chosen": -191.9010467529297,
214
+ "logps/rejected": -193.20651245117188,
215
+ "loss": 0.4293,
216
+ "rewards/accuracies": 0.7875000238418579,
217
+ "rewards/chosen": 0.045776158571243286,
218
+ "rewards/margins": 1.6033704280853271,
219
+ "rewards/rejected": -1.5575940608978271,
220
+ "step": 65
221
+ },
222
+ {
223
+ "epoch": 0.5072463768115942,
224
+ "grad_norm": 26.318712537847357,
225
+ "learning_rate": 9.465575080933957e-07,
226
+ "logits/chosen": -2.1214632987976074,
227
+ "logits/rejected": -2.081162214279175,
228
+ "logps/chosen": -172.66419982910156,
229
+ "logps/rejected": -208.19149780273438,
230
+ "loss": 0.3963,
231
+ "rewards/accuracies": 0.831250011920929,
232
+ "rewards/chosen": 0.17696644365787506,
233
+ "rewards/margins": 1.8183233737945557,
234
+ "rewards/rejected": -1.6413570642471313,
235
+ "step": 70
236
+ },
237
+ {
238
+ "epoch": 0.5434782608695652,
239
+ "grad_norm": 32.35315227999789,
240
+ "learning_rate": 9.374773040194878e-07,
241
+ "logits/chosen": -2.1271562576293945,
242
+ "logits/rejected": -2.0680038928985596,
243
+ "logps/chosen": -205.5594482421875,
244
+ "logps/rejected": -210.7834014892578,
245
+ "loss": 0.4353,
246
+ "rewards/accuracies": 0.824999988079071,
247
+ "rewards/chosen": 0.16769471764564514,
248
+ "rewards/margins": 1.786273717880249,
249
+ "rewards/rejected": -1.6185792684555054,
250
+ "step": 75
251
+ },
252
+ {
253
+ "epoch": 0.5797101449275363,
254
+ "grad_norm": 29.96593011887677,
255
+ "learning_rate": 9.277358314405818e-07,
256
+ "logits/chosen": -2.046504497528076,
257
+ "logits/rejected": -2.0122628211975098,
258
+ "logps/chosen": -189.37771606445312,
259
+ "logps/rejected": -206.10659790039062,
260
+ "loss": 0.3941,
261
+ "rewards/accuracies": 0.8125,
262
+ "rewards/chosen": -0.2374907284975052,
263
+ "rewards/margins": 1.910300612449646,
264
+ "rewards/rejected": -2.1477913856506348,
265
+ "step": 80
266
+ },
267
+ {
268
+ "epoch": 0.5797101449275363,
269
+ "eval_logits/chosen": -1.9900643825531006,
270
+ "eval_logits/rejected": -1.9449083805084229,
271
+ "eval_logps/chosen": -202.59437561035156,
272
+ "eval_logps/rejected": -221.12725830078125,
273
+ "eval_loss": 0.4217630624771118,
274
+ "eval_rewards/accuracies": 0.8044354915618896,
275
+ "eval_rewards/chosen": -0.22749020159244537,
276
+ "eval_rewards/margins": 2.0644266605377197,
277
+ "eval_rewards/rejected": -2.291916847229004,
278
+ "eval_runtime": 247.3121,
279
+ "eval_samples_per_second": 15.859,
280
+ "eval_steps_per_second": 0.251,
281
+ "step": 80
282
+ },
283
+ {
284
+ "epoch": 0.6159420289855072,
285
+ "grad_norm": 34.207898226903815,
286
+ "learning_rate": 9.173478150725651e-07,
287
+ "logits/chosen": -2.0169568061828613,
288
+ "logits/rejected": -1.9391515254974365,
289
+ "logps/chosen": -209.73440551757812,
290
+ "logps/rejected": -215.14205932617188,
291
+ "loss": 0.422,
292
+ "rewards/accuracies": 0.831250011920929,
293
+ "rewards/chosen": 0.1130056157708168,
294
+ "rewards/margins": 2.3998470306396484,
295
+ "rewards/rejected": -2.2868411540985107,
296
+ "step": 85
297
+ },
298
+ {
299
+ "epoch": 0.6521739130434783,
300
+ "grad_norm": 28.850805260840193,
301
+ "learning_rate": 9.063289569141251e-07,
302
+ "logits/chosen": -2.1180360317230225,
303
+ "logits/rejected": -2.080235242843628,
304
+ "logps/chosen": -214.2292022705078,
305
+ "logps/rejected": -223.37564086914062,
306
+ "loss": 0.4135,
307
+ "rewards/accuracies": 0.84375,
308
+ "rewards/chosen": 0.5250645279884338,
309
+ "rewards/margins": 2.2779579162597656,
310
+ "rewards/rejected": -1.7528936862945557,
311
+ "step": 90
312
+ },
313
+ {
314
+ "epoch": 0.6884057971014492,
315
+ "grad_norm": 34.73111091101309,
316
+ "learning_rate": 8.946959125124051e-07,
317
+ "logits/chosen": -2.2047152519226074,
318
+ "logits/rejected": -2.1581788063049316,
319
+ "logps/chosen": -209.3234405517578,
320
+ "logps/rejected": -195.6946563720703,
321
+ "loss": 0.4077,
322
+ "rewards/accuracies": 0.824999988079071,
323
+ "rewards/chosen": 0.25443512201309204,
324
+ "rewards/margins": 2.2029635906219482,
325
+ "rewards/rejected": -1.948528528213501,
326
+ "step": 95
327
+ },
328
+ {
329
+ "epoch": 0.7246376811594203,
330
+ "grad_norm": 32.202226649445414,
331
+ "learning_rate": 8.824662657873238e-07,
332
+ "logits/chosen": -2.256134510040283,
333
+ "logits/rejected": -2.2395310401916504,
334
+ "logps/chosen": -176.19464111328125,
335
+ "logps/rejected": -209.0999298095703,
336
+ "loss": 0.3767,
337
+ "rewards/accuracies": 0.800000011920929,
338
+ "rewards/chosen": -0.3457742929458618,
339
+ "rewards/margins": 1.9383150339126587,
340
+ "rewards/rejected": -2.2840893268585205,
341
+ "step": 100
342
+ },
343
+ {
344
+ "epoch": 0.7608695652173914,
345
+ "grad_norm": 31.35693490170537,
346
+ "learning_rate": 8.696585024526135e-07,
347
+ "logits/chosen": -2.3358142375946045,
348
+ "logits/rejected": -2.307068347930908,
349
+ "logps/chosen": -191.8298797607422,
350
+ "logps/rejected": -227.55612182617188,
351
+ "loss": 0.4017,
352
+ "rewards/accuracies": 0.8125,
353
+ "rewards/chosen": -0.2825666069984436,
354
+ "rewards/margins": 2.5402400493621826,
355
+ "rewards/rejected": -2.8228065967559814,
356
+ "step": 105
357
+ },
358
+ {
359
+ "epoch": 0.7971014492753623,
360
+ "grad_norm": 36.29919961993296,
361
+ "learning_rate": 8.562919820737535e-07,
362
+ "logits/chosen": -2.2931602001190186,
363
+ "logits/rejected": -2.2568557262420654,
364
+ "logps/chosen": -207.34158325195312,
365
+ "logps/rejected": -208.9143524169922,
366
+ "loss": 0.3767,
367
+ "rewards/accuracies": 0.84375,
368
+ "rewards/chosen": -0.3017815351486206,
369
+ "rewards/margins": 2.3830084800720215,
370
+ "rewards/rejected": -2.6847901344299316,
371
+ "step": 110
372
+ },
373
+ {
374
+ "epoch": 0.8333333333333334,
375
+ "grad_norm": 28.819478551304137,
376
+ "learning_rate": 8.423869088050315e-07,
377
+ "logits/chosen": -2.2734763622283936,
378
+ "logits/rejected": -2.2522785663604736,
379
+ "logps/chosen": -195.6924591064453,
380
+ "logps/rejected": -221.85104370117188,
381
+ "loss": 0.3892,
382
+ "rewards/accuracies": 0.8125,
383
+ "rewards/chosen": -0.12773282825946808,
384
+ "rewards/margins": 2.3888843059539795,
385
+ "rewards/rejected": -2.5166170597076416,
386
+ "step": 115
387
+ },
388
+ {
389
+ "epoch": 0.8695652173913043,
390
+ "grad_norm": 37.085951928654964,
391
+ "learning_rate": 8.2796430084997e-07,
392
+ "logits/chosen": -2.2070839405059814,
393
+ "logits/rejected": -2.1736464500427246,
394
+ "logps/chosen": -196.70889282226562,
395
+ "logps/rejected": -207.93417358398438,
396
+ "loss": 0.3717,
397
+ "rewards/accuracies": 0.862500011920929,
398
+ "rewards/chosen": 0.12862932682037354,
399
+ "rewards/margins": 2.698848009109497,
400
+ "rewards/rejected": -2.570218563079834,
401
+ "step": 120
402
+ },
403
+ {
404
+ "epoch": 0.8695652173913043,
405
+ "eval_logits/chosen": -2.122933864593506,
406
+ "eval_logits/rejected": -2.0901710987091064,
407
+ "eval_logps/chosen": -202.4208221435547,
408
+ "eval_logps/rejected": -223.09356689453125,
409
+ "eval_loss": 0.438678115606308,
410
+ "eval_rewards/accuracies": 0.8286290168762207,
411
+ "eval_rewards/chosen": -0.2101391851902008,
412
+ "eval_rewards/margins": 2.278407573699951,
413
+ "eval_rewards/rejected": -2.488546848297119,
414
+ "eval_runtime": 247.3754,
415
+ "eval_samples_per_second": 15.854,
416
+ "eval_steps_per_second": 0.251,
417
+ "step": 120
418
+ }
419
+ ],
420
+ "logging_steps": 5,
421
+ "max_steps": 414,
422
+ "num_input_tokens_seen": 0,
423
+ "num_train_epochs": 3,
424
+ "save_steps": 40,
425
+ "stateful_callbacks": {
426
+ "TrainerControl": {
427
+ "args": {
428
+ "should_epoch_stop": false,
429
+ "should_evaluate": false,
430
+ "should_log": false,
431
+ "should_save": true,
432
+ "should_training_stop": false
433
+ },
434
+ "attributes": {}
435
+ }
436
+ },
437
+ "total_flos": 1414680891359232.0,
438
+ "train_batch_size": 8,
439
+ "trial_name": null,
440
+ "trial_params": null
441
+ }
checkpoint-120/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af79ff24bae38840a3c8efe3d28d0cc2a77ca640996f1dd8521f5747a2625682
3
+ size 7096
checkpoint-160/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
checkpoint-160/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/data/align-anything/hantao/models/llava-v1.6-mistral-7b-hf",
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "hidden_size": 4096,
7
+ "ignore_index": -100,
8
+ "image_grid_pinpoints": [
9
+ [
10
+ 336,
11
+ 672
12
+ ],
13
+ [
14
+ 672,
15
+ 336
16
+ ],
17
+ [
18
+ 672,
19
+ 672
20
+ ],
21
+ [
22
+ 1008,
23
+ 336
24
+ ],
25
+ [
26
+ 336,
27
+ 1008
28
+ ]
29
+ ],
30
+ "image_seq_length": 576,
31
+ "image_token_index": 32000,
32
+ "model_type": "llava_next",
33
+ "projector_hidden_act": "gelu",
34
+ "text_config": {
35
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
36
+ "architectures": [
37
+ "MistralForCausalLM"
38
+ ],
39
+ "intermediate_size": 14336,
40
+ "max_position_embeddings": 32768,
41
+ "model_type": "mistral",
42
+ "num_key_value_heads": 8,
43
+ "rms_norm_eps": 1e-05,
44
+ "rope_theta": 1000000.0,
45
+ "sliding_window": null,
46
+ "torch_dtype": "bfloat16",
47
+ "vocab_size": 32064
48
+ },
49
+ "tie_word_embeddings": false,
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.45.2",
52
+ "use_cache": false,
53
+ "use_image_newline_parameter": true,
54
+ "vision_config": {
55
+ "hidden_size": 1024,
56
+ "image_size": 336,
57
+ "intermediate_size": 4096,
58
+ "model_type": "clip_vision_model",
59
+ "num_attention_heads": 16,
60
+ "num_hidden_layers": 24,
61
+ "patch_size": 14,
62
+ "projection_dim": 768,
63
+ "vocab_size": 32000
64
+ },
65
+ "vision_feature_layer": -2,
66
+ "vision_feature_select_strategy": "default",
67
+ "vocab_size": 32064
68
+ }
checkpoint-160/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.45.2"
6
+ }
checkpoint-160/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e5da867f9d0130e643755f51786e2ee6be28ec716a4ec5d51bb5966d53d3119
3
+ size 4921618624
checkpoint-160/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c64f563545c1d629bc913dfbf65b32ead3f13fd8037b23a4c12f305a383ab1
3
+ size 4915917672
checkpoint-160/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83ddddabd722f7f44910af6d7a8e33882b6237d9f98759b20d8486434a8fb10f
3
+ size 4915917680
checkpoint-160/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1221e691075d89324fd24160979b901e8c5d17d09554a049568326b3370a5920
3
+ size 380134008
checkpoint-160/model.safetensors.index.json ADDED
@@ -0,0 +1,694 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15133495296
4
+ },
5
+ "weight_map": {
6
+ "image_newline": "model-00001-of-00004.safetensors",
7
+ "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
8
+ "language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
9
+ "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
10
+ "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
11
+ "language_model.model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
12
+ "language_model.model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
13
+ "language_model.model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
14
+ "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
+ "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
17
+ "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
18
+ "language_model.model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
19
+ "language_model.model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
20
+ "language_model.model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
21
+ "language_model.model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
22
+ "language_model.model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
23
+ "language_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
24
+ "language_model.model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
25
+ "language_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
26
+ "language_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
27
+ "language_model.model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
28
+ "language_model.model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
29
+ "language_model.model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
30
+ "language_model.model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
31
+ "language_model.model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
32
+ "language_model.model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
33
+ "language_model.model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
34
+ "language_model.model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
35
+ "language_model.model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
36
+ "language_model.model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "language_model.model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
38
+ "language_model.model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
39
+ "language_model.model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
40
+ "language_model.model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
41
+ "language_model.model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
42
+ "language_model.model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
43
+ "language_model.model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
44
+ "language_model.model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
45
+ "language_model.model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
46
+ "language_model.model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
47
+ "language_model.model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
48
+ "language_model.model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
49
+ "language_model.model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
50
+ "language_model.model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "language_model.model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
+ "language_model.model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
53
+ "language_model.model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
54
+ "language_model.model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
55
+ "language_model.model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
56
+ "language_model.model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
57
+ "language_model.model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
58
+ "language_model.model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
59
+ "language_model.model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
60
+ "language_model.model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
61
+ "language_model.model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
62
+ "language_model.model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
63
+ "language_model.model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
64
+ "language_model.model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
65
+ "language_model.model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
66
+ "language_model.model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
67
+ "language_model.model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
68
+ "language_model.model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
69
+ "language_model.model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
70
+ "language_model.model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
71
+ "language_model.model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
72
+ "language_model.model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
73
+ "language_model.model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
74
+ "language_model.model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
75
+ "language_model.model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
76
+ "language_model.model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
77
+ "language_model.model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
78
+ "language_model.model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
79
+ "language_model.model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
80
+ "language_model.model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
81
+ "language_model.model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
82
+ "language_model.model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
83
+ "language_model.model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
84
+ "language_model.model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
85
+ "language_model.model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
86
+ "language_model.model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
+ "language_model.model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "language_model.model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
89
+ "language_model.model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
90
+ "language_model.model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
91
+ "language_model.model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
92
+ "language_model.model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
93
+ "language_model.model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
94
+ "language_model.model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
95
+ "language_model.model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
96
+ "language_model.model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
97
+ "language_model.model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
98
+ "language_model.model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
99
+ "language_model.model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors",
100
+ "language_model.model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
101
+ "language_model.model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
102
+ "language_model.model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
103
+ "language_model.model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
104
+ "language_model.model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
105
+ "language_model.model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
106
+ "language_model.model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
107
+ "language_model.model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
108
+ "language_model.model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors",
109
+ "language_model.model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
110
+ "language_model.model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
111
+ "language_model.model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
112
+ "language_model.model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
113
+ "language_model.model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
114
+ "language_model.model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
115
+ "language_model.model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
116
+ "language_model.model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
117
+ "language_model.model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
118
+ "language_model.model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
119
+ "language_model.model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
120
+ "language_model.model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
121
+ "language_model.model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
122
+ "language_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
123
+ "language_model.model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
124
+ "language_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
125
+ "language_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
126
+ "language_model.model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
127
+ "language_model.model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
128
+ "language_model.model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
129
+ "language_model.model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
130
+ "language_model.model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
131
+ "language_model.model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
132
+ "language_model.model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
133
+ "language_model.model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
134
+ "language_model.model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
135
+ "language_model.model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
136
+ "language_model.model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
137
+ "language_model.model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
138
+ "language_model.model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
139
+ "language_model.model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
140
+ "language_model.model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
141
+ "language_model.model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
142
+ "language_model.model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
143
+ "language_model.model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
144
+ "language_model.model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
145
+ "language_model.model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
146
+ "language_model.model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
147
+ "language_model.model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
148
+ "language_model.model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
149
+ "language_model.model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
150
+ "language_model.model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
151
+ "language_model.model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
152
+ "language_model.model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
153
+ "language_model.model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
154
+ "language_model.model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
155
+ "language_model.model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
156
+ "language_model.model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
157
+ "language_model.model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
158
+ "language_model.model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
159
+ "language_model.model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
160
+ "language_model.model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
161
+ "language_model.model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
162
+ "language_model.model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
163
+ "language_model.model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
164
+ "language_model.model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
165
+ "language_model.model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
166
+ "language_model.model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
167
+ "language_model.model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
168
+ "language_model.model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
169
+ "language_model.model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
170
+ "language_model.model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
171
+ "language_model.model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
172
+ "language_model.model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
173
+ "language_model.model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
174
+ "language_model.model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
175
+ "language_model.model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
176
+ "language_model.model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
177
+ "language_model.model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
178
+ "language_model.model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
179
+ "language_model.model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
180
+ "language_model.model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "language_model.model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
182
+ "language_model.model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
183
+ "language_model.model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
184
+ "language_model.model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
185
+ "language_model.model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
186
+ "language_model.model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
187
+ "language_model.model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
188
+ "language_model.model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
189
+ "language_model.model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
190
+ "language_model.model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
191
+ "language_model.model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
192
+ "language_model.model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
193
+ "language_model.model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
194
+ "language_model.model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
+ "language_model.model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
+ "language_model.model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
197
+ "language_model.model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
198
+ "language_model.model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors",
199
+ "language_model.model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
200
+ "language_model.model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
201
+ "language_model.model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
202
+ "language_model.model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
203
+ "language_model.model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
204
+ "language_model.model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
205
+ "language_model.model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
206
+ "language_model.model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
207
+ "language_model.model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors",
208
+ "language_model.model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
209
+ "language_model.model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
210
+ "language_model.model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
211
+ "language_model.model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
212
+ "language_model.model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
213
+ "language_model.model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
214
+ "language_model.model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
215
+ "language_model.model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
216
+ "language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
217
+ "language_model.model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
218
+ "language_model.model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
219
+ "language_model.model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
220
+ "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
221
+ "language_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
222
+ "language_model.model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
223
+ "language_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
224
+ "language_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
225
+ "language_model.model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors",
226
+ "language_model.model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
227
+ "language_model.model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
228
+ "language_model.model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
229
+ "language_model.model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
230
+ "language_model.model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
+ "language_model.model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
+ "language_model.model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
233
+ "language_model.model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
234
+ "language_model.model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors",
235
+ "language_model.model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
236
+ "language_model.model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
237
+ "language_model.model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
238
+ "language_model.model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
239
+ "language_model.model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
240
+ "language_model.model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
241
+ "language_model.model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
242
+ "language_model.model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
243
+ "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
244
+ "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
245
+ "language_model.model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
246
+ "language_model.model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
247
+ "language_model.model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
248
+ "language_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
249
+ "language_model.model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
250
+ "language_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
251
+ "language_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
252
+ "language_model.model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
253
+ "language_model.model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
254
+ "language_model.model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
255
+ "language_model.model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
256
+ "language_model.model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
257
+ "language_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
258
+ "language_model.model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
259
+ "language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
260
+ "language_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
261
+ "language_model.model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
262
+ "language_model.model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
263
+ "language_model.model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
264
+ "language_model.model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
265
+ "language_model.model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
266
+ "language_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
+ "language_model.model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
+ "language_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
269
+ "language_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
270
+ "language_model.model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
271
+ "language_model.model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
272
+ "language_model.model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
273
+ "language_model.model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
274
+ "language_model.model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
275
+ "language_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
276
+ "language_model.model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
277
+ "language_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
278
+ "language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
279
+ "language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors",
280
+ "language_model.model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
281
+ "language_model.model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
282
+ "language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
283
+ "language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
284
+ "language_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
285
+ "language_model.model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
286
+ "language_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
287
+ "language_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
288
+ "language_model.model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
289
+ "language_model.model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
290
+ "language_model.model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
291
+ "language_model.model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
292
+ "language_model.model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
293
+ "language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
294
+ "language_model.model.layers.9.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
295
+ "language_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
296
+ "language_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
297
+ "language_model.model.norm.weight": "model-00004-of-00004.safetensors",
298
+ "multi_modal_projector.linear_1.bias": "model-00001-of-00004.safetensors",
299
+ "multi_modal_projector.linear_1.weight": "model-00001-of-00004.safetensors",
300
+ "multi_modal_projector.linear_2.bias": "model-00001-of-00004.safetensors",
301
+ "multi_modal_projector.linear_2.weight": "model-00001-of-00004.safetensors",
302
+ "vision_tower.vision_model.embeddings.class_embedding": "model-00001-of-00004.safetensors",
303
+ "vision_tower.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00004.safetensors",
304
+ "vision_tower.vision_model.embeddings.position_embedding.weight": "model-00001-of-00004.safetensors",
305
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00004.safetensors",
306
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00004.safetensors",
307
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00004.safetensors",
308
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00004.safetensors",
309
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00004.safetensors",
310
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00004.safetensors",
311
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00004.safetensors",
312
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00004.safetensors",
313
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
315
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
316
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
317
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
318
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
319
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
320
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
321
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00004.safetensors",
322
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00004.safetensors",
323
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00004.safetensors",
324
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00004.safetensors",
325
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00004.safetensors",
326
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00004.safetensors",
327
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00004.safetensors",
328
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00004.safetensors",
329
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
330
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
331
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
332
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
333
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
334
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
335
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
336
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
337
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00004.safetensors",
338
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00004.safetensors",
339
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00004.safetensors",
340
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00004.safetensors",
341
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00004.safetensors",
342
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00004.safetensors",
343
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00004.safetensors",
344
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00004.safetensors",
345
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
346
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
347
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
348
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
349
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
350
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
351
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
352
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
353
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00004.safetensors",
354
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00004.safetensors",
355
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00004.safetensors",
356
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00004.safetensors",
357
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00004.safetensors",
358
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00004.safetensors",
359
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00004.safetensors",
360
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00004.safetensors",
361
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
362
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
363
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
364
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
365
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
366
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
367
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
368
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
369
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00004.safetensors",
370
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00004.safetensors",
371
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00004.safetensors",
372
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00004.safetensors",
373
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00004.safetensors",
374
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00004.safetensors",
375
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00004.safetensors",
376
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00004.safetensors",
377
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
378
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
379
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
380
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
381
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
382
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
383
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
384
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
385
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00004.safetensors",
386
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00004.safetensors",
387
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00004.safetensors",
388
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00004.safetensors",
389
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00004.safetensors",
390
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00004.safetensors",
391
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00004.safetensors",
392
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00004.safetensors",
393
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
394
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
395
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
396
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
397
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
398
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
399
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
400
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
401
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00004.safetensors",
402
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00004.safetensors",
403
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00004.safetensors",
404
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00004.safetensors",
405
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00004.safetensors",
406
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00004.safetensors",
407
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00004.safetensors",
408
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00004.safetensors",
409
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
410
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
411
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
412
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
413
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
414
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
415
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
416
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
417
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00004.safetensors",
418
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00004.safetensors",
419
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00004.safetensors",
420
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00004.safetensors",
421
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00004.safetensors",
422
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00004.safetensors",
423
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00004.safetensors",
424
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00004.safetensors",
425
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
426
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
427
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
428
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
429
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
430
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
431
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
432
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
433
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00004.safetensors",
434
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00004.safetensors",
435
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00004.safetensors",
436
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00004.safetensors",
437
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00004.safetensors",
438
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00004.safetensors",
439
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00004.safetensors",
440
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00004.safetensors",
441
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
442
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
443
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
444
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
445
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
446
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
447
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
448
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
449
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00004.safetensors",
450
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00004.safetensors",
451
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00004.safetensors",
452
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00004.safetensors",
453
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00004.safetensors",
454
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00004.safetensors",
455
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00004.safetensors",
456
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00004.safetensors",
457
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
458
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
459
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
460
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
461
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
462
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
463
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
464
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
465
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00004.safetensors",
466
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00004.safetensors",
467
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00004.safetensors",
468
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00004.safetensors",
469
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00004.safetensors",
470
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00004.safetensors",
471
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00004.safetensors",
472
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00004.safetensors",
473
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
474
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
475
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
476
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
477
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
478
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
479
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
480
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
481
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00004.safetensors",
482
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00004.safetensors",
483
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00004.safetensors",
484
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00004.safetensors",
485
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00004.safetensors",
486
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00004.safetensors",
487
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00004.safetensors",
488
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00004.safetensors",
489
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
490
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
491
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
492
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
493
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
494
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
495
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
496
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
497
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00004.safetensors",
498
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00004.safetensors",
499
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00004.safetensors",
500
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00004.safetensors",
501
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00004.safetensors",
502
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00004.safetensors",
503
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00004.safetensors",
504
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00004.safetensors",
505
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
506
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
507
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
508
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
509
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
510
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
511
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
512
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
513
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00004.safetensors",
514
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00004.safetensors",
515
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00004.safetensors",
516
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00004.safetensors",
517
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00004.safetensors",
518
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00004.safetensors",
519
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00004.safetensors",
520
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00004.safetensors",
521
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
522
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
523
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
524
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
525
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
526
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
527
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
528
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
529
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00004.safetensors",
530
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00004.safetensors",
531
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00004.safetensors",
532
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00004.safetensors",
533
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00004.safetensors",
534
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00004.safetensors",
535
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00004.safetensors",
536
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00004.safetensors",
537
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
538
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
539
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
540
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
541
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
542
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
543
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
544
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
545
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00004.safetensors",
546
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00004.safetensors",
547
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00004.safetensors",
548
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00004.safetensors",
549
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00004.safetensors",
550
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00004.safetensors",
551
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00004.safetensors",
552
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00004.safetensors",
553
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
554
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
555
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
556
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
557
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
558
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
559
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
560
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
561
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00004.safetensors",
562
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00004.safetensors",
563
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00004.safetensors",
564
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00004.safetensors",
565
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00004.safetensors",
566
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00004.safetensors",
567
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00004.safetensors",
568
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00004.safetensors",
569
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
570
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
571
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
572
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
573
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
574
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
575
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
576
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
577
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00004.safetensors",
578
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00004.safetensors",
579
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00004.safetensors",
580
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00004.safetensors",
581
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00004.safetensors",
582
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00004.safetensors",
583
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00004.safetensors",
584
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00004.safetensors",
585
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
586
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
587
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
588
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
589
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
590
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
591
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
592
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
593
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00004.safetensors",
594
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00004.safetensors",
595
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00004.safetensors",
596
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00004.safetensors",
597
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00004.safetensors",
598
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00004.safetensors",
599
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00004.safetensors",
600
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00004.safetensors",
601
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
602
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
603
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
604
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
605
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
606
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
607
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
608
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
609
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00004.safetensors",
610
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00004.safetensors",
611
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00004.safetensors",
612
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00004.safetensors",
613
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00004.safetensors",
614
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00004.safetensors",
615
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00004.safetensors",
616
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00004.safetensors",
617
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
618
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
619
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
620
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
621
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
622
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
623
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
624
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
625
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00004.safetensors",
626
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00004.safetensors",
627
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00004.safetensors",
628
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00004.safetensors",
629
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00004.safetensors",
630
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00004.safetensors",
631
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00004.safetensors",
632
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00004.safetensors",
633
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
634
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
635
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
636
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
637
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
638
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
639
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
640
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
641
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00004.safetensors",
642
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00004.safetensors",
643
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00004.safetensors",
644
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00004.safetensors",
645
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00004.safetensors",
646
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00004.safetensors",
647
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00004.safetensors",
648
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00004.safetensors",
649
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
650
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
651
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
652
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
653
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
654
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
655
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
656
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
657
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00004.safetensors",
658
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00004.safetensors",
659
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00004.safetensors",
660
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00004.safetensors",
661
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00004.safetensors",
662
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00004.safetensors",
663
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00004.safetensors",
664
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00004.safetensors",
665
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
666
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
667
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
668
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
669
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
670
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
671
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
672
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
673
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00004.safetensors",
674
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00004.safetensors",
675
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00004.safetensors",
676
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00004.safetensors",
677
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00004.safetensors",
678
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00004.safetensors",
679
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00004.safetensors",
680
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00004.safetensors",
681
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
682
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
683
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
684
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
685
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
686
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
687
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
688
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
689
+ "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
690
+ "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors",
691
+ "vision_tower.vision_model.pre_layrnorm.bias": "model-00001-of-00004.safetensors",
692
+ "vision_tower.vision_model.pre_layrnorm.weight": "model-00001-of-00004.safetensors"
693
+ }
694
+ }
checkpoint-160/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
checkpoint-160/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-160/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-160/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
checkpoint-160/tokenizer_config.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ '<s>' }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '`[INST] `' + content + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 1000000000000000019884624838656,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "split_special_tokens": false,
67
+ "tokenizer_class": "LlamaTokenizer",
68
+ "unk_token": "<unk>",
69
+ "use_default_system_prompt": false
70
+ }
checkpoint-160/trainer_state.json ADDED
@@ -0,0 +1,577 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.1594202898550725,
5
+ "eval_steps": 40,
6
+ "global_step": 160,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.036231884057971016,
13
+ "grad_norm": 61.66789827681496,
14
+ "learning_rate": 5e-07,
15
+ "logits/chosen": -2.7321553230285645,
16
+ "logits/rejected": -2.7100937366485596,
17
+ "logps/chosen": -182.5845489501953,
18
+ "logps/rejected": -189.55001831054688,
19
+ "loss": 0.689,
20
+ "rewards/accuracies": 0.3187499940395355,
21
+ "rewards/chosen": -0.0021577859297394753,
22
+ "rewards/margins": 0.005646524019539356,
23
+ "rewards/rejected": -0.007804309483617544,
24
+ "step": 5
25
+ },
26
+ {
27
+ "epoch": 0.07246376811594203,
28
+ "grad_norm": 44.96703657621111,
29
+ "learning_rate": 1e-06,
30
+ "logits/chosen": -2.753889799118042,
31
+ "logits/rejected": -2.7519516944885254,
32
+ "logps/chosen": -197.34320068359375,
33
+ "logps/rejected": -184.00961303710938,
34
+ "loss": 0.6275,
35
+ "rewards/accuracies": 0.675000011920929,
36
+ "rewards/chosen": 0.030874451622366905,
37
+ "rewards/margins": 0.18904080986976624,
38
+ "rewards/rejected": -0.15816636383533478,
39
+ "step": 10
40
+ },
41
+ {
42
+ "epoch": 0.10869565217391304,
43
+ "grad_norm": 51.54402636298773,
44
+ "learning_rate": 9.996221126793764e-07,
45
+ "logits/chosen": -2.694437265396118,
46
+ "logits/rejected": -2.691904067993164,
47
+ "logps/chosen": -203.14883422851562,
48
+ "logps/rejected": -204.52386474609375,
49
+ "loss": 0.5851,
50
+ "rewards/accuracies": 0.75,
51
+ "rewards/chosen": 0.6205412745475769,
52
+ "rewards/margins": 0.9350436925888062,
53
+ "rewards/rejected": -0.314502477645874,
54
+ "step": 15
55
+ },
56
+ {
57
+ "epoch": 0.14492753623188406,
58
+ "grad_norm": 35.11282014809116,
59
+ "learning_rate": 9.984890219128145e-07,
60
+ "logits/chosen": -2.609405517578125,
61
+ "logits/rejected": -2.5795176029205322,
62
+ "logps/chosen": -188.33395385742188,
63
+ "logps/rejected": -192.52633666992188,
64
+ "loss": 0.5144,
65
+ "rewards/accuracies": 0.699999988079071,
66
+ "rewards/chosen": 0.8838651776313782,
67
+ "rewards/margins": 1.275496244430542,
68
+ "rewards/rejected": -0.39163118600845337,
69
+ "step": 20
70
+ },
71
+ {
72
+ "epoch": 0.18115942028985507,
73
+ "grad_norm": 37.05591291167376,
74
+ "learning_rate": 9.966024404228493e-07,
75
+ "logits/chosen": -2.4429798126220703,
76
+ "logits/rejected": -2.4225034713745117,
77
+ "logps/chosen": -179.79977416992188,
78
+ "logps/rejected": -179.25279235839844,
79
+ "loss": 0.5036,
80
+ "rewards/accuracies": 0.7562500238418579,
81
+ "rewards/chosen": 0.40532931685447693,
82
+ "rewards/margins": 0.8947975039482117,
83
+ "rewards/rejected": -0.4894680976867676,
84
+ "step": 25
85
+ },
86
+ {
87
+ "epoch": 0.21739130434782608,
88
+ "grad_norm": 33.27618771325657,
89
+ "learning_rate": 9.939652198703783e-07,
90
+ "logits/chosen": -2.319044589996338,
91
+ "logits/rejected": -2.320253372192383,
92
+ "logps/chosen": -188.62039184570312,
93
+ "logps/rejected": -193.8806915283203,
94
+ "loss": 0.4987,
95
+ "rewards/accuracies": 0.762499988079071,
96
+ "rewards/chosen": 0.6481167078018188,
97
+ "rewards/margins": 1.2183058261871338,
98
+ "rewards/rejected": -0.5701891183853149,
99
+ "step": 30
100
+ },
101
+ {
102
+ "epoch": 0.2536231884057971,
103
+ "grad_norm": 38.645073192304736,
104
+ "learning_rate": 9.905813465442354e-07,
105
+ "logits/chosen": -2.242053747177124,
106
+ "logits/rejected": -2.2155580520629883,
107
+ "logps/chosen": -203.94546508789062,
108
+ "logps/rejected": -194.8705291748047,
109
+ "loss": 0.5091,
110
+ "rewards/accuracies": 0.7250000238418579,
111
+ "rewards/chosen": 0.8871867060661316,
112
+ "rewards/margins": 1.2737131118774414,
113
+ "rewards/rejected": -0.3865264356136322,
114
+ "step": 35
115
+ },
116
+ {
117
+ "epoch": 0.2898550724637681,
118
+ "grad_norm": 30.937642063399068,
119
+ "learning_rate": 9.864559353357187e-07,
120
+ "logits/chosen": -2.0998620986938477,
121
+ "logits/rejected": -2.094836473464966,
122
+ "logps/chosen": -182.6898956298828,
123
+ "logps/rejected": -185.73983764648438,
124
+ "loss": 0.4889,
125
+ "rewards/accuracies": 0.7749999761581421,
126
+ "rewards/chosen": 1.0410809516906738,
127
+ "rewards/margins": 1.0222995281219482,
128
+ "rewards/rejected": 0.018781563267111778,
129
+ "step": 40
130
+ },
131
+ {
132
+ "epoch": 0.2898550724637681,
133
+ "eval_logits/chosen": -2.035081624984741,
134
+ "eval_logits/rejected": -1.9875677824020386,
135
+ "eval_logps/chosen": -188.77523803710938,
136
+ "eval_logps/rejected": -200.09498596191406,
137
+ "eval_loss": 0.4642000198364258,
138
+ "eval_rewards/accuracies": 0.7943548560142517,
139
+ "eval_rewards/chosen": 1.1544201374053955,
140
+ "eval_rewards/margins": 1.3431099653244019,
141
+ "eval_rewards/rejected": -0.18868987262248993,
142
+ "eval_runtime": 248.6586,
143
+ "eval_samples_per_second": 15.773,
144
+ "eval_steps_per_second": 0.249,
145
+ "step": 40
146
+ },
147
+ {
148
+ "epoch": 0.32608695652173914,
149
+ "grad_norm": 29.205255063362404,
150
+ "learning_rate": 9.815952220071804e-07,
151
+ "logits/chosen": -1.9733244180679321,
152
+ "logits/rejected": -1.9242098331451416,
153
+ "logps/chosen": -195.65652465820312,
154
+ "logps/rejected": -221.744384765625,
155
+ "loss": 0.4587,
156
+ "rewards/accuracies": 0.762499988079071,
157
+ "rewards/chosen": 1.3803393840789795,
158
+ "rewards/margins": 1.8610279560089111,
159
+ "rewards/rejected": -0.4806883931159973,
160
+ "step": 45
161
+ },
162
+ {
163
+ "epoch": 0.36231884057971014,
164
+ "grad_norm": 29.678687938529745,
165
+ "learning_rate": 9.76006553766365e-07,
166
+ "logits/chosen": -1.8085625171661377,
167
+ "logits/rejected": -1.772071123123169,
168
+ "logps/chosen": -198.9405517578125,
169
+ "logps/rejected": -203.29978942871094,
170
+ "loss": 0.4531,
171
+ "rewards/accuracies": 0.768750011920929,
172
+ "rewards/chosen": 0.8211376070976257,
173
+ "rewards/margins": 1.646414041519165,
174
+ "rewards/rejected": -0.8252763748168945,
175
+ "step": 50
176
+ },
177
+ {
178
+ "epoch": 0.39855072463768115,
179
+ "grad_norm": 30.73732601293307,
180
+ "learning_rate": 9.696983781607415e-07,
181
+ "logits/chosen": -1.7908868789672852,
182
+ "logits/rejected": -1.7599233388900757,
183
+ "logps/chosen": -183.01809692382812,
184
+ "logps/rejected": -172.02801513671875,
185
+ "loss": 0.4595,
186
+ "rewards/accuracies": 0.8374999761581421,
187
+ "rewards/chosen": 0.607209324836731,
188
+ "rewards/margins": 1.649515151977539,
189
+ "rewards/rejected": -1.042305827140808,
190
+ "step": 55
191
+ },
192
+ {
193
+ "epoch": 0.43478260869565216,
194
+ "grad_norm": 41.49881496846909,
195
+ "learning_rate": 9.626802303086209e-07,
196
+ "logits/chosen": -1.9050066471099854,
197
+ "logits/rejected": -1.8625679016113281,
198
+ "logps/chosen": -186.8179931640625,
199
+ "logps/rejected": -194.26707458496094,
200
+ "loss": 0.4276,
201
+ "rewards/accuracies": 0.8187500238418579,
202
+ "rewards/chosen": 0.4140642285346985,
203
+ "rewards/margins": 1.712418794631958,
204
+ "rewards/rejected": -1.2983543872833252,
205
+ "step": 60
206
+ },
207
+ {
208
+ "epoch": 0.47101449275362317,
209
+ "grad_norm": 36.480881476794956,
210
+ "learning_rate": 9.549627184863528e-07,
211
+ "logits/chosen": -2.1638290882110596,
212
+ "logits/rejected": -2.0793471336364746,
213
+ "logps/chosen": -191.9010467529297,
214
+ "logps/rejected": -193.20651245117188,
215
+ "loss": 0.4293,
216
+ "rewards/accuracies": 0.7875000238418579,
217
+ "rewards/chosen": 0.045776158571243286,
218
+ "rewards/margins": 1.6033704280853271,
219
+ "rewards/rejected": -1.5575940608978271,
220
+ "step": 65
221
+ },
222
+ {
223
+ "epoch": 0.5072463768115942,
224
+ "grad_norm": 26.318712537847357,
225
+ "learning_rate": 9.465575080933957e-07,
226
+ "logits/chosen": -2.1214632987976074,
227
+ "logits/rejected": -2.081162214279175,
228
+ "logps/chosen": -172.66419982910156,
229
+ "logps/rejected": -208.19149780273438,
230
+ "loss": 0.3963,
231
+ "rewards/accuracies": 0.831250011920929,
232
+ "rewards/chosen": 0.17696644365787506,
233
+ "rewards/margins": 1.8183233737945557,
234
+ "rewards/rejected": -1.6413570642471313,
235
+ "step": 70
236
+ },
237
+ {
238
+ "epoch": 0.5434782608695652,
239
+ "grad_norm": 32.35315227999789,
240
+ "learning_rate": 9.374773040194878e-07,
241
+ "logits/chosen": -2.1271562576293945,
242
+ "logits/rejected": -2.0680038928985596,
243
+ "logps/chosen": -205.5594482421875,
244
+ "logps/rejected": -210.7834014892578,
245
+ "loss": 0.4353,
246
+ "rewards/accuracies": 0.824999988079071,
247
+ "rewards/chosen": 0.16769471764564514,
248
+ "rewards/margins": 1.786273717880249,
249
+ "rewards/rejected": -1.6185792684555054,
250
+ "step": 75
251
+ },
252
+ {
253
+ "epoch": 0.5797101449275363,
254
+ "grad_norm": 29.96593011887677,
255
+ "learning_rate": 9.277358314405818e-07,
256
+ "logits/chosen": -2.046504497528076,
257
+ "logits/rejected": -2.0122628211975098,
258
+ "logps/chosen": -189.37771606445312,
259
+ "logps/rejected": -206.10659790039062,
260
+ "loss": 0.3941,
261
+ "rewards/accuracies": 0.8125,
262
+ "rewards/chosen": -0.2374907284975052,
263
+ "rewards/margins": 1.910300612449646,
264
+ "rewards/rejected": -2.1477913856506348,
265
+ "step": 80
266
+ },
267
+ {
268
+ "epoch": 0.5797101449275363,
269
+ "eval_logits/chosen": -1.9900643825531006,
270
+ "eval_logits/rejected": -1.9449083805084229,
271
+ "eval_logps/chosen": -202.59437561035156,
272
+ "eval_logps/rejected": -221.12725830078125,
273
+ "eval_loss": 0.4217630624771118,
274
+ "eval_rewards/accuracies": 0.8044354915618896,
275
+ "eval_rewards/chosen": -0.22749020159244537,
276
+ "eval_rewards/margins": 2.0644266605377197,
277
+ "eval_rewards/rejected": -2.291916847229004,
278
+ "eval_runtime": 247.3121,
279
+ "eval_samples_per_second": 15.859,
280
+ "eval_steps_per_second": 0.251,
281
+ "step": 80
282
+ },
283
+ {
284
+ "epoch": 0.6159420289855072,
285
+ "grad_norm": 34.207898226903815,
286
+ "learning_rate": 9.173478150725651e-07,
287
+ "logits/chosen": -2.0169568061828613,
288
+ "logits/rejected": -1.9391515254974365,
289
+ "logps/chosen": -209.73440551757812,
290
+ "logps/rejected": -215.14205932617188,
291
+ "loss": 0.422,
292
+ "rewards/accuracies": 0.831250011920929,
293
+ "rewards/chosen": 0.1130056157708168,
294
+ "rewards/margins": 2.3998470306396484,
295
+ "rewards/rejected": -2.2868411540985107,
296
+ "step": 85
297
+ },
298
+ {
299
+ "epoch": 0.6521739130434783,
300
+ "grad_norm": 28.850805260840193,
301
+ "learning_rate": 9.063289569141251e-07,
302
+ "logits/chosen": -2.1180360317230225,
303
+ "logits/rejected": -2.080235242843628,
304
+ "logps/chosen": -214.2292022705078,
305
+ "logps/rejected": -223.37564086914062,
306
+ "loss": 0.4135,
307
+ "rewards/accuracies": 0.84375,
308
+ "rewards/chosen": 0.5250645279884338,
309
+ "rewards/margins": 2.2779579162597656,
310
+ "rewards/rejected": -1.7528936862945557,
311
+ "step": 90
312
+ },
313
+ {
314
+ "epoch": 0.6884057971014492,
315
+ "grad_norm": 34.73111091101309,
316
+ "learning_rate": 8.946959125124051e-07,
317
+ "logits/chosen": -2.2047152519226074,
318
+ "logits/rejected": -2.1581788063049316,
319
+ "logps/chosen": -209.3234405517578,
320
+ "logps/rejected": -195.6946563720703,
321
+ "loss": 0.4077,
322
+ "rewards/accuracies": 0.824999988079071,
323
+ "rewards/chosen": 0.25443512201309204,
324
+ "rewards/margins": 2.2029635906219482,
325
+ "rewards/rejected": -1.948528528213501,
326
+ "step": 95
327
+ },
328
+ {
329
+ "epoch": 0.7246376811594203,
330
+ "grad_norm": 32.202226649445414,
331
+ "learning_rate": 8.824662657873238e-07,
332
+ "logits/chosen": -2.256134510040283,
333
+ "logits/rejected": -2.2395310401916504,
334
+ "logps/chosen": -176.19464111328125,
335
+ "logps/rejected": -209.0999298095703,
336
+ "loss": 0.3767,
337
+ "rewards/accuracies": 0.800000011920929,
338
+ "rewards/chosen": -0.3457742929458618,
339
+ "rewards/margins": 1.9383150339126587,
340
+ "rewards/rejected": -2.2840893268585205,
341
+ "step": 100
342
+ },
343
+ {
344
+ "epoch": 0.7608695652173914,
345
+ "grad_norm": 31.35693490170537,
346
+ "learning_rate": 8.696585024526135e-07,
347
+ "logits/chosen": -2.3358142375946045,
348
+ "logits/rejected": -2.307068347930908,
349
+ "logps/chosen": -191.8298797607422,
350
+ "logps/rejected": -227.55612182617188,
351
+ "loss": 0.4017,
352
+ "rewards/accuracies": 0.8125,
353
+ "rewards/chosen": -0.2825666069984436,
354
+ "rewards/margins": 2.5402400493621826,
355
+ "rewards/rejected": -2.8228065967559814,
356
+ "step": 105
357
+ },
358
+ {
359
+ "epoch": 0.7971014492753623,
360
+ "grad_norm": 36.29919961993296,
361
+ "learning_rate": 8.562919820737535e-07,
362
+ "logits/chosen": -2.2931602001190186,
363
+ "logits/rejected": -2.2568557262420654,
364
+ "logps/chosen": -207.34158325195312,
365
+ "logps/rejected": -208.9143524169922,
366
+ "loss": 0.3767,
367
+ "rewards/accuracies": 0.84375,
368
+ "rewards/chosen": -0.3017815351486206,
369
+ "rewards/margins": 2.3830084800720215,
370
+ "rewards/rejected": -2.6847901344299316,
371
+ "step": 110
372
+ },
373
+ {
374
+ "epoch": 0.8333333333333334,
375
+ "grad_norm": 28.819478551304137,
376
+ "learning_rate": 8.423869088050315e-07,
377
+ "logits/chosen": -2.2734763622283936,
378
+ "logits/rejected": -2.2522785663604736,
379
+ "logps/chosen": -195.6924591064453,
380
+ "logps/rejected": -221.85104370117188,
381
+ "loss": 0.3892,
382
+ "rewards/accuracies": 0.8125,
383
+ "rewards/chosen": -0.12773282825946808,
384
+ "rewards/margins": 2.3888843059539795,
385
+ "rewards/rejected": -2.5166170597076416,
386
+ "step": 115
387
+ },
388
+ {
389
+ "epoch": 0.8695652173913043,
390
+ "grad_norm": 37.085951928654964,
391
+ "learning_rate": 8.2796430084997e-07,
392
+ "logits/chosen": -2.2070839405059814,
393
+ "logits/rejected": -2.1736464500427246,
394
+ "logps/chosen": -196.70889282226562,
395
+ "logps/rejected": -207.93417358398438,
396
+ "loss": 0.3717,
397
+ "rewards/accuracies": 0.862500011920929,
398
+ "rewards/chosen": 0.12862932682037354,
399
+ "rewards/margins": 2.698848009109497,
400
+ "rewards/rejected": -2.570218563079834,
401
+ "step": 120
402
+ },
403
+ {
404
+ "epoch": 0.8695652173913043,
405
+ "eval_logits/chosen": -2.122933864593506,
406
+ "eval_logits/rejected": -2.0901710987091064,
407
+ "eval_logps/chosen": -202.4208221435547,
408
+ "eval_logps/rejected": -223.09356689453125,
409
+ "eval_loss": 0.438678115606308,
410
+ "eval_rewards/accuracies": 0.8286290168762207,
411
+ "eval_rewards/chosen": -0.2101391851902008,
412
+ "eval_rewards/margins": 2.278407573699951,
413
+ "eval_rewards/rejected": -2.488546848297119,
414
+ "eval_runtime": 247.3754,
415
+ "eval_samples_per_second": 15.854,
416
+ "eval_steps_per_second": 0.251,
417
+ "step": 120
418
+ },
419
+ {
420
+ "epoch": 0.9057971014492754,
421
+ "grad_norm": 27.843666767086823,
422
+ "learning_rate": 8.130459586912753e-07,
423
+ "logits/chosen": -2.0930094718933105,
424
+ "logits/rejected": -2.0791330337524414,
425
+ "logps/chosen": -220.4320526123047,
426
+ "logps/rejected": -218.3594512939453,
427
+ "loss": 0.4575,
428
+ "rewards/accuracies": 0.800000011920929,
429
+ "rewards/chosen": -0.9592973589897156,
430
+ "rewards/margins": 1.66092050075531,
431
+ "rewards/rejected": -2.620218276977539,
432
+ "step": 125
433
+ },
434
+ {
435
+ "epoch": 0.9420289855072463,
436
+ "grad_norm": 28.902451958691426,
437
+ "learning_rate": 7.97654432138333e-07,
438
+ "logits/chosen": -2.1393418312072754,
439
+ "logits/rejected": -2.103482723236084,
440
+ "logps/chosen": -214.5059051513672,
441
+ "logps/rejected": -248.8673858642578,
442
+ "loss": 0.388,
443
+ "rewards/accuracies": 0.856249988079071,
444
+ "rewards/chosen": -0.1911260336637497,
445
+ "rewards/margins": 2.8205361366271973,
446
+ "rewards/rejected": -3.011662006378174,
447
+ "step": 130
448
+ },
449
+ {
450
+ "epoch": 0.9782608695652174,
451
+ "grad_norm": 26.45174989804514,
452
+ "learning_rate": 7.81812986242061e-07,
453
+ "logits/chosen": -2.2018837928771973,
454
+ "logits/rejected": -2.1459240913391113,
455
+ "logps/chosen": -193.5723419189453,
456
+ "logps/rejected": -233.81838989257812,
457
+ "loss": 0.365,
458
+ "rewards/accuracies": 0.831250011920929,
459
+ "rewards/chosen": 0.10008885711431503,
460
+ "rewards/margins": 3.072911262512207,
461
+ "rewards/rejected": -2.972822427749634,
462
+ "step": 135
463
+ },
464
+ {
465
+ "epoch": 1.0144927536231885,
466
+ "grad_norm": 16.755101670108846,
467
+ "learning_rate": 7.655455661286375e-07,
468
+ "logits/chosen": -2.190566301345825,
469
+ "logits/rejected": -2.1852922439575195,
470
+ "logps/chosen": -190.56027221679688,
471
+ "logps/rejected": -234.6202392578125,
472
+ "loss": 0.2524,
473
+ "rewards/accuracies": 0.918749988079071,
474
+ "rewards/chosen": 0.05487387627363205,
475
+ "rewards/margins": 3.58606219291687,
476
+ "rewards/rejected": -3.531188488006592,
477
+ "step": 140
478
+ },
479
+ {
480
+ "epoch": 1.0507246376811594,
481
+ "grad_norm": 16.022443175608785,
482
+ "learning_rate": 7.488767608052628e-07,
483
+ "logits/chosen": -2.2735958099365234,
484
+ "logits/rejected": -2.199552059173584,
485
+ "logps/chosen": -190.21511840820312,
486
+ "logps/rejected": -237.59854125976562,
487
+ "loss": 0.171,
488
+ "rewards/accuracies": 0.9375,
489
+ "rewards/chosen": 0.7184330821037292,
490
+ "rewards/margins": 4.115548133850098,
491
+ "rewards/rejected": -3.3971149921417236,
492
+ "step": 145
493
+ },
494
+ {
495
+ "epoch": 1.0869565217391304,
496
+ "grad_norm": 16.37925141930124,
497
+ "learning_rate": 7.318317659926636e-07,
498
+ "logits/chosen": -2.250054121017456,
499
+ "logits/rejected": -2.221043109893799,
500
+ "logps/chosen": -174.0218048095703,
501
+ "logps/rejected": -237.14120483398438,
502
+ "loss": 0.1436,
503
+ "rewards/accuracies": 0.9375,
504
+ "rewards/chosen": 1.0139648914337158,
505
+ "rewards/margins": 4.169236660003662,
506
+ "rewards/rejected": -3.1552722454071045,
507
+ "step": 150
508
+ },
509
+ {
510
+ "epoch": 1.1231884057971016,
511
+ "grad_norm": 13.693805739113818,
512
+ "learning_rate": 7.144363460405189e-07,
513
+ "logits/chosen": -2.3314082622528076,
514
+ "logits/rejected": -2.2712883949279785,
515
+ "logps/chosen": -192.2017059326172,
516
+ "logps/rejected": -236.91586303710938,
517
+ "loss": 0.1415,
518
+ "rewards/accuracies": 0.9624999761581421,
519
+ "rewards/chosen": 1.1952037811279297,
520
+ "rewards/margins": 4.705626487731934,
521
+ "rewards/rejected": -3.510422945022583,
522
+ "step": 155
523
+ },
524
+ {
525
+ "epoch": 1.1594202898550725,
526
+ "grad_norm": 13.203700040695288,
527
+ "learning_rate": 6.967167949833762e-07,
528
+ "logits/chosen": -2.324735164642334,
529
+ "logits/rejected": -2.283231258392334,
530
+ "logps/chosen": -194.0807342529297,
531
+ "logps/rejected": -247.2218475341797,
532
+ "loss": 0.1459,
533
+ "rewards/accuracies": 0.949999988079071,
534
+ "rewards/chosen": 0.3761358857154846,
535
+ "rewards/margins": 4.9711785316467285,
536
+ "rewards/rejected": -4.595042705535889,
537
+ "step": 160
538
+ },
539
+ {
540
+ "epoch": 1.1594202898550725,
541
+ "eval_logits/chosen": -2.3007450103759766,
542
+ "eval_logits/rejected": -2.273284673690796,
543
+ "eval_logps/chosen": -204.34878540039062,
544
+ "eval_logps/rejected": -232.13629150390625,
545
+ "eval_loss": 0.42876046895980835,
546
+ "eval_rewards/accuracies": 0.8286290168762207,
547
+ "eval_rewards/chosen": -0.40293240547180176,
548
+ "eval_rewards/margins": 2.9898877143859863,
549
+ "eval_rewards/rejected": -3.392819881439209,
550
+ "eval_runtime": 247.3207,
551
+ "eval_samples_per_second": 15.858,
552
+ "eval_steps_per_second": 0.251,
553
+ "step": 160
554
+ }
555
+ ],
556
+ "logging_steps": 5,
557
+ "max_steps": 414,
558
+ "num_input_tokens_seen": 0,
559
+ "num_train_epochs": 3,
560
+ "save_steps": 40,
561
+ "stateful_callbacks": {
562
+ "TrainerControl": {
563
+ "args": {
564
+ "should_epoch_stop": false,
565
+ "should_evaluate": false,
566
+ "should_log": false,
567
+ "should_save": true,
568
+ "should_training_stop": false
569
+ },
570
+ "attributes": {}
571
+ }
572
+ },
573
+ "total_flos": 1886397149478912.0,
574
+ "train_batch_size": 8,
575
+ "trial_name": null,
576
+ "trial_params": null
577
+ }
checkpoint-160/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af79ff24bae38840a3c8efe3d28d0cc2a77ca640996f1dd8521f5747a2625682
3
+ size 7096
checkpoint-200/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
checkpoint-200/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/data/align-anything/hantao/models/llava-v1.6-mistral-7b-hf",
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "hidden_size": 4096,
7
+ "ignore_index": -100,
8
+ "image_grid_pinpoints": [
9
+ [
10
+ 336,
11
+ 672
12
+ ],
13
+ [
14
+ 672,
15
+ 336
16
+ ],
17
+ [
18
+ 672,
19
+ 672
20
+ ],
21
+ [
22
+ 1008,
23
+ 336
24
+ ],
25
+ [
26
+ 336,
27
+ 1008
28
+ ]
29
+ ],
30
+ "image_seq_length": 576,
31
+ "image_token_index": 32000,
32
+ "model_type": "llava_next",
33
+ "projector_hidden_act": "gelu",
34
+ "text_config": {
35
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
36
+ "architectures": [
37
+ "MistralForCausalLM"
38
+ ],
39
+ "intermediate_size": 14336,
40
+ "max_position_embeddings": 32768,
41
+ "model_type": "mistral",
42
+ "num_key_value_heads": 8,
43
+ "rms_norm_eps": 1e-05,
44
+ "rope_theta": 1000000.0,
45
+ "sliding_window": null,
46
+ "torch_dtype": "bfloat16",
47
+ "vocab_size": 32064
48
+ },
49
+ "tie_word_embeddings": false,
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.45.2",
52
+ "use_cache": false,
53
+ "use_image_newline_parameter": true,
54
+ "vision_config": {
55
+ "hidden_size": 1024,
56
+ "image_size": 336,
57
+ "intermediate_size": 4096,
58
+ "model_type": "clip_vision_model",
59
+ "num_attention_heads": 16,
60
+ "num_hidden_layers": 24,
61
+ "patch_size": 14,
62
+ "projection_dim": 768,
63
+ "vocab_size": 32000
64
+ },
65
+ "vision_feature_layer": -2,
66
+ "vision_feature_select_strategy": "default",
67
+ "vocab_size": 32064
68
+ }
checkpoint-200/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.45.2"
6
+ }
checkpoint-200/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:537a3da88cf529a9efeb07b422fd7bdac0d807d05846dc56c25b76d92af41cbf
3
+ size 4921618624
checkpoint-200/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d006fef53cce2f2e43034cb2db577d07562d64646856b232f7e753dc0814fd4d
3
+ size 4915917672
checkpoint-200/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04c639af878f3974616a77f6a988beb204d98f25d90fb2b3b7c0bf37f9bc1980
3
+ size 4915917680
checkpoint-200/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e3b51d06805ebb8ccae561ddb4f54919d8269efc2a5a65af5addedb31e5778b
3
+ size 380134008
checkpoint-200/model.safetensors.index.json ADDED
@@ -0,0 +1,694 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15133495296
4
+ },
5
+ "weight_map": {
6
+ "image_newline": "model-00001-of-00004.safetensors",
7
+ "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
8
+ "language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
9
+ "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
10
+ "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
11
+ "language_model.model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
12
+ "language_model.model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
13
+ "language_model.model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
14
+ "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
+ "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
17
+ "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
18
+ "language_model.model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
19
+ "language_model.model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
20
+ "language_model.model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
21
+ "language_model.model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
22
+ "language_model.model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
23
+ "language_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
24
+ "language_model.model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
25
+ "language_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
26
+ "language_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
27
+ "language_model.model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
28
+ "language_model.model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
29
+ "language_model.model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
30
+ "language_model.model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
31
+ "language_model.model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
32
+ "language_model.model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
33
+ "language_model.model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
34
+ "language_model.model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
35
+ "language_model.model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
36
+ "language_model.model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "language_model.model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
38
+ "language_model.model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
39
+ "language_model.model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
40
+ "language_model.model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
41
+ "language_model.model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
42
+ "language_model.model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
43
+ "language_model.model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
44
+ "language_model.model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
45
+ "language_model.model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
46
+ "language_model.model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
47
+ "language_model.model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
48
+ "language_model.model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
49
+ "language_model.model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
50
+ "language_model.model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "language_model.model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
+ "language_model.model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
53
+ "language_model.model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
54
+ "language_model.model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
55
+ "language_model.model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
56
+ "language_model.model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
57
+ "language_model.model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
58
+ "language_model.model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
59
+ "language_model.model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
60
+ "language_model.model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
61
+ "language_model.model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
62
+ "language_model.model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
63
+ "language_model.model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
64
+ "language_model.model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
65
+ "language_model.model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
66
+ "language_model.model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
67
+ "language_model.model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
68
+ "language_model.model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
69
+ "language_model.model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
70
+ "language_model.model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
71
+ "language_model.model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
72
+ "language_model.model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
73
+ "language_model.model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
74
+ "language_model.model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
75
+ "language_model.model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
76
+ "language_model.model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
77
+ "language_model.model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
78
+ "language_model.model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
79
+ "language_model.model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
80
+ "language_model.model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
81
+ "language_model.model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
82
+ "language_model.model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
83
+ "language_model.model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
84
+ "language_model.model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
85
+ "language_model.model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
86
+ "language_model.model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
+ "language_model.model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "language_model.model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
89
+ "language_model.model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
90
+ "language_model.model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
91
+ "language_model.model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
92
+ "language_model.model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
93
+ "language_model.model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
94
+ "language_model.model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
95
+ "language_model.model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
96
+ "language_model.model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
97
+ "language_model.model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
98
+ "language_model.model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
99
+ "language_model.model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors",
100
+ "language_model.model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
101
+ "language_model.model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
102
+ "language_model.model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
103
+ "language_model.model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
104
+ "language_model.model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
105
+ "language_model.model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
106
+ "language_model.model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
107
+ "language_model.model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
108
+ "language_model.model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors",
109
+ "language_model.model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
110
+ "language_model.model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
111
+ "language_model.model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
112
+ "language_model.model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
113
+ "language_model.model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
114
+ "language_model.model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
115
+ "language_model.model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
116
+ "language_model.model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
117
+ "language_model.model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
118
+ "language_model.model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
119
+ "language_model.model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
120
+ "language_model.model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
121
+ "language_model.model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
122
+ "language_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
123
+ "language_model.model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
124
+ "language_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
125
+ "language_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
126
+ "language_model.model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
127
+ "language_model.model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
128
+ "language_model.model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
129
+ "language_model.model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
130
+ "language_model.model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
131
+ "language_model.model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
132
+ "language_model.model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
133
+ "language_model.model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
134
+ "language_model.model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
135
+ "language_model.model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
136
+ "language_model.model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
137
+ "language_model.model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
138
+ "language_model.model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
139
+ "language_model.model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
140
+ "language_model.model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
141
+ "language_model.model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
142
+ "language_model.model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
143
+ "language_model.model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
144
+ "language_model.model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
145
+ "language_model.model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
146
+ "language_model.model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
147
+ "language_model.model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
148
+ "language_model.model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
149
+ "language_model.model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
150
+ "language_model.model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
151
+ "language_model.model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
152
+ "language_model.model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
153
+ "language_model.model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
154
+ "language_model.model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
155
+ "language_model.model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
156
+ "language_model.model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
157
+ "language_model.model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
158
+ "language_model.model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
159
+ "language_model.model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
160
+ "language_model.model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
161
+ "language_model.model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
162
+ "language_model.model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
163
+ "language_model.model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
164
+ "language_model.model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
165
+ "language_model.model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
166
+ "language_model.model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
167
+ "language_model.model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
168
+ "language_model.model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
169
+ "language_model.model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
170
+ "language_model.model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
171
+ "language_model.model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
172
+ "language_model.model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
173
+ "language_model.model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
174
+ "language_model.model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
175
+ "language_model.model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
176
+ "language_model.model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
177
+ "language_model.model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
178
+ "language_model.model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
179
+ "language_model.model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
180
+ "language_model.model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "language_model.model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
182
+ "language_model.model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
183
+ "language_model.model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
184
+ "language_model.model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
185
+ "language_model.model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
186
+ "language_model.model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
187
+ "language_model.model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
188
+ "language_model.model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
189
+ "language_model.model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
190
+ "language_model.model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
191
+ "language_model.model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
192
+ "language_model.model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
193
+ "language_model.model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
194
+ "language_model.model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
+ "language_model.model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
+ "language_model.model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
197
+ "language_model.model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
198
+ "language_model.model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors",
199
+ "language_model.model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
200
+ "language_model.model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
201
+ "language_model.model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
202
+ "language_model.model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
203
+ "language_model.model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
204
+ "language_model.model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
205
+ "language_model.model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
206
+ "language_model.model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
207
+ "language_model.model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors",
208
+ "language_model.model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
209
+ "language_model.model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
210
+ "language_model.model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
211
+ "language_model.model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
212
+ "language_model.model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
213
+ "language_model.model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
214
+ "language_model.model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
215
+ "language_model.model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
216
+ "language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
217
+ "language_model.model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
218
+ "language_model.model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
219
+ "language_model.model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
220
+ "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
221
+ "language_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
222
+ "language_model.model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
223
+ "language_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
224
+ "language_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
225
+ "language_model.model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors",
226
+ "language_model.model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
227
+ "language_model.model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
228
+ "language_model.model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
229
+ "language_model.model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
230
+ "language_model.model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
+ "language_model.model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
+ "language_model.model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
233
+ "language_model.model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
234
+ "language_model.model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors",
235
+ "language_model.model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
236
+ "language_model.model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
237
+ "language_model.model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
238
+ "language_model.model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
239
+ "language_model.model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
240
+ "language_model.model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
241
+ "language_model.model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
242
+ "language_model.model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
243
+ "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
244
+ "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
245
+ "language_model.model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
246
+ "language_model.model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
247
+ "language_model.model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
248
+ "language_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
249
+ "language_model.model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
250
+ "language_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
251
+ "language_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
252
+ "language_model.model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
253
+ "language_model.model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
254
+ "language_model.model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
255
+ "language_model.model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
256
+ "language_model.model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
257
+ "language_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
258
+ "language_model.model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
259
+ "language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
260
+ "language_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
261
+ "language_model.model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
262
+ "language_model.model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
263
+ "language_model.model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
264
+ "language_model.model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
265
+ "language_model.model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
266
+ "language_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
+ "language_model.model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
+ "language_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
269
+ "language_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
270
+ "language_model.model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
271
+ "language_model.model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
272
+ "language_model.model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
273
+ "language_model.model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
274
+ "language_model.model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
275
+ "language_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
276
+ "language_model.model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
277
+ "language_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
278
+ "language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
279
+ "language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors",
280
+ "language_model.model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
281
+ "language_model.model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
282
+ "language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
283
+ "language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
284
+ "language_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
285
+ "language_model.model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
286
+ "language_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
287
+ "language_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
288
+ "language_model.model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
289
+ "language_model.model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
290
+ "language_model.model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
291
+ "language_model.model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
292
+ "language_model.model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
293
+ "language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
294
+ "language_model.model.layers.9.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
295
+ "language_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
296
+ "language_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
297
+ "language_model.model.norm.weight": "model-00004-of-00004.safetensors",
298
+ "multi_modal_projector.linear_1.bias": "model-00001-of-00004.safetensors",
299
+ "multi_modal_projector.linear_1.weight": "model-00001-of-00004.safetensors",
300
+ "multi_modal_projector.linear_2.bias": "model-00001-of-00004.safetensors",
301
+ "multi_modal_projector.linear_2.weight": "model-00001-of-00004.safetensors",
302
+ "vision_tower.vision_model.embeddings.class_embedding": "model-00001-of-00004.safetensors",
303
+ "vision_tower.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00004.safetensors",
304
+ "vision_tower.vision_model.embeddings.position_embedding.weight": "model-00001-of-00004.safetensors",
305
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00004.safetensors",
306
+ "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00004.safetensors",
307
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00004.safetensors",
308
+ "vision_tower.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00004.safetensors",
309
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00004.safetensors",
310
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00004.safetensors",
311
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00004.safetensors",
312
+ "vision_tower.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00004.safetensors",
313
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
315
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
316
+ "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
317
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
318
+ "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
319
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
320
+ "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
321
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00004.safetensors",
322
+ "vision_tower.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00004.safetensors",
323
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00004.safetensors",
324
+ "vision_tower.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00004.safetensors",
325
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00004.safetensors",
326
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00004.safetensors",
327
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00004.safetensors",
328
+ "vision_tower.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00004.safetensors",
329
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
330
+ "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
331
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
332
+ "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
333
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
334
+ "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
335
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
336
+ "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
337
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00004.safetensors",
338
+ "vision_tower.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00004.safetensors",
339
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00004.safetensors",
340
+ "vision_tower.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00004.safetensors",
341
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00004.safetensors",
342
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00004.safetensors",
343
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00004.safetensors",
344
+ "vision_tower.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00004.safetensors",
345
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
346
+ "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
347
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
348
+ "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
349
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
350
+ "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
351
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
352
+ "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
353
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00004.safetensors",
354
+ "vision_tower.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00004.safetensors",
355
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00004.safetensors",
356
+ "vision_tower.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00004.safetensors",
357
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00004.safetensors",
358
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00004.safetensors",
359
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00004.safetensors",
360
+ "vision_tower.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00004.safetensors",
361
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
362
+ "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
363
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
364
+ "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
365
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
366
+ "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
367
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
368
+ "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
369
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00004.safetensors",
370
+ "vision_tower.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00004.safetensors",
371
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00004.safetensors",
372
+ "vision_tower.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00004.safetensors",
373
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00004.safetensors",
374
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00004.safetensors",
375
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00004.safetensors",
376
+ "vision_tower.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00004.safetensors",
377
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
378
+ "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
379
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
380
+ "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
381
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
382
+ "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
383
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
384
+ "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
385
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00004.safetensors",
386
+ "vision_tower.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00004.safetensors",
387
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00004.safetensors",
388
+ "vision_tower.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00004.safetensors",
389
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00004.safetensors",
390
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00004.safetensors",
391
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00004.safetensors",
392
+ "vision_tower.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00004.safetensors",
393
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
394
+ "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
395
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
396
+ "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
397
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
398
+ "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
399
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
400
+ "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
401
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00004.safetensors",
402
+ "vision_tower.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00004.safetensors",
403
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00004.safetensors",
404
+ "vision_tower.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00004.safetensors",
405
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00004.safetensors",
406
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00004.safetensors",
407
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00004.safetensors",
408
+ "vision_tower.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00004.safetensors",
409
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
410
+ "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
411
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
412
+ "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
413
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
414
+ "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
415
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
416
+ "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
417
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00004.safetensors",
418
+ "vision_tower.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00004.safetensors",
419
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00004.safetensors",
420
+ "vision_tower.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00004.safetensors",
421
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00004.safetensors",
422
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00004.safetensors",
423
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00004.safetensors",
424
+ "vision_tower.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00004.safetensors",
425
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
426
+ "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
427
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
428
+ "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
429
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
430
+ "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
431
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
432
+ "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
433
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00004.safetensors",
434
+ "vision_tower.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00004.safetensors",
435
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00004.safetensors",
436
+ "vision_tower.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00004.safetensors",
437
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00004.safetensors",
438
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00004.safetensors",
439
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00004.safetensors",
440
+ "vision_tower.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00004.safetensors",
441
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
442
+ "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
443
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
444
+ "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
445
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
446
+ "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
447
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
448
+ "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
449
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00004.safetensors",
450
+ "vision_tower.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00004.safetensors",
451
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00004.safetensors",
452
+ "vision_tower.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00004.safetensors",
453
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00004.safetensors",
454
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00004.safetensors",
455
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00004.safetensors",
456
+ "vision_tower.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00004.safetensors",
457
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
458
+ "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
459
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
460
+ "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
461
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
462
+ "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
463
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
464
+ "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
465
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00004.safetensors",
466
+ "vision_tower.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00004.safetensors",
467
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00004.safetensors",
468
+ "vision_tower.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00004.safetensors",
469
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00004.safetensors",
470
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00004.safetensors",
471
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00004.safetensors",
472
+ "vision_tower.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00004.safetensors",
473
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
474
+ "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
475
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
476
+ "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
477
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
478
+ "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
479
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
480
+ "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
481
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00004.safetensors",
482
+ "vision_tower.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00004.safetensors",
483
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00004.safetensors",
484
+ "vision_tower.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00004.safetensors",
485
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00004.safetensors",
486
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00004.safetensors",
487
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00004.safetensors",
488
+ "vision_tower.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00004.safetensors",
489
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
490
+ "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
491
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
492
+ "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
493
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
494
+ "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
495
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
496
+ "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
497
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00004.safetensors",
498
+ "vision_tower.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00004.safetensors",
499
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00004.safetensors",
500
+ "vision_tower.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00004.safetensors",
501
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00004.safetensors",
502
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00004.safetensors",
503
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00004.safetensors",
504
+ "vision_tower.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00004.safetensors",
505
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
506
+ "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
507
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
508
+ "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
509
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
510
+ "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
511
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
512
+ "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
513
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00004.safetensors",
514
+ "vision_tower.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00004.safetensors",
515
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00004.safetensors",
516
+ "vision_tower.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00004.safetensors",
517
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00004.safetensors",
518
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00004.safetensors",
519
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00004.safetensors",
520
+ "vision_tower.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00004.safetensors",
521
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
522
+ "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
523
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
524
+ "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
525
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
526
+ "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
527
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
528
+ "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
529
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00004.safetensors",
530
+ "vision_tower.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00004.safetensors",
531
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00004.safetensors",
532
+ "vision_tower.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00004.safetensors",
533
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00004.safetensors",
534
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00004.safetensors",
535
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00004.safetensors",
536
+ "vision_tower.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00004.safetensors",
537
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
538
+ "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
539
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
540
+ "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
541
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
542
+ "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
543
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
544
+ "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
545
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00004.safetensors",
546
+ "vision_tower.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00004.safetensors",
547
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00004.safetensors",
548
+ "vision_tower.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00004.safetensors",
549
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00004.safetensors",
550
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00004.safetensors",
551
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00004.safetensors",
552
+ "vision_tower.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00004.safetensors",
553
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
554
+ "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
555
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
556
+ "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
557
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
558
+ "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
559
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
560
+ "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
561
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00004.safetensors",
562
+ "vision_tower.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00004.safetensors",
563
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00004.safetensors",
564
+ "vision_tower.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00004.safetensors",
565
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00004.safetensors",
566
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00004.safetensors",
567
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00004.safetensors",
568
+ "vision_tower.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00004.safetensors",
569
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
570
+ "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
571
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
572
+ "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
573
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
574
+ "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
575
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
576
+ "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
577
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00004.safetensors",
578
+ "vision_tower.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00004.safetensors",
579
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00004.safetensors",
580
+ "vision_tower.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00004.safetensors",
581
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00004.safetensors",
582
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00004.safetensors",
583
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00004.safetensors",
584
+ "vision_tower.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00004.safetensors",
585
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
586
+ "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
587
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
588
+ "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
589
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
590
+ "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
591
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
592
+ "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
593
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00004.safetensors",
594
+ "vision_tower.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00004.safetensors",
595
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00004.safetensors",
596
+ "vision_tower.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00004.safetensors",
597
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00004.safetensors",
598
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00004.safetensors",
599
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00004.safetensors",
600
+ "vision_tower.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00004.safetensors",
601
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
602
+ "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
603
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
604
+ "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
605
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
606
+ "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
607
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
608
+ "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
609
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00004.safetensors",
610
+ "vision_tower.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00004.safetensors",
611
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00004.safetensors",
612
+ "vision_tower.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00004.safetensors",
613
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00004.safetensors",
614
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00004.safetensors",
615
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00004.safetensors",
616
+ "vision_tower.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00004.safetensors",
617
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
618
+ "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
619
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
620
+ "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
621
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
622
+ "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
623
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
624
+ "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
625
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00004.safetensors",
626
+ "vision_tower.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00004.safetensors",
627
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00004.safetensors",
628
+ "vision_tower.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00004.safetensors",
629
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00004.safetensors",
630
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00004.safetensors",
631
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00004.safetensors",
632
+ "vision_tower.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00004.safetensors",
633
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
634
+ "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
635
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
636
+ "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
637
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
638
+ "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
639
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
640
+ "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
641
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00004.safetensors",
642
+ "vision_tower.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00004.safetensors",
643
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00004.safetensors",
644
+ "vision_tower.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00004.safetensors",
645
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00004.safetensors",
646
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00004.safetensors",
647
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00004.safetensors",
648
+ "vision_tower.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00004.safetensors",
649
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
650
+ "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
651
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
652
+ "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
653
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
654
+ "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
655
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
656
+ "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
657
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00004.safetensors",
658
+ "vision_tower.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00004.safetensors",
659
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00004.safetensors",
660
+ "vision_tower.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00004.safetensors",
661
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00004.safetensors",
662
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00004.safetensors",
663
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00004.safetensors",
664
+ "vision_tower.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00004.safetensors",
665
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
666
+ "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
667
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
668
+ "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
669
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
670
+ "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
671
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
672
+ "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
673
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00004.safetensors",
674
+ "vision_tower.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00004.safetensors",
675
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00004.safetensors",
676
+ "vision_tower.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00004.safetensors",
677
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00004.safetensors",
678
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00004.safetensors",
679
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00004.safetensors",
680
+ "vision_tower.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00004.safetensors",
681
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
682
+ "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
683
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00004.safetensors",
684
+ "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00004.safetensors",
685
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
686
+ "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
687
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
688
+ "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
689
+ "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
690
+ "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors",
691
+ "vision_tower.vision_model.pre_layrnorm.bias": "model-00001-of-00004.safetensors",
692
+ "vision_tower.vision_model.pre_layrnorm.weight": "model-00001-of-00004.safetensors"
693
+ }
694
+ }
checkpoint-200/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
checkpoint-200/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-200/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-200/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
checkpoint-200/tokenizer_config.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ '<s>' }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '`[INST] `' + content + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 1000000000000000019884624838656,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "split_special_tokens": false,
67
+ "tokenizer_class": "LlamaTokenizer",
68
+ "unk_token": "<unk>",
69
+ "use_default_system_prompt": false
70
+ }
checkpoint-200/trainer_state.json ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.4492753623188406,
5
+ "eval_steps": 40,
6
+ "global_step": 200,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.036231884057971016,
13
+ "grad_norm": 61.66789827681496,
14
+ "learning_rate": 5e-07,
15
+ "logits/chosen": -2.7321553230285645,
16
+ "logits/rejected": -2.7100937366485596,
17
+ "logps/chosen": -182.5845489501953,
18
+ "logps/rejected": -189.55001831054688,
19
+ "loss": 0.689,
20
+ "rewards/accuracies": 0.3187499940395355,
21
+ "rewards/chosen": -0.0021577859297394753,
22
+ "rewards/margins": 0.005646524019539356,
23
+ "rewards/rejected": -0.007804309483617544,
24
+ "step": 5
25
+ },
26
+ {
27
+ "epoch": 0.07246376811594203,
28
+ "grad_norm": 44.96703657621111,
29
+ "learning_rate": 1e-06,
30
+ "logits/chosen": -2.753889799118042,
31
+ "logits/rejected": -2.7519516944885254,
32
+ "logps/chosen": -197.34320068359375,
33
+ "logps/rejected": -184.00961303710938,
34
+ "loss": 0.6275,
35
+ "rewards/accuracies": 0.675000011920929,
36
+ "rewards/chosen": 0.030874451622366905,
37
+ "rewards/margins": 0.18904080986976624,
38
+ "rewards/rejected": -0.15816636383533478,
39
+ "step": 10
40
+ },
41
+ {
42
+ "epoch": 0.10869565217391304,
43
+ "grad_norm": 51.54402636298773,
44
+ "learning_rate": 9.996221126793764e-07,
45
+ "logits/chosen": -2.694437265396118,
46
+ "logits/rejected": -2.691904067993164,
47
+ "logps/chosen": -203.14883422851562,
48
+ "logps/rejected": -204.52386474609375,
49
+ "loss": 0.5851,
50
+ "rewards/accuracies": 0.75,
51
+ "rewards/chosen": 0.6205412745475769,
52
+ "rewards/margins": 0.9350436925888062,
53
+ "rewards/rejected": -0.314502477645874,
54
+ "step": 15
55
+ },
56
+ {
57
+ "epoch": 0.14492753623188406,
58
+ "grad_norm": 35.11282014809116,
59
+ "learning_rate": 9.984890219128145e-07,
60
+ "logits/chosen": -2.609405517578125,
61
+ "logits/rejected": -2.5795176029205322,
62
+ "logps/chosen": -188.33395385742188,
63
+ "logps/rejected": -192.52633666992188,
64
+ "loss": 0.5144,
65
+ "rewards/accuracies": 0.699999988079071,
66
+ "rewards/chosen": 0.8838651776313782,
67
+ "rewards/margins": 1.275496244430542,
68
+ "rewards/rejected": -0.39163118600845337,
69
+ "step": 20
70
+ },
71
+ {
72
+ "epoch": 0.18115942028985507,
73
+ "grad_norm": 37.05591291167376,
74
+ "learning_rate": 9.966024404228493e-07,
75
+ "logits/chosen": -2.4429798126220703,
76
+ "logits/rejected": -2.4225034713745117,
77
+ "logps/chosen": -179.79977416992188,
78
+ "logps/rejected": -179.25279235839844,
79
+ "loss": 0.5036,
80
+ "rewards/accuracies": 0.7562500238418579,
81
+ "rewards/chosen": 0.40532931685447693,
82
+ "rewards/margins": 0.8947975039482117,
83
+ "rewards/rejected": -0.4894680976867676,
84
+ "step": 25
85
+ },
86
+ {
87
+ "epoch": 0.21739130434782608,
88
+ "grad_norm": 33.27618771325657,
89
+ "learning_rate": 9.939652198703783e-07,
90
+ "logits/chosen": -2.319044589996338,
91
+ "logits/rejected": -2.320253372192383,
92
+ "logps/chosen": -188.62039184570312,
93
+ "logps/rejected": -193.8806915283203,
94
+ "loss": 0.4987,
95
+ "rewards/accuracies": 0.762499988079071,
96
+ "rewards/chosen": 0.6481167078018188,
97
+ "rewards/margins": 1.2183058261871338,
98
+ "rewards/rejected": -0.5701891183853149,
99
+ "step": 30
100
+ },
101
+ {
102
+ "epoch": 0.2536231884057971,
103
+ "grad_norm": 38.645073192304736,
104
+ "learning_rate": 9.905813465442354e-07,
105
+ "logits/chosen": -2.242053747177124,
106
+ "logits/rejected": -2.2155580520629883,
107
+ "logps/chosen": -203.94546508789062,
108
+ "logps/rejected": -194.8705291748047,
109
+ "loss": 0.5091,
110
+ "rewards/accuracies": 0.7250000238418579,
111
+ "rewards/chosen": 0.8871867060661316,
112
+ "rewards/margins": 1.2737131118774414,
113
+ "rewards/rejected": -0.3865264356136322,
114
+ "step": 35
115
+ },
116
+ {
117
+ "epoch": 0.2898550724637681,
118
+ "grad_norm": 30.937642063399068,
119
+ "learning_rate": 9.864559353357187e-07,
120
+ "logits/chosen": -2.0998620986938477,
121
+ "logits/rejected": -2.094836473464966,
122
+ "logps/chosen": -182.6898956298828,
123
+ "logps/rejected": -185.73983764648438,
124
+ "loss": 0.4889,
125
+ "rewards/accuracies": 0.7749999761581421,
126
+ "rewards/chosen": 1.0410809516906738,
127
+ "rewards/margins": 1.0222995281219482,
128
+ "rewards/rejected": 0.018781563267111778,
129
+ "step": 40
130
+ },
131
+ {
132
+ "epoch": 0.2898550724637681,
133
+ "eval_logits/chosen": -2.035081624984741,
134
+ "eval_logits/rejected": -1.9875677824020386,
135
+ "eval_logps/chosen": -188.77523803710938,
136
+ "eval_logps/rejected": -200.09498596191406,
137
+ "eval_loss": 0.4642000198364258,
138
+ "eval_rewards/accuracies": 0.7943548560142517,
139
+ "eval_rewards/chosen": 1.1544201374053955,
140
+ "eval_rewards/margins": 1.3431099653244019,
141
+ "eval_rewards/rejected": -0.18868987262248993,
142
+ "eval_runtime": 248.6586,
143
+ "eval_samples_per_second": 15.773,
144
+ "eval_steps_per_second": 0.249,
145
+ "step": 40
146
+ },
147
+ {
148
+ "epoch": 0.32608695652173914,
149
+ "grad_norm": 29.205255063362404,
150
+ "learning_rate": 9.815952220071804e-07,
151
+ "logits/chosen": -1.9733244180679321,
152
+ "logits/rejected": -1.9242098331451416,
153
+ "logps/chosen": -195.65652465820312,
154
+ "logps/rejected": -221.744384765625,
155
+ "loss": 0.4587,
156
+ "rewards/accuracies": 0.762499988079071,
157
+ "rewards/chosen": 1.3803393840789795,
158
+ "rewards/margins": 1.8610279560089111,
159
+ "rewards/rejected": -0.4806883931159973,
160
+ "step": 45
161
+ },
162
+ {
163
+ "epoch": 0.36231884057971014,
164
+ "grad_norm": 29.678687938529745,
165
+ "learning_rate": 9.76006553766365e-07,
166
+ "logits/chosen": -1.8085625171661377,
167
+ "logits/rejected": -1.772071123123169,
168
+ "logps/chosen": -198.9405517578125,
169
+ "logps/rejected": -203.29978942871094,
170
+ "loss": 0.4531,
171
+ "rewards/accuracies": 0.768750011920929,
172
+ "rewards/chosen": 0.8211376070976257,
173
+ "rewards/margins": 1.646414041519165,
174
+ "rewards/rejected": -0.8252763748168945,
175
+ "step": 50
176
+ },
177
+ {
178
+ "epoch": 0.39855072463768115,
179
+ "grad_norm": 30.73732601293307,
180
+ "learning_rate": 9.696983781607415e-07,
181
+ "logits/chosen": -1.7908868789672852,
182
+ "logits/rejected": -1.7599233388900757,
183
+ "logps/chosen": -183.01809692382812,
184
+ "logps/rejected": -172.02801513671875,
185
+ "loss": 0.4595,
186
+ "rewards/accuracies": 0.8374999761581421,
187
+ "rewards/chosen": 0.607209324836731,
188
+ "rewards/margins": 1.649515151977539,
189
+ "rewards/rejected": -1.042305827140808,
190
+ "step": 55
191
+ },
192
+ {
193
+ "epoch": 0.43478260869565216,
194
+ "grad_norm": 41.49881496846909,
195
+ "learning_rate": 9.626802303086209e-07,
196
+ "logits/chosen": -1.9050066471099854,
197
+ "logits/rejected": -1.8625679016113281,
198
+ "logps/chosen": -186.8179931640625,
199
+ "logps/rejected": -194.26707458496094,
200
+ "loss": 0.4276,
201
+ "rewards/accuracies": 0.8187500238418579,
202
+ "rewards/chosen": 0.4140642285346985,
203
+ "rewards/margins": 1.712418794631958,
204
+ "rewards/rejected": -1.2983543872833252,
205
+ "step": 60
206
+ },
207
+ {
208
+ "epoch": 0.47101449275362317,
209
+ "grad_norm": 36.480881476794956,
210
+ "learning_rate": 9.549627184863528e-07,
211
+ "logits/chosen": -2.1638290882110596,
212
+ "logits/rejected": -2.0793471336364746,
213
+ "logps/chosen": -191.9010467529297,
214
+ "logps/rejected": -193.20651245117188,
215
+ "loss": 0.4293,
216
+ "rewards/accuracies": 0.7875000238418579,
217
+ "rewards/chosen": 0.045776158571243286,
218
+ "rewards/margins": 1.6033704280853271,
219
+ "rewards/rejected": -1.5575940608978271,
220
+ "step": 65
221
+ },
222
+ {
223
+ "epoch": 0.5072463768115942,
224
+ "grad_norm": 26.318712537847357,
225
+ "learning_rate": 9.465575080933957e-07,
226
+ "logits/chosen": -2.1214632987976074,
227
+ "logits/rejected": -2.081162214279175,
228
+ "logps/chosen": -172.66419982910156,
229
+ "logps/rejected": -208.19149780273438,
230
+ "loss": 0.3963,
231
+ "rewards/accuracies": 0.831250011920929,
232
+ "rewards/chosen": 0.17696644365787506,
233
+ "rewards/margins": 1.8183233737945557,
234
+ "rewards/rejected": -1.6413570642471313,
235
+ "step": 70
236
+ },
237
+ {
238
+ "epoch": 0.5434782608695652,
239
+ "grad_norm": 32.35315227999789,
240
+ "learning_rate": 9.374773040194878e-07,
241
+ "logits/chosen": -2.1271562576293945,
242
+ "logits/rejected": -2.0680038928985596,
243
+ "logps/chosen": -205.5594482421875,
244
+ "logps/rejected": -210.7834014892578,
245
+ "loss": 0.4353,
246
+ "rewards/accuracies": 0.824999988079071,
247
+ "rewards/chosen": 0.16769471764564514,
248
+ "rewards/margins": 1.786273717880249,
249
+ "rewards/rejected": -1.6185792684555054,
250
+ "step": 75
251
+ },
252
+ {
253
+ "epoch": 0.5797101449275363,
254
+ "grad_norm": 29.96593011887677,
255
+ "learning_rate": 9.277358314405818e-07,
256
+ "logits/chosen": -2.046504497528076,
257
+ "logits/rejected": -2.0122628211975098,
258
+ "logps/chosen": -189.37771606445312,
259
+ "logps/rejected": -206.10659790039062,
260
+ "loss": 0.3941,
261
+ "rewards/accuracies": 0.8125,
262
+ "rewards/chosen": -0.2374907284975052,
263
+ "rewards/margins": 1.910300612449646,
264
+ "rewards/rejected": -2.1477913856506348,
265
+ "step": 80
266
+ },
267
+ {
268
+ "epoch": 0.5797101449275363,
269
+ "eval_logits/chosen": -1.9900643825531006,
270
+ "eval_logits/rejected": -1.9449083805084229,
271
+ "eval_logps/chosen": -202.59437561035156,
272
+ "eval_logps/rejected": -221.12725830078125,
273
+ "eval_loss": 0.4217630624771118,
274
+ "eval_rewards/accuracies": 0.8044354915618896,
275
+ "eval_rewards/chosen": -0.22749020159244537,
276
+ "eval_rewards/margins": 2.0644266605377197,
277
+ "eval_rewards/rejected": -2.291916847229004,
278
+ "eval_runtime": 247.3121,
279
+ "eval_samples_per_second": 15.859,
280
+ "eval_steps_per_second": 0.251,
281
+ "step": 80
282
+ },
283
+ {
284
+ "epoch": 0.6159420289855072,
285
+ "grad_norm": 34.207898226903815,
286
+ "learning_rate": 9.173478150725651e-07,
287
+ "logits/chosen": -2.0169568061828613,
288
+ "logits/rejected": -1.9391515254974365,
289
+ "logps/chosen": -209.73440551757812,
290
+ "logps/rejected": -215.14205932617188,
291
+ "loss": 0.422,
292
+ "rewards/accuracies": 0.831250011920929,
293
+ "rewards/chosen": 0.1130056157708168,
294
+ "rewards/margins": 2.3998470306396484,
295
+ "rewards/rejected": -2.2868411540985107,
296
+ "step": 85
297
+ },
298
+ {
299
+ "epoch": 0.6521739130434783,
300
+ "grad_norm": 28.850805260840193,
301
+ "learning_rate": 9.063289569141251e-07,
302
+ "logits/chosen": -2.1180360317230225,
303
+ "logits/rejected": -2.080235242843628,
304
+ "logps/chosen": -214.2292022705078,
305
+ "logps/rejected": -223.37564086914062,
306
+ "loss": 0.4135,
307
+ "rewards/accuracies": 0.84375,
308
+ "rewards/chosen": 0.5250645279884338,
309
+ "rewards/margins": 2.2779579162597656,
310
+ "rewards/rejected": -1.7528936862945557,
311
+ "step": 90
312
+ },
313
+ {
314
+ "epoch": 0.6884057971014492,
315
+ "grad_norm": 34.73111091101309,
316
+ "learning_rate": 8.946959125124051e-07,
317
+ "logits/chosen": -2.2047152519226074,
318
+ "logits/rejected": -2.1581788063049316,
319
+ "logps/chosen": -209.3234405517578,
320
+ "logps/rejected": -195.6946563720703,
321
+ "loss": 0.4077,
322
+ "rewards/accuracies": 0.824999988079071,
323
+ "rewards/chosen": 0.25443512201309204,
324
+ "rewards/margins": 2.2029635906219482,
325
+ "rewards/rejected": -1.948528528213501,
326
+ "step": 95
327
+ },
328
+ {
329
+ "epoch": 0.7246376811594203,
330
+ "grad_norm": 32.202226649445414,
331
+ "learning_rate": 8.824662657873238e-07,
332
+ "logits/chosen": -2.256134510040283,
333
+ "logits/rejected": -2.2395310401916504,
334
+ "logps/chosen": -176.19464111328125,
335
+ "logps/rejected": -209.0999298095703,
336
+ "loss": 0.3767,
337
+ "rewards/accuracies": 0.800000011920929,
338
+ "rewards/chosen": -0.3457742929458618,
339
+ "rewards/margins": 1.9383150339126587,
340
+ "rewards/rejected": -2.2840893268585205,
341
+ "step": 100
342
+ },
343
+ {
344
+ "epoch": 0.7608695652173914,
345
+ "grad_norm": 31.35693490170537,
346
+ "learning_rate": 8.696585024526135e-07,
347
+ "logits/chosen": -2.3358142375946045,
348
+ "logits/rejected": -2.307068347930908,
349
+ "logps/chosen": -191.8298797607422,
350
+ "logps/rejected": -227.55612182617188,
351
+ "loss": 0.4017,
352
+ "rewards/accuracies": 0.8125,
353
+ "rewards/chosen": -0.2825666069984436,
354
+ "rewards/margins": 2.5402400493621826,
355
+ "rewards/rejected": -2.8228065967559814,
356
+ "step": 105
357
+ },
358
+ {
359
+ "epoch": 0.7971014492753623,
360
+ "grad_norm": 36.29919961993296,
361
+ "learning_rate": 8.562919820737535e-07,
362
+ "logits/chosen": -2.2931602001190186,
363
+ "logits/rejected": -2.2568557262420654,
364
+ "logps/chosen": -207.34158325195312,
365
+ "logps/rejected": -208.9143524169922,
366
+ "loss": 0.3767,
367
+ "rewards/accuracies": 0.84375,
368
+ "rewards/chosen": -0.3017815351486206,
369
+ "rewards/margins": 2.3830084800720215,
370
+ "rewards/rejected": -2.6847901344299316,
371
+ "step": 110
372
+ },
373
+ {
374
+ "epoch": 0.8333333333333334,
375
+ "grad_norm": 28.819478551304137,
376
+ "learning_rate": 8.423869088050315e-07,
377
+ "logits/chosen": -2.2734763622283936,
378
+ "logits/rejected": -2.2522785663604736,
379
+ "logps/chosen": -195.6924591064453,
380
+ "logps/rejected": -221.85104370117188,
381
+ "loss": 0.3892,
382
+ "rewards/accuracies": 0.8125,
383
+ "rewards/chosen": -0.12773282825946808,
384
+ "rewards/margins": 2.3888843059539795,
385
+ "rewards/rejected": -2.5166170597076416,
386
+ "step": 115
387
+ },
388
+ {
389
+ "epoch": 0.8695652173913043,
390
+ "grad_norm": 37.085951928654964,
391
+ "learning_rate": 8.2796430084997e-07,
392
+ "logits/chosen": -2.2070839405059814,
393
+ "logits/rejected": -2.1736464500427246,
394
+ "logps/chosen": -196.70889282226562,
395
+ "logps/rejected": -207.93417358398438,
396
+ "loss": 0.3717,
397
+ "rewards/accuracies": 0.862500011920929,
398
+ "rewards/chosen": 0.12862932682037354,
399
+ "rewards/margins": 2.698848009109497,
400
+ "rewards/rejected": -2.570218563079834,
401
+ "step": 120
402
+ },
403
+ {
404
+ "epoch": 0.8695652173913043,
405
+ "eval_logits/chosen": -2.122933864593506,
406
+ "eval_logits/rejected": -2.0901710987091064,
407
+ "eval_logps/chosen": -202.4208221435547,
408
+ "eval_logps/rejected": -223.09356689453125,
409
+ "eval_loss": 0.438678115606308,
410
+ "eval_rewards/accuracies": 0.8286290168762207,
411
+ "eval_rewards/chosen": -0.2101391851902008,
412
+ "eval_rewards/margins": 2.278407573699951,
413
+ "eval_rewards/rejected": -2.488546848297119,
414
+ "eval_runtime": 247.3754,
415
+ "eval_samples_per_second": 15.854,
416
+ "eval_steps_per_second": 0.251,
417
+ "step": 120
418
+ },
419
+ {
420
+ "epoch": 0.9057971014492754,
421
+ "grad_norm": 27.843666767086823,
422
+ "learning_rate": 8.130459586912753e-07,
423
+ "logits/chosen": -2.0930094718933105,
424
+ "logits/rejected": -2.0791330337524414,
425
+ "logps/chosen": -220.4320526123047,
426
+ "logps/rejected": -218.3594512939453,
427
+ "loss": 0.4575,
428
+ "rewards/accuracies": 0.800000011920929,
429
+ "rewards/chosen": -0.9592973589897156,
430
+ "rewards/margins": 1.66092050075531,
431
+ "rewards/rejected": -2.620218276977539,
432
+ "step": 125
433
+ },
434
+ {
435
+ "epoch": 0.9420289855072463,
436
+ "grad_norm": 28.902451958691426,
437
+ "learning_rate": 7.97654432138333e-07,
438
+ "logits/chosen": -2.1393418312072754,
439
+ "logits/rejected": -2.103482723236084,
440
+ "logps/chosen": -214.5059051513672,
441
+ "logps/rejected": -248.8673858642578,
442
+ "loss": 0.388,
443
+ "rewards/accuracies": 0.856249988079071,
444
+ "rewards/chosen": -0.1911260336637497,
445
+ "rewards/margins": 2.8205361366271973,
446
+ "rewards/rejected": -3.011662006378174,
447
+ "step": 130
448
+ },
449
+ {
450
+ "epoch": 0.9782608695652174,
451
+ "grad_norm": 26.45174989804514,
452
+ "learning_rate": 7.81812986242061e-07,
453
+ "logits/chosen": -2.2018837928771973,
454
+ "logits/rejected": -2.1459240913391113,
455
+ "logps/chosen": -193.5723419189453,
456
+ "logps/rejected": -233.81838989257812,
457
+ "loss": 0.365,
458
+ "rewards/accuracies": 0.831250011920929,
459
+ "rewards/chosen": 0.10008885711431503,
460
+ "rewards/margins": 3.072911262512207,
461
+ "rewards/rejected": -2.972822427749634,
462
+ "step": 135
463
+ },
464
+ {
465
+ "epoch": 1.0144927536231885,
466
+ "grad_norm": 16.755101670108846,
467
+ "learning_rate": 7.655455661286375e-07,
468
+ "logits/chosen": -2.190566301345825,
469
+ "logits/rejected": -2.1852922439575195,
470
+ "logps/chosen": -190.56027221679688,
471
+ "logps/rejected": -234.6202392578125,
472
+ "loss": 0.2524,
473
+ "rewards/accuracies": 0.918749988079071,
474
+ "rewards/chosen": 0.05487387627363205,
475
+ "rewards/margins": 3.58606219291687,
476
+ "rewards/rejected": -3.531188488006592,
477
+ "step": 140
478
+ },
479
+ {
480
+ "epoch": 1.0507246376811594,
481
+ "grad_norm": 16.022443175608785,
482
+ "learning_rate": 7.488767608052628e-07,
483
+ "logits/chosen": -2.2735958099365234,
484
+ "logits/rejected": -2.199552059173584,
485
+ "logps/chosen": -190.21511840820312,
486
+ "logps/rejected": -237.59854125976562,
487
+ "loss": 0.171,
488
+ "rewards/accuracies": 0.9375,
489
+ "rewards/chosen": 0.7184330821037292,
490
+ "rewards/margins": 4.115548133850098,
491
+ "rewards/rejected": -3.3971149921417236,
492
+ "step": 145
493
+ },
494
+ {
495
+ "epoch": 1.0869565217391304,
496
+ "grad_norm": 16.37925141930124,
497
+ "learning_rate": 7.318317659926636e-07,
498
+ "logits/chosen": -2.250054121017456,
499
+ "logits/rejected": -2.221043109893799,
500
+ "logps/chosen": -174.0218048095703,
501
+ "logps/rejected": -237.14120483398438,
502
+ "loss": 0.1436,
503
+ "rewards/accuracies": 0.9375,
504
+ "rewards/chosen": 1.0139648914337158,
505
+ "rewards/margins": 4.169236660003662,
506
+ "rewards/rejected": -3.1552722454071045,
507
+ "step": 150
508
+ },
509
+ {
510
+ "epoch": 1.1231884057971016,
511
+ "grad_norm": 13.693805739113818,
512
+ "learning_rate": 7.144363460405189e-07,
513
+ "logits/chosen": -2.3314082622528076,
514
+ "logits/rejected": -2.2712883949279785,
515
+ "logps/chosen": -192.2017059326172,
516
+ "logps/rejected": -236.91586303710938,
517
+ "loss": 0.1415,
518
+ "rewards/accuracies": 0.9624999761581421,
519
+ "rewards/chosen": 1.1952037811279297,
520
+ "rewards/margins": 4.705626487731934,
521
+ "rewards/rejected": -3.510422945022583,
522
+ "step": 155
523
+ },
524
+ {
525
+ "epoch": 1.1594202898550725,
526
+ "grad_norm": 13.203700040695288,
527
+ "learning_rate": 6.967167949833762e-07,
528
+ "logits/chosen": -2.324735164642334,
529
+ "logits/rejected": -2.283231258392334,
530
+ "logps/chosen": -194.0807342529297,
531
+ "logps/rejected": -247.2218475341797,
532
+ "loss": 0.1459,
533
+ "rewards/accuracies": 0.949999988079071,
534
+ "rewards/chosen": 0.3761358857154846,
535
+ "rewards/margins": 4.9711785316467285,
536
+ "rewards/rejected": -4.595042705535889,
537
+ "step": 160
538
+ },
539
+ {
540
+ "epoch": 1.1594202898550725,
541
+ "eval_logits/chosen": -2.3007450103759766,
542
+ "eval_logits/rejected": -2.273284673690796,
543
+ "eval_logps/chosen": -204.34878540039062,
544
+ "eval_logps/rejected": -232.13629150390625,
545
+ "eval_loss": 0.42876046895980835,
546
+ "eval_rewards/accuracies": 0.8286290168762207,
547
+ "eval_rewards/chosen": -0.40293240547180176,
548
+ "eval_rewards/margins": 2.9898877143859863,
549
+ "eval_rewards/rejected": -3.392819881439209,
550
+ "eval_runtime": 247.3207,
551
+ "eval_samples_per_second": 15.858,
552
+ "eval_steps_per_second": 0.251,
553
+ "step": 160
554
+ },
555
+ {
556
+ "epoch": 1.1956521739130435,
557
+ "grad_norm": 17.287170066985027,
558
+ "learning_rate": 6.786998967959219e-07,
559
+ "logits/chosen": -2.313910961151123,
560
+ "logits/rejected": -2.274190902709961,
561
+ "logps/chosen": -200.9847869873047,
562
+ "logps/rejected": -228.14138793945312,
563
+ "loss": 0.1497,
564
+ "rewards/accuracies": 0.9375,
565
+ "rewards/chosen": 0.39603787660598755,
566
+ "rewards/margins": 4.507396697998047,
567
+ "rewards/rejected": -4.111359119415283,
568
+ "step": 165
569
+ },
570
+ {
571
+ "epoch": 1.2318840579710144,
572
+ "grad_norm": 16.43946047429659,
573
+ "learning_rate": 6.604128849076838e-07,
574
+ "logits/chosen": -2.340721607208252,
575
+ "logits/rejected": -2.305487871170044,
576
+ "logps/chosen": -200.94406127929688,
577
+ "logps/rejected": -238.47921752929688,
578
+ "loss": 0.1533,
579
+ "rewards/accuracies": 0.9437500238418579,
580
+ "rewards/chosen": 0.9461654424667358,
581
+ "rewards/margins": 4.6484904289245605,
582
+ "rewards/rejected": -3.702324390411377,
583
+ "step": 170
584
+ },
585
+ {
586
+ "epoch": 1.2681159420289856,
587
+ "grad_norm": 13.545354366232296,
588
+ "learning_rate": 6.418834010383609e-07,
589
+ "logits/chosen": -2.400474786758423,
590
+ "logits/rejected": -2.343231439590454,
591
+ "logps/chosen": -172.61148071289062,
592
+ "logps/rejected": -230.24197387695312,
593
+ "loss": 0.1459,
594
+ "rewards/accuracies": 0.9437500238418579,
595
+ "rewards/chosen": 0.5564432144165039,
596
+ "rewards/margins": 4.526711463928223,
597
+ "rewards/rejected": -3.9702675342559814,
598
+ "step": 175
599
+ },
600
+ {
601
+ "epoch": 1.3043478260869565,
602
+ "grad_norm": 16.208760712520707,
603
+ "learning_rate": 6.231394534160007e-07,
604
+ "logits/chosen": -2.3880228996276855,
605
+ "logits/rejected": -2.377924680709839,
606
+ "logps/chosen": -186.1154327392578,
607
+ "logps/rejected": -228.6842498779297,
608
+ "loss": 0.1408,
609
+ "rewards/accuracies": 0.96875,
610
+ "rewards/chosen": 1.147924542427063,
611
+ "rewards/margins": 4.681893348693848,
612
+ "rewards/rejected": -3.533968687057495,
613
+ "step": 180
614
+ },
615
+ {
616
+ "epoch": 1.3405797101449275,
617
+ "grad_norm": 16.391749704364813,
618
+ "learning_rate": 6.042093744411828e-07,
619
+ "logits/chosen": -2.3000502586364746,
620
+ "logits/rejected": -2.280015468597412,
621
+ "logps/chosen": -184.9687042236328,
622
+ "logps/rejected": -230.6714630126953,
623
+ "loss": 0.1426,
624
+ "rewards/accuracies": 0.9437500238418579,
625
+ "rewards/chosen": 1.1101362705230713,
626
+ "rewards/margins": 4.539022922515869,
627
+ "rewards/rejected": -3.4288864135742188,
628
+ "step": 185
629
+ },
630
+ {
631
+ "epoch": 1.3768115942028984,
632
+ "grad_norm": 15.103230138315158,
633
+ "learning_rate": 5.851217778611993e-07,
634
+ "logits/chosen": -2.2735018730163574,
635
+ "logits/rejected": -2.272218942642212,
636
+ "logps/chosen": -196.11097717285156,
637
+ "logps/rejected": -217.5700225830078,
638
+ "loss": 0.1351,
639
+ "rewards/accuracies": 0.925000011920929,
640
+ "rewards/chosen": 0.8358832597732544,
641
+ "rewards/margins": 4.623397350311279,
642
+ "rewards/rejected": -3.7875142097473145,
643
+ "step": 190
644
+ },
645
+ {
646
+ "epoch": 1.4130434782608696,
647
+ "grad_norm": 23.621795316466446,
648
+ "learning_rate": 5.659055155189651e-07,
649
+ "logits/chosen": -2.325028896331787,
650
+ "logits/rejected": -2.252711296081543,
651
+ "logps/chosen": -193.0414276123047,
652
+ "logps/rejected": -231.9776611328125,
653
+ "loss": 0.1542,
654
+ "rewards/accuracies": 0.9437500238418579,
655
+ "rewards/chosen": 0.06639621406793594,
656
+ "rewards/margins": 4.913178443908691,
657
+ "rewards/rejected": -4.8467817306518555,
658
+ "step": 195
659
+ },
660
+ {
661
+ "epoch": 1.4492753623188406,
662
+ "grad_norm": 11.913033785498357,
663
+ "learning_rate": 5.465896337420358e-07,
664
+ "logits/chosen": -2.3189964294433594,
665
+ "logits/rejected": -2.2462618350982666,
666
+ "logps/chosen": -207.55685424804688,
667
+ "logps/rejected": -269.5002746582031,
668
+ "loss": 0.1455,
669
+ "rewards/accuracies": 0.9437500238418579,
670
+ "rewards/chosen": 0.3675020635128021,
671
+ "rewards/margins": 5.308381080627441,
672
+ "rewards/rejected": -4.940878868103027,
673
+ "step": 200
674
+ },
675
+ {
676
+ "epoch": 1.4492753623188406,
677
+ "eval_logits/chosen": -2.2696547508239746,
678
+ "eval_logits/rejected": -2.246581554412842,
679
+ "eval_logps/chosen": -205.6576690673828,
680
+ "eval_logps/rejected": -234.53866577148438,
681
+ "eval_loss": 0.42552247643470764,
682
+ "eval_rewards/accuracies": 0.8165322542190552,
683
+ "eval_rewards/chosen": -0.5338226556777954,
684
+ "eval_rewards/margins": 3.099236011505127,
685
+ "eval_rewards/rejected": -3.633058547973633,
686
+ "eval_runtime": 247.2511,
687
+ "eval_samples_per_second": 15.862,
688
+ "eval_steps_per_second": 0.251,
689
+ "step": 200
690
+ }
691
+ ],
692
+ "logging_steps": 5,
693
+ "max_steps": 414,
694
+ "num_input_tokens_seen": 0,
695
+ "num_train_epochs": 3,
696
+ "save_steps": 40,
697
+ "stateful_callbacks": {
698
+ "TrainerControl": {
699
+ "args": {
700
+ "should_epoch_stop": false,
701
+ "should_evaluate": false,
702
+ "should_log": false,
703
+ "should_save": true,
704
+ "should_training_stop": false
705
+ },
706
+ "attributes": {}
707
+ }
708
+ },
709
+ "total_flos": 2358113407598592.0,
710
+ "train_batch_size": 8,
711
+ "trial_name": null,
712
+ "trial_params": null
713
+ }
checkpoint-200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af79ff24bae38840a3c8efe3d28d0cc2a77ca640996f1dd8521f5747a2625682
3
+ size 7096
checkpoint-240/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
checkpoint-240/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/data/align-anything/hantao/models/llava-v1.6-mistral-7b-hf",
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "hidden_size": 4096,
7
+ "ignore_index": -100,
8
+ "image_grid_pinpoints": [
9
+ [
10
+ 336,
11
+ 672
12
+ ],
13
+ [
14
+ 672,
15
+ 336
16
+ ],
17
+ [
18
+ 672,
19
+ 672
20
+ ],
21
+ [
22
+ 1008,
23
+ 336
24
+ ],
25
+ [
26
+ 336,
27
+ 1008
28
+ ]
29
+ ],
30
+ "image_seq_length": 576,
31
+ "image_token_index": 32000,
32
+ "model_type": "llava_next",
33
+ "projector_hidden_act": "gelu",
34
+ "text_config": {
35
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
36
+ "architectures": [
37
+ "MistralForCausalLM"
38
+ ],
39
+ "intermediate_size": 14336,
40
+ "max_position_embeddings": 32768,
41
+ "model_type": "mistral",
42
+ "num_key_value_heads": 8,
43
+ "rms_norm_eps": 1e-05,
44
+ "rope_theta": 1000000.0,
45
+ "sliding_window": null,
46
+ "torch_dtype": "bfloat16",
47
+ "vocab_size": 32064
48
+ },
49
+ "tie_word_embeddings": false,
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.45.2",
52
+ "use_cache": false,
53
+ "use_image_newline_parameter": true,
54
+ "vision_config": {
55
+ "hidden_size": 1024,
56
+ "image_size": 336,
57
+ "intermediate_size": 4096,
58
+ "model_type": "clip_vision_model",
59
+ "num_attention_heads": 16,
60
+ "num_hidden_layers": 24,
61
+ "patch_size": 14,
62
+ "projection_dim": 768,
63
+ "vocab_size": 32000
64
+ },
65
+ "vision_feature_layer": -2,
66
+ "vision_feature_select_strategy": "default",
67
+ "vocab_size": 32064
68
+ }