iyashnayi commited on
Commit
4e29401
·
verified ·
1 Parent(s): f7d87d9

iyashnayi/SocioLens-llama-3.2-3B

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/yashnayi00-university-of-new-haven/huggingface/runs/ifkobwii)
31
 
32
 
33
  This model was trained with SFT.
@@ -35,7 +35,7 @@ This model was trained with SFT.
35
  ### Framework versions
36
 
37
  - TRL: 0.16.1
38
- - Transformers: 4.51.2
39
  - Pytorch: 2.2.0+cu118
40
  - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/yashnayi00-university-of-new-haven/huggingface/runs/r1mh79zm)
31
 
32
 
33
  This model was trained with SFT.
 
35
  ### Framework versions
36
 
37
  - TRL: 0.16.1
38
+ - Transformers: 4.51.3
39
  - Pytorch: 2.2.0+cu118
40
  - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
adapter_config.json CHANGED
@@ -14,11 +14,8 @@
14
  "r": 16,
15
  "revision": null,
16
  "target_modules": [
17
- "up_proj",
18
- "down_proj",
19
- "gate_proj",
20
- "k_proj",
21
  "q_proj",
 
22
  "v_proj",
23
  "o_proj"
24
  ],
 
14
  "r": 16,
15
  "revision": null,
16
  "target_modules": [
 
 
 
 
17
  "q_proj",
18
+ "k_proj",
19
  "v_proj",
20
  "o_proj"
21
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35b1eace7775b6995772099264429042406c610ecd29d2ba2bb1650777c9a277
3
- size 97307544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58628071bf6251f15c43ab108a9dfcc619577e514e8a4009144d6e1305d5537e
3
+ size 36730224
config.json CHANGED
@@ -45,7 +45,7 @@
45
  "rope_theta": 500000.0,
46
  "tie_word_embeddings": true,
47
  "torch_dtype": "float16",
48
- "transformers_version": "4.51.2",
49
  "use_cache": false,
50
  "vocab_size": 128256
51
  }
 
45
  "rope_theta": 500000.0,
46
  "tie_word_embeddings": true,
47
  "torch_dtype": "float16",
48
+ "transformers_version": "4.51.3",
49
  "use_cache": false,
50
  "vocab_size": 128256
51
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9b6842936cd09712f4382b4202f568e5c6db596ed73030459cd3980c3d1ead3
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67f5db76b3df48d7826532eb3fd17580a466d67b16ae396b9bdcd8afc3c629b2
3
  size 5624