hexuan21 commited on
Commit
42f1065
·
verified ·
1 Parent(s): a8cd62c

Model save

Browse files
Files changed (2) hide show
  1. README.md +3 -3
  2. generation_config.json +1 -4
README.md CHANGED
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - gradient_accumulation_steps: 8
43
  - total_train_batch_size: 64
44
  - total_eval_batch_size: 8
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_ratio: 0.03
48
  - num_epochs: 1.0
@@ -53,7 +53,7 @@ The following hyperparameters were used during training:
53
 
54
  ### Framework versions
55
 
56
- - Transformers 4.45.0
57
- - Pytorch 2.5.1+cu124
58
  - Datasets 2.18.0
59
  - Tokenizers 0.20.3
 
42
  - gradient_accumulation_steps: 8
43
  - total_train_batch_size: 64
44
  - total_eval_batch_size: 8
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_ratio: 0.03
48
  - num_epochs: 1.0
 
53
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.46.3
57
+ - Pytorch 2.5.1+cu118
58
  - Datasets 2.18.0
59
  - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -1,16 +1,13 @@
1
  {
2
- "attn_implementation": "flash_attention_2",
3
  "bos_token_id": 151643,
4
  "do_sample": true,
5
  "eos_token_id": [
6
  151645,
7
  151643
8
  ],
9
- "num_labels": 5,
10
  "pad_token_id": 151643,
11
- "problem_type": "regression",
12
  "temperature": 0.01,
13
  "top_k": 1,
14
  "top_p": 0.001,
15
- "transformers_version": "4.45.0"
16
  }
 
1
  {
 
2
  "bos_token_id": 151643,
3
  "do_sample": true,
4
  "eos_token_id": [
5
  151645,
6
  151643
7
  ],
 
8
  "pad_token_id": 151643,
 
9
  "temperature": 0.01,
10
  "top_k": 1,
11
  "top_p": 0.001,
12
+ "transformers_version": "4.46.3"
13
  }