RemVdH commited on
Commit
460819c
·
verified ·
1 Parent(s): b3723de

End of training

Browse files
Files changed (2) hide show
  1. README.md +21 -18
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -2,10 +2,11 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - generated_from_trainer
6
  base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
7
  model-index:
8
- - name: outputs/lora-out
9
  results: []
10
  ---
11
 
@@ -28,7 +29,7 @@ strict: false
28
  datasets:
29
  - path: RemVdH/databricks-dolly-3k
30
  type: alpaca
31
- dataset_prepared_path:
32
  val_set_size: 0.05
33
  output_dir: ./outputs/lora-out
34
 
@@ -45,6 +46,8 @@ lora_dropout: 0.05
45
  lora_target_linear: true
46
  lora_fan_in_fan_out:
47
 
 
 
48
  wandb_project:
49
  wandb_entity:
50
  wandb_watch:
@@ -86,11 +89,11 @@ special_tokens:
86
 
87
  </details><br>
88
 
89
- # outputs/lora-out
90
 
91
  This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) on the None dataset.
92
  It achieves the following results on the evaluation set:
93
- - Loss: 1.7524
94
 
95
  ## Model description
96
 
@@ -125,20 +128,20 @@ The following hyperparameters were used during training:
125
  | Training Loss | Epoch | Step | Validation Loss |
126
  |:-------------:|:------:|:----:|:---------------:|
127
  | 1.9375 | 0.0465 | 1 | 2.1187 |
128
- | 1.9298 | 0.2791 | 6 | 2.0222 |
129
- | 1.8244 | 0.5581 | 12 | 1.8084 |
130
- | 1.7278 | 0.8372 | 18 | 1.7874 |
131
- | 1.7525 | 1.1047 | 24 | 1.7821 |
132
- | 1.8701 | 1.3837 | 30 | 1.7774 |
133
- | 1.6449 | 1.6628 | 36 | 1.7599 |
134
- | 1.7854 | 1.9419 | 42 | 1.7577 |
135
- | 1.6611 | 2.1977 | 48 | 1.7541 |
136
- | 1.6157 | 2.4767 | 54 | 1.7543 |
137
- | 1.7079 | 2.7558 | 60 | 1.7552 |
138
- | 1.7268 | 3.0116 | 66 | 1.7492 |
139
- | 1.8108 | 3.2907 | 72 | 1.7522 |
140
- | 1.6888 | 3.5698 | 78 | 1.7493 |
141
- | 1.7847 | 3.8488 | 84 | 1.7524 |
142
 
143
 
144
  ### Framework versions
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
+ - axolotl
6
  - generated_from_trainer
7
  base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
8
  model-index:
9
+ - name: test-model-ft-tinylama
10
  results: []
11
  ---
12
 
 
29
  datasets:
30
  - path: RemVdH/databricks-dolly-3k
31
  type: alpaca
32
+ dataset_prepared_path: last_run_prepared
33
  val_set_size: 0.05
34
  output_dir: ./outputs/lora-out
35
 
 
46
  lora_target_linear: true
47
  lora_fan_in_fan_out:
48
 
49
+ hub_model_id: RemVdH/test-model-ft-tinylama
50
+
51
  wandb_project:
52
  wandb_entity:
53
  wandb_watch:
 
89
 
90
  </details><br>
91
 
92
+ # test-model-ft-tinylama
93
 
94
  This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) on the None dataset.
95
  It achieves the following results on the evaluation set:
96
+ - Loss: 1.7487
97
 
98
  ## Model description
99
 
 
128
  | Training Loss | Epoch | Step | Validation Loss |
129
  |:-------------:|:------:|:----:|:---------------:|
130
  | 1.9375 | 0.0465 | 1 | 2.1187 |
131
+ | 1.9299 | 0.2791 | 6 | 2.0168 |
132
+ | 1.825 | 0.5581 | 12 | 1.8114 |
133
+ | 1.7291 | 0.8372 | 18 | 1.7892 |
134
+ | 1.7519 | 1.1047 | 24 | 1.7811 |
135
+ | 1.8679 | 1.3837 | 30 | 1.7753 |
136
+ | 1.6452 | 1.6628 | 36 | 1.7567 |
137
+ | 1.7842 | 1.9419 | 42 | 1.7574 |
138
+ | 1.6599 | 2.1977 | 48 | 1.7538 |
139
+ | 1.6158 | 2.4767 | 54 | 1.7543 |
140
+ | 1.7082 | 2.7558 | 60 | 1.7560 |
141
+ | 1.7263 | 3.0116 | 66 | 1.7518 |
142
+ | 1.8113 | 3.2907 | 72 | 1.7511 |
143
+ | 1.6883 | 3.5698 | 78 | 1.7497 |
144
+ | 1.7864 | 3.8488 | 84 | 1.7487 |
145
 
146
 
147
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d69db749b108e8961ff48c48274c973f35d04cd7d9ab69ce66c9ff0b941c006d
3
  size 101036698
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243c991e5798d9cbb6504705fa0d5f46258ebfc1d4780dc4fff8d407258591bf
3
  size 101036698