groderg commited on
Commit
49b770e
·
verified ·
1 Parent(s): 548972b

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +228 -131
README.md CHANGED
@@ -1,155 +1,252 @@
 
1
  ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: facebook/dinov2-large
5
  tags:
 
 
6
  - generated_from_trainer
7
- metrics:
8
- - accuracy
9
  model-index:
10
  - name: DinoVdo-large-2025_01_27_45863-bs32_freeze
11
  results: []
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
 
17
- # DinoVdo-large-2025_01_27_45863-bs32_freeze
18
 
19
- This model is a fine-tuned version of [facebook/dinov2-large](https://huggingface.co/facebook/dinov2-large) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
  - Loss: 0.1236
22
  - F1 Micro: 0.8136
23
  - F1 Macro: 0.7060
24
  - Accuracy: 0.3071
25
- - Learning Rate: 0.0000
26
 
27
- ## Model description
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- More information needed
30
 
31
- ## Intended uses & limitations
 
32
 
33
- More information needed
34
 
35
- ## Training and evaluation data
 
 
36
 
37
- More information needed
 
38
 
39
- ## Training procedure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
- ### Training hyperparameters
 
 
42
 
43
  The following hyperparameters were used during training:
44
- - learning_rate: 0.001
45
- - train_batch_size: 32
46
- - eval_batch_size: 32
47
- - seed: 42
48
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
- - lr_scheduler_type: linear
50
- - num_epochs: 150
51
- - mixed_precision_training: Native AMP
52
-
53
- ### Training results
54
-
55
- | Training Loss | Epoch | Step | Validation Loss | F1 Micro | F1 Macro | Accuracy | Rate |
56
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|:--------:|:------:|
57
- | No log | 1.0 | 273 | 0.1678 | 0.7433 | 0.5138 | 0.2293 | 0.001 |
58
- | 0.2722 | 2.0 | 546 | 0.1537 | 0.7614 | 0.5771 | 0.2436 | 0.001 |
59
- | 0.2722 | 3.0 | 819 | 0.1483 | 0.7763 | 0.6194 | 0.2415 | 0.001 |
60
- | 0.169 | 4.0 | 1092 | 0.1464 | 0.7808 | 0.6276 | 0.2555 | 0.001 |
61
- | 0.169 | 5.0 | 1365 | 0.1452 | 0.7788 | 0.6421 | 0.2520 | 0.001 |
62
- | 0.1612 | 6.0 | 1638 | 0.1440 | 0.7802 | 0.6147 | 0.2579 | 0.001 |
63
- | 0.1612 | 7.0 | 1911 | 0.1452 | 0.7787 | 0.6141 | 0.2503 | 0.001 |
64
- | 0.1594 | 8.0 | 2184 | 0.1446 | 0.7776 | 0.6193 | 0.2534 | 0.001 |
65
- | 0.1594 | 9.0 | 2457 | 0.1479 | 0.7813 | 0.6363 | 0.2454 | 0.001 |
66
- | 0.1567 | 10.0 | 2730 | 0.1425 | 0.7866 | 0.6366 | 0.2607 | 0.001 |
67
- | 0.1559 | 11.0 | 3003 | 0.1454 | 0.7908 | 0.6566 | 0.2618 | 0.001 |
68
- | 0.1559 | 12.0 | 3276 | 0.1404 | 0.7895 | 0.6439 | 0.2600 | 0.001 |
69
- | 0.1549 | 13.0 | 3549 | 0.1414 | 0.7883 | 0.6490 | 0.2531 | 0.001 |
70
- | 0.1549 | 14.0 | 3822 | 0.1406 | 0.7894 | 0.6348 | 0.2649 | 0.001 |
71
- | 0.155 | 15.0 | 4095 | 0.1387 | 0.7906 | 0.6470 | 0.2632 | 0.001 |
72
- | 0.155 | 16.0 | 4368 | 0.1401 | 0.7858 | 0.6331 | 0.2604 | 0.001 |
73
- | 0.1531 | 17.0 | 4641 | 0.1381 | 0.7955 | 0.6651 | 0.2520 | 0.001 |
74
- | 0.1531 | 18.0 | 4914 | 0.1387 | 0.7914 | 0.6498 | 0.2702 | 0.001 |
75
- | 0.1549 | 19.0 | 5187 | 0.1374 | 0.7940 | 0.6356 | 0.2632 | 0.001 |
76
- | 0.1549 | 20.0 | 5460 | 0.1386 | 0.7850 | 0.6393 | 0.2551 | 0.001 |
77
- | 0.1524 | 21.0 | 5733 | 0.1357 | 0.7943 | 0.6496 | 0.2646 | 0.001 |
78
- | 0.1519 | 22.0 | 6006 | 0.1372 | 0.7972 | 0.6400 | 0.2723 | 0.001 |
79
- | 0.1519 | 23.0 | 6279 | 0.1361 | 0.7938 | 0.6595 | 0.2604 | 0.001 |
80
- | 0.1528 | 24.0 | 6552 | 0.1363 | 0.7954 | 0.6418 | 0.2649 | 0.001 |
81
- | 0.1528 | 25.0 | 6825 | 0.1359 | 0.7963 | 0.6500 | 0.2733 | 0.001 |
82
- | 0.1524 | 26.0 | 7098 | 0.1348 | 0.7984 | 0.6555 | 0.2691 | 0.001 |
83
- | 0.1524 | 27.0 | 7371 | 0.1367 | 0.7944 | 0.6535 | 0.2688 | 0.001 |
84
- | 0.1521 | 28.0 | 7644 | 0.1357 | 0.7923 | 0.6398 | 0.2677 | 0.001 |
85
- | 0.1521 | 29.0 | 7917 | 0.1405 | 0.7924 | 0.6637 | 0.2639 | 0.001 |
86
- | 0.1525 | 30.0 | 8190 | 0.1373 | 0.7875 | 0.6405 | 0.2723 | 0.001 |
87
- | 0.1525 | 31.0 | 8463 | 0.1354 | 0.7986 | 0.6562 | 0.2719 | 0.001 |
88
- | 0.1522 | 32.0 | 8736 | 0.1369 | 0.7930 | 0.6463 | 0.2649 | 0.001 |
89
- | 0.1486 | 33.0 | 9009 | 0.1320 | 0.8015 | 0.6660 | 0.2747 | 0.0001 |
90
- | 0.1486 | 34.0 | 9282 | 0.1300 | 0.8043 | 0.6756 | 0.2834 | 0.0001 |
91
- | 0.1419 | 35.0 | 9555 | 0.1295 | 0.8066 | 0.6772 | 0.2810 | 0.0001 |
92
- | 0.1419 | 36.0 | 9828 | 0.1309 | 0.8046 | 0.6795 | 0.2817 | 0.0001 |
93
- | 0.1396 | 37.0 | 10101 | 0.1279 | 0.8054 | 0.6792 | 0.2824 | 0.0001 |
94
- | 0.1396 | 38.0 | 10374 | 0.1290 | 0.8078 | 0.6814 | 0.2848 | 0.0001 |
95
- | 0.1366 | 39.0 | 10647 | 0.1272 | 0.8116 | 0.6833 | 0.2939 | 0.0001 |
96
- | 0.1366 | 40.0 | 10920 | 0.1293 | 0.8116 | 0.6879 | 0.2908 | 0.0001 |
97
- | 0.1361 | 41.0 | 11193 | 0.1270 | 0.8089 | 0.6864 | 0.2918 | 0.0001 |
98
- | 0.1361 | 42.0 | 11466 | 0.1262 | 0.8109 | 0.6837 | 0.2914 | 0.0001 |
99
- | 0.135 | 43.0 | 11739 | 0.1261 | 0.8123 | 0.6984 | 0.2949 | 0.0001 |
100
- | 0.134 | 44.0 | 12012 | 0.1283 | 0.8106 | 0.6834 | 0.2935 | 0.0001 |
101
- | 0.134 | 45.0 | 12285 | 0.1262 | 0.8113 | 0.7010 | 0.2932 | 0.0001 |
102
- | 0.1331 | 46.0 | 12558 | 0.1246 | 0.8147 | 0.6964 | 0.2960 | 0.0001 |
103
- | 0.1331 | 47.0 | 12831 | 0.1253 | 0.8126 | 0.6923 | 0.2988 | 0.0001 |
104
- | 0.1325 | 48.0 | 13104 | 0.1263 | 0.8133 | 0.6954 | 0.2977 | 0.0001 |
105
- | 0.1325 | 49.0 | 13377 | 0.1253 | 0.8158 | 0.6952 | 0.3037 | 0.0001 |
106
- | 0.1312 | 50.0 | 13650 | 0.1263 | 0.8136 | 0.7008 | 0.3005 | 0.0001 |
107
- | 0.1312 | 51.0 | 13923 | 0.1246 | 0.8158 | 0.7019 | 0.3009 | 0.0001 |
108
- | 0.1301 | 52.0 | 14196 | 0.1253 | 0.8092 | 0.6949 | 0.2911 | 0.0001 |
109
- | 0.1301 | 53.0 | 14469 | 0.1244 | 0.8154 | 0.7019 | 0.3023 | 1e-05 |
110
- | 0.1306 | 54.0 | 14742 | 0.1249 | 0.8154 | 0.7040 | 0.3009 | 1e-05 |
111
- | 0.1282 | 55.0 | 15015 | 0.1237 | 0.8144 | 0.6998 | 0.3005 | 1e-05 |
112
- | 0.1282 | 56.0 | 15288 | 0.1235 | 0.8168 | 0.7004 | 0.3033 | 1e-05 |
113
- | 0.1281 | 57.0 | 15561 | 0.1239 | 0.8157 | 0.7002 | 0.3030 | 1e-05 |
114
- | 0.1281 | 58.0 | 15834 | 0.1234 | 0.8157 | 0.6995 | 0.3026 | 1e-05 |
115
- | 0.129 | 59.0 | 16107 | 0.1235 | 0.8150 | 0.7012 | 0.3047 | 1e-05 |
116
- | 0.129 | 60.0 | 16380 | 0.1239 | 0.8128 | 0.6932 | 0.2981 | 1e-05 |
117
- | 0.1284 | 61.0 | 16653 | 0.1240 | 0.8174 | 0.7076 | 0.3009 | 1e-05 |
118
- | 0.1284 | 62.0 | 16926 | 0.1233 | 0.8152 | 0.7032 | 0.3019 | 1e-05 |
119
- | 0.127 | 63.0 | 17199 | 0.1233 | 0.8158 | 0.7023 | 0.3023 | 1e-05 |
120
- | 0.127 | 64.0 | 17472 | 0.1235 | 0.8152 | 0.6999 | 0.3047 | 1e-05 |
121
- | 0.1279 | 65.0 | 17745 | 0.1232 | 0.8145 | 0.7001 | 0.2977 | 1e-05 |
122
- | 0.1273 | 66.0 | 18018 | 0.1229 | 0.8145 | 0.7004 | 0.3012 | 1e-05 |
123
- | 0.1273 | 67.0 | 18291 | 0.1238 | 0.8159 | 0.6993 | 0.3016 | 1e-05 |
124
- | 0.1272 | 68.0 | 18564 | 0.1229 | 0.8176 | 0.7039 | 0.2998 | 1e-05 |
125
- | 0.1272 | 69.0 | 18837 | 0.1230 | 0.8157 | 0.7006 | 0.3058 | 1e-05 |
126
- | 0.1274 | 70.0 | 19110 | 0.1228 | 0.8170 | 0.7009 | 0.3040 | 1e-05 |
127
- | 0.1274 | 71.0 | 19383 | 0.1230 | 0.8158 | 0.7043 | 0.3019 | 1e-05 |
128
- | 0.1272 | 72.0 | 19656 | 0.1232 | 0.8171 | 0.7025 | 0.3016 | 1e-05 |
129
- | 0.1272 | 73.0 | 19929 | 0.1229 | 0.8174 | 0.7049 | 0.3054 | 1e-05 |
130
- | 0.1267 | 74.0 | 20202 | 0.1230 | 0.8141 | 0.6942 | 0.3009 | 1e-05 |
131
- | 0.1267 | 75.0 | 20475 | 0.1232 | 0.8161 | 0.7001 | 0.3033 | 0.0000 |
132
- | 0.1269 | 76.0 | 20748 | 0.1227 | 0.8171 | 0.7020 | 0.3058 | 0.0000 |
133
- | 0.1261 | 77.0 | 21021 | 0.1228 | 0.8191 | 0.7060 | 0.3079 | 0.0000 |
134
- | 0.1261 | 78.0 | 21294 | 0.1237 | 0.8166 | 0.7072 | 0.3019 | 0.0000 |
135
- | 0.1268 | 79.0 | 21567 | 0.1233 | 0.8156 | 0.6992 | 0.3047 | 0.0000 |
136
- | 0.1268 | 80.0 | 21840 | 0.1233 | 0.8172 | 0.6994 | 0.3026 | 0.0000 |
137
- | 0.1271 | 81.0 | 22113 | 0.1224 | 0.8176 | 0.7037 | 0.3054 | 0.0000 |
138
- | 0.1271 | 82.0 | 22386 | 0.1227 | 0.8151 | 0.6972 | 0.3002 | 0.0000 |
139
- | 0.1263 | 83.0 | 22659 | 0.1232 | 0.8146 | 0.6939 | 0.2995 | 0.0000 |
140
- | 0.1263 | 84.0 | 22932 | 0.1226 | 0.8177 | 0.7017 | 0.3026 | 0.0000 |
141
- | 0.1265 | 85.0 | 23205 | 0.1230 | 0.8181 | 0.7059 | 0.3072 | 0.0000 |
142
- | 0.1265 | 86.0 | 23478 | 0.1234 | 0.8168 | 0.7059 | 0.3033 | 0.0000 |
143
- | 0.1265 | 87.0 | 23751 | 0.1236 | 0.8143 | 0.6944 | 0.3005 | 0.0000 |
144
- | 0.1266 | 88.0 | 24024 | 0.1226 | 0.8186 | 0.7029 | 0.3079 | 0.0000 |
145
- | 0.1266 | 89.0 | 24297 | 0.1230 | 0.8179 | 0.7084 | 0.3075 | 0.0000 |
146
- | 0.1263 | 90.0 | 24570 | 0.1232 | 0.8195 | 0.7105 | 0.3037 | 0.0000 |
147
- | 0.1263 | 91.0 | 24843 | 0.1227 | 0.8155 | 0.6972 | 0.3019 | 0.0000 |
148
-
149
-
150
- ### Framework versions
151
-
152
- - Transformers 4.48.0
153
- - Pytorch 2.5.1+cu124
154
- - Datasets 3.0.2
155
- - Tokenizers 0.21.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
+ language:
4
+ - eng
5
+ license: cc0-1.0
6
  tags:
7
+ - multilabel-image-classification
8
+ - multilabel
9
  - generated_from_trainer
10
+ base_model: DinoVdo-large-2025_01_27_45863-bs32_freeze
 
11
  model-index:
12
  - name: DinoVdo-large-2025_01_27_45863-bs32_freeze
13
  results: []
14
  ---
15
 
16
+ DinoVdo is a fine-tuned version of [DinoVdo-large-2025_01_27_45863-bs32_freeze](https://huggingface.co/DinoVdo-large-2025_01_27_45863-bs32_freeze). It achieves the following results on the test set:
 
17
 
 
18
 
 
 
19
  - Loss: 0.1236
20
  - F1 Micro: 0.8136
21
  - F1 Macro: 0.7060
22
  - Accuracy: 0.3071
 
23
 
24
+ | Class | F1 per class |
25
+ |----------|-------|
26
+ | Acropore_branched | 0.9015 |
27
+ | Acropore_digitised | 0.6437 |
28
+ | Acropore_sub_massive | 0.3853 |
29
+ | Acropore_tabular | 0.9293 |
30
+ | Algae_assembly | 0.7615 |
31
+ | Algae_drawn_up | 0.4765 |
32
+ | Algae_limestone | 0.7542 |
33
+ | Algae_sodding | 0.8600 |
34
+ | Atra/Leucospilota | 0.8369 |
35
+ | Bleached_coral | 0.6405 |
36
+ | Blurred | 0.6294 |
37
+ | Dead_coral | 0.7340 |
38
+ | Fish | 0.7477 |
39
+ | Homo_sapiens | 0.7788 |
40
+ | Human_object | 0.7629 |
41
+ | Living_coral | 0.6290 |
42
+ | Millepore | 0.8251 |
43
+ | No_acropore_encrusting | 0.6364 |
44
+ | No_acropore_foliaceous | 0.8077 |
45
+ | No_acropore_massive | 0.7208 |
46
+ | No_acropore_solitary | 0.4468 |
47
+ | No_acropore_sub_massive | 0.6970 |
48
+ | Rock | 0.8818 |
49
+ | Rubble | 0.7686 |
50
+ | Sand | 0.9235 |
51
+ | Sea_cucumber | 0.8234 |
52
+ | Sea_urchins | 0.7079 |
53
+ | Sponge | 0.3861 |
54
+ | Syringodium_isoetifolium | 0.9720 |
55
+ | Thalassodendron_ciliatum | 0.9886 |
56
+ | Useless | 0.9745 |
57
+
58
 
59
+ ---
60
 
61
+ # Model description
62
+ DinoVdo is a model built on top of DinoVdo-large-2025_01_27_45863-bs32_freeze model for underwater multilabel image classification.The classification head is a combination of linear, ReLU, batch normalization, and dropout layers.
63
 
64
+ The source code for training the model can be found in this [Git repository](https://github.com/SeatizenDOI/DinoVdeau).
65
 
66
+ - **Developed by:** [lombardata](https://huggingface.co/lombardata), credits to [César Leblanc](https://huggingface.co/CesarLeblanc) and [Victor Illien](https://huggingface.co/groderg)
67
+
68
+ ---
69
 
70
+ # Intended uses & limitations
71
+ You can use the raw model for classify diverse marine species, encompassing coral morphotypes classes taken from the Global Coral Reef Monitoring Network (GCRMN), habitats classes and seagrass species.
72
 
73
+ ---
74
+
75
+ # Training and evaluation data
76
+ Details on the number of images for each class are given in the following table:
77
+ | Class | train | test | val | Total |
78
+ |:-------------------------|--------:|-------:|------:|--------:|
79
+ | Acropore_branched | 1480 | 469 | 459 | 2408 |
80
+ | Acropore_digitised | 571 | 156 | 161 | 888 |
81
+ | Acropore_sub_massive | 150 | 52 | 41 | 243 |
82
+ | Acropore_tabular | 999 | 292 | 298 | 1589 |
83
+ | Algae_assembly | 2554 | 842 | 842 | 4238 |
84
+ | Algae_drawn_up | 367 | 130 | 123 | 620 |
85
+ | Algae_limestone | 1651 | 562 | 559 | 2772 |
86
+ | Algae_sodding | 3142 | 994 | 981 | 5117 |
87
+ | Atra/Leucospilota | 1084 | 349 | 359 | 1792 |
88
+ | Bleached_coral | 219 | 69 | 72 | 360 |
89
+ | Blurred | 191 | 68 | 61 | 320 |
90
+ | Dead_coral | 1980 | 648 | 636 | 3264 |
91
+ | Fish | 2018 | 661 | 642 | 3321 |
92
+ | Homo_sapiens | 161 | 63 | 58 | 282 |
93
+ | Human_object | 156 | 55 | 59 | 270 |
94
+ | Living_coral | 397 | 151 | 153 | 701 |
95
+ | Millepore | 386 | 127 | 124 | 637 |
96
+ | No_acropore_encrusting | 442 | 141 | 142 | 725 |
97
+ | No_acropore_foliaceous | 204 | 47 | 35 | 286 |
98
+ | No_acropore_massive | 1030 | 341 | 334 | 1705 |
99
+ | No_acropore_solitary | 202 | 55 | 46 | 303 |
100
+ | No_acropore_sub_massive | 1402 | 428 | 426 | 2256 |
101
+ | Rock | 4481 | 1495 | 1481 | 7457 |
102
+ | Rubble | 3092 | 1015 | 1016 | 5123 |
103
+ | Sand | 5839 | 1945 | 1935 | 9719 |
104
+ | Sea_cucumber | 1407 | 437 | 450 | 2294 |
105
+ | Sea_urchins | 328 | 110 | 107 | 545 |
106
+ | Sponge | 267 | 98 | 105 | 470 |
107
+ | Syringodium_isoetifolium | 1213 | 392 | 390 | 1995 |
108
+ | Thalassodendron_ciliatum | 781 | 262 | 260 | 1303 |
109
+ | Useless | 579 | 193 | 193 | 965 |
110
+
111
+ ---
112
 
113
+ # Training procedure
114
+
115
+ ## Training hyperparameters
116
 
117
  The following hyperparameters were used during training:
118
+
119
+ - **Number of Epochs**: 91.0
120
+ - **Learning Rate**: 0.001
121
+ - **Train Batch Size**: 32
122
+ - **Eval Batch Size**: 32
123
+ - **Optimizer**: Adam
124
+ - **LR Scheduler Type**: ReduceLROnPlateau with a patience of 5 epochs and a factor of 0.1
125
+ - **Freeze Encoder**: Yes
126
+ - **Data Augmentation**: Yes
127
+
128
+
129
+ ## Data Augmentation
130
+ Data were augmented using the following transformations :
131
+
132
+ Train Transforms
133
+ - **PreProcess**: No additional parameters
134
+ - **Resize**: probability=1.00
135
+ - **RandomHorizontalFlip**: probability=0.25
136
+ - **RandomVerticalFlip**: probability=0.25
137
+ - **ColorJiggle**: probability=0.25
138
+ - **RandomPerspective**: probability=0.25
139
+ - **Normalize**: probability=1.00
140
+
141
+ Val Transforms
142
+ - **PreProcess**: No additional parameters
143
+ - **Resize**: probability=1.00
144
+ - **Normalize**: probability=1.00
145
+
146
+
147
+
148
+ ## Training results
149
+ Epoch | Validation Loss | Accuracy | F1 Macro | F1 Micro | Learning Rate
150
+ --- | --- | --- | --- | --- | ---
151
+ 1 | 0.16775080561637878 | 0.2293 | 0.7433 | 0.5138 | 0.001
152
+ 2 | 0.15366077423095703 | 0.2436 | 0.7614 | 0.5771 | 0.001
153
+ 3 | 0.14831368625164032 | 0.2415 | 0.7763 | 0.6194 | 0.001
154
+ 4 | 0.14640773832798004 | 0.2555 | 0.7808 | 0.6276 | 0.001
155
+ 5 | 0.14515382051467896 | 0.2520 | 0.7788 | 0.6421 | 0.001
156
+ 6 | 0.14404040575027466 | 0.2579 | 0.7802 | 0.6147 | 0.001
157
+ 7 | 0.14518576860427856 | 0.2503 | 0.7787 | 0.6141 | 0.001
158
+ 8 | 0.14463570713996887 | 0.2534 | 0.7776 | 0.6193 | 0.001
159
+ 9 | 0.14786836504936218 | 0.2454 | 0.7813 | 0.6363 | 0.001
160
+ 10 | 0.14251072704792023 | 0.2607 | 0.7866 | 0.6366 | 0.001
161
+ 11 | 0.14535894989967346 | 0.2618 | 0.7908 | 0.6566 | 0.001
162
+ 12 | 0.14042578637599945 | 0.2600 | 0.7895 | 0.6439 | 0.001
163
+ 13 | 0.1413952261209488 | 0.2531 | 0.7883 | 0.6490 | 0.001
164
+ 14 | 0.14056158065795898 | 0.2649 | 0.7894 | 0.6348 | 0.001
165
+ 15 | 0.13873133063316345 | 0.2632 | 0.7906 | 0.6470 | 0.001
166
+ 16 | 0.14008578658103943 | 0.2604 | 0.7858 | 0.6331 | 0.001
167
+ 17 | 0.13811993598937988 | 0.2520 | 0.7955 | 0.6651 | 0.001
168
+ 18 | 0.13870170712471008 | 0.2702 | 0.7914 | 0.6498 | 0.001
169
+ 19 | 0.1373777985572815 | 0.2632 | 0.7940 | 0.6356 | 0.001
170
+ 20 | 0.13864819705486298 | 0.2551 | 0.7850 | 0.6393 | 0.001
171
+ 21 | 0.13566707074642181 | 0.2646 | 0.7943 | 0.6496 | 0.001
172
+ 22 | 0.1371580958366394 | 0.2723 | 0.7972 | 0.6400 | 0.001
173
+ 23 | 0.13614478707313538 | 0.2604 | 0.7938 | 0.6595 | 0.001
174
+ 24 | 0.1362675279378891 | 0.2649 | 0.7954 | 0.6418 | 0.001
175
+ 25 | 0.1358671337366104 | 0.2733 | 0.7963 | 0.6500 | 0.001
176
+ 26 | 0.13484793901443481 | 0.2691 | 0.7984 | 0.6555 | 0.001
177
+ 27 | 0.13669784367084503 | 0.2688 | 0.7944 | 0.6535 | 0.001
178
+ 28 | 0.13569706678390503 | 0.2677 | 0.7923 | 0.6398 | 0.001
179
+ 29 | 0.14052562415599823 | 0.2639 | 0.7924 | 0.6637 | 0.001
180
+ 30 | 0.13725879788398743 | 0.2723 | 0.7875 | 0.6405 | 0.001
181
+ 31 | 0.13544805347919464 | 0.2719 | 0.7986 | 0.6562 | 0.001
182
+ 32 | 0.13693773746490479 | 0.2649 | 0.7930 | 0.6463 | 0.001
183
+ 33 | 0.13195939362049103 | 0.2747 | 0.8015 | 0.6660 | 0.0001
184
+ 34 | 0.1299898624420166 | 0.2834 | 0.8043 | 0.6756 | 0.0001
185
+ 35 | 0.12946264445781708 | 0.2810 | 0.8066 | 0.6772 | 0.0001
186
+ 36 | 0.13086125254631042 | 0.2817 | 0.8046 | 0.6795 | 0.0001
187
+ 37 | 0.1278763711452484 | 0.2824 | 0.8054 | 0.6792 | 0.0001
188
+ 38 | 0.12904110550880432 | 0.2848 | 0.8078 | 0.6814 | 0.0001
189
+ 39 | 0.12716704607009888 | 0.2939 | 0.8116 | 0.6833 | 0.0001
190
+ 40 | 0.1293308585882187 | 0.2908 | 0.8116 | 0.6879 | 0.0001
191
+ 41 | 0.12695887684822083 | 0.2918 | 0.8089 | 0.6864 | 0.0001
192
+ 42 | 0.12624548375606537 | 0.2914 | 0.8109 | 0.6837 | 0.0001
193
+ 43 | 0.1261172592639923 | 0.2949 | 0.8123 | 0.6984 | 0.0001
194
+ 44 | 0.12830273807048798 | 0.2935 | 0.8106 | 0.6834 | 0.0001
195
+ 45 | 0.12624593079090118 | 0.2932 | 0.8113 | 0.7010 | 0.0001
196
+ 46 | 0.12462077289819717 | 0.2960 | 0.8147 | 0.6964 | 0.0001
197
+ 47 | 0.12529432773590088 | 0.2988 | 0.8126 | 0.6923 | 0.0001
198
+ 48 | 0.12631145119667053 | 0.2977 | 0.8133 | 0.6954 | 0.0001
199
+ 49 | 0.1252526491880417 | 0.3037 | 0.8158 | 0.6952 | 0.0001
200
+ 50 | 0.12632089853286743 | 0.3005 | 0.8136 | 0.7008 | 0.0001
201
+ 51 | 0.1246422603726387 | 0.3009 | 0.8158 | 0.7019 | 0.0001
202
+ 52 | 0.12534211575984955 | 0.2911 | 0.8092 | 0.6949 | 0.0001
203
+ 53 | 0.12436465919017792 | 0.3023 | 0.8154 | 0.7019 | 1e-05
204
+ 54 | 0.12488020956516266 | 0.3009 | 0.8154 | 0.7040 | 1e-05
205
+ 55 | 0.12366042286157608 | 0.3005 | 0.8144 | 0.6998 | 1e-05
206
+ 56 | 0.12352865934371948 | 0.3033 | 0.8168 | 0.7004 | 1e-05
207
+ 57 | 0.1239086389541626 | 0.3030 | 0.8157 | 0.7002 | 1e-05
208
+ 58 | 0.12343526631593704 | 0.3026 | 0.8157 | 0.6995 | 1e-05
209
+ 59 | 0.12345146387815475 | 0.3047 | 0.8150 | 0.7012 | 1e-05
210
+ 60 | 0.1239377036690712 | 0.2981 | 0.8128 | 0.6932 | 1e-05
211
+ 61 | 0.12398885935544968 | 0.3009 | 0.8174 | 0.7076 | 1e-05
212
+ 62 | 0.12334412336349487 | 0.3019 | 0.8152 | 0.7032 | 1e-05
213
+ 63 | 0.12325507402420044 | 0.3023 | 0.8158 | 0.7023 | 1e-05
214
+ 64 | 0.12346883863210678 | 0.3047 | 0.8152 | 0.6999 | 1e-05
215
+ 65 | 0.12324482202529907 | 0.2977 | 0.8145 | 0.7001 | 1e-05
216
+ 66 | 0.12292143702507019 | 0.3012 | 0.8145 | 0.7004 | 1e-05
217
+ 67 | 0.12375594675540924 | 0.3016 | 0.8159 | 0.6993 | 1e-05
218
+ 68 | 0.1228519007563591 | 0.2998 | 0.8176 | 0.7039 | 1e-05
219
+ 69 | 0.12302352488040924 | 0.3058 | 0.8157 | 0.7006 | 1e-05
220
+ 70 | 0.12284138053655624 | 0.3040 | 0.8170 | 0.7009 | 1e-05
221
+ 71 | 0.12295401096343994 | 0.3019 | 0.8158 | 0.7043 | 1e-05
222
+ 72 | 0.123215451836586 | 0.3016 | 0.8171 | 0.7025 | 1e-05
223
+ 73 | 0.12291014939546585 | 0.3054 | 0.8174 | 0.7049 | 1e-05
224
+ 74 | 0.12304174154996872 | 0.3009 | 0.8141 | 0.6942 | 1e-05
225
+ 75 | 0.1232200339436531 | 0.3033 | 0.8161 | 0.7001 | 1.0000000000000002e-06
226
+ 76 | 0.12267689406871796 | 0.3058 | 0.8171 | 0.7020 | 1.0000000000000002e-06
227
+ 77 | 0.12284990400075912 | 0.3079 | 0.8191 | 0.7060 | 1.0000000000000002e-06
228
+ 78 | 0.123690165579319 | 0.3019 | 0.8166 | 0.7072 | 1.0000000000000002e-06
229
+ 79 | 0.12329532951116562 | 0.3047 | 0.8156 | 0.6992 | 1.0000000000000002e-06
230
+ 80 | 0.12325812131166458 | 0.3026 | 0.8172 | 0.6994 | 1.0000000000000002e-06
231
+ 81 | 0.12240613251924515 | 0.3054 | 0.8176 | 0.7037 | 1.0000000000000002e-06
232
+ 82 | 0.12270382046699524 | 0.3002 | 0.8151 | 0.6972 | 1.0000000000000002e-06
233
+ 83 | 0.12315402179956436 | 0.2995 | 0.8146 | 0.6939 | 1.0000000000000002e-06
234
+ 84 | 0.1225922629237175 | 0.3026 | 0.8177 | 0.7017 | 1.0000000000000002e-06
235
+ 85 | 0.1230376735329628 | 0.3072 | 0.8181 | 0.7059 | 1.0000000000000002e-06
236
+ 86 | 0.12335028499364853 | 0.3033 | 0.8168 | 0.7059 | 1.0000000000000002e-06
237
+ 87 | 0.12355341017246246 | 0.3005 | 0.8143 | 0.6944 | 1.0000000000000002e-06
238
+ 88 | 0.12264065444469452 | 0.3079 | 0.8186 | 0.7029 | 1.0000000000000002e-07
239
+ 89 | 0.1229795441031456 | 0.3075 | 0.8179 | 0.7084 | 1.0000000000000002e-07
240
+ 90 | 0.12317965924739838 | 0.3037 | 0.8195 | 0.7105 | 1.0000000000000002e-07
241
+ 91 | 0.12267619371414185 | 0.3019 | 0.8155 | 0.6972 | 1.0000000000000002e-07
242
+
243
+
244
+ ---
245
+
246
+ # Framework Versions
247
+
248
+ - **Transformers**: 4.48.0
249
+ - **Pytorch**: 2.5.1+cu124
250
+ - **Datasets**: 3.0.2
251
+ - **Tokenizers**: 0.21.0
252
+