encku commited on
Commit
594141c
·
verified ·
1 Parent(s): b79cfc6

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - transformers
6
+ - image-classification
7
+ base_model: google/vit-large-patch16-384
8
+ widget:
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
10
+ example_title: Tiger
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
12
+ example_title: Teapot
13
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
14
+ example_title: Palace
15
+ ---
16
+
17
+ # Model Trained Using AutoTrain
18
+
19
+ - Problem type: Image Classification
20
+
21
+ ## Validation Metrics
22
+ loss: 0.0007826363435015082
23
+
24
+ f1_macro: 0.9998667225525736
25
+
26
+ f1_micro: 0.9998677353649181
27
+
28
+ f1_weighted: 0.9998677517242137
29
+
30
+ precision_macro: 0.9998635579506288
31
+
32
+ precision_micro: 0.9998677353649181
33
+
34
+ precision_weighted: 0.999868129449558
35
+
36
+ recall_macro: 0.9998702523325717
37
+
38
+ recall_micro: 0.9998677353649181
39
+
40
+ recall_weighted: 0.9998677353649181
41
+
42
+ accuracy: 0.9998677353649181
config.json ADDED
@@ -0,0 +1,235 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-large-patch16-384",
3
+ "_num_labels": 103,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "6974202725334",
14
+ "1": "C001",
15
+ "2": "C002",
16
+ "3": "C003",
17
+ "4": "C004",
18
+ "5": "C005",
19
+ "6": "C006",
20
+ "7": "C007",
21
+ "8": "C008",
22
+ "9": "C009",
23
+ "10": "C010",
24
+ "11": "C012",
25
+ "12": "C014",
26
+ "13": "C015",
27
+ "14": "C017",
28
+ "15": "C018",
29
+ "16": "C020",
30
+ "17": "C021",
31
+ "18": "C022",
32
+ "19": "C023",
33
+ "20": "C024",
34
+ "21": "C025",
35
+ "22": "C026",
36
+ "23": "C027",
37
+ "24": "C028",
38
+ "25": "C029",
39
+ "26": "C030",
40
+ "27": "C031",
41
+ "28": "C032",
42
+ "29": "C033",
43
+ "30": "C034",
44
+ "31": "C035",
45
+ "32": "C036",
46
+ "33": "C037",
47
+ "34": "C038",
48
+ "35": "C039",
49
+ "36": "C040",
50
+ "37": "C041",
51
+ "38": "C042",
52
+ "39": "C043",
53
+ "40": "C044",
54
+ "41": "TBRG067",
55
+ "42": "TBRG068",
56
+ "43": "TBRG072",
57
+ "44": "TBRG073",
58
+ "45": "TBRG074",
59
+ "46": "TBRG075",
60
+ "47": "TBRG085",
61
+ "48": "TBRG086",
62
+ "49": "TBRG087",
63
+ "50": "TBRG090",
64
+ "51": "TBRG092",
65
+ "52": "TBRG093",
66
+ "53": "TBRG096",
67
+ "54": "TBRG097",
68
+ "55": "TBRG098",
69
+ "56": "TBRG100",
70
+ "57": "TBRG156",
71
+ "58": "TBRG157",
72
+ "59": "TBRG158",
73
+ "60": "TBRG159",
74
+ "61": "TT00277",
75
+ "62": "TT00523",
76
+ "63": "TT00677",
77
+ "64": "TT00685",
78
+ "65": "TT00765",
79
+ "66": "TT00792",
80
+ "67": "TT00793",
81
+ "68": "TT00810",
82
+ "69": "TT00811",
83
+ "70": "TT00812",
84
+ "71": "TT00853",
85
+ "72": "TT00854",
86
+ "73": "TT00857",
87
+ "74": "TT00859",
88
+ "75": "TT00875",
89
+ "76": "TT00876",
90
+ "77": "TT00944",
91
+ "78": "TT00947",
92
+ "79": "TT00964",
93
+ "80": "TT00980",
94
+ "81": "TT01001",
95
+ "82": "TT01037",
96
+ "83": "TT01069",
97
+ "84": "TT01070",
98
+ "85": "TT01071",
99
+ "86": "TT01072",
100
+ "87": "TT01142",
101
+ "88": "TT01148",
102
+ "89": "TT01149",
103
+ "90": "TT01162",
104
+ "91": "TT01169",
105
+ "92": "TT01172",
106
+ "93": "TT01174",
107
+ "94": "TT01178",
108
+ "95": "TT01231",
109
+ "96": "TT01276",
110
+ "97": "TT01277",
111
+ "98": "TT01300",
112
+ "99": "TT01307",
113
+ "100": "TT01431",
114
+ "101": "TT01460",
115
+ "102": "TT01482"
116
+ },
117
+ "image_size": 384,
118
+ "initializer_range": 0.02,
119
+ "intermediate_size": 4096,
120
+ "label2id": {
121
+ "6974202725334": 0,
122
+ "C001": 1,
123
+ "C002": 2,
124
+ "C003": 3,
125
+ "C004": 4,
126
+ "C005": 5,
127
+ "C006": 6,
128
+ "C007": 7,
129
+ "C008": 8,
130
+ "C009": 9,
131
+ "C010": 10,
132
+ "C012": 11,
133
+ "C014": 12,
134
+ "C015": 13,
135
+ "C017": 14,
136
+ "C018": 15,
137
+ "C020": 16,
138
+ "C021": 17,
139
+ "C022": 18,
140
+ "C023": 19,
141
+ "C024": 20,
142
+ "C025": 21,
143
+ "C026": 22,
144
+ "C027": 23,
145
+ "C028": 24,
146
+ "C029": 25,
147
+ "C030": 26,
148
+ "C031": 27,
149
+ "C032": 28,
150
+ "C033": 29,
151
+ "C034": 30,
152
+ "C035": 31,
153
+ "C036": 32,
154
+ "C037": 33,
155
+ "C038": 34,
156
+ "C039": 35,
157
+ "C040": 36,
158
+ "C041": 37,
159
+ "C042": 38,
160
+ "C043": 39,
161
+ "C044": 40,
162
+ "TBRG067": 41,
163
+ "TBRG068": 42,
164
+ "TBRG072": 43,
165
+ "TBRG073": 44,
166
+ "TBRG074": 45,
167
+ "TBRG075": 46,
168
+ "TBRG085": 47,
169
+ "TBRG086": 48,
170
+ "TBRG087": 49,
171
+ "TBRG090": 50,
172
+ "TBRG092": 51,
173
+ "TBRG093": 52,
174
+ "TBRG096": 53,
175
+ "TBRG097": 54,
176
+ "TBRG098": 55,
177
+ "TBRG100": 56,
178
+ "TBRG156": 57,
179
+ "TBRG157": 58,
180
+ "TBRG158": 59,
181
+ "TBRG159": 60,
182
+ "TT00277": 61,
183
+ "TT00523": 62,
184
+ "TT00677": 63,
185
+ "TT00685": 64,
186
+ "TT00765": 65,
187
+ "TT00792": 66,
188
+ "TT00793": 67,
189
+ "TT00810": 68,
190
+ "TT00811": 69,
191
+ "TT00812": 70,
192
+ "TT00853": 71,
193
+ "TT00854": 72,
194
+ "TT00857": 73,
195
+ "TT00859": 74,
196
+ "TT00875": 75,
197
+ "TT00876": 76,
198
+ "TT00944": 77,
199
+ "TT00947": 78,
200
+ "TT00964": 79,
201
+ "TT00980": 80,
202
+ "TT01001": 81,
203
+ "TT01037": 82,
204
+ "TT01069": 83,
205
+ "TT01070": 84,
206
+ "TT01071": 85,
207
+ "TT01072": 86,
208
+ "TT01142": 87,
209
+ "TT01148": 88,
210
+ "TT01149": 89,
211
+ "TT01162": 90,
212
+ "TT01169": 91,
213
+ "TT01172": 92,
214
+ "TT01174": 93,
215
+ "TT01178": 94,
216
+ "TT01231": 95,
217
+ "TT01276": 96,
218
+ "TT01277": 97,
219
+ "TT01300": 98,
220
+ "TT01307": 99,
221
+ "TT01431": 100,
222
+ "TT01460": 101,
223
+ "TT01482": 102
224
+ },
225
+ "layer_norm_eps": 1e-12,
226
+ "model_type": "vit",
227
+ "num_attention_heads": 16,
228
+ "num_channels": 3,
229
+ "num_hidden_layers": 24,
230
+ "patch_size": 16,
231
+ "problem_type": "single_label_classification",
232
+ "qkv_bias": true,
233
+ "torch_dtype": "float32",
234
+ "transformers_version": "4.48.0"
235
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c223d811c4432b298394839c20173971e4125e7c5cabda36335fcda543f318b0
3
+ size 1215231868
preprocessor_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "ViTImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "resample": 2,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "height": 384,
21
+ "width": 384
22
+ }
23
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d15ec7b6e8d8d9fa0f6119b21a7a77a56e4e58ba2b24f23d7484cf5cb8353c6
3
+ size 5713
training_params.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "project-name/autotrain-data",
3
+ "model": "google/vit-large-patch16-384",
4
+ "username": null,
5
+ "lr": 0.00005,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "project-name",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": null,
21
+ "save_total_limit": 1,
22
+ "token": null,
23
+ "push_to_hub": false,
24
+ "eval_strategy": "epoch",
25
+ "image_column": "autotrain_image",
26
+ "target_column": "autotrain_label",
27
+ "log": "none",
28
+ "early_stopping_patience": 5,
29
+ "early_stopping_threshold": 0.01
30
+ }