LucasThil commited on
Commit
9dbb164
·
verified ·
1 Parent(s): cfad5ee

Training in progress, epoch 1

Browse files
config.json ADDED
@@ -0,0 +1,453 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/dinov2-small",
3
+ "apply_layernorm": true,
4
+ "architectures": [
5
+ "Dinov2ForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "drop_path_rate": 0.0,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 384,
12
+ "id2label": {
13
+ "0": "n01443537",
14
+ "1": "n01629819",
15
+ "2": "n01641577",
16
+ "3": "n01644900",
17
+ "4": "n01698640",
18
+ "5": "n01742172",
19
+ "6": "n01768244",
20
+ "7": "n01770393",
21
+ "8": "n01774384",
22
+ "9": "n01774750",
23
+ "10": "n01784675",
24
+ "11": "n01882714",
25
+ "12": "n01910747",
26
+ "13": "n01917289",
27
+ "14": "n01944390",
28
+ "15": "n01950731",
29
+ "16": "n01983481",
30
+ "17": "n01984695",
31
+ "18": "n02002724",
32
+ "19": "n02056570",
33
+ "20": "n02058221",
34
+ "21": "n02074367",
35
+ "22": "n02094433",
36
+ "23": "n02099601",
37
+ "24": "n02099712",
38
+ "25": "n02106662",
39
+ "26": "n02113799",
40
+ "27": "n02123045",
41
+ "28": "n02123394",
42
+ "29": "n02124075",
43
+ "30": "n02125311",
44
+ "31": "n02129165",
45
+ "32": "n02132136",
46
+ "33": "n02165456",
47
+ "34": "n02226429",
48
+ "35": "n02231487",
49
+ "36": "n02233338",
50
+ "37": "n02236044",
51
+ "38": "n02268443",
52
+ "39": "n02279972",
53
+ "40": "n02281406",
54
+ "41": "n02321529",
55
+ "42": "n02364673",
56
+ "43": "n02395406",
57
+ "44": "n02403003",
58
+ "45": "n02410509",
59
+ "46": "n02415577",
60
+ "47": "n02423022",
61
+ "48": "n02437312",
62
+ "49": "n02480495",
63
+ "50": "n02481823",
64
+ "51": "n02486410",
65
+ "52": "n02504458",
66
+ "53": "n02509815",
67
+ "54": "n02666347",
68
+ "55": "n02669723",
69
+ "56": "n02699494",
70
+ "57": "n02769748",
71
+ "58": "n02788148",
72
+ "59": "n02791270",
73
+ "60": "n02793495",
74
+ "61": "n02795169",
75
+ "62": "n02802426",
76
+ "63": "n02808440",
77
+ "64": "n02814533",
78
+ "65": "n02814860",
79
+ "66": "n02815834",
80
+ "67": "n02823428",
81
+ "68": "n02837789",
82
+ "69": "n02841315",
83
+ "70": "n02843684",
84
+ "71": "n02883205",
85
+ "72": "n02892201",
86
+ "73": "n02909870",
87
+ "74": "n02917067",
88
+ "75": "n02927161",
89
+ "76": "n02948072",
90
+ "77": "n02950826",
91
+ "78": "n02963159",
92
+ "79": "n02977058",
93
+ "80": "n02988304",
94
+ "81": "n03014705",
95
+ "82": "n03026506",
96
+ "83": "n03042490",
97
+ "84": "n03085013",
98
+ "85": "n03089624",
99
+ "86": "n03100240",
100
+ "87": "n03126707",
101
+ "88": "n03160309",
102
+ "89": "n03179701",
103
+ "90": "n03201208",
104
+ "91": "n03255030",
105
+ "92": "n03355925",
106
+ "93": "n03373237",
107
+ "94": "n03388043",
108
+ "95": "n03393912",
109
+ "96": "n03400231",
110
+ "97": "n03404251",
111
+ "98": "n03424325",
112
+ "99": "n03444034",
113
+ "100": "n03447447",
114
+ "101": "n03544143",
115
+ "102": "n03584254",
116
+ "103": "n03599486",
117
+ "104": "n03617480",
118
+ "105": "n03637318",
119
+ "106": "n03649909",
120
+ "107": "n03662601",
121
+ "108": "n03670208",
122
+ "109": "n03706229",
123
+ "110": "n03733131",
124
+ "111": "n03763968",
125
+ "112": "n03770439",
126
+ "113": "n03796401",
127
+ "114": "n03814639",
128
+ "115": "n03837869",
129
+ "116": "n03838899",
130
+ "117": "n03854065",
131
+ "118": "n03891332",
132
+ "119": "n03902125",
133
+ "120": "n03930313",
134
+ "121": "n03937543",
135
+ "122": "n03970156",
136
+ "123": "n03977966",
137
+ "124": "n03980874",
138
+ "125": "n03983396",
139
+ "126": "n03992509",
140
+ "127": "n04008634",
141
+ "128": "n04023962",
142
+ "129": "n04070727",
143
+ "130": "n04074963",
144
+ "131": "n04099969",
145
+ "132": "n04118538",
146
+ "133": "n04133789",
147
+ "134": "n04146614",
148
+ "135": "n04149813",
149
+ "136": "n04179913",
150
+ "137": "n04251144",
151
+ "138": "n04254777",
152
+ "139": "n04259630",
153
+ "140": "n04265275",
154
+ "141": "n04275548",
155
+ "142": "n04285008",
156
+ "143": "n04311004",
157
+ "144": "n04328186",
158
+ "145": "n04356056",
159
+ "146": "n04366367",
160
+ "147": "n04371430",
161
+ "148": "n04376876",
162
+ "149": "n04398044",
163
+ "150": "n04399382",
164
+ "151": "n04417672",
165
+ "152": "n04456115",
166
+ "153": "n04465666",
167
+ "154": "n04486054",
168
+ "155": "n04487081",
169
+ "156": "n04501370",
170
+ "157": "n04507155",
171
+ "158": "n04532106",
172
+ "159": "n04532670",
173
+ "160": "n04540053",
174
+ "161": "n04560804",
175
+ "162": "n04562935",
176
+ "163": "n04596742",
177
+ "164": "n04598010",
178
+ "165": "n06596364",
179
+ "166": "n07056680",
180
+ "167": "n07583066",
181
+ "168": "n07614500",
182
+ "169": "n07615774",
183
+ "170": "n07646821",
184
+ "171": "n07647870",
185
+ "172": "n07657664",
186
+ "173": "n07695742",
187
+ "174": "n07711569",
188
+ "175": "n07715103",
189
+ "176": "n07720875",
190
+ "177": "n07749582",
191
+ "178": "n07753592",
192
+ "179": "n07768694",
193
+ "180": "n07871810",
194
+ "181": "n07873807",
195
+ "182": "n07875152",
196
+ "183": "n07920052",
197
+ "184": "n07975909",
198
+ "185": "n08496334",
199
+ "186": "n08620881",
200
+ "187": "n08742578",
201
+ "188": "n09193705",
202
+ "189": "n09246464",
203
+ "190": "n09256479",
204
+ "191": "n09332890",
205
+ "192": "n09428293",
206
+ "193": "n12267677",
207
+ "194": "n12520864",
208
+ "195": "n13001041",
209
+ "196": "n13652335",
210
+ "197": "n13652994",
211
+ "198": "n13719102",
212
+ "199": "n14991210"
213
+ },
214
+ "image_size": 518,
215
+ "initializer_range": 0.02,
216
+ "label2id": {
217
+ "n01443537": 0,
218
+ "n01629819": 1,
219
+ "n01641577": 2,
220
+ "n01644900": 3,
221
+ "n01698640": 4,
222
+ "n01742172": 5,
223
+ "n01768244": 6,
224
+ "n01770393": 7,
225
+ "n01774384": 8,
226
+ "n01774750": 9,
227
+ "n01784675": 10,
228
+ "n01882714": 11,
229
+ "n01910747": 12,
230
+ "n01917289": 13,
231
+ "n01944390": 14,
232
+ "n01950731": 15,
233
+ "n01983481": 16,
234
+ "n01984695": 17,
235
+ "n02002724": 18,
236
+ "n02056570": 19,
237
+ "n02058221": 20,
238
+ "n02074367": 21,
239
+ "n02094433": 22,
240
+ "n02099601": 23,
241
+ "n02099712": 24,
242
+ "n02106662": 25,
243
+ "n02113799": 26,
244
+ "n02123045": 27,
245
+ "n02123394": 28,
246
+ "n02124075": 29,
247
+ "n02125311": 30,
248
+ "n02129165": 31,
249
+ "n02132136": 32,
250
+ "n02165456": 33,
251
+ "n02226429": 34,
252
+ "n02231487": 35,
253
+ "n02233338": 36,
254
+ "n02236044": 37,
255
+ "n02268443": 38,
256
+ "n02279972": 39,
257
+ "n02281406": 40,
258
+ "n02321529": 41,
259
+ "n02364673": 42,
260
+ "n02395406": 43,
261
+ "n02403003": 44,
262
+ "n02410509": 45,
263
+ "n02415577": 46,
264
+ "n02423022": 47,
265
+ "n02437312": 48,
266
+ "n02480495": 49,
267
+ "n02481823": 50,
268
+ "n02486410": 51,
269
+ "n02504458": 52,
270
+ "n02509815": 53,
271
+ "n02666347": 54,
272
+ "n02669723": 55,
273
+ "n02699494": 56,
274
+ "n02769748": 57,
275
+ "n02788148": 58,
276
+ "n02791270": 59,
277
+ "n02793495": 60,
278
+ "n02795169": 61,
279
+ "n02802426": 62,
280
+ "n02808440": 63,
281
+ "n02814533": 64,
282
+ "n02814860": 65,
283
+ "n02815834": 66,
284
+ "n02823428": 67,
285
+ "n02837789": 68,
286
+ "n02841315": 69,
287
+ "n02843684": 70,
288
+ "n02883205": 71,
289
+ "n02892201": 72,
290
+ "n02909870": 73,
291
+ "n02917067": 74,
292
+ "n02927161": 75,
293
+ "n02948072": 76,
294
+ "n02950826": 77,
295
+ "n02963159": 78,
296
+ "n02977058": 79,
297
+ "n02988304": 80,
298
+ "n03014705": 81,
299
+ "n03026506": 82,
300
+ "n03042490": 83,
301
+ "n03085013": 84,
302
+ "n03089624": 85,
303
+ "n03100240": 86,
304
+ "n03126707": 87,
305
+ "n03160309": 88,
306
+ "n03179701": 89,
307
+ "n03201208": 90,
308
+ "n03255030": 91,
309
+ "n03355925": 92,
310
+ "n03373237": 93,
311
+ "n03388043": 94,
312
+ "n03393912": 95,
313
+ "n03400231": 96,
314
+ "n03404251": 97,
315
+ "n03424325": 98,
316
+ "n03444034": 99,
317
+ "n03447447": 100,
318
+ "n03544143": 101,
319
+ "n03584254": 102,
320
+ "n03599486": 103,
321
+ "n03617480": 104,
322
+ "n03637318": 105,
323
+ "n03649909": 106,
324
+ "n03662601": 107,
325
+ "n03670208": 108,
326
+ "n03706229": 109,
327
+ "n03733131": 110,
328
+ "n03763968": 111,
329
+ "n03770439": 112,
330
+ "n03796401": 113,
331
+ "n03814639": 114,
332
+ "n03837869": 115,
333
+ "n03838899": 116,
334
+ "n03854065": 117,
335
+ "n03891332": 118,
336
+ "n03902125": 119,
337
+ "n03930313": 120,
338
+ "n03937543": 121,
339
+ "n03970156": 122,
340
+ "n03977966": 123,
341
+ "n03980874": 124,
342
+ "n03983396": 125,
343
+ "n03992509": 126,
344
+ "n04008634": 127,
345
+ "n04023962": 128,
346
+ "n04070727": 129,
347
+ "n04074963": 130,
348
+ "n04099969": 131,
349
+ "n04118538": 132,
350
+ "n04133789": 133,
351
+ "n04146614": 134,
352
+ "n04149813": 135,
353
+ "n04179913": 136,
354
+ "n04251144": 137,
355
+ "n04254777": 138,
356
+ "n04259630": 139,
357
+ "n04265275": 140,
358
+ "n04275548": 141,
359
+ "n04285008": 142,
360
+ "n04311004": 143,
361
+ "n04328186": 144,
362
+ "n04356056": 145,
363
+ "n04366367": 146,
364
+ "n04371430": 147,
365
+ "n04376876": 148,
366
+ "n04398044": 149,
367
+ "n04399382": 150,
368
+ "n04417672": 151,
369
+ "n04456115": 152,
370
+ "n04465666": 153,
371
+ "n04486054": 154,
372
+ "n04487081": 155,
373
+ "n04501370": 156,
374
+ "n04507155": 157,
375
+ "n04532106": 158,
376
+ "n04532670": 159,
377
+ "n04540053": 160,
378
+ "n04560804": 161,
379
+ "n04562935": 162,
380
+ "n04596742": 163,
381
+ "n04598010": 164,
382
+ "n06596364": 165,
383
+ "n07056680": 166,
384
+ "n07583066": 167,
385
+ "n07614500": 168,
386
+ "n07615774": 169,
387
+ "n07646821": 170,
388
+ "n07647870": 171,
389
+ "n07657664": 172,
390
+ "n07695742": 173,
391
+ "n07711569": 174,
392
+ "n07715103": 175,
393
+ "n07720875": 176,
394
+ "n07749582": 177,
395
+ "n07753592": 178,
396
+ "n07768694": 179,
397
+ "n07871810": 180,
398
+ "n07873807": 181,
399
+ "n07875152": 182,
400
+ "n07920052": 183,
401
+ "n07975909": 184,
402
+ "n08496334": 185,
403
+ "n08620881": 186,
404
+ "n08742578": 187,
405
+ "n09193705": 188,
406
+ "n09246464": 189,
407
+ "n09256479": 190,
408
+ "n09332890": 191,
409
+ "n09428293": 192,
410
+ "n12267677": 193,
411
+ "n12520864": 194,
412
+ "n13001041": 195,
413
+ "n13652335": 196,
414
+ "n13652994": 197,
415
+ "n13719102": 198,
416
+ "n14991210": 199
417
+ },
418
+ "layer_norm_eps": 1e-06,
419
+ "layerscale_value": 1.0,
420
+ "mlp_ratio": 4,
421
+ "model_type": "dinov2",
422
+ "num_attention_heads": 6,
423
+ "num_channels": 3,
424
+ "num_hidden_layers": 12,
425
+ "out_features": [
426
+ "stage12"
427
+ ],
428
+ "out_indices": [
429
+ 12
430
+ ],
431
+ "patch_size": 14,
432
+ "problem_type": "single_label_classification",
433
+ "qkv_bias": true,
434
+ "reshape_hidden_states": true,
435
+ "stage_names": [
436
+ "stem",
437
+ "stage1",
438
+ "stage2",
439
+ "stage3",
440
+ "stage4",
441
+ "stage5",
442
+ "stage6",
443
+ "stage7",
444
+ "stage8",
445
+ "stage9",
446
+ "stage10",
447
+ "stage11",
448
+ "stage12"
449
+ ],
450
+ "torch_dtype": "float32",
451
+ "transformers_version": "4.40.2",
452
+ "use_swiglu_ffn": false
453
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc361cf48b717bd60ccc57ddb2dd7e510bdb648a0cb91ff6e13bde5211c638bf
3
+ size 88866896
preprocessor_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_center_crop",
8
+ "crop_size",
9
+ "do_rescale",
10
+ "rescale_factor",
11
+ "do_normalize",
12
+ "image_mean",
13
+ "image_std",
14
+ "do_convert_rgb",
15
+ "return_tensors",
16
+ "data_format",
17
+ "input_data_format"
18
+ ],
19
+ "crop_size": {
20
+ "height": 224,
21
+ "width": 224
22
+ },
23
+ "do_center_crop": true,
24
+ "do_convert_rgb": true,
25
+ "do_normalize": true,
26
+ "do_rescale": true,
27
+ "do_resize": true,
28
+ "image_mean": [
29
+ 0.485,
30
+ 0.456,
31
+ 0.406
32
+ ],
33
+ "image_processor_type": "BitImageProcessor",
34
+ "image_std": [
35
+ 0.229,
36
+ 0.224,
37
+ 0.225
38
+ ],
39
+ "resample": 3,
40
+ "rescale_factor": 0.00392156862745098,
41
+ "size": {
42
+ "shortest_edge": 256
43
+ }
44
+ }
runs/Jun27_07-46-30_ubuntu-128gb-fsn1-1-ML/events.out.tfevents.1719474477.ubuntu-128gb-fsn1-1-ML.384837.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba4a01ffc14f8d5eb4cd01e3081c196d55bbfeebfe604fc3bd92c39ec1c6aca7
3
+ size 17907
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d91b9c972064c0d105f39999c3499224ab8fba2246a847c7f4eb4253787a515
3
+ size 5048