Bongyun commited on
Commit
f3d2d10
·
verified ·
1 Parent(s): 566efda
README.md CHANGED
@@ -1,10 +1,13 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: google/vit-base-patch16-224-in21k
4
  tags:
5
  - image-classification
6
  - ViT
7
  - generated_from_trainer
 
 
8
  model-index:
9
  - name: vit-base-beans-demo-v5
10
  results: []
@@ -16,6 +19,9 @@ should probably proofread and complete it, then remove this comment. -->
16
  # vit-base-beans-demo-v5
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the beans dataset.
 
 
 
19
 
20
  ## Model description
21
 
@@ -38,14 +44,22 @@ The following hyperparameters were used during training:
38
  - train_batch_size: 16
39
  - eval_batch_size: 8
40
  - seed: 42
41
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - num_epochs: 4
44
  - mixed_precision_training: Native AMP
45
 
 
 
 
 
 
 
 
 
46
  ### Framework versions
47
 
48
- - Transformers 4.42.4
49
- - Pytorch 2.3.1+cu121
50
- - Datasets 2.20.0
51
- - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: google/vit-base-patch16-224-in21k
5
  tags:
6
  - image-classification
7
  - ViT
8
  - generated_from_trainer
9
+ metrics:
10
+ - accuracy
11
  model-index:
12
  - name: vit-base-beans-demo-v5
13
  results: []
 
19
  # vit-base-beans-demo-v5
20
 
21
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the beans dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.0227
24
+ - Accuracy: 0.9925
25
 
26
  ## Model description
27
 
 
44
  - train_batch_size: 16
45
  - eval_batch_size: 8
46
  - seed: 42
47
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - num_epochs: 4
50
  - mixed_precision_training: Native AMP
51
 
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
55
+ |:-------------:|:------:|:----:|:---------------:|:--------:|
56
+ | 0.0783 | 1.5385 | 100 | 0.0551 | 0.9850 |
57
+ | 0.0124 | 3.0769 | 200 | 0.0227 | 0.9925 |
58
+
59
+
60
  ### Framework versions
61
 
62
+ - Transformers 4.46.2
63
+ - Pytorch 2.5.0+cu121
64
+ - Datasets 3.1.0
65
+ - Tokenizers 0.20.3
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 4.0,
3
  "total_flos": 3.205097416476426e+17,
4
- "train_loss": 0.11968801841139794,
5
- "train_runtime": 105.0874,
6
- "train_samples_per_second": 39.358,
7
- "train_steps_per_second": 2.474
8
  }
 
1
  {
2
  "epoch": 4.0,
3
  "total_flos": 3.205097416476426e+17,
4
+ "train_loss": 0.12550590594227498,
5
+ "train_runtime": 104.2364,
6
+ "train_samples_per_second": 39.679,
7
+ "train_steps_per_second": 2.494
8
  }
config.json CHANGED
@@ -30,5 +30,5 @@
30
  "problem_type": "single_label_classification",
31
  "qkv_bias": true,
32
  "torch_dtype": "float32",
33
- "transformers_version": "4.42.4"
34
  }
 
30
  "problem_type": "single_label_classification",
31
  "qkv_bias": true,
32
  "torch_dtype": "float32",
33
+ "transformers_version": "4.46.2"
34
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24b062e7e864e24a09178386c6ae7ddeef3e8fe6e8db2dcd796aaf28738a9586
3
  size 343227052
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38b87efad3dccfb9ec1d1621491fd143e65044773c30145f9f6393e94fbe82fd
3
  size 343227052
preprocessor_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "do_normalize": true,
3
  "do_rescale": true,
4
- "do_resize": "google/vit-base-patch16-224-in21k",
5
  "image_mean": [
6
  0.5,
7
  0.5,
 
1
  {
2
  "do_normalize": true,
3
  "do_rescale": true,
4
+ "do_resize": true,
5
  "image_mean": [
6
  0.5,
7
  0.5,
runs/Nov13_04-02-46_1f52fa223e2d/events.out.tfevents.1731470572.1f52fa223e2d.833.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34c20ca800a31afbfc4e22c87279e968f74d9de614aba2b2e7aec68c9ef55b7c
3
+ size 11559
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 4.0,
3
  "total_flos": 3.205097416476426e+17,
4
- "train_loss": 0.11968801841139794,
5
- "train_runtime": 105.0874,
6
- "train_samples_per_second": 39.358,
7
- "train_steps_per_second": 2.474
8
  }
 
1
  {
2
  "epoch": 4.0,
3
  "total_flos": 3.205097416476426e+17,
4
+ "train_loss": 0.12550590594227498,
5
+ "train_runtime": 104.2364,
6
+ "train_samples_per_second": 39.679,
7
+ "train_steps_per_second": 2.494
8
  }
trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.05026606097817421,
3
  "best_model_checkpoint": "./vit-base-beans-demo-v5/checkpoint-200",
4
  "epoch": 4.0,
5
  "eval_steps": 100,
@@ -10,212 +10,212 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.15384615384615385,
13
- "grad_norm": 1.2843873500823975,
14
  "learning_rate": 0.00019230769230769233,
15
- "loss": 0.8434,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.3076923076923077,
20
- "grad_norm": 3.7151455879211426,
21
  "learning_rate": 0.00018461538461538463,
22
- "loss": 0.3085,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.46153846153846156,
27
- "grad_norm": 0.8530910611152649,
28
  "learning_rate": 0.00017692307692307693,
29
- "loss": 0.248,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.6153846153846154,
34
- "grad_norm": 1.1804994344711304,
35
  "learning_rate": 0.00016923076923076923,
36
- "loss": 0.2334,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.7692307692307693,
41
- "grad_norm": 0.5084353089332581,
42
  "learning_rate": 0.00016153846153846155,
43
- "loss": 0.1626,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.9230769230769231,
48
- "grad_norm": 6.173972129821777,
49
  "learning_rate": 0.00015384615384615385,
50
- "loss": 0.2912,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 1.0769230769230769,
55
- "grad_norm": 2.3577582836151123,
56
  "learning_rate": 0.00014615384615384615,
57
- "loss": 0.1226,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 1.2307692307692308,
62
- "grad_norm": 0.14449569582939148,
63
  "learning_rate": 0.00013846153846153847,
64
- "loss": 0.0796,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.3846153846153846,
69
- "grad_norm": 4.846245288848877,
70
  "learning_rate": 0.00013076923076923077,
71
- "loss": 0.0813,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 1.5384615384615383,
76
- "grad_norm": 3.6284821033477783,
77
  "learning_rate": 0.0001230769230769231,
78
- "loss": 0.0645,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 1.5384615384615383,
83
- "eval_accuracy": 0.9699248120300752,
84
- "eval_loss": 0.09509244561195374,
85
- "eval_runtime": 1.3146,
86
- "eval_samples_per_second": 101.169,
87
- "eval_steps_per_second": 12.931,
88
  "step": 100
89
  },
90
  {
91
  "epoch": 1.6923076923076923,
92
- "grad_norm": 2.209476947784424,
93
  "learning_rate": 0.00011538461538461538,
94
- "loss": 0.0932,
95
  "step": 110
96
  },
97
  {
98
  "epoch": 1.8461538461538463,
99
- "grad_norm": 0.10745280981063843,
100
  "learning_rate": 0.0001076923076923077,
101
- "loss": 0.1311,
102
  "step": 120
103
  },
104
  {
105
  "epoch": 2.0,
106
- "grad_norm": 0.1176445409655571,
107
  "learning_rate": 0.0001,
108
- "loss": 0.0584,
109
  "step": 130
110
  },
111
  {
112
  "epoch": 2.1538461538461537,
113
- "grad_norm": 0.11741472780704498,
114
  "learning_rate": 9.230769230769232e-05,
115
- "loss": 0.0464,
116
  "step": 140
117
  },
118
  {
119
  "epoch": 2.3076923076923075,
120
- "grad_norm": 1.5773084163665771,
121
  "learning_rate": 8.461538461538461e-05,
122
- "loss": 0.0727,
123
  "step": 150
124
  },
125
  {
126
  "epoch": 2.4615384615384617,
127
- "grad_norm": 0.5665645003318787,
128
  "learning_rate": 7.692307692307693e-05,
129
- "loss": 0.0669,
130
  "step": 160
131
  },
132
  {
133
  "epoch": 2.6153846153846154,
134
- "grad_norm": 0.20813684165477753,
135
  "learning_rate": 6.923076923076924e-05,
136
- "loss": 0.0246,
137
  "step": 170
138
  },
139
  {
140
  "epoch": 2.769230769230769,
141
- "grad_norm": 0.12770630419254303,
142
  "learning_rate": 6.153846153846155e-05,
143
- "loss": 0.0285,
144
  "step": 180
145
  },
146
  {
147
  "epoch": 2.9230769230769234,
148
- "grad_norm": 0.06980925053358078,
149
  "learning_rate": 5.384615384615385e-05,
150
- "loss": 0.0185,
151
  "step": 190
152
  },
153
  {
154
  "epoch": 3.076923076923077,
155
- "grad_norm": 0.07084397971630096,
156
  "learning_rate": 4.615384615384616e-05,
157
- "loss": 0.0316,
158
  "step": 200
159
  },
160
  {
161
  "epoch": 3.076923076923077,
162
- "eval_accuracy": 0.9849624060150376,
163
- "eval_loss": 0.05026606097817421,
164
- "eval_runtime": 2.2826,
165
- "eval_samples_per_second": 58.267,
166
- "eval_steps_per_second": 7.448,
167
  "step": 200
168
  },
169
  {
170
  "epoch": 3.230769230769231,
171
- "grad_norm": 0.061542343348264694,
172
  "learning_rate": 3.846153846153846e-05,
173
- "loss": 0.0128,
174
  "step": 210
175
  },
176
  {
177
  "epoch": 3.3846153846153846,
178
- "grad_norm": 0.4158130884170532,
179
  "learning_rate": 3.0769230769230774e-05,
180
- "loss": 0.0173,
181
  "step": 220
182
  },
183
  {
184
  "epoch": 3.5384615384615383,
185
- "grad_norm": 0.07470071315765381,
186
  "learning_rate": 2.307692307692308e-05,
187
- "loss": 0.0119,
188
  "step": 230
189
  },
190
  {
191
  "epoch": 3.6923076923076925,
192
- "grad_norm": 0.0632583349943161,
193
  "learning_rate": 1.5384615384615387e-05,
194
- "loss": 0.0391,
195
  "step": 240
196
  },
197
  {
198
  "epoch": 3.8461538461538463,
199
- "grad_norm": 0.46634915471076965,
200
  "learning_rate": 7.692307692307694e-06,
201
- "loss": 0.012,
202
  "step": 250
203
  },
204
  {
205
  "epoch": 4.0,
206
- "grad_norm": 0.056492287665605545,
207
  "learning_rate": 0.0,
208
- "loss": 0.0116,
209
  "step": 260
210
  },
211
  {
212
  "epoch": 4.0,
213
  "step": 260,
214
  "total_flos": 3.205097416476426e+17,
215
- "train_loss": 0.11968801841139794,
216
- "train_runtime": 105.0874,
217
- "train_samples_per_second": 39.358,
218
- "train_steps_per_second": 2.474
219
  }
220
  ],
221
  "logging_steps": 10,
@@ -230,7 +230,7 @@
230
  "should_evaluate": false,
231
  "should_log": false,
232
  "should_save": true,
233
- "should_training_stop": false
234
  },
235
  "attributes": {}
236
  }
 
1
  {
2
+ "best_metric": 0.02266957052052021,
3
  "best_model_checkpoint": "./vit-base-beans-demo-v5/checkpoint-200",
4
  "epoch": 4.0,
5
  "eval_steps": 100,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.15384615384615385,
13
+ "grad_norm": 1.8716892004013062,
14
  "learning_rate": 0.00019230769230769233,
15
+ "loss": 0.8364,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.3076923076923077,
20
+ "grad_norm": 2.2301363945007324,
21
  "learning_rate": 0.00018461538461538463,
22
+ "loss": 0.3741,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.46153846153846156,
27
+ "grad_norm": 2.373594284057617,
28
  "learning_rate": 0.00017692307692307693,
29
+ "loss": 0.2396,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.6153846153846154,
34
+ "grad_norm": 0.6037946343421936,
35
  "learning_rate": 0.00016923076923076923,
36
+ "loss": 0.2642,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.7692307692307693,
41
+ "grad_norm": 0.3296580910682678,
42
  "learning_rate": 0.00016153846153846155,
43
+ "loss": 0.2242,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.9230769230769231,
48
+ "grad_norm": 3.0283963680267334,
49
  "learning_rate": 0.00015384615384615385,
50
+ "loss": 0.3345,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 1.0769230769230769,
55
+ "grad_norm": 0.7587921619415283,
56
  "learning_rate": 0.00014615384615384615,
57
+ "loss": 0.1903,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 1.2307692307692308,
62
+ "grad_norm": 0.2090693563222885,
63
  "learning_rate": 0.00013846153846153847,
64
+ "loss": 0.0633,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.3846153846153846,
69
+ "grad_norm": 0.17675693333148956,
70
  "learning_rate": 0.00013076923076923077,
71
+ "loss": 0.0942,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 1.5384615384615383,
76
+ "grad_norm": 0.648328423500061,
77
  "learning_rate": 0.0001230769230769231,
78
+ "loss": 0.0783,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 1.5384615384615383,
83
+ "eval_accuracy": 0.9849624060150376,
84
+ "eval_loss": 0.055133331567049026,
85
+ "eval_runtime": 2.1339,
86
+ "eval_samples_per_second": 62.326,
87
+ "eval_steps_per_second": 7.966,
88
  "step": 100
89
  },
90
  {
91
  "epoch": 1.6923076923076923,
92
+ "grad_norm": 0.66573166847229,
93
  "learning_rate": 0.00011538461538461538,
94
+ "loss": 0.1371,
95
  "step": 110
96
  },
97
  {
98
  "epoch": 1.8461538461538463,
99
+ "grad_norm": 3.307586908340454,
100
  "learning_rate": 0.0001076923076923077,
101
+ "loss": 0.0776,
102
  "step": 120
103
  },
104
  {
105
  "epoch": 2.0,
106
+ "grad_norm": 0.10508356243371964,
107
  "learning_rate": 0.0001,
108
+ "loss": 0.0821,
109
  "step": 130
110
  },
111
  {
112
  "epoch": 2.1538461538461537,
113
+ "grad_norm": 0.14109393954277039,
114
  "learning_rate": 9.230769230769232e-05,
115
+ "loss": 0.0622,
116
  "step": 140
117
  },
118
  {
119
  "epoch": 2.3076923076923075,
120
+ "grad_norm": 0.2330102026462555,
121
  "learning_rate": 8.461538461538461e-05,
122
+ "loss": 0.0204,
123
  "step": 150
124
  },
125
  {
126
  "epoch": 2.4615384615384617,
127
+ "grad_norm": 0.10599979013204575,
128
  "learning_rate": 7.692307692307693e-05,
129
+ "loss": 0.0219,
130
  "step": 160
131
  },
132
  {
133
  "epoch": 2.6153846153846154,
134
+ "grad_norm": 0.07637907564640045,
135
  "learning_rate": 6.923076923076924e-05,
136
+ "loss": 0.0149,
137
  "step": 170
138
  },
139
  {
140
  "epoch": 2.769230769230769,
141
+ "grad_norm": 0.06972856819629669,
142
  "learning_rate": 6.153846153846155e-05,
143
+ "loss": 0.0383,
144
  "step": 180
145
  },
146
  {
147
  "epoch": 2.9230769230769234,
148
+ "grad_norm": 0.06536921858787537,
149
  "learning_rate": 5.384615384615385e-05,
150
+ "loss": 0.0135,
151
  "step": 190
152
  },
153
  {
154
  "epoch": 3.076923076923077,
155
+ "grad_norm": 0.06367834657430649,
156
  "learning_rate": 4.615384615384616e-05,
157
+ "loss": 0.0124,
158
  "step": 200
159
  },
160
  {
161
  "epoch": 3.076923076923077,
162
+ "eval_accuracy": 0.9924812030075187,
163
+ "eval_loss": 0.02266957052052021,
164
+ "eval_runtime": 1.2444,
165
+ "eval_samples_per_second": 106.875,
166
+ "eval_steps_per_second": 13.661,
167
  "step": 200
168
  },
169
  {
170
  "epoch": 3.230769230769231,
171
+ "grad_norm": 0.06352599710226059,
172
  "learning_rate": 3.846153846153846e-05,
173
+ "loss": 0.0133,
174
  "step": 210
175
  },
176
  {
177
  "epoch": 3.3846153846153846,
178
+ "grad_norm": 0.06076899170875549,
179
  "learning_rate": 3.0769230769230774e-05,
180
+ "loss": 0.0123,
181
  "step": 220
182
  },
183
  {
184
  "epoch": 3.5384615384615383,
185
+ "grad_norm": 0.06032305210828781,
186
  "learning_rate": 2.307692307692308e-05,
187
+ "loss": 0.0114,
188
  "step": 230
189
  },
190
  {
191
  "epoch": 3.6923076923076925,
192
+ "grad_norm": 0.05871906504034996,
193
  "learning_rate": 1.5384615384615387e-05,
194
+ "loss": 0.0111,
195
  "step": 240
196
  },
197
  {
198
  "epoch": 3.8461538461538463,
199
+ "grad_norm": 0.06305041909217834,
200
  "learning_rate": 7.692307692307694e-06,
201
+ "loss": 0.0112,
202
  "step": 250
203
  },
204
  {
205
  "epoch": 4.0,
206
+ "grad_norm": 0.060099028050899506,
207
  "learning_rate": 0.0,
208
+ "loss": 0.0244,
209
  "step": 260
210
  },
211
  {
212
  "epoch": 4.0,
213
  "step": 260,
214
  "total_flos": 3.205097416476426e+17,
215
+ "train_loss": 0.12550590594227498,
216
+ "train_runtime": 104.2364,
217
+ "train_samples_per_second": 39.679,
218
+ "train_steps_per_second": 2.494
219
  }
220
  ],
221
  "logging_steps": 10,
 
230
  "should_evaluate": false,
231
  "should_log": false,
232
  "should_save": true,
233
+ "should_training_stop": true
234
  },
235
  "attributes": {}
236
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0678555ad6faa886947d13cb58047927698f16f53c2db26bf31733267fecff14
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83e883a042c6d83197805380350a3ce4a9ce367789ac2974287d6580a562bfd4
3
+ size 5304