CRISMARHO commited on
Commit
3b52256
1 Parent(s): 4e1afa1

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ base_model: google/mobilenet_v2_1.0_224
7
+ widget:
8
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
9
+ example_title: Tiger
10
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
11
+ example_title: Teapot
12
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
13
+ example_title: Palace
14
+ ---
15
+
16
+ # Model Trained Using AutoTrain
17
+
18
+ - Problem type: Image Classification
19
+
20
+ ## Validation Metrics
21
+ loss: 0.38848769664764404
22
+
23
+ f1: 1.0
24
+
25
+ precision: 1.0
26
+
27
+ recall: 1.0
28
+
29
+ auc: 1.0
30
+
31
+ accuracy: 1.0
checkpoint-9/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/mobilenet_v2_1.0_224",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "MobileNetV2ForImageClassification"
6
+ ],
7
+ "classifier_dropout_prob": 0.2,
8
+ "depth_divisible_by": 8,
9
+ "depth_multiplier": 1.0,
10
+ "expand_ratio": 6,
11
+ "finegrained_output": true,
12
+ "first_layer_is_expansion": true,
13
+ "hidden_act": "relu6",
14
+ "id2label": {
15
+ "0": "maduras",
16
+ "1": "normales"
17
+ },
18
+ "image_size": 224,
19
+ "initializer_range": 0.02,
20
+ "label2id": {
21
+ "maduras": 0,
22
+ "normales": 1
23
+ },
24
+ "layer_norm_eps": 0.001,
25
+ "min_depth": 8,
26
+ "model_type": "mobilenet_v2",
27
+ "num_channels": 3,
28
+ "output_stride": 32,
29
+ "problem_type": "single_label_classification",
30
+ "semantic_loss_ignore_index": 255,
31
+ "tf_padding": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.42.2"
34
+ }
checkpoint-9/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544bedc9b1390f28885e561bc08d3a5fcfe1c0fec6f35cef134a3446e31fa694
3
+ size 9080216
checkpoint-9/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:239da59a52e6f5acb228aa7211d9048998bc516f293a6a724007bce75c24c1ea
3
+ size 17939322
checkpoint-9/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006efa233fdc7e9cfbdb8a2da4ac1cb8be36da6ea93b3d2e48ce6f6bdd419dd1
3
+ size 13990
checkpoint-9/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8788b0e47bb2b1903df73ccb3d223b85d3ba41bb31eb1abc2d47cf3e26a88d9b
3
+ size 1064
checkpoint-9/trainer_state.json ADDED
@@ -0,0 +1,222 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.38848769664764404,
3
+ "best_model_checkpoint": "clasificador-de-bananas/checkpoint-9",
4
+ "epoch": 9.0,
5
+ "eval_steps": 500,
6
+ "global_step": 9,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 14.218433380126953,
14
+ "learning_rate": 5e-05,
15
+ "loss": 0.7738,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 1.0,
21
+ "eval_auc": 1.0,
22
+ "eval_f1": 1.0,
23
+ "eval_loss": 0.4954526126384735,
24
+ "eval_precision": 1.0,
25
+ "eval_recall": 1.0,
26
+ "eval_runtime": 0.1112,
27
+ "eval_samples_per_second": 17.979,
28
+ "eval_steps_per_second": 8.99,
29
+ "step": 1
30
+ },
31
+ {
32
+ "epoch": 2.0,
33
+ "grad_norm": 12.876874923706055,
34
+ "learning_rate": 4.4444444444444447e-05,
35
+ "loss": 0.694,
36
+ "step": 2
37
+ },
38
+ {
39
+ "epoch": 2.0,
40
+ "eval_accuracy": 1.0,
41
+ "eval_auc": 1.0,
42
+ "eval_f1": 1.0,
43
+ "eval_loss": 0.4974704682826996,
44
+ "eval_precision": 1.0,
45
+ "eval_recall": 1.0,
46
+ "eval_runtime": 0.1041,
47
+ "eval_samples_per_second": 19.218,
48
+ "eval_steps_per_second": 9.609,
49
+ "step": 2
50
+ },
51
+ {
52
+ "epoch": 3.0,
53
+ "grad_norm": 12.629737854003906,
54
+ "learning_rate": 3.888888888888889e-05,
55
+ "loss": 0.6465,
56
+ "step": 3
57
+ },
58
+ {
59
+ "epoch": 3.0,
60
+ "eval_accuracy": 1.0,
61
+ "eval_auc": 1.0,
62
+ "eval_f1": 1.0,
63
+ "eval_loss": 0.48398110270500183,
64
+ "eval_precision": 1.0,
65
+ "eval_recall": 1.0,
66
+ "eval_runtime": 0.0886,
67
+ "eval_samples_per_second": 22.569,
68
+ "eval_steps_per_second": 11.285,
69
+ "step": 3
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "grad_norm": 12.785294532775879,
74
+ "learning_rate": 3.3333333333333335e-05,
75
+ "loss": 0.6718,
76
+ "step": 4
77
+ },
78
+ {
79
+ "epoch": 4.0,
80
+ "eval_accuracy": 0.5,
81
+ "eval_auc": 1.0,
82
+ "eval_f1": 0.6666666666666666,
83
+ "eval_loss": 0.5571763515472412,
84
+ "eval_precision": 0.5,
85
+ "eval_recall": 1.0,
86
+ "eval_runtime": 0.0927,
87
+ "eval_samples_per_second": 21.569,
88
+ "eval_steps_per_second": 10.784,
89
+ "step": 4
90
+ },
91
+ {
92
+ "epoch": 5.0,
93
+ "grad_norm": 14.149540901184082,
94
+ "learning_rate": 2.777777777777778e-05,
95
+ "loss": 0.6452,
96
+ "step": 5
97
+ },
98
+ {
99
+ "epoch": 5.0,
100
+ "eval_accuracy": 1.0,
101
+ "eval_auc": 1.0,
102
+ "eval_f1": 1.0,
103
+ "eval_loss": 0.42564940452575684,
104
+ "eval_precision": 1.0,
105
+ "eval_recall": 1.0,
106
+ "eval_runtime": 0.0948,
107
+ "eval_samples_per_second": 21.094,
108
+ "eval_steps_per_second": 10.547,
109
+ "step": 5
110
+ },
111
+ {
112
+ "epoch": 6.0,
113
+ "grad_norm": 18.8388729095459,
114
+ "learning_rate": 2.2222222222222223e-05,
115
+ "loss": 0.6702,
116
+ "step": 6
117
+ },
118
+ {
119
+ "epoch": 6.0,
120
+ "eval_accuracy": 1.0,
121
+ "eval_auc": 1.0,
122
+ "eval_f1": 1.0,
123
+ "eval_loss": 0.4299352467060089,
124
+ "eval_precision": 1.0,
125
+ "eval_recall": 1.0,
126
+ "eval_runtime": 0.1843,
127
+ "eval_samples_per_second": 10.849,
128
+ "eval_steps_per_second": 5.425,
129
+ "step": 6
130
+ },
131
+ {
132
+ "epoch": 7.0,
133
+ "grad_norm": 16.019344329833984,
134
+ "learning_rate": 1.6666666666666667e-05,
135
+ "loss": 0.5318,
136
+ "step": 7
137
+ },
138
+ {
139
+ "epoch": 7.0,
140
+ "eval_accuracy": 0.5,
141
+ "eval_auc": 1.0,
142
+ "eval_f1": 0.6666666666666666,
143
+ "eval_loss": 0.5418829917907715,
144
+ "eval_precision": 0.5,
145
+ "eval_recall": 1.0,
146
+ "eval_runtime": 0.1051,
147
+ "eval_samples_per_second": 19.035,
148
+ "eval_steps_per_second": 9.517,
149
+ "step": 7
150
+ },
151
+ {
152
+ "epoch": 8.0,
153
+ "grad_norm": 12.85544204711914,
154
+ "learning_rate": 1.1111111111111112e-05,
155
+ "loss": 0.5421,
156
+ "step": 8
157
+ },
158
+ {
159
+ "epoch": 8.0,
160
+ "eval_accuracy": 1.0,
161
+ "eval_auc": 1.0,
162
+ "eval_f1": 1.0,
163
+ "eval_loss": 0.4823116064071655,
164
+ "eval_precision": 1.0,
165
+ "eval_recall": 1.0,
166
+ "eval_runtime": 0.091,
167
+ "eval_samples_per_second": 21.981,
168
+ "eval_steps_per_second": 10.99,
169
+ "step": 8
170
+ },
171
+ {
172
+ "epoch": 9.0,
173
+ "grad_norm": 9.829968452453613,
174
+ "learning_rate": 5.555555555555556e-06,
175
+ "loss": 0.4502,
176
+ "step": 9
177
+ },
178
+ {
179
+ "epoch": 9.0,
180
+ "eval_accuracy": 1.0,
181
+ "eval_auc": 1.0,
182
+ "eval_f1": 1.0,
183
+ "eval_loss": 0.38848769664764404,
184
+ "eval_precision": 1.0,
185
+ "eval_recall": 1.0,
186
+ "eval_runtime": 0.1243,
187
+ "eval_samples_per_second": 16.095,
188
+ "eval_steps_per_second": 8.048,
189
+ "step": 9
190
+ }
191
+ ],
192
+ "logging_steps": 1,
193
+ "max_steps": 10,
194
+ "num_input_tokens_seen": 0,
195
+ "num_train_epochs": 10,
196
+ "save_steps": 500,
197
+ "stateful_callbacks": {
198
+ "EarlyStoppingCallback": {
199
+ "args": {
200
+ "early_stopping_patience": 5,
201
+ "early_stopping_threshold": 0.01
202
+ },
203
+ "attributes": {
204
+ "early_stopping_patience_counter": 0
205
+ }
206
+ },
207
+ "TrainerControl": {
208
+ "args": {
209
+ "should_epoch_stop": false,
210
+ "should_evaluate": false,
211
+ "should_log": false,
212
+ "should_save": true,
213
+ "should_training_stop": false
214
+ },
215
+ "attributes": {}
216
+ }
217
+ },
218
+ "total_flos": 189101405896704.0,
219
+ "train_batch_size": 8,
220
+ "trial_name": null,
221
+ "trial_params": null
222
+ }
checkpoint-9/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bb2b54f30d644674f2fd8e04f2f07d822e4465f9d2ae3472a82bb26e613692b
3
+ size 5176
config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/mobilenet_v2_1.0_224",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "MobileNetV2ForImageClassification"
6
+ ],
7
+ "classifier_dropout_prob": 0.2,
8
+ "depth_divisible_by": 8,
9
+ "depth_multiplier": 1.0,
10
+ "expand_ratio": 6,
11
+ "finegrained_output": true,
12
+ "first_layer_is_expansion": true,
13
+ "hidden_act": "relu6",
14
+ "id2label": {
15
+ "0": "maduras",
16
+ "1": "normales"
17
+ },
18
+ "image_size": 224,
19
+ "initializer_range": 0.02,
20
+ "label2id": {
21
+ "maduras": 0,
22
+ "normales": 1
23
+ },
24
+ "layer_norm_eps": 0.001,
25
+ "min_depth": 8,
26
+ "model_type": "mobilenet_v2",
27
+ "num_channels": 3,
28
+ "output_stride": 32,
29
+ "problem_type": "single_label_classification",
30
+ "semantic_loss_ignore_index": 255,
31
+ "tf_padding": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.42.2"
34
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544bedc9b1390f28885e561bc08d3a5fcfe1c0fec6f35cef134a3446e31fa694
3
+ size 9080216
preprocessor_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_normalize": true,
8
+ "do_rescale": true,
9
+ "do_resize": true,
10
+ "image_mean": [
11
+ 0.5,
12
+ 0.5,
13
+ 0.5
14
+ ],
15
+ "image_processor_type": "MobileNetV2ImageProcessor",
16
+ "image_std": [
17
+ 0.5,
18
+ 0.5,
19
+ 0.5
20
+ ],
21
+ "resample": 2,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "shortest_edge": 256
25
+ }
26
+ }
runs/Jul15_17-03-49_r-crismarho-prueba-finetuning-3a20v3sx-d0aae-x42nn/events.out.tfevents.1721063032.r-crismarho-prueba-finetuning-3a20v3sx-d0aae-x42nn.102.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33570a519da30b8178150cd53850370f3362ca3e5760edca74beeff15e06af94
3
- size 4996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b60d5c9a7442a30f0b1a20ebbdbfec429b40eb56dcc5f2803430235c4a8882e
3
+ size 12504
runs/Jul15_17-03-49_r-crismarho-prueba-finetuning-3a20v3sx-d0aae-x42nn/events.out.tfevents.1721063052.r-crismarho-prueba-finetuning-3a20v3sx-d0aae-x42nn.102.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ffc47b3e50f9870fe8315056114540b8a5ba64b4462fa2c0e5b043c7b953ec5
3
+ size 597
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bb2b54f30d644674f2fd8e04f2f07d822e4465f9d2ae3472a82bb26e613692b
3
+ size 5176
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "clasificador-de-bananas/autotrain-data",
3
+ "model": "google/mobilenet_v2_1.0_224",
4
+ "username": "CRISMARHO",
5
+ "lr": 5e-05,
6
+ "epochs": 10,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "clasificador-de-bananas",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "none",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "eval_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }