Kushagra07 commited on
Commit
50c00a4
1 Parent(s): c8a377f

Training in progress, epoch 1

Browse files
README.md ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/vit-base-patch16-224
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - imagefolder
8
+ metrics:
9
+ - accuracy
10
+ - recall
11
+ - f1
12
+ - precision
13
+ model-index:
14
+ - name: vit-base-patch16-224-finetuned-ind-17-imbalanced-aadhaarmask-new-parameter
15
+ results:
16
+ - task:
17
+ name: Image Classification
18
+ type: image-classification
19
+ dataset:
20
+ name: imagefolder
21
+ type: imagefolder
22
+ config: default
23
+ split: train
24
+ args: default
25
+ metrics:
26
+ - name: Accuracy
27
+ type: accuracy
28
+ value: 0.8339719029374202
29
+ - name: Recall
30
+ type: recall
31
+ value: 0.8339719029374202
32
+ - name: F1
33
+ type: f1
34
+ value: 0.8319571049551264
35
+ - name: Precision
36
+ type: precision
37
+ value: 0.8325133593723552
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # vit-base-patch16-224-finetuned-ind-17-imbalanced-aadhaarmask-new-parameter
44
+
45
+ This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Loss: 0.3507
48
+ - Accuracy: 0.8340
49
+ - Recall: 0.8340
50
+ - F1: 0.8320
51
+ - Precision: 0.8325
52
+
53
+ ## Model description
54
+
55
+ More information needed
56
+
57
+ ## Intended uses & limitations
58
+
59
+ More information needed
60
+
61
+ ## Training and evaluation data
62
+
63
+ More information needed
64
+
65
+ ## Training procedure
66
+
67
+ ### Training hyperparameters
68
+
69
+ The following hyperparameters were used during training:
70
+ - learning_rate: 5e-05
71
+ - train_batch_size: 8
72
+ - eval_batch_size: 8
73
+ - seed: 42
74
+ - gradient_accumulation_steps: 4
75
+ - total_train_batch_size: 32
76
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
77
+ - lr_scheduler_type: linear
78
+ - lr_scheduler_warmup_ratio: 0.1
79
+ - num_epochs: 10
80
+
81
+ ### Training results
82
+
83
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Recall | F1 | Precision |
84
+ |:-------------:|:------:|:----:|:---------------:|:--------:|:------:|:------:|:---------:|
85
+ | No log | 0.9974 | 293 | 0.6168 | 0.7923 | 0.7923 | 0.7737 | 0.7684 |
86
+ | No log | 1.9983 | 587 | 0.4599 | 0.8110 | 0.8110 | 0.8056 | 0.8085 |
87
+ | No log | 2.9991 | 881 | 0.4305 | 0.8233 | 0.8233 | 0.8211 | 0.8250 |
88
+ | No log | 4.0 | 1175 | 0.3966 | 0.8365 | 0.8365 | 0.8323 | 0.8452 |
89
+ | No log | 4.9974 | 1468 | 0.4100 | 0.8221 | 0.8221 | 0.8195 | 0.8219 |
90
+ | No log | 5.9983 | 1762 | 0.3890 | 0.8412 | 0.8412 | 0.8375 | 0.8466 |
91
+ | No log | 6.9991 | 2056 | 0.3659 | 0.8357 | 0.8357 | 0.8335 | 0.8386 |
92
+ | No log | 8.0 | 2350 | 0.3562 | 0.8395 | 0.8395 | 0.8379 | 0.8403 |
93
+ | No log | 8.9974 | 2643 | 0.3613 | 0.8382 | 0.8382 | 0.8373 | 0.8391 |
94
+ | 0.4339 | 9.9745 | 2930 | 0.3405 | 0.8455 | 0.8455 | 0.8447 | 0.8467 |
95
+
96
+
97
+ ### Framework versions
98
+
99
+ - Transformers 4.40.1
100
+ - Pytorch 2.2.0a0+81ea7a4
101
+ - Datasets 2.19.0
102
+ - Tokenizers 0.19.1
config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "AadhaarBack",
13
+ "1": "AadhaarBackMasked",
14
+ "10": "PanCard",
15
+ "11": "PassportFirst",
16
+ "12": "PassportLast",
17
+ "13": "PassportRegular",
18
+ "14": "VoterCardBack",
19
+ "15": "VoterCardFront",
20
+ "16": "VoterCardRegular",
21
+ "2": "AadhaarFront",
22
+ "3": "AadhaarFrontMasked",
23
+ "4": "AadhaarRegular",
24
+ "5": "AadhaarRegularMasked",
25
+ "6": "AadhaarSelfGenerated",
26
+ "7": "DrivingLicenseNew",
27
+ "8": "DrivingLicenseOld",
28
+ "9": "MultipleOVDs"
29
+ },
30
+ "image_size": 224,
31
+ "initializer_range": 0.02,
32
+ "intermediate_size": 3072,
33
+ "label2id": {
34
+ "AadhaarBack": "0",
35
+ "AadhaarBackMasked": "1",
36
+ "AadhaarFront": "2",
37
+ "AadhaarFrontMasked": "3",
38
+ "AadhaarRegular": "4",
39
+ "AadhaarRegularMasked": "5",
40
+ "AadhaarSelfGenerated": "6",
41
+ "DrivingLicenseNew": "7",
42
+ "DrivingLicenseOld": "8",
43
+ "MultipleOVDs": "9",
44
+ "PanCard": "10",
45
+ "PassportFirst": "11",
46
+ "PassportLast": "12",
47
+ "PassportRegular": "13",
48
+ "VoterCardBack": "14",
49
+ "VoterCardFront": "15",
50
+ "VoterCardRegular": "16"
51
+ },
52
+ "layer_norm_eps": 1e-12,
53
+ "model_type": "vit",
54
+ "num_attention_heads": 12,
55
+ "num_channels": 3,
56
+ "num_hidden_layers": 12,
57
+ "patch_size": 16,
58
+ "problem_type": "single_label_classification",
59
+ "qkv_bias": true,
60
+ "torch_dtype": "float32",
61
+ "transformers_version": "4.40.1"
62
+ }
emissions.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ timestamp,project_name,run_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
+ 2024-05-02T12:01:50,codecarbon,bad44f8c-0329-471d-bdd3-6315ce4270b2,1361.8545150756836,0.00014171788704263056,1.0406242772177081e-07,42.5,79.6684356058394,11.667008399963379,0.016076764230098996,0.03912815380249812,0.004410080944630984,0.05961499897722809,Canada,CAN,quebec,,,Linux-5.15.0-105-generic-x86_64-with-glibc2.35,3.10.12,2.3.5,32,13th Gen Intel(R) Core(TM) i9-13900K,1,1 x NVIDIA GeForce RTX 4060 Ti,-71.2,46.8,31.112022399902344,machine,N,1.0
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f15ff5ff647d5112289eba79872672f6c86730f4cd5a1e558acf04607572bdc
3
+ size 343270116
preprocessor_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_rescale",
8
+ "rescale_factor",
9
+ "do_normalize",
10
+ "image_mean",
11
+ "image_std",
12
+ "return_tensors",
13
+ "data_format",
14
+ "input_data_format"
15
+ ],
16
+ "do_normalize": true,
17
+ "do_rescale": true,
18
+ "do_resize": true,
19
+ "image_mean": [
20
+ 0.5,
21
+ 0.5,
22
+ 0.5
23
+ ],
24
+ "image_processor_type": "ViTImageProcessor",
25
+ "image_std": [
26
+ 0.5,
27
+ 0.5,
28
+ 0.5
29
+ ],
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "size": {
33
+ "height": 224,
34
+ "width": 224
35
+ }
36
+ }
runs/May02_11-39-06_60f4804cf903/events.out.tfevents.1714649948.60f4804cf903.8371.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29bbb69f014fd967ffe9f474b72e338ec545c55a065ca0bda130f5e7b3b14461
3
+ size 11045
runs/May02_11-39-06_60f4804cf903/events.out.tfevents.1714651369.60f4804cf903.8371.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:888904878fd1c261473a1f109df64fef2a94b1c9280583008a077fde377faa1c
3
+ size 560
runs/May02_12-13-38_60f4804cf903/events.out.tfevents.1714652021.60f4804cf903.9256.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eba9ad016ff478d0ec9bc30191b7f0254be41e577b089270b867755c13b974c
3
+ size 6231
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90116e41f2173db950fcb15f88d1676e19dc4b9acd665e5e8eaa03f6ea775035
3
+ size 5112