prithivMLmods commited on
Commit
ddb0bbd
·
verified ·
1 Parent(s): 715ec8e

Upload folder using huggingface_hub

Browse files
checkpoint-537/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "No Tumor",
7
+ "1": "Glioma",
8
+ "2": "Meningioma",
9
+ "3": "Pituitary"
10
+ },
11
+ "initializer_factor": 1.0,
12
+ "label2id": {
13
+ "Glioma": 1,
14
+ "Meningioma": 2,
15
+ "No Tumor": 0,
16
+ "Pituitary": 3
17
+ },
18
+ "model_type": "siglip",
19
+ "problem_type": "single_label_classification",
20
+ "text_config": {
21
+ "attention_dropout": 0.0,
22
+ "hidden_act": "gelu_pytorch_tanh",
23
+ "hidden_size": 768,
24
+ "intermediate_size": 3072,
25
+ "layer_norm_eps": 1e-06,
26
+ "max_position_embeddings": 64,
27
+ "model_type": "siglip_text_model",
28
+ "num_attention_heads": 12,
29
+ "num_hidden_layers": 12,
30
+ "projection_size": 768,
31
+ "torch_dtype": "float32",
32
+ "vocab_size": 256000
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.51.0.dev0",
36
+ "vision_config": {
37
+ "attention_dropout": 0.0,
38
+ "hidden_act": "gelu_pytorch_tanh",
39
+ "hidden_size": 768,
40
+ "image_size": 224,
41
+ "intermediate_size": 3072,
42
+ "layer_norm_eps": 1e-06,
43
+ "model_type": "siglip_vision_model",
44
+ "num_attention_heads": 12,
45
+ "num_channels": 3,
46
+ "num_hidden_layers": 12,
47
+ "patch_size": 16,
48
+ "torch_dtype": "float32"
49
+ }
50
+ }
checkpoint-537/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c52ebc80aa16b342f91a59576fdf796e943f73a17c8f495e6d618580c08282e5
3
+ size 371574144
checkpoint-537/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0972b46767445a1bfd1cfc36dbd0b4c895120b5010bdd5c96b526659dacc06
3
+ size 686568058
checkpoint-537/preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
checkpoint-537/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfaa9f5cce2ba6255d46c29746a49cc93367c9ad64bdd2d4cd1d543d0a57d741
3
+ size 14244
checkpoint-537/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ada34c7b0bc3dacf1a511c856a4bbe6871e7bd8d7dcddb1ff89e99f2b138041
3
+ size 1064
checkpoint-537/trainer_state.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 537,
3
+ "best_metric": 0.052135273814201355,
4
+ "best_model_checkpoint": "siglip2-finetune-full/checkpoint-537",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 537,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_accuracy": 0.9198179271708683,
15
+ "eval_loss": 0.2214546948671341,
16
+ "eval_model_preparation_time": 0.0025,
17
+ "eval_runtime": 78.9185,
18
+ "eval_samples_per_second": 72.379,
19
+ "eval_steps_per_second": 9.047,
20
+ "step": 179
21
+ },
22
+ {
23
+ "epoch": 2.0,
24
+ "eval_accuracy": 0.9529061624649859,
25
+ "eval_loss": 0.13619443774223328,
26
+ "eval_model_preparation_time": 0.0025,
27
+ "eval_runtime": 78.4297,
28
+ "eval_samples_per_second": 72.83,
29
+ "eval_steps_per_second": 9.104,
30
+ "step": 358
31
+ },
32
+ {
33
+ "epoch": 2.793296089385475,
34
+ "grad_norm": 22.69907569885254,
35
+ "learning_rate": 2.337278106508876e-06,
36
+ "loss": 0.2664,
37
+ "step": 500
38
+ },
39
+ {
40
+ "epoch": 3.0,
41
+ "eval_accuracy": 0.9812675070028011,
42
+ "eval_loss": 0.052135273814201355,
43
+ "eval_model_preparation_time": 0.0025,
44
+ "eval_runtime": 78.377,
45
+ "eval_samples_per_second": 72.879,
46
+ "eval_steps_per_second": 9.11,
47
+ "step": 537
48
+ }
49
+ ],
50
+ "logging_steps": 500,
51
+ "max_steps": 895,
52
+ "num_input_tokens_seen": 0,
53
+ "num_train_epochs": 5,
54
+ "save_steps": 500,
55
+ "stateful_callbacks": {
56
+ "TrainerControl": {
57
+ "args": {
58
+ "should_epoch_stop": false,
59
+ "should_evaluate": false,
60
+ "should_log": false,
61
+ "should_save": true,
62
+ "should_training_stop": false
63
+ },
64
+ "attributes": {}
65
+ }
66
+ },
67
+ "total_flos": 1.4352579795383747e+18,
68
+ "train_batch_size": 32,
69
+ "trial_name": null,
70
+ "trial_params": null
71
+ }
checkpoint-537/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:167d1ac48636feaf88b05554f05a3d6348ac59b24df87be1d59d81eed83e94a4
3
+ size 5304
checkpoint-716/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "No Tumor",
7
+ "1": "Glioma",
8
+ "2": "Meningioma",
9
+ "3": "Pituitary"
10
+ },
11
+ "initializer_factor": 1.0,
12
+ "label2id": {
13
+ "Glioma": 1,
14
+ "Meningioma": 2,
15
+ "No Tumor": 0,
16
+ "Pituitary": 3
17
+ },
18
+ "model_type": "siglip",
19
+ "problem_type": "single_label_classification",
20
+ "text_config": {
21
+ "attention_dropout": 0.0,
22
+ "hidden_act": "gelu_pytorch_tanh",
23
+ "hidden_size": 768,
24
+ "intermediate_size": 3072,
25
+ "layer_norm_eps": 1e-06,
26
+ "max_position_embeddings": 64,
27
+ "model_type": "siglip_text_model",
28
+ "num_attention_heads": 12,
29
+ "num_hidden_layers": 12,
30
+ "projection_size": 768,
31
+ "torch_dtype": "float32",
32
+ "vocab_size": 256000
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.51.0.dev0",
36
+ "vision_config": {
37
+ "attention_dropout": 0.0,
38
+ "hidden_act": "gelu_pytorch_tanh",
39
+ "hidden_size": 768,
40
+ "image_size": 224,
41
+ "intermediate_size": 3072,
42
+ "layer_norm_eps": 1e-06,
43
+ "model_type": "siglip_vision_model",
44
+ "num_attention_heads": 12,
45
+ "num_channels": 3,
46
+ "num_hidden_layers": 12,
47
+ "patch_size": 16,
48
+ "torch_dtype": "float32"
49
+ }
50
+ }
checkpoint-716/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08ed37ac828314564b75876c243e1189b49f28409dab196c49151bd69c166905
3
+ size 371574144
checkpoint-716/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:738f51692d2eb41381d20fdc381de11bd85e142af7438ca7aaafe9777686fd4b
3
+ size 686568058
checkpoint-716/preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
checkpoint-716/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3d2e4b33055e316d2be7ba57d85c5a9cbc0717e6b1883beade4f0338f43eae8
3
+ size 14244
checkpoint-716/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:effb6efd2194b8e30faf98b70b38dd36e57731038d93adc03939951740697e20
3
+ size 1064
checkpoint-716/trainer_state.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 716,
3
+ "best_metric": 0.028042811900377274,
4
+ "best_model_checkpoint": "siglip2-finetune-full/checkpoint-716",
5
+ "epoch": 4.0,
6
+ "eval_steps": 500,
7
+ "global_step": 716,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_accuracy": 0.9198179271708683,
15
+ "eval_loss": 0.2214546948671341,
16
+ "eval_model_preparation_time": 0.0025,
17
+ "eval_runtime": 78.9185,
18
+ "eval_samples_per_second": 72.379,
19
+ "eval_steps_per_second": 9.047,
20
+ "step": 179
21
+ },
22
+ {
23
+ "epoch": 2.0,
24
+ "eval_accuracy": 0.9529061624649859,
25
+ "eval_loss": 0.13619443774223328,
26
+ "eval_model_preparation_time": 0.0025,
27
+ "eval_runtime": 78.4297,
28
+ "eval_samples_per_second": 72.83,
29
+ "eval_steps_per_second": 9.104,
30
+ "step": 358
31
+ },
32
+ {
33
+ "epoch": 2.793296089385475,
34
+ "grad_norm": 22.69907569885254,
35
+ "learning_rate": 2.337278106508876e-06,
36
+ "loss": 0.2664,
37
+ "step": 500
38
+ },
39
+ {
40
+ "epoch": 3.0,
41
+ "eval_accuracy": 0.9812675070028011,
42
+ "eval_loss": 0.052135273814201355,
43
+ "eval_model_preparation_time": 0.0025,
44
+ "eval_runtime": 78.377,
45
+ "eval_samples_per_second": 72.879,
46
+ "eval_steps_per_second": 9.11,
47
+ "step": 537
48
+ },
49
+ {
50
+ "epoch": 4.0,
51
+ "eval_accuracy": 0.9891456582633054,
52
+ "eval_loss": 0.028042811900377274,
53
+ "eval_model_preparation_time": 0.0025,
54
+ "eval_runtime": 78.2982,
55
+ "eval_samples_per_second": 72.952,
56
+ "eval_steps_per_second": 9.119,
57
+ "step": 716
58
+ }
59
+ ],
60
+ "logging_steps": 500,
61
+ "max_steps": 895,
62
+ "num_input_tokens_seen": 0,
63
+ "num_train_epochs": 5,
64
+ "save_steps": 500,
65
+ "stateful_callbacks": {
66
+ "TrainerControl": {
67
+ "args": {
68
+ "should_epoch_stop": false,
69
+ "should_evaluate": false,
70
+ "should_log": false,
71
+ "should_save": true,
72
+ "should_training_stop": false
73
+ },
74
+ "attributes": {}
75
+ }
76
+ },
77
+ "total_flos": 1.9136773060511662e+18,
78
+ "train_batch_size": 32,
79
+ "trial_name": null,
80
+ "trial_params": null
81
+ }
checkpoint-716/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:167d1ac48636feaf88b05554f05a3d6348ac59b24df87be1d59d81eed83e94a4
3
+ size 5304
checkpoint-895/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "No Tumor",
7
+ "1": "Glioma",
8
+ "2": "Meningioma",
9
+ "3": "Pituitary"
10
+ },
11
+ "initializer_factor": 1.0,
12
+ "label2id": {
13
+ "Glioma": 1,
14
+ "Meningioma": 2,
15
+ "No Tumor": 0,
16
+ "Pituitary": 3
17
+ },
18
+ "model_type": "siglip",
19
+ "problem_type": "single_label_classification",
20
+ "text_config": {
21
+ "attention_dropout": 0.0,
22
+ "hidden_act": "gelu_pytorch_tanh",
23
+ "hidden_size": 768,
24
+ "intermediate_size": 3072,
25
+ "layer_norm_eps": 1e-06,
26
+ "max_position_embeddings": 64,
27
+ "model_type": "siglip_text_model",
28
+ "num_attention_heads": 12,
29
+ "num_hidden_layers": 12,
30
+ "projection_size": 768,
31
+ "torch_dtype": "float32",
32
+ "vocab_size": 256000
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.51.0.dev0",
36
+ "vision_config": {
37
+ "attention_dropout": 0.0,
38
+ "hidden_act": "gelu_pytorch_tanh",
39
+ "hidden_size": 768,
40
+ "image_size": 224,
41
+ "intermediate_size": 3072,
42
+ "layer_norm_eps": 1e-06,
43
+ "model_type": "siglip_vision_model",
44
+ "num_attention_heads": 12,
45
+ "num_channels": 3,
46
+ "num_hidden_layers": 12,
47
+ "patch_size": 16,
48
+ "torch_dtype": "float32"
49
+ }
50
+ }
checkpoint-895/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2715062cb9a38068d4eec629b1177199d354a341ae27b1231c8075857a6ad366
3
+ size 371574144
checkpoint-895/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daa8e051e1bf85670b67e903036b914fda2f1ba53794171170da0348949b56ef
3
+ size 686568058
checkpoint-895/preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
checkpoint-895/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10917ea1a57a5c1cb7b3d93e6da56769cdba572bf1d96f190e2d80e84b0aa201
3
+ size 14244
checkpoint-895/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7414976fa4a1a7ee6171757beffaf1e95991d80178aae69444879fcef7361692
3
+ size 1064
checkpoint-895/trainer_state.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 895,
3
+ "best_metric": 0.02312469482421875,
4
+ "best_model_checkpoint": "siglip2-finetune-full/checkpoint-895",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 895,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_accuracy": 0.9198179271708683,
15
+ "eval_loss": 0.2214546948671341,
16
+ "eval_model_preparation_time": 0.0025,
17
+ "eval_runtime": 78.9185,
18
+ "eval_samples_per_second": 72.379,
19
+ "eval_steps_per_second": 9.047,
20
+ "step": 179
21
+ },
22
+ {
23
+ "epoch": 2.0,
24
+ "eval_accuracy": 0.9529061624649859,
25
+ "eval_loss": 0.13619443774223328,
26
+ "eval_model_preparation_time": 0.0025,
27
+ "eval_runtime": 78.4297,
28
+ "eval_samples_per_second": 72.83,
29
+ "eval_steps_per_second": 9.104,
30
+ "step": 358
31
+ },
32
+ {
33
+ "epoch": 2.793296089385475,
34
+ "grad_norm": 22.69907569885254,
35
+ "learning_rate": 2.337278106508876e-06,
36
+ "loss": 0.2664,
37
+ "step": 500
38
+ },
39
+ {
40
+ "epoch": 3.0,
41
+ "eval_accuracy": 0.9812675070028011,
42
+ "eval_loss": 0.052135273814201355,
43
+ "eval_model_preparation_time": 0.0025,
44
+ "eval_runtime": 78.377,
45
+ "eval_samples_per_second": 72.879,
46
+ "eval_steps_per_second": 9.11,
47
+ "step": 537
48
+ },
49
+ {
50
+ "epoch": 4.0,
51
+ "eval_accuracy": 0.9891456582633054,
52
+ "eval_loss": 0.028042811900377274,
53
+ "eval_model_preparation_time": 0.0025,
54
+ "eval_runtime": 78.2982,
55
+ "eval_samples_per_second": 72.952,
56
+ "eval_steps_per_second": 9.119,
57
+ "step": 716
58
+ },
59
+ {
60
+ "epoch": 5.0,
61
+ "eval_accuracy": 0.9917717086834734,
62
+ "eval_loss": 0.02312469482421875,
63
+ "eval_model_preparation_time": 0.0025,
64
+ "eval_runtime": 79.1299,
65
+ "eval_samples_per_second": 72.185,
66
+ "eval_steps_per_second": 9.023,
67
+ "step": 895
68
+ }
69
+ ],
70
+ "logging_steps": 500,
71
+ "max_steps": 895,
72
+ "num_input_tokens_seen": 0,
73
+ "num_train_epochs": 5,
74
+ "save_steps": 500,
75
+ "stateful_callbacks": {
76
+ "TrainerControl": {
77
+ "args": {
78
+ "should_epoch_stop": false,
79
+ "should_evaluate": false,
80
+ "should_log": false,
81
+ "should_save": true,
82
+ "should_training_stop": true
83
+ },
84
+ "attributes": {}
85
+ }
86
+ },
87
+ "total_flos": 2.392096632563958e+18,
88
+ "train_batch_size": 32,
89
+ "trial_name": null,
90
+ "trial_params": null
91
+ }
checkpoint-895/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:167d1ac48636feaf88b05554f05a3d6348ac59b24df87be1d59d81eed83e94a4
3
+ size 5304
config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "No Tumor",
7
+ "1": "Glioma",
8
+ "2": "Meningioma",
9
+ "3": "Pituitary"
10
+ },
11
+ "initializer_factor": 1.0,
12
+ "label2id": {
13
+ "Glioma": 1,
14
+ "Meningioma": 2,
15
+ "No Tumor": 0,
16
+ "Pituitary": 3
17
+ },
18
+ "model_type": "siglip",
19
+ "problem_type": "single_label_classification",
20
+ "text_config": {
21
+ "attention_dropout": 0.0,
22
+ "hidden_act": "gelu_pytorch_tanh",
23
+ "hidden_size": 768,
24
+ "intermediate_size": 3072,
25
+ "layer_norm_eps": 1e-06,
26
+ "max_position_embeddings": 64,
27
+ "model_type": "siglip_text_model",
28
+ "num_attention_heads": 12,
29
+ "num_hidden_layers": 12,
30
+ "projection_size": 768,
31
+ "torch_dtype": "float32",
32
+ "vocab_size": 256000
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.51.0.dev0",
36
+ "vision_config": {
37
+ "attention_dropout": 0.0,
38
+ "hidden_act": "gelu_pytorch_tanh",
39
+ "hidden_size": 768,
40
+ "image_size": 224,
41
+ "intermediate_size": 3072,
42
+ "layer_norm_eps": 1e-06,
43
+ "model_type": "siglip_vision_model",
44
+ "num_attention_heads": 12,
45
+ "num_channels": 3,
46
+ "num_hidden_layers": 12,
47
+ "patch_size": 16,
48
+ "torch_dtype": "float32"
49
+ }
50
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2715062cb9a38068d4eec629b1177199d354a341ae27b1231c8075857a6ad366
3
+ size 371574144
preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:167d1ac48636feaf88b05554f05a3d6348ac59b24df87be1d59d81eed83e94a4
3
+ size 5304