HorcruxNo13 commited on
Commit
67542fa
1 Parent(s): 18fd3eb

Training in progress, epoch 0

Browse files
config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "_name_or_path": "microsoft/beit-base-patch16-224",
 
3
  "architectures": [
4
  "BeitForImageClassification"
5
  ],
@@ -13,15 +14,15 @@
13
  "hidden_dropout_prob": 0.0,
14
  "hidden_size": 768,
15
  "id2label": {
16
- "0": "Normal",
17
- "1": "Abnormal"
18
  },
19
  "image_size": 224,
20
  "initializer_range": 0.02,
21
  "intermediate_size": 3072,
22
  "label2id": {
23
- "Abnormal": 1,
24
- "Normal": 0
25
  },
26
  "layer_norm_eps": 1e-12,
27
  "layer_scale_init_value": 0.1,
@@ -29,11 +30,11 @@
29
  "num_attention_heads": 12,
30
  "num_channels": 3,
31
  "num_hidden_layers": 12,
 
 
 
32
  "out_indices": [
33
- 3,
34
- 5,
35
- 7,
36
- 11
37
  ],
38
  "patch_size": 16,
39
  "pool_scales": [
@@ -43,9 +44,25 @@
43
  6
44
  ],
45
  "problem_type": "single_label_classification",
 
46
  "semantic_loss_ignore_index": 255,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  "torch_dtype": "float32",
48
- "transformers_version": "4.34.0",
49
  "use_absolute_position_embeddings": false,
50
  "use_auxiliary_head": true,
51
  "use_mask_token": false,
 
1
  {
2
  "_name_or_path": "microsoft/beit-base-patch16-224",
3
+ "add_fpn": false,
4
  "architectures": [
5
  "BeitForImageClassification"
6
  ],
 
14
  "hidden_dropout_prob": 0.0,
15
  "hidden_size": 768,
16
  "id2label": {
17
+ "0": "Absent",
18
+ "1": "Present"
19
  },
20
  "image_size": 224,
21
  "initializer_range": 0.02,
22
  "intermediate_size": 3072,
23
  "label2id": {
24
+ "Absent": 0,
25
+ "Present": 1
26
  },
27
  "layer_norm_eps": 1e-12,
28
  "layer_scale_init_value": 0.1,
 
30
  "num_attention_heads": 12,
31
  "num_channels": 3,
32
  "num_hidden_layers": 12,
33
+ "out_features": [
34
+ "stage12"
35
+ ],
36
  "out_indices": [
37
+ 12
 
 
 
38
  ],
39
  "patch_size": 16,
40
  "pool_scales": [
 
44
  6
45
  ],
46
  "problem_type": "single_label_classification",
47
+ "reshape_hidden_states": true,
48
  "semantic_loss_ignore_index": 255,
49
+ "stage_names": [
50
+ "stem",
51
+ "stage1",
52
+ "stage2",
53
+ "stage3",
54
+ "stage4",
55
+ "stage5",
56
+ "stage6",
57
+ "stage7",
58
+ "stage8",
59
+ "stage9",
60
+ "stage10",
61
+ "stage11",
62
+ "stage12"
63
+ ],
64
  "torch_dtype": "float32",
65
+ "transformers_version": "4.38.2",
66
  "use_absolute_position_embeddings": false,
67
  "use_auxiliary_head": true,
68
  "use_mask_token": false,
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ea9189622ea7e227c8bdec69c7bb8f711c954452b29d4cbb3e75573c74e8f5e
3
+ size 343080328
runs/Mar27_15-55-15_dbda3bb1da36/events.out.tfevents.1711554952.dbda3bb1da36.335.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:177eca21528817bf17d509b6afe42dff04880dc40f243ce5a116fd36c22253f4
3
+ size 4184
runs/Mar27_15-56-28_dbda3bb1da36/events.out.tfevents.1711554998.dbda3bb1da36.335.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:befbbd2bc128cdc070fcf63151f277048783c6031d93b8102ac915a35217844c
3
+ size 5928
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:efbbd180dd3e6860ad7e7fe6edebf3f47f293ba339c7b3ee420bb443f89da7c3
3
- size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d2c6e5c19c7d1274ad99ba3a3c074144bc73e58a639bb1d2d4e66cb12c6046b
3
+ size 4920