abhiman23897 commited on
Commit
34c1014
·
1 Parent(s): 7677557

Training in progress, epoch 1

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "allenai/longformer-base-4096",
3
+ "architectures": [
4
+ "LongformerForTokenClassification"
5
+ ],
6
+ "attention_mode": "longformer",
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "attention_window": [
9
+ 512,
10
+ 512,
11
+ 512,
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512,
19
+ 512,
20
+ 512
21
+ ],
22
+ "bos_token_id": 0,
23
+ "eos_token_id": 2,
24
+ "gradient_checkpointing": false,
25
+ "hidden_act": "gelu",
26
+ "hidden_dropout_prob": 0.1,
27
+ "hidden_size": 768,
28
+ "id2label": {
29
+ "0": "O",
30
+ "1": "B-total-participants",
31
+ "2": "I-total-participants",
32
+ "3": "B-intervention-participants",
33
+ "4": "I-intervention-participants",
34
+ "5": "B-control-participants",
35
+ "6": "I-control-participants",
36
+ "7": "B-age",
37
+ "8": "I-age",
38
+ "9": "B-eligibility",
39
+ "10": "I-eligibility",
40
+ "11": "B-ethinicity",
41
+ "12": "I-ethinicity",
42
+ "13": "B-condition",
43
+ "14": "I-condition",
44
+ "15": "B-location",
45
+ "16": "I-location",
46
+ "17": "B-intervention",
47
+ "18": "I-intervention",
48
+ "19": "B-control",
49
+ "20": "I-control",
50
+ "21": "B-outcome",
51
+ "22": "I-outcome",
52
+ "23": "B-outcome-measure",
53
+ "24": "I-outcome-measure",
54
+ "25": "B-iv-bin-abs",
55
+ "26": "I-iv-bin-abs",
56
+ "27": "B-cv-bin-abs",
57
+ "28": "I-cv-bin-abs",
58
+ "29": "B-iv-bin-percent",
59
+ "30": "I-iv-bin-percent",
60
+ "31": "B-cv-bin-percent",
61
+ "32": "I-cv-bin-percent",
62
+ "33": "B-iv-cont-mean",
63
+ "34": "I-iv-cont-mean",
64
+ "35": "B-cv-cont-mean",
65
+ "36": "I-cv-cont-mean",
66
+ "37": "B-iv-cont-median",
67
+ "38": "I-iv-cont-median",
68
+ "39": "B-cv-cont-median",
69
+ "40": "I-cv-cont-median",
70
+ "41": "B-iv-cont-sd",
71
+ "42": "I-iv-cont-sd",
72
+ "43": "B-cv-cont-sd",
73
+ "44": "I-cv-cont-sd",
74
+ "45": "B-iv-cont-q1",
75
+ "46": "I-iv-cont-q1",
76
+ "47": "B-cv-cont-q1",
77
+ "48": "I-cv-cont-q1",
78
+ "49": "B-iv-cont-q3",
79
+ "50": "I-iv-cont-q3",
80
+ "51": "B-cv-cont-q3",
81
+ "52": "I-cv-cont-q3"
82
+ },
83
+ "ignore_attention_mask": false,
84
+ "initializer_range": 0.02,
85
+ "intermediate_size": 3072,
86
+ "label2id": {
87
+ "B-age": 7,
88
+ "B-condition": 13,
89
+ "B-control": 19,
90
+ "B-control-participants": 5,
91
+ "B-cv-bin-abs": 27,
92
+ "B-cv-bin-percent": 31,
93
+ "B-cv-cont-mean": 35,
94
+ "B-cv-cont-median": 39,
95
+ "B-cv-cont-q1": 47,
96
+ "B-cv-cont-q3": 51,
97
+ "B-cv-cont-sd": 43,
98
+ "B-eligibility": 9,
99
+ "B-ethinicity": 11,
100
+ "B-intervention": 17,
101
+ "B-intervention-participants": 3,
102
+ "B-iv-bin-abs": 25,
103
+ "B-iv-bin-percent": 29,
104
+ "B-iv-cont-mean": 33,
105
+ "B-iv-cont-median": 37,
106
+ "B-iv-cont-q1": 45,
107
+ "B-iv-cont-q3": 49,
108
+ "B-iv-cont-sd": 41,
109
+ "B-location": 15,
110
+ "B-outcome": 21,
111
+ "B-outcome-measure": 23,
112
+ "B-total-participants": 1,
113
+ "I-age": 8,
114
+ "I-condition": 14,
115
+ "I-control": 20,
116
+ "I-control-participants": 6,
117
+ "I-cv-bin-abs": 28,
118
+ "I-cv-bin-percent": 32,
119
+ "I-cv-cont-mean": 36,
120
+ "I-cv-cont-median": 40,
121
+ "I-cv-cont-q1": 48,
122
+ "I-cv-cont-q3": 52,
123
+ "I-cv-cont-sd": 44,
124
+ "I-eligibility": 10,
125
+ "I-ethinicity": 12,
126
+ "I-intervention": 18,
127
+ "I-intervention-participants": 4,
128
+ "I-iv-bin-abs": 26,
129
+ "I-iv-bin-percent": 30,
130
+ "I-iv-cont-mean": 34,
131
+ "I-iv-cont-median": 38,
132
+ "I-iv-cont-q1": 46,
133
+ "I-iv-cont-q3": 50,
134
+ "I-iv-cont-sd": 42,
135
+ "I-location": 16,
136
+ "I-outcome": 22,
137
+ "I-outcome-measure": 24,
138
+ "I-total-participants": 2,
139
+ "O": 0
140
+ },
141
+ "layer_norm_eps": 1e-05,
142
+ "max_position_embeddings": 4098,
143
+ "model_type": "longformer",
144
+ "num_attention_heads": 12,
145
+ "num_hidden_layers": 12,
146
+ "onnx_export": false,
147
+ "pad_token_id": 1,
148
+ "sep_token_id": 2,
149
+ "torch_dtype": "float32",
150
+ "transformers_version": "4.30.2",
151
+ "type_vocab_size": 1,
152
+ "vocab_size": 50265
153
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d04124bdb6f9793ecae76164a996e95ea79e083bd4d92e1cbeb1269222a4ad2
3
+ size 592533574
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "bos_token": "<s>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "cls_token": "<s>",
6
+ "eos_token": "</s>",
7
+ "errors": "replace",
8
+ "mask_token": "<mask>",
9
+ "model_max_length": 4096,
10
+ "pad_token": "<pad>",
11
+ "sep_token": "</s>",
12
+ "tokenizer_class": "LongformerTokenizer",
13
+ "trim_offsets": true,
14
+ "unk_token": "<unk>"
15
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87532b54df2b8992b142d6c361697304cd068cb815f61b8bd644ed3dbc231cd0
3
+ size 4344
vocab.json ADDED
The diff for this file is too large to render. See raw diff