qianyihuang1203 commited on
Commit
4882538
1 Parent(s): 37e63bf

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,33 +1,36 @@
1
  {
2
- "_name_or_path": "autoevaluate/binary-classification",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForSequenceClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
 
11
  "id2label": {
12
  "0": "NEGATIVE",
13
  "1": "POSITIVE"
14
  },
15
  "initializer_range": 0.02,
 
16
  "label2id": {
17
  "NEGATIVE": 0,
18
  "POSITIVE": 1
19
  },
20
- "max_position_embeddings": 512,
21
- "model_type": "distilbert",
22
- "n_heads": 12,
23
- "n_layers": 6,
24
- "pad_token_id": 0,
 
 
25
  "problem_type": "single_label_classification",
26
- "qa_dropout": 0.1,
27
- "seq_classif_dropout": 0.2,
28
- "sinusoidal_pos_embds": false,
29
- "tie_weights_": true,
30
  "torch_dtype": "float32",
31
  "transformers_version": "4.40.1",
32
- "vocab_size": 30522
 
 
33
  }
 
1
  {
2
+ "_name_or_path": "roberta-base",
 
3
  "architectures": [
4
+ "RobertaForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
  "id2label": {
14
  "0": "NEGATIVE",
15
  "1": "POSITIVE"
16
  },
17
  "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
  "label2id": {
20
  "NEGATIVE": 0,
21
  "POSITIVE": 1
22
  },
23
+ "layer_norm_eps": 1e-05,
24
+ "max_position_embeddings": 514,
25
+ "model_type": "roberta",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "pad_token_id": 1,
29
+ "position_embedding_type": "absolute",
30
  "problem_type": "single_label_classification",
 
 
 
 
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.40.1",
33
+ "type_vocab_size": 1,
34
+ "use_cache": true,
35
+ "vocab_size": 50265
36
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24bc1dda45a4b2b01c6a703a7980fd85595aac2ebd9bfcfc4d26832474c4c359
3
- size 267832560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:217bb08dd5daa9b589b24d57957975c7e79f7e2b62f06e81054179a7b1d06779
3
+ size 498612824
runs/Apr25_23-21-43_f3663bd2dc7e/events.out.tfevents.1714087304.f3663bd2dc7e.2696.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7ace90aefab2a4314475a264f5ee4f120c648a5c7de95db1de96647c5835dda
3
+ size 4983
runs/Apr26_00-40-51_b1a258b2a825/events.out.tfevents.1714092053.b1a258b2a825.1507.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca116cfa03dc1a8ffa34ca6729bdbdd91b81ac4948383b4b94519417bfc3376e
3
+ size 88
runs/Apr26_00-42-52_b1a258b2a825/events.out.tfevents.1714092173.b1a258b2a825.5990.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6971e613cd83679e64c292f13fcda5c0139edcc668e2f69332d05af935657c6e
3
+ size 4983
runs/Apr26_02-52-42_cd90b4e5a9bf/events.out.tfevents.1714099963.cd90b4e5a9bf.2562.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a255d7d7d66ce56a6fe23cb608a37efef882fa667a52d862b26e72e8677e2bd2
3
+ size 4712
runs/Apr26_03-03-33_cd90b4e5a9bf/events.out.tfevents.1714100614.cd90b4e5a9bf.2562.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b37533ac4adeecc89b408c5756ed2b81475826df4759b3aa14830eba0613f31
3
+ size 88
runs/Apr26_03-07-48_cd90b4e5a9bf/events.out.tfevents.1714100868.cd90b4e5a9bf.34134.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb65a65699fd0b174c0f2b6b68004fa6b0a4ebfb089b6351f07540891882354c
3
+ size 88
runs/Apr26_03-12-03_cd90b4e5a9bf/events.out.tfevents.1714101123.cd90b4e5a9bf.35306.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b7115fc5e8ba830b400cf8fcb70387105fc530d8035b505e128717b458a948f
3
+ size 88
runs/Apr26_03-14-22_cd90b4e5a9bf/events.out.tfevents.1714101263.cd90b4e5a9bf.36469.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f967023a507f0af8b00fb58e5d650c58756e9bb925b9228f916051b39de160e2
3
+ size 88
runs/Apr26_03-23-19_3ce9a1a77357/events.out.tfevents.1714101800.3ce9a1a77357.6030.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e0d00494247d483f7388c3b73b6ace350570f3af3711a4b2c21572a0a1f44cc
3
+ size 5788
special_tokens_map.json CHANGED
@@ -1,36 +1,50 @@
1
  {
 
 
 
 
 
 
 
2
  "cls_token": {
3
- "content": "[CLS]",
4
  "lstrip": false,
5
- "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
- "mask_token": {
10
- "content": "[MASK]",
11
  "lstrip": false,
 
 
 
 
 
 
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "[PAD]",
18
  "lstrip": false,
19
- "normalized": false,
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
  "sep_token": {
24
- "content": "[SEP]",
25
  "lstrip": false,
26
- "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
  "unk_token": {
31
- "content": "[UNK]",
32
  "lstrip": false,
33
- "normalized": false,
34
  "rstrip": false,
35
  "single_word": false
36
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
  "cls_token": {
10
+ "content": "<s>",
11
  "lstrip": false,
12
+ "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "eos_token": {
17
+ "content": "</s>",
18
  "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
  "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
  "pad_token": {
31
+ "content": "<pad>",
32
  "lstrip": false,
33
+ "normalized": true,
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
  "sep_token": {
38
+ "content": "</s>",
39
  "lstrip": false,
40
+ "normalized": true,
41
  "rstrip": false,
42
  "single_word": false
43
  },
44
  "unk_token": {
45
+ "content": "<unk>",
46
  "lstrip": false,
47
+ "normalized": true,
48
  "rstrip": false,
49
  "single_word": false
50
  }
tokenizer_config.json CHANGED
@@ -1,59 +1,56 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[PAD]",
5
  "lstrip": false,
6
- "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
- "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
- "content": "[CLS]",
21
  "lstrip": false,
22
- "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
- "content": "[SEP]",
29
  "lstrip": false,
30
- "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_lower_case": true,
47
- "mask_token": "[MASK]",
48
- "max_length": 512,
49
  "model_max_length": 512,
50
- "pad_token": "[PAD]",
51
- "sep_token": "[SEP]",
52
- "stride": 0,
53
- "strip_accents": null,
54
- "tokenize_chinese_chars": true,
55
- "tokenizer_class": "DistilBertTokenizer",
56
- "truncation_side": "right",
57
- "truncation_strategy": "longest_first",
58
- "unk_token": "[UNK]"
59
  }
 
1
  {
2
+ "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
+ "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
+ "1": {
13
+ "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
19
  },
20
+ "2": {
21
+ "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
+ "3": {
29
+ "content": "<unk>",
30
  "lstrip": false,
31
+ "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
  }
44
  },
45
+ "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "mask_token": "<mask>",
51
  "model_max_length": 512,
52
+ "pad_token": "<pad>",
53
+ "sep_token": "</s>",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "unk_token": "<unk>"
 
 
 
 
 
56
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c00d8eb79bc8a33e63744bd13abc7c81d1b4326a0445edb61730e65feef3175
3
- size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3d7f128d33824d966ae197c4d93f09c68ef493b51ba9493ec1ab88ac5159ea3
3
+ size 4984
vocab.json ADDED
The diff for this file is too large to render. See raw diff