dgktrnh commited on
Commit
f730f63
1 Parent(s): 0e07b03

Upload RobertaForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +30 -4
  2. model.safetensors +2 -2
config.json CHANGED
@@ -6,14 +6,41 @@
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
- "classifier_input_length": 768,
10
  "eos_token_id": 2,
11
  "gradient_checkpointing": false,
12
  "hidden_act": "gelu",
13
  "hidden_dropout_prob": 0.1,
14
  "hidden_size": 768,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  "initializer_range": 0.02,
16
  "intermediate_size": 3072,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  "layer_norm_eps": 1e-05,
18
  "max_position_embeddings": 258,
19
  "model_type": "roberta",
@@ -21,12 +48,11 @@
21
  "num_hidden_layers": 12,
22
  "pad_token_id": 1,
23
  "position_embedding_type": "absolute",
24
- "problem_type": "single_label_classification",
25
  "tokenizer_class": "PhobertTokenizer",
26
  "torch_dtype": "float32",
27
- "transformers_version": "4.42.0.dev0",
28
  "type_vocab_size": 1,
29
  "use_cache": true,
30
- "use_cls": true,
31
  "vocab_size": 64001
32
  }
 
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
 
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "LABEL_0",
16
+ "1": "LABEL_1",
17
+ "2": "LABEL_2",
18
+ "3": "LABEL_3",
19
+ "4": "LABEL_4",
20
+ "5": "LABEL_5",
21
+ "6": "LABEL_6",
22
+ "7": "LABEL_7",
23
+ "8": "LABEL_8",
24
+ "9": "LABEL_9",
25
+ "10": "LABEL_10",
26
+ "11": "LABEL_11"
27
+ },
28
  "initializer_range": 0.02,
29
  "intermediate_size": 3072,
30
+ "label2id": {
31
+ "LABEL_0": 0,
32
+ "LABEL_1": 1,
33
+ "LABEL_10": 10,
34
+ "LABEL_11": 11,
35
+ "LABEL_2": 2,
36
+ "LABEL_3": 3,
37
+ "LABEL_4": 4,
38
+ "LABEL_5": 5,
39
+ "LABEL_6": 6,
40
+ "LABEL_7": 7,
41
+ "LABEL_8": 8,
42
+ "LABEL_9": 9
43
+ },
44
  "layer_norm_eps": 1e-05,
45
  "max_position_embeddings": 258,
46
  "model_type": "roberta",
 
48
  "num_hidden_layers": 12,
49
  "pad_token_id": 1,
50
  "position_embedding_type": "absolute",
51
+ "problem_type": "multi_label_classification",
52
  "tokenizer_class": "PhobertTokenizer",
53
  "torch_dtype": "float32",
54
+ "transformers_version": "4.41.1",
55
  "type_vocab_size": 1,
56
  "use_cache": true,
 
57
  "vocab_size": 64001
58
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9634866557c69f85460b843cb757eb8f8e8009cdeb45c2ea30f720cdc610a080
3
- size 632033008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c19c6bf1ab8921eabb4f5826b65f9cd4a5923c23bff4a911ae3df909fda765
3
+ size 540054144