dandankim commited on
Commit
6aa952c
·
verified ·
1 Parent(s): bc7bc76

dandankim/token_classifier2

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
- base_model: distilbert-base-uncased
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -19,13 +19,13 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # distilbert-token-classifier
21
 
22
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.0738
25
- - Precision: 0.9657
26
- - Recall: 0.9761
27
- - F1: 0.9709
28
- - Accuracy: 0.9845
29
 
30
  ## Model description
31
 
@@ -44,27 +44,28 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 2e-05
48
- - train_batch_size: 16
49
- - eval_batch_size: 16
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
- - lr_scheduler_warmup_ratio: 0.1
54
  - num_epochs: 10
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
59
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
60
- | 0.898 | 1.0 | 149 | 0.1613 | 0.8995 | 0.8729 | 0.8860 | 0.9637 |
61
- | 0.1161 | 2.0 | 298 | 0.0749 | 0.9682 | 0.9501 | 0.9591 | 0.9790 |
62
- | 0.0717 | 3.0 | 447 | 0.0762 | 0.9723 | 0.9630 | 0.9676 | 0.9814 |
63
- | 0.0412 | 4.0 | 596 | 0.0661 | 0.9699 | 0.9694 | 0.9697 | 0.9845 |
64
- | 0.0295 | 5.0 | 745 | 0.0630 | 0.9596 | 0.9739 | 0.9667 | 0.9835 |
65
- | 0.0262 | 6.0 | 894 | 0.0658 | 0.9696 | 0.9732 | 0.9714 | 0.9851 |
66
- | 0.0191 | 7.0 | 1043 | 0.0695 | 0.9696 | 0.9707 | 0.9702 | 0.9846 |
67
- | 0.0165 | 8.0 | 1192 | 0.0738 | 0.9657 | 0.9761 | 0.9709 | 0.9845 |
 
68
 
69
 
70
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ license: mit
4
+ base_model: FacebookAI/roberta-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
19
 
20
  # distilbert-token-classifier
21
 
22
+ This model is a fine-tuned version of [FacebookAI/roberta-base](https://huggingface.co/FacebookAI/roberta-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.0728
25
+ - Precision: 0.9694
26
+ - Recall: 0.9767
27
+ - F1: 0.9730
28
+ - Accuracy: 0.9846
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 1e-05
48
+ - train_batch_size: 20
49
+ - eval_batch_size: 20
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
+ - lr_scheduler_warmup_ratio: 0.2
54
  - num_epochs: 10
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
59
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
60
+ | 1.0798 | 1.0 | 119 | 0.5221 | 0.5989 | 0.3881 | 0.4710 | 0.8427 |
61
+ | 0.2561 | 2.0 | 238 | 0.1148 | 0.9162 | 0.9214 | 0.9188 | 0.9716 |
62
+ | 0.0901 | 3.0 | 357 | 0.0863 | 0.9729 | 0.9584 | 0.9656 | 0.9799 |
63
+ | 0.0735 | 4.0 | 476 | 0.0699 | 0.9658 | 0.9701 | 0.9680 | 0.9827 |
64
+ | 0.0528 | 5.0 | 595 | 0.0674 | 0.9545 | 0.9761 | 0.9652 | 0.9831 |
65
+ | 0.0505 | 6.0 | 714 | 0.0659 | 0.9689 | 0.9757 | 0.9723 | 0.9841 |
66
+ | 0.0394 | 7.0 | 833 | 0.0696 | 0.9633 | 0.9771 | 0.9701 | 0.9839 |
67
+ | 0.0278 | 8.0 | 952 | 0.0728 | 0.9640 | 0.9772 | 0.9706 | 0.9837 |
68
+ | 0.0241 | 9.0 | 1071 | 0.0728 | 0.9694 | 0.9767 | 0.9730 | 0.9846 |
69
 
70
 
71
  ### Framework versions
config.json CHANGED
@@ -1,13 +1,15 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForTokenClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
 
11
  "id2label": {
12
  "0": "O",
13
  "1": "B-BASE_TOKEN",
@@ -28,6 +30,7 @@
28
  "16": "I-LEVERAGE"
29
  },
30
  "initializer_range": 0.02,
 
31
  "label2id": {
32
  "B-ACTION": 5,
33
  "B-BASE_TOKEN": 1,
@@ -47,16 +50,16 @@
47
  "I-TARGET_SELL_PRICE": 10,
48
  "O": 0
49
  },
50
- "max_position_embeddings": 512,
51
- "model_type": "distilbert",
52
- "n_heads": 12,
53
- "n_layers": 6,
54
- "pad_token_id": 0,
55
- "qa_dropout": 0.1,
56
- "seq_classif_dropout": 0.2,
57
- "sinusoidal_pos_embds": false,
58
- "tie_weights_": true,
59
  "torch_dtype": "float32",
60
  "transformers_version": "4.48.3",
61
- "vocab_size": 30522
 
 
62
  }
 
1
  {
2
+ "_name_or_path": "FacebookAI/roberta-base",
 
3
  "architectures": [
4
+ "RobertaForTokenClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
  "id2label": {
14
  "0": "O",
15
  "1": "B-BASE_TOKEN",
 
30
  "16": "I-LEVERAGE"
31
  },
32
  "initializer_range": 0.02,
33
+ "intermediate_size": 3072,
34
  "label2id": {
35
  "B-ACTION": 5,
36
  "B-BASE_TOKEN": 1,
 
50
  "I-TARGET_SELL_PRICE": 10,
51
  "O": 0
52
  },
53
+ "layer_norm_eps": 1e-05,
54
+ "max_position_embeddings": 514,
55
+ "model_type": "roberta",
56
+ "num_attention_heads": 12,
57
+ "num_hidden_layers": 12,
58
+ "pad_token_id": 1,
59
+ "position_embedding_type": "absolute",
 
 
60
  "torch_dtype": "float32",
61
  "transformers_version": "4.48.3",
62
+ "type_vocab_size": 1,
63
+ "use_cache": true,
64
+ "vocab_size": 50265
65
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b21e75c082fe70716c9e78e7ac9075627dd6dcc9c27b08c05ece728bd4287d78
3
- size 265516156
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbabdbdc4ff986f2b2cc36998af941e433cdbd5cce644f6a494fd47a325dc7ed
3
+ size 496296380
runs/Feb27_13-41-58_db74930b30d9/events.out.tfevents.1740663720.db74930b30d9.911.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02c804e0537ffbb3c81d2e9fe51e5c9b398a83cc59c673b7dc69e39b5da30830
3
+ size 6268
runs/Feb27_13-44-28_db74930b30d9/events.out.tfevents.1740663868.db74930b30d9.911.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1378fe2a31bcb98abc09af566e6e27d6130abae04527a11fc01eb612d80ceda
3
+ size 13724
runs/Feb27_13-54-15_db74930b30d9/events.out.tfevents.1740664455.db74930b30d9.911.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:693334c44f3bc157cc75601cba18bf98c0b9c0b9c24637a9f57501700f513f0c
3
+ size 10871
runs/Feb27_13-56-26_db74930b30d9/events.out.tfevents.1740664586.db74930b30d9.911.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b22b62145da73f857f38e570743a01edbe1eafb748c5cb1eb82d24f44ce2a6f3
3
+ size 13976
runs/Feb27_13-58-49_db74930b30d9/events.out.tfevents.1740664729.db74930b30d9.911.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee1b259c4e382f83d580fc207bf01eee4e3edf7e34e75aacbf233c4c61613b4d
3
+ size 7154
runs/Feb27_13-59-45_db74930b30d9/events.out.tfevents.1740664786.db74930b30d9.911.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52ce82e251f839e376efe66569d3ace925bffb3340a39eba6391557b456c844c
3
+ size 14871
runs/Feb27_14-04-10_db74930b30d9/events.out.tfevents.1740665051.db74930b30d9.911.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4479d29a5c8d62f7e9e16369c48f1e63bc4bb2d71e1dd4a0595f14f66f4bf8d
3
+ size 16540
runs/Feb27_14-12-12_db74930b30d9/events.out.tfevents.1740665532.db74930b30d9.911.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:933617470ce9c5c109b22a0e2d9eae51ee436803c1e447d1e44a5e344b98613b
3
+ size 9756
runs/Feb27_14-14-07_db74930b30d9/events.out.tfevents.1740665648.db74930b30d9.911.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34bc05bfb5bd99c99855eb6cb6087bc31d2c3ca5187689c184380fca18bc35ea
3
+ size 15002
special_tokens_map.json CHANGED
@@ -1,7 +1,15 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,56 +1,58 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[PAD]",
5
  "lstrip": false,
6
- "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
- "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
- "content": "[CLS]",
21
  "lstrip": false,
22
- "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
- "content": "[SEP]",
29
  "lstrip": false,
30
- "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": false,
45
- "cls_token": "[CLS]",
46
- "do_lower_case": true,
 
47
  "extra_special_tokens": {},
48
- "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
- "pad_token": "[PAD]",
51
- "sep_token": "[SEP]",
52
- "strip_accents": null,
53
- "tokenize_chinese_chars": true,
54
- "tokenizer_class": "DistilBertTokenizer",
55
- "unk_token": "[UNK]"
56
  }
 
1
  {
2
+ "add_prefix_space": true,
3
  "added_tokens_decoder": {
4
  "0": {
5
+ "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
+ "1": {
13
+ "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
19
  },
20
+ "2": {
21
+ "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
+ "3": {
29
+ "content": "<unk>",
30
  "lstrip": false,
31
+ "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
  }
44
  },
45
+ "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
  "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
  "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
 
58
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7787e2f7d843089369530f9da3cbba46eb271f002f6deb2d26ddbced101d2372
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a6647940cf51d5452e3002f44c907fd51093c4f0d0208669a7d09ee4aa86113
3
  size 5368
vocab.json ADDED
The diff for this file is too large to render. See raw diff