AriyanH22 commited on
Commit
73275bb
·
verified ·
1 Parent(s): e94db72

End of training

Browse files
Files changed (7) hide show
  1. README.md +23 -16
  2. config.json +1 -2
  3. merges.txt +1 -1
  4. pytorch_model.bin +2 -2
  5. tokenizer.json +1 -0
  6. tokenizer_config.json +1 -1
  7. training_args.bin +2 -2
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
  metrics:
@@ -19,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 1.1860
23
- - Accuracy: 0.8377
24
- - F1: 0.9093
25
- - Precision: 0.8493
26
- - Recall: 0.9785
27
 
28
  ## Model description
29
 
@@ -42,27 +43,33 @@ More information needed
42
  ### Training hyperparameters
43
 
44
  The following hyperparameters were used during training:
45
- - learning_rate: 5e-05
46
  - train_batch_size: 1
47
- - eval_batch_size: 2
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
- - num_epochs: 3
52
- - mixed_precision_training: Native AMP
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:---------:|:------:|
58
- | 0.0053 | 1.0 | 5152 | 1.4254 | 0.8315 | 0.9080 | 0.8315 | 1.0 |
59
- | 0.0107 | 2.0 | 10304 | 1.2232 | 0.8315 | 0.9080 | 0.8315 | 1.0 |
60
- | 1.2863 | 3.0 | 15456 | 1.1860 | 0.8377 | 0.9093 | 0.8493 | 0.9785 |
 
 
 
 
 
 
 
61
 
62
 
63
  ### Framework versions
64
 
65
- - Transformers 4.27.4
66
- - Pytorch 2.2.0+cu121
67
- - Datasets 2.18.0
68
- - Tokenizers 0.13.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: allenai/longformer-base-4096
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
20
 
21
  This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4961
24
+ - Accuracy: 0.9068
25
+ - F1: 0.9452
26
+ - Precision: 0.9265
27
+ - Recall: 0.9646
28
 
29
  ## Model description
30
 
 
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
+ - learning_rate: 2e-05
47
  - train_batch_size: 1
48
+ - eval_batch_size: 1
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - num_epochs: 10
 
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:---------:|:------:|
58
+ | 0.8485 | 1.0 | 5152 | 0.8574 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
59
+ | 0.7968 | 2.0 | 10304 | 0.8441 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
60
+ | 0.847 | 3.0 | 15456 | 0.8049 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
61
+ | 0.8677 | 4.0 | 20608 | 0.7919 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
62
+ | 0.8778 | 5.0 | 25760 | 0.8980 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
63
+ | 0.7563 | 6.0 | 30912 | 0.8299 | 0.8323 | 0.9085 | 0.8323 | 1.0 |
64
+ | 0.661 | 7.0 | 36064 | 0.6065 | 0.8882 | 0.9357 | 0.8973 | 0.9776 |
65
+ | 0.8207 | 8.0 | 41216 | 0.5387 | 0.8975 | 0.9410 | 0.9038 | 0.9813 |
66
+ | 0.6872 | 9.0 | 46368 | 0.5960 | 0.8602 | 0.9212 | 0.8680 | 0.9813 |
67
+ | 0.4596 | 10.0 | 51520 | 0.4961 | 0.9068 | 0.9452 | 0.9265 | 0.9646 |
68
 
69
 
70
  ### Framework versions
71
 
72
+ - Transformers 4.32.0.dev0
73
+ - Pytorch 2.0.1
74
+ - Datasets 2.14.4
75
+ - Tokenizers 0.13.3
config.json CHANGED
@@ -43,11 +43,10 @@
43
  "num_hidden_layers": 12,
44
  "onnx_export": false,
45
  "pad_token_id": 1,
46
- "position_embedding_type": "absolute",
47
  "problem_type": "single_label_classification",
48
  "sep_token_id": 2,
49
  "torch_dtype": "float32",
50
- "transformers_version": "4.27.4",
51
  "type_vocab_size": 1,
52
  "vocab_size": 50265
53
  }
 
43
  "num_hidden_layers": 12,
44
  "onnx_export": false,
45
  "pad_token_id": 1,
 
46
  "problem_type": "single_label_classification",
47
  "sep_token_id": 2,
48
  "torch_dtype": "float32",
49
+ "transformers_version": "4.32.0.dev0",
50
  "type_vocab_size": 1,
51
  "vocab_size": 50265
52
  }
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:178624f5e416d2324eb092aca88314c7e41c27ae3cd6683746e1fa059dbcf8d8
3
- size 594739726
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f743a83cf538002482f27db4cbfc382cb110c6dcd8ecc9a5125adb8a051f37e8
3
+ size 594739281
tokenizer.json CHANGED
@@ -82,6 +82,7 @@
82
  "continuing_subword_prefix": "",
83
  "end_of_word_suffix": "",
84
  "fuse_unk": false,
 
85
  "vocab": {
86
  "<s>": 0,
87
  "<pad>": 1,
 
82
  "continuing_subword_prefix": "",
83
  "end_of_word_suffix": "",
84
  "fuse_unk": false,
85
+ "byte_fallback": false,
86
  "vocab": {
87
  "<s>": 0,
88
  "<pad>": 1,
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<s>",
 
4
  "cls_token": "<s>",
5
  "eos_token": "</s>",
6
  "errors": "replace",
@@ -9,7 +10,6 @@
9
  "model_max_length": 4096,
10
  "pad_token": "<pad>",
11
  "sep_token": "</s>",
12
- "special_tokens_map_file": null,
13
  "tokenizer_class": "LongformerTokenizer",
14
  "trim_offsets": true,
15
  "unk_token": "<unk>"
 
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<s>",
4
+ "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
6
  "eos_token": "</s>",
7
  "errors": "replace",
 
10
  "model_max_length": 4096,
11
  "pad_token": "<pad>",
12
  "sep_token": "</s>",
 
13
  "tokenizer_class": "LongformerTokenizer",
14
  "trim_offsets": true,
15
  "unk_token": "<unk>"
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01c47c7ef56337c9b0d198aefe3555b2889a6b6917880f0aec799c150894cb3d
3
- size 3960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4788a8133052e4ee70ab0ba139a04d730dc16b230624e8e6d542478eabdc68b6
3
+ size 4091