shirzady1934 commited on
Commit
a9bc9c4
1 Parent(s): 6ccfbff

End of training

Browse files
Files changed (4) hide show
  1. README.md +35 -35
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.3192
21
- - Accuracy: 0.8500
22
 
23
  ## Model description
24
 
@@ -38,8 +38,8 @@ More information needed
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0005
41
- - train_batch_size: 16
42
- - eval_batch_size: 16
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
@@ -49,41 +49,41 @@ The following hyperparameters were used during training:
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | No log | 1.0 | 23 | 0.4749 | 0.75 |
53
- | No log | 2.0 | 46 | 0.4396 | 0.7750 |
54
- | No log | 3.0 | 69 | 0.4988 | 0.7750 |
55
- | No log | 4.0 | 92 | 0.4534 | 0.8000 |
56
- | No log | 5.0 | 115 | 0.4505 | 0.8250 |
57
- | No log | 6.0 | 138 | 0.4108 | 0.8250 |
58
- | No log | 7.0 | 161 | 0.4701 | 0.8000 |
59
- | No log | 8.0 | 184 | 0.4327 | 0.8250 |
60
- | No log | 9.0 | 207 | 0.5293 | 0.8000 |
61
- | No log | 10.0 | 230 | 0.5596 | 0.7750 |
62
- | No log | 11.0 | 253 | 0.4872 | 0.9000 |
63
- | No log | 12.0 | 276 | 0.3860 | 0.8500 |
64
- | No log | 13.0 | 299 | 0.4549 | 0.9000 |
65
- | No log | 14.0 | 322 | 0.4340 | 0.8250 |
66
- | No log | 15.0 | 345 | 0.4540 | 0.7750 |
67
- | No log | 16.0 | 368 | 0.5259 | 0.75 |
68
- | No log | 17.0 | 391 | 0.3192 | 0.8500 |
69
- | No log | 18.0 | 414 | 0.3699 | 0.875 |
70
- | No log | 19.0 | 437 | 0.3577 | 0.875 |
71
- | No log | 20.0 | 460 | 0.4405 | 0.8250 |
72
- | No log | 21.0 | 483 | 0.5207 | 0.8250 |
73
- | 0.1396 | 22.0 | 506 | 0.4686 | 0.8000 |
74
- | 0.1396 | 23.0 | 529 | 0.4614 | 0.875 |
75
- | 0.1396 | 24.0 | 552 | 0.4442 | 0.8250 |
76
- | 0.1396 | 25.0 | 575 | 0.4242 | 0.8250 |
77
- | 0.1396 | 26.0 | 598 | 0.4943 | 0.8000 |
78
- | 0.1396 | 27.0 | 621 | 0.4973 | 0.8500 |
79
- | 0.1396 | 28.0 | 644 | 0.4542 | 0.875 |
80
- | 0.1396 | 29.0 | 667 | 0.4671 | 0.875 |
81
- | 0.1396 | 30.0 | 690 | 0.4679 | 0.875 |
82
 
83
 
84
  ### Framework versions
85
 
86
- - Transformers 4.36.2
87
  - Pytorch 1.13.1+cu117
88
  - Datasets 2.16.1
89
  - Tokenizers 0.15.0
 
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.3450
21
+ - Accuracy: 0.875
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0005
41
+ - train_batch_size: 32
42
+ - eval_batch_size: 32
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
 
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 12 | 1.3014 | 0.6000 |
53
+ | No log | 2.0 | 24 | 1.0521 | 0.5 |
54
+ | No log | 3.0 | 36 | 0.9390 | 0.6500 |
55
+ | No log | 4.0 | 48 | 0.9100 | 0.5250 |
56
+ | No log | 5.0 | 60 | 0.8639 | 0.7000 |
57
+ | No log | 6.0 | 72 | 0.7679 | 0.75 |
58
+ | No log | 7.0 | 84 | 0.6577 | 0.75 |
59
+ | No log | 8.0 | 96 | 0.6747 | 0.7250 |
60
+ | No log | 9.0 | 108 | 0.5993 | 0.8250 |
61
+ | No log | 10.0 | 120 | 0.5482 | 0.8000 |
62
+ | No log | 11.0 | 132 | 0.4691 | 0.9250 |
63
+ | No log | 12.0 | 144 | 0.7105 | 0.7750 |
64
+ | No log | 13.0 | 156 | 0.4716 | 0.9250 |
65
+ | No log | 14.0 | 168 | 0.4583 | 0.8000 |
66
+ | No log | 15.0 | 180 | 0.3940 | 0.8500 |
67
+ | No log | 16.0 | 192 | 0.3841 | 0.875 |
68
+ | No log | 17.0 | 204 | 0.3450 | 0.875 |
69
+ | No log | 18.0 | 216 | 0.3995 | 0.8500 |
70
+ | No log | 19.0 | 228 | 0.4778 | 0.8250 |
71
+ | No log | 20.0 | 240 | 0.4217 | 0.8500 |
72
+ | No log | 21.0 | 252 | 0.3841 | 0.875 |
73
+ | No log | 22.0 | 264 | 0.3971 | 0.875 |
74
+ | No log | 23.0 | 276 | 0.4158 | 0.8250 |
75
+ | No log | 24.0 | 288 | 0.3915 | 0.8500 |
76
+ | No log | 25.0 | 300 | 0.4234 | 0.8250 |
77
+ | No log | 26.0 | 312 | 0.3857 | 0.8500 |
78
+ | No log | 27.0 | 324 | 0.4306 | 0.8250 |
79
+ | No log | 28.0 | 336 | 0.4166 | 0.8250 |
80
+ | No log | 29.0 | 348 | 0.4101 | 0.8500 |
81
+ | No log | 30.0 | 360 | 0.4044 | 0.8500 |
82
 
83
 
84
  ### Framework versions
85
 
86
+ - Transformers 4.37.2
87
  - Pytorch 1.13.1+cu117
88
  - Datasets 2.16.1
89
  - Tokenizers 0.15.0
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.36.2",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
 
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "torch_dtype": "float32",
23
+ "transformers_version": "4.37.2",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6b278e98c0a756838e5cba2f200c2e281724b0e7e8259e44b234769f1af74a2
3
  size 499796292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac3e9b32fd383d77b15506b2c8829f310c5be4c65d2b4b43838b788dd6f25890
3
  size 499796292
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ead920aa307d80c1ade03639b074dc91ce07cc31c0a8fc6fff62acf3c4c14985
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66d8c0f0956438e072965a382f44ed34ee2ea1df5762f28d25393fe1fafde091
3
  size 4283