SimoneJLaudani commited on
Commit
fd7f47e
·
verified ·
1 Parent(s): 0f5f860

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -18,13 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # trainerH
20
 
21
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.6940
24
- - Precision: 0.7930
25
- - Recall: 0.7871
26
- - F1: 0.7867
27
- - Accuracy: 0.7871
28
 
29
  ## Model description
30
 
@@ -55,18 +55,18 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
- | 0.3386 | 1.07 | 30 | 0.6489 | 0.7405 | 0.7395 | 0.7358 | 0.7395 |
59
- | 0.2218 | 2.14 | 60 | 0.6254 | 0.7593 | 0.7591 | 0.7570 | 0.7591 |
60
- | 0.1485 | 3.21 | 90 | 0.6242 | 0.7842 | 0.7815 | 0.7801 | 0.7815 |
61
- | 0.0913 | 4.29 | 120 | 0.6697 | 0.7739 | 0.7703 | 0.7693 | 0.7703 |
62
- | 0.0646 | 5.36 | 150 | 0.6619 | 0.7893 | 0.7871 | 0.7862 | 0.7871 |
63
- | 0.0484 | 6.43 | 180 | 0.7026 | 0.8078 | 0.8011 | 0.7998 | 0.8011 |
64
- | 0.0423 | 7.5 | 210 | 0.6913 | 0.7904 | 0.7871 | 0.7864 | 0.7871 |
65
 
66
 
67
  ### Framework versions
68
 
69
- - Transformers 4.39.0
70
  - Pytorch 2.2.1+cu121
71
  - Datasets 2.18.0
72
  - Tokenizers 0.15.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: distilbert-base-cased
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
18
 
19
  # trainerH
20
 
21
+ This model is a fine-tuned version of [distilbert-base-cased](https://huggingface.co/distilbert-base-cased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.6797
24
+ - Precision: 0.7850
25
+ - Recall: 0.7787
26
+ - F1: 0.7762
27
+ - Accuracy: 0.7787
28
 
29
  ## Model description
30
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | 1.8653 | 1.07 | 30 | 1.7135 | 0.3177 | 0.3501 | 0.2302 | 0.3501 |
59
+ | 1.5399 | 2.14 | 60 | 1.3143 | 0.6622 | 0.5490 | 0.5241 | 0.5490 |
60
+ | 1.1791 | 3.21 | 90 | 1.0361 | 0.7271 | 0.7115 | 0.7074 | 0.7115 |
61
+ | 0.891 | 4.29 | 120 | 0.8605 | 0.7451 | 0.7255 | 0.7214 | 0.7255 |
62
+ | 0.6817 | 5.36 | 150 | 0.7495 | 0.7595 | 0.7507 | 0.7470 | 0.7507 |
63
+ | 0.539 | 6.43 | 180 | 0.6965 | 0.7987 | 0.7927 | 0.7918 | 0.7927 |
64
+ | 0.4777 | 7.5 | 210 | 0.6830 | 0.7787 | 0.7703 | 0.7680 | 0.7703 |
65
 
66
 
67
  ### Framework versions
68
 
69
+ - Transformers 4.39.1
70
  - Pytorch 2.2.1+cu121
71
  - Datasets 2.18.0
72
  - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
@@ -31,6 +31,7 @@
31
  "model_type": "distilbert",
32
  "n_heads": 12,
33
  "n_layers": 6,
 
34
  "pad_token_id": 0,
35
  "problem_type": "single_label_classification",
36
  "qa_dropout": 0.1,
@@ -38,6 +39,6 @@
38
  "sinusoidal_pos_embds": false,
39
  "tie_weights_": true,
40
  "torch_dtype": "float32",
41
- "transformers_version": "4.39.0",
42
- "vocab_size": 30522
43
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
 
31
  "model_type": "distilbert",
32
  "n_heads": 12,
33
  "n_layers": 6,
34
+ "output_past": true,
35
  "pad_token_id": 0,
36
  "problem_type": "single_label_classification",
37
  "qa_dropout": 0.1,
 
39
  "sinusoidal_pos_embds": false,
40
  "tie_weights_": true,
41
  "torch_dtype": "float32",
42
+ "transformers_version": "4.39.1",
43
+ "vocab_size": 28996
44
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce29d90050f24a4348e3c8c6a1fb154c62876e55674e6fde6618d948511c7340
3
- size 267847948
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60a2f4755de6c20a6344b9b8db8468e590f4af44b2af9651727b8779f6df153
3
+ size 263160068
runs/Mar23_13-11-27_3015ac71de9b/events.out.tfevents.1711199496.3015ac71de9b.221.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6faebb6c676704bf96506c099336ec7c2d7a7f8ab64ff1a51ab0abff5d9df04b
3
+ size 9861
runs/Mar23_13-11-27_3015ac71de9b/events.out.tfevents.1711199793.3015ac71de9b.221.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a192b00b3cc3a3ecfb731445a4bc574cbf9511a0ce9fd55e050989542eb4b91
3
+ size 560
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -43,11 +43,9 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
- "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
53
  "strip_accents": null,
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_lower_case": false,
 
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c9e704479209c1027a5e5144c082532e5fe882e9363aab60ead9e7e7bef0cd4
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3c382a07ee673191df35f19c87872a95e79134a0ec6ae6a4d5efdbdba8601d6
3
  size 4920
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff