shorecode commited on
Commit
1d0b25a
1 Parent(s): 77d476a
Files changed (4) hide show
  1. README.md +23 -7
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,7 +1,6 @@
1
  ---
2
- base_model: google/t5-efficient-tiny-nh8
3
- library_name: peft
4
- license: apache-2.0
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,7 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # t5-efficient-tiny-nh8-summarizer
16
 
17
- This model is a fine-tuned version of [google/t5-efficient-tiny-nh8](https://huggingface.co/google/t5-efficient-tiny-nh8) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
@@ -33,7 +34,7 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 0.0002
37
  - train_batch_size: 63
38
  - eval_batch_size: 63
39
  - seed: 42
@@ -44,12 +45,27 @@ The following hyperparameters were used during training:
44
 
45
  ### Training results
46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
 
48
 
49
  ### Framework versions
50
 
51
- - PEFT 0.14.0
52
  - Transformers 4.47.0
53
  - Pytorch 2.4.0+cu121
54
  - Datasets 3.0.0
55
- - Tokenizers 0.21.0
 
1
  ---
2
+ library_name: transformers
3
+ base_model: tarekziade/wikipedia-summaries-t5-efficient-tiny
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # t5-efficient-tiny-nh8-summarizer
15
 
16
+ This model is a fine-tuned version of [tarekziade/wikipedia-summaries-t5-efficient-tiny](https://huggingface.co/tarekziade/wikipedia-summaries-t5-efficient-tiny) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: nan
19
 
20
  ## Model description
21
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
+ - learning_rate: 3.0000000000000004e-05
38
  - train_batch_size: 63
39
  - eval_batch_size: 63
40
  - seed: 42
 
45
 
46
  ### Training results
47
 
48
+ | Training Loss | Epoch | Step | Validation Loss |
49
+ |:-------------:|:------:|:----:|:---------------:|
50
+ | 0.0 | 0.2096 | 200 | nan |
51
+ | 0.0 | 0.4193 | 400 | nan |
52
+ | 0.0 | 0.6289 | 600 | nan |
53
+ | 0.0 | 0.8386 | 800 | nan |
54
+ | 0.0 | 1.0482 | 1000 | nan |
55
+ | 0.0 | 1.2579 | 1200 | nan |
56
+ | 0.0 | 1.4675 | 1400 | nan |
57
+ | 0.0 | 1.6771 | 1600 | nan |
58
+ | 0.0 | 1.8868 | 1800 | nan |
59
+ | 0.0 | 2.0964 | 2000 | nan |
60
+ | 0.0 | 2.3061 | 2200 | nan |
61
+ | 0.0 | 2.5157 | 2400 | nan |
62
+ | 0.0 | 2.7254 | 2600 | nan |
63
+ | 0.0 | 2.9350 | 2800 | nan |
64
 
65
 
66
  ### Framework versions
67
 
 
68
  - Transformers 4.47.0
69
  - Pytorch 2.4.0+cu121
70
  - Datasets 3.0.0
71
+ - Tokenizers 0.21.0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "shorecode/t5-efficient-tiny-nh8-summarizer",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "tarekziade/wikipedia-summaries-t5-efficient-tiny",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a59a8c9c0ed288ff84d9af8d349bc7f8a93fef22d16d02f70e19f317c75f18e
3
  size 62293080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:872e0ba189e975d6c71e7b192f221dbf797621d8b412b3a6071b633e589892f6
3
  size 62293080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54ba631ad7f037db6625df8b1984c9558bdff959c75d58765bcdf550a535821b
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3174668a328939c678d3aa9dcc142da4000c6dfa8bd3f7227b2a4a7d69e8be40
3
  size 5304