tarsssss commited on
Commit
946c722
·
1 Parent(s): 6afa6ee

Training in progress epoch 0

Browse files
Files changed (4) hide show
  1. README.md +5 -149
  2. config.json +1 -1
  3. tf_model.h5 +1 -1
  4. tokenizer_config.json +4 -0
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: apache-2.0
3
- base_model: t5-small
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
@@ -13,11 +11,11 @@ probably proofread and complete it, then remove this comment. -->
13
 
14
  # tarsssss/eng-jagoy-t5-002
15
 
16
- This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 4.7104
19
- - Validation Loss: 5.2694
20
- - Epoch: 142
21
 
22
  ## Model description
23
 
@@ -43,149 +41,7 @@ The following hyperparameters were used during training:
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
- | 7.9230 | 7.4880 | 0 |
47
- | 7.4261 | 7.2365 | 1 |
48
- | 7.2130 | 7.0830 | 2 |
49
- | 7.0761 | 6.9717 | 3 |
50
- | 6.9678 | 6.8724 | 4 |
51
- | 6.8904 | 6.7979 | 5 |
52
- | 6.8053 | 6.7267 | 6 |
53
- | 6.7518 | 6.6660 | 7 |
54
- | 6.6899 | 6.6093 | 8 |
55
- | 6.6337 | 6.5595 | 9 |
56
- | 6.5867 | 6.5166 | 10 |
57
- | 6.5419 | 6.4748 | 11 |
58
- | 6.5020 | 6.4362 | 12 |
59
- | 6.4560 | 6.3967 | 13 |
60
- | 6.4189 | 6.3589 | 14 |
61
- | 6.3782 | 6.3209 | 15 |
62
- | 6.3324 | 6.2874 | 16 |
63
- | 6.2983 | 6.2530 | 17 |
64
- | 6.2466 | 6.2201 | 18 |
65
- | 6.2155 | 6.1900 | 19 |
66
- | 6.1906 | 6.1594 | 20 |
67
- | 6.1562 | 6.1335 | 21 |
68
- | 6.1244 | 6.1083 | 22 |
69
- | 6.1023 | 6.0840 | 23 |
70
- | 6.0687 | 6.0593 | 24 |
71
- | 6.0526 | 6.0386 | 25 |
72
- | 6.0281 | 6.0160 | 26 |
73
- | 5.9981 | 5.9931 | 27 |
74
- | 5.9756 | 5.9769 | 28 |
75
- | 5.9500 | 5.9537 | 29 |
76
- | 5.9325 | 5.9356 | 30 |
77
- | 5.9008 | 5.9193 | 31 |
78
- | 5.8793 | 5.9051 | 32 |
79
- | 5.8718 | 5.8847 | 33 |
80
- | 5.8474 | 5.8641 | 34 |
81
- | 5.8370 | 5.8531 | 35 |
82
- | 5.8135 | 5.8424 | 36 |
83
- | 5.7981 | 5.8243 | 37 |
84
- | 5.7759 | 5.8142 | 38 |
85
- | 5.7582 | 5.8050 | 39 |
86
- | 5.7482 | 5.7910 | 40 |
87
- | 5.7354 | 5.7740 | 41 |
88
- | 5.7066 | 5.7762 | 42 |
89
- | 5.6981 | 5.7576 | 43 |
90
- | 5.6797 | 5.7470 | 44 |
91
- | 5.6631 | 5.7400 | 45 |
92
- | 5.6511 | 5.7255 | 46 |
93
- | 5.6433 | 5.7211 | 47 |
94
- | 5.6196 | 5.7050 | 48 |
95
- | 5.6173 | 5.6980 | 49 |
96
- | 5.5903 | 5.6902 | 50 |
97
- | 5.5723 | 5.6808 | 51 |
98
- | 5.5701 | 5.6754 | 52 |
99
- | 5.5537 | 5.6724 | 53 |
100
- | 5.5478 | 5.6545 | 54 |
101
- | 5.5182 | 5.6566 | 55 |
102
- | 5.5112 | 5.6417 | 56 |
103
- | 5.4994 | 5.6290 | 57 |
104
- | 5.4895 | 5.6216 | 58 |
105
- | 5.4610 | 5.6147 | 59 |
106
- | 5.4567 | 5.6035 | 60 |
107
- | 5.4433 | 5.5957 | 61 |
108
- | 5.4386 | 5.5855 | 62 |
109
- | 5.4253 | 5.5851 | 63 |
110
- | 5.4214 | 5.5804 | 64 |
111
- | 5.4117 | 5.5714 | 65 |
112
- | 5.3940 | 5.5632 | 66 |
113
- | 5.3756 | 5.5616 | 67 |
114
- | 5.3668 | 5.5540 | 68 |
115
- | 5.3483 | 5.5465 | 69 |
116
- | 5.3523 | 5.5380 | 70 |
117
- | 5.3367 | 5.5371 | 71 |
118
- | 5.3398 | 5.5187 | 72 |
119
- | 5.3236 | 5.5241 | 73 |
120
- | 5.3003 | 5.5139 | 74 |
121
- | 5.3053 | 5.5093 | 75 |
122
- | 5.2933 | 5.5120 | 76 |
123
- | 5.2735 | 5.5064 | 77 |
124
- | 5.2715 | 5.4971 | 78 |
125
- | 5.2476 | 5.4956 | 79 |
126
- | 5.2534 | 5.4750 | 80 |
127
- | 5.2274 | 5.4845 | 81 |
128
- | 5.2343 | 5.4761 | 82 |
129
- | 5.2161 | 5.4686 | 83 |
130
- | 5.1994 | 5.4656 | 84 |
131
- | 5.1993 | 5.4551 | 85 |
132
- | 5.1786 | 5.4680 | 86 |
133
- | 5.1800 | 5.4557 | 87 |
134
- | 5.1552 | 5.4460 | 88 |
135
- | 5.1547 | 5.4377 | 89 |
136
- | 5.1450 | 5.4365 | 90 |
137
- | 5.1429 | 5.4228 | 91 |
138
- | 5.1303 | 5.4245 | 92 |
139
- | 5.1164 | 5.4177 | 93 |
140
- | 5.1200 | 5.4135 | 94 |
141
- | 5.1115 | 5.4120 | 95 |
142
- | 5.0928 | 5.3963 | 96 |
143
- | 5.0825 | 5.3945 | 97 |
144
- | 5.0760 | 5.4078 | 98 |
145
- | 5.0733 | 5.3935 | 99 |
146
- | 5.0611 | 5.3944 | 100 |
147
- | 5.0457 | 5.3882 | 101 |
148
- | 5.0296 | 5.3774 | 102 |
149
- | 5.0311 | 5.3812 | 103 |
150
- | 5.0189 | 5.3756 | 104 |
151
- | 5.0076 | 5.3756 | 105 |
152
- | 5.0037 | 5.3751 | 106 |
153
- | 4.9992 | 5.3582 | 107 |
154
- | 4.9774 | 5.3603 | 108 |
155
- | 4.9789 | 5.3679 | 109 |
156
- | 4.9717 | 5.3566 | 110 |
157
- | 4.9684 | 5.3489 | 111 |
158
- | 4.9578 | 5.3454 | 112 |
159
- | 4.9519 | 5.3457 | 113 |
160
- | 4.9327 | 5.3380 | 114 |
161
- | 4.9413 | 5.3324 | 115 |
162
- | 4.9093 | 5.3461 | 116 |
163
- | 4.9126 | 5.3269 | 117 |
164
- | 4.9096 | 5.3283 | 118 |
165
- | 4.8899 | 5.3252 | 119 |
166
- | 4.8936 | 5.3227 | 120 |
167
- | 4.8873 | 5.3213 | 121 |
168
- | 4.8657 | 5.3216 | 122 |
169
- | 4.8581 | 5.3195 | 123 |
170
- | 4.8619 | 5.3218 | 124 |
171
- | 4.8534 | 5.3062 | 125 |
172
- | 4.8466 | 5.3037 | 126 |
173
- | 4.8239 | 5.3075 | 127 |
174
- | 4.8287 | 5.3052 | 128 |
175
- | 4.8165 | 5.2992 | 129 |
176
- | 4.8136 | 5.2968 | 130 |
177
- | 4.8005 | 5.2902 | 131 |
178
- | 4.7898 | 5.2949 | 132 |
179
- | 4.7949 | 5.2858 | 133 |
180
- | 4.7786 | 5.2868 | 134 |
181
- | 4.7706 | 5.2928 | 135 |
182
- | 4.7579 | 5.2866 | 136 |
183
- | 4.7477 | 5.2772 | 137 |
184
- | 4.7517 | 5.2827 | 138 |
185
- | 4.7391 | 5.2849 | 139 |
186
- | 4.7268 | 5.2823 | 140 |
187
- | 4.7275 | 5.2832 | 141 |
188
- | 4.7104 | 5.2694 | 142 |
189
 
190
 
191
  ### Framework versions
 
1
  ---
 
 
2
  tags:
3
  - generated_from_keras_callback
4
  model-index:
 
11
 
12
  # tarsssss/eng-jagoy-t5-002
13
 
14
+ This model was trained from scratch on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Train Loss: 4.8828
17
+ - Validation Loss: 4.4230
18
+ - Epoch: 0
19
 
20
  ## Model description
21
 
 
41
 
42
  | Train Loss | Validation Loss | Epoch |
43
  |:----------:|:---------------:|:-----:|
44
+ | 4.8828 | 4.4230 | 0 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
 
47
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "t5-small",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "./eng-jagoy-t5-002",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b85344689c0acb3d4ede8ce58079a92f9af7d98bdbd95ea1ff7158e9e3f9ba25
3
  size 373902664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fecdab4ca7e83d87db3bc3aa082cbfdabf14dbad513f2e95b68884523e3651ba
3
  size 373902664
tokenizer_config.json CHANGED
@@ -104,8 +104,12 @@
104
  "clean_up_tokenization_spaces": true,
105
  "eos_token": "</s>",
106
  "extra_ids": 100,
 
107
  "model_max_length": 512,
108
  "pad_token": "<pad>",
 
109
  "tokenizer_class": "T5Tokenizer",
 
 
110
  "unk_token": "<unk>"
111
  }
 
104
  "clean_up_tokenization_spaces": true,
105
  "eos_token": "</s>",
106
  "extra_ids": 100,
107
+ "max_length": 128,
108
  "model_max_length": 512,
109
  "pad_token": "<pad>",
110
+ "stride": 0,
111
  "tokenizer_class": "T5Tokenizer",
112
+ "truncation_side": "right",
113
+ "truncation_strategy": "longest_first",
114
  "unk_token": "<unk>"
115
  }