Training in progress epoch 0
Browse files- README.md +5 -149
- config.json +1 -1
- tf_model.h5 +1 -1
- tokenizer_config.json +4 -0
README.md
CHANGED
@@ -1,6 +1,4 @@
|
|
1 |
---
|
2 |
-
license: apache-2.0
|
3 |
-
base_model: t5-small
|
4 |
tags:
|
5 |
- generated_from_keras_callback
|
6 |
model-index:
|
@@ -13,11 +11,11 @@ probably proofread and complete it, then remove this comment. -->
|
|
13 |
|
14 |
# tarsssss/eng-jagoy-t5-002
|
15 |
|
16 |
-
This model
|
17 |
It achieves the following results on the evaluation set:
|
18 |
-
- Train Loss: 4.
|
19 |
-
- Validation Loss:
|
20 |
-
- Epoch:
|
21 |
|
22 |
## Model description
|
23 |
|
@@ -43,149 +41,7 @@ The following hyperparameters were used during training:
|
|
43 |
|
44 |
| Train Loss | Validation Loss | Epoch |
|
45 |
|:----------:|:---------------:|:-----:|
|
46 |
-
|
|
47 |
-
| 7.4261 | 7.2365 | 1 |
|
48 |
-
| 7.2130 | 7.0830 | 2 |
|
49 |
-
| 7.0761 | 6.9717 | 3 |
|
50 |
-
| 6.9678 | 6.8724 | 4 |
|
51 |
-
| 6.8904 | 6.7979 | 5 |
|
52 |
-
| 6.8053 | 6.7267 | 6 |
|
53 |
-
| 6.7518 | 6.6660 | 7 |
|
54 |
-
| 6.6899 | 6.6093 | 8 |
|
55 |
-
| 6.6337 | 6.5595 | 9 |
|
56 |
-
| 6.5867 | 6.5166 | 10 |
|
57 |
-
| 6.5419 | 6.4748 | 11 |
|
58 |
-
| 6.5020 | 6.4362 | 12 |
|
59 |
-
| 6.4560 | 6.3967 | 13 |
|
60 |
-
| 6.4189 | 6.3589 | 14 |
|
61 |
-
| 6.3782 | 6.3209 | 15 |
|
62 |
-
| 6.3324 | 6.2874 | 16 |
|
63 |
-
| 6.2983 | 6.2530 | 17 |
|
64 |
-
| 6.2466 | 6.2201 | 18 |
|
65 |
-
| 6.2155 | 6.1900 | 19 |
|
66 |
-
| 6.1906 | 6.1594 | 20 |
|
67 |
-
| 6.1562 | 6.1335 | 21 |
|
68 |
-
| 6.1244 | 6.1083 | 22 |
|
69 |
-
| 6.1023 | 6.0840 | 23 |
|
70 |
-
| 6.0687 | 6.0593 | 24 |
|
71 |
-
| 6.0526 | 6.0386 | 25 |
|
72 |
-
| 6.0281 | 6.0160 | 26 |
|
73 |
-
| 5.9981 | 5.9931 | 27 |
|
74 |
-
| 5.9756 | 5.9769 | 28 |
|
75 |
-
| 5.9500 | 5.9537 | 29 |
|
76 |
-
| 5.9325 | 5.9356 | 30 |
|
77 |
-
| 5.9008 | 5.9193 | 31 |
|
78 |
-
| 5.8793 | 5.9051 | 32 |
|
79 |
-
| 5.8718 | 5.8847 | 33 |
|
80 |
-
| 5.8474 | 5.8641 | 34 |
|
81 |
-
| 5.8370 | 5.8531 | 35 |
|
82 |
-
| 5.8135 | 5.8424 | 36 |
|
83 |
-
| 5.7981 | 5.8243 | 37 |
|
84 |
-
| 5.7759 | 5.8142 | 38 |
|
85 |
-
| 5.7582 | 5.8050 | 39 |
|
86 |
-
| 5.7482 | 5.7910 | 40 |
|
87 |
-
| 5.7354 | 5.7740 | 41 |
|
88 |
-
| 5.7066 | 5.7762 | 42 |
|
89 |
-
| 5.6981 | 5.7576 | 43 |
|
90 |
-
| 5.6797 | 5.7470 | 44 |
|
91 |
-
| 5.6631 | 5.7400 | 45 |
|
92 |
-
| 5.6511 | 5.7255 | 46 |
|
93 |
-
| 5.6433 | 5.7211 | 47 |
|
94 |
-
| 5.6196 | 5.7050 | 48 |
|
95 |
-
| 5.6173 | 5.6980 | 49 |
|
96 |
-
| 5.5903 | 5.6902 | 50 |
|
97 |
-
| 5.5723 | 5.6808 | 51 |
|
98 |
-
| 5.5701 | 5.6754 | 52 |
|
99 |
-
| 5.5537 | 5.6724 | 53 |
|
100 |
-
| 5.5478 | 5.6545 | 54 |
|
101 |
-
| 5.5182 | 5.6566 | 55 |
|
102 |
-
| 5.5112 | 5.6417 | 56 |
|
103 |
-
| 5.4994 | 5.6290 | 57 |
|
104 |
-
| 5.4895 | 5.6216 | 58 |
|
105 |
-
| 5.4610 | 5.6147 | 59 |
|
106 |
-
| 5.4567 | 5.6035 | 60 |
|
107 |
-
| 5.4433 | 5.5957 | 61 |
|
108 |
-
| 5.4386 | 5.5855 | 62 |
|
109 |
-
| 5.4253 | 5.5851 | 63 |
|
110 |
-
| 5.4214 | 5.5804 | 64 |
|
111 |
-
| 5.4117 | 5.5714 | 65 |
|
112 |
-
| 5.3940 | 5.5632 | 66 |
|
113 |
-
| 5.3756 | 5.5616 | 67 |
|
114 |
-
| 5.3668 | 5.5540 | 68 |
|
115 |
-
| 5.3483 | 5.5465 | 69 |
|
116 |
-
| 5.3523 | 5.5380 | 70 |
|
117 |
-
| 5.3367 | 5.5371 | 71 |
|
118 |
-
| 5.3398 | 5.5187 | 72 |
|
119 |
-
| 5.3236 | 5.5241 | 73 |
|
120 |
-
| 5.3003 | 5.5139 | 74 |
|
121 |
-
| 5.3053 | 5.5093 | 75 |
|
122 |
-
| 5.2933 | 5.5120 | 76 |
|
123 |
-
| 5.2735 | 5.5064 | 77 |
|
124 |
-
| 5.2715 | 5.4971 | 78 |
|
125 |
-
| 5.2476 | 5.4956 | 79 |
|
126 |
-
| 5.2534 | 5.4750 | 80 |
|
127 |
-
| 5.2274 | 5.4845 | 81 |
|
128 |
-
| 5.2343 | 5.4761 | 82 |
|
129 |
-
| 5.2161 | 5.4686 | 83 |
|
130 |
-
| 5.1994 | 5.4656 | 84 |
|
131 |
-
| 5.1993 | 5.4551 | 85 |
|
132 |
-
| 5.1786 | 5.4680 | 86 |
|
133 |
-
| 5.1800 | 5.4557 | 87 |
|
134 |
-
| 5.1552 | 5.4460 | 88 |
|
135 |
-
| 5.1547 | 5.4377 | 89 |
|
136 |
-
| 5.1450 | 5.4365 | 90 |
|
137 |
-
| 5.1429 | 5.4228 | 91 |
|
138 |
-
| 5.1303 | 5.4245 | 92 |
|
139 |
-
| 5.1164 | 5.4177 | 93 |
|
140 |
-
| 5.1200 | 5.4135 | 94 |
|
141 |
-
| 5.1115 | 5.4120 | 95 |
|
142 |
-
| 5.0928 | 5.3963 | 96 |
|
143 |
-
| 5.0825 | 5.3945 | 97 |
|
144 |
-
| 5.0760 | 5.4078 | 98 |
|
145 |
-
| 5.0733 | 5.3935 | 99 |
|
146 |
-
| 5.0611 | 5.3944 | 100 |
|
147 |
-
| 5.0457 | 5.3882 | 101 |
|
148 |
-
| 5.0296 | 5.3774 | 102 |
|
149 |
-
| 5.0311 | 5.3812 | 103 |
|
150 |
-
| 5.0189 | 5.3756 | 104 |
|
151 |
-
| 5.0076 | 5.3756 | 105 |
|
152 |
-
| 5.0037 | 5.3751 | 106 |
|
153 |
-
| 4.9992 | 5.3582 | 107 |
|
154 |
-
| 4.9774 | 5.3603 | 108 |
|
155 |
-
| 4.9789 | 5.3679 | 109 |
|
156 |
-
| 4.9717 | 5.3566 | 110 |
|
157 |
-
| 4.9684 | 5.3489 | 111 |
|
158 |
-
| 4.9578 | 5.3454 | 112 |
|
159 |
-
| 4.9519 | 5.3457 | 113 |
|
160 |
-
| 4.9327 | 5.3380 | 114 |
|
161 |
-
| 4.9413 | 5.3324 | 115 |
|
162 |
-
| 4.9093 | 5.3461 | 116 |
|
163 |
-
| 4.9126 | 5.3269 | 117 |
|
164 |
-
| 4.9096 | 5.3283 | 118 |
|
165 |
-
| 4.8899 | 5.3252 | 119 |
|
166 |
-
| 4.8936 | 5.3227 | 120 |
|
167 |
-
| 4.8873 | 5.3213 | 121 |
|
168 |
-
| 4.8657 | 5.3216 | 122 |
|
169 |
-
| 4.8581 | 5.3195 | 123 |
|
170 |
-
| 4.8619 | 5.3218 | 124 |
|
171 |
-
| 4.8534 | 5.3062 | 125 |
|
172 |
-
| 4.8466 | 5.3037 | 126 |
|
173 |
-
| 4.8239 | 5.3075 | 127 |
|
174 |
-
| 4.8287 | 5.3052 | 128 |
|
175 |
-
| 4.8165 | 5.2992 | 129 |
|
176 |
-
| 4.8136 | 5.2968 | 130 |
|
177 |
-
| 4.8005 | 5.2902 | 131 |
|
178 |
-
| 4.7898 | 5.2949 | 132 |
|
179 |
-
| 4.7949 | 5.2858 | 133 |
|
180 |
-
| 4.7786 | 5.2868 | 134 |
|
181 |
-
| 4.7706 | 5.2928 | 135 |
|
182 |
-
| 4.7579 | 5.2866 | 136 |
|
183 |
-
| 4.7477 | 5.2772 | 137 |
|
184 |
-
| 4.7517 | 5.2827 | 138 |
|
185 |
-
| 4.7391 | 5.2849 | 139 |
|
186 |
-
| 4.7268 | 5.2823 | 140 |
|
187 |
-
| 4.7275 | 5.2832 | 141 |
|
188 |
-
| 4.7104 | 5.2694 | 142 |
|
189 |
|
190 |
|
191 |
### Framework versions
|
|
|
1 |
---
|
|
|
|
|
2 |
tags:
|
3 |
- generated_from_keras_callback
|
4 |
model-index:
|
|
|
11 |
|
12 |
# tarsssss/eng-jagoy-t5-002
|
13 |
|
14 |
+
This model was trained from scratch on an unknown dataset.
|
15 |
It achieves the following results on the evaluation set:
|
16 |
+
- Train Loss: 4.8828
|
17 |
+
- Validation Loss: 4.4230
|
18 |
+
- Epoch: 0
|
19 |
|
20 |
## Model description
|
21 |
|
|
|
41 |
|
42 |
| Train Loss | Validation Loss | Epoch |
|
43 |
|:----------:|:---------------:|:-----:|
|
44 |
+
| 4.8828 | 4.4230 | 0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
|
46 |
|
47 |
### Framework versions
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "t5-
|
3 |
"architectures": [
|
4 |
"T5ForConditionalGeneration"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "./eng-jagoy-t5-002",
|
3 |
"architectures": [
|
4 |
"T5ForConditionalGeneration"
|
5 |
],
|
tf_model.h5
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 373902664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fecdab4ca7e83d87db3bc3aa082cbfdabf14dbad513f2e95b68884523e3651ba
|
3 |
size 373902664
|
tokenizer_config.json
CHANGED
@@ -104,8 +104,12 @@
|
|
104 |
"clean_up_tokenization_spaces": true,
|
105 |
"eos_token": "</s>",
|
106 |
"extra_ids": 100,
|
|
|
107 |
"model_max_length": 512,
|
108 |
"pad_token": "<pad>",
|
|
|
109 |
"tokenizer_class": "T5Tokenizer",
|
|
|
|
|
110 |
"unk_token": "<unk>"
|
111 |
}
|
|
|
104 |
"clean_up_tokenization_spaces": true,
|
105 |
"eos_token": "</s>",
|
106 |
"extra_ids": 100,
|
107 |
+
"max_length": 128,
|
108 |
"model_max_length": 512,
|
109 |
"pad_token": "<pad>",
|
110 |
+
"stride": 0,
|
111 |
"tokenizer_class": "T5Tokenizer",
|
112 |
+
"truncation_side": "right",
|
113 |
+
"truncation_strategy": "longest_first",
|
114 |
"unk_token": "<unk>"
|
115 |
}
|