File size: 5,675 Bytes
1b4baf5
 
 
 
 
 
 
 
 
 
 
 
 
 
b8416cc
1b4baf5
30e0d08
 
 
1b4baf5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b8416cc
1b4baf5
 
 
 
 
 
b8416cc
ee52bb1
a3655be
b605a02
5f95213
6155222
b877e47
ee8e49a
061dbfb
d7d0cfa
014179c
3289c55
a5dc932
2ea699b
2ab0f07
1a9e037
72deebd
06c45bd
de845f3
a879685
3b50ddb
f19aed0
dc92914
4cad0a7
aadf174
def5168
5322feb
484894b
c67e1c0
9a04309
82c446c
eec2cef
24630d7
7f9a0be
6fdd0a5
7b5ada4
888f660
e1b9e19
066a5b6
bd2dd5a
2876822
71fb4a6
1188ab9
f991cf7
576eb72
5b81015
8390e6b
bf7b5d0
848bf81
f029ac3
2edd3f1
b23cd79
5314d49
88ebb32
5807efe
d24555c
15a4fc9
b663e02
ee5d73b
50738c8
6a0b5c5
21e600c
19038ac
baec5ac
ae0ef70
2dd58b2
bfc615b
f8558b1
387aaf6
098c30c
021c4e3
2810726
8f10797
80bd71f
aca554f
a780c01
fbe68d2
ff729a5
4b0a0db
793f44b
50a8b8e
1498b0e
307cc41
fa52427
a86fd45
a85b7ef
a9bbfcd
d36c85a
cff68c9
9664bc5
2a180d0
5e5f982
a813c31
2c31a46
5c1ced2
3cbc054
f44e34b
c4df154
14e4917
30e0d08
1b4baf5
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
---
license: mit
tags:
- generated_from_keras_callback
model-index:
- name: Ashraf-kasem/custom_gpt2_frames_text_continue
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# Ashraf-kasem/custom_gpt2_frames_text_continue

This model is a fine-tuned version of [Ashraf-kasem/custom_gpt2_frames_text_continue](https://huggingface.co/Ashraf-kasem/custom_gpt2_frames_text_continue) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.6337
- Validation Loss: 2.3028
- Epoch: 99

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'LinearWarmup', 'config': {'after_warmup_lr_sched': {'initial_learning_rate': 5e-05, 'decay_steps': 628900, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'warmup_steps': 125780, 'warmup_learning_rate': 0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
- training_precision: mixed_float16

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 1.0060     | 2.0768          | 0     |
| 1.0147     | 2.0771          | 1     |
| 1.0238     | 2.0821          | 2     |
| 1.0331     | 2.0851          | 3     |
| 1.0422     | 2.0870          | 4     |
| 1.0525     | 2.0945          | 5     |
| 1.0618     | 2.1005          | 6     |
| 1.0718     | 2.1014          | 7     |
| 1.0823     | 2.1056          | 8     |
| 1.0921     | 2.1099          | 9     |
| 1.1028     | 2.1106          | 10    |
| 1.1127     | 2.1127          | 11    |
| 1.1230     | 2.1183          | 12    |
| 1.1329     | 2.1207          | 13    |
| 1.1423     | 2.1270          | 14    |
| 1.1521     | 2.1234          | 15    |
| 1.1614     | 2.1283          | 16    |
| 1.1700     | 2.1236          | 17    |
| 1.1784     | 2.1320          | 18    |
| 1.1864     | 2.1359          | 19    |
| 1.1873     | 2.1272          | 20    |
| 1.1766     | 2.1250          | 21    |
| 1.1652     | 2.1260          | 22    |
| 1.1537     | 2.1224          | 23    |
| 1.1415     | 2.1278          | 24    |
| 1.1296     | 2.1254          | 25    |
| 1.1178     | 2.1213          | 26    |
| 1.1059     | 2.1301          | 27    |
| 1.0950     | 2.1253          | 28    |
| 1.0838     | 2.1264          | 29    |
| 1.0729     | 2.1273          | 30    |
| 1.0625     | 2.1355          | 31    |
| 1.0519     | 2.1345          | 32    |
| 1.0414     | 2.1364          | 33    |
| 1.0317     | 2.1324          | 34    |
| 1.0217     | 2.1410          | 35    |
| 1.0126     | 2.1428          | 36    |
| 1.0027     | 2.1427          | 37    |
| 0.9936     | 2.1494          | 38    |
| 0.9846     | 2.1502          | 39    |
| 0.9752     | 2.1490          | 40    |
| 0.9665     | 2.1501          | 41    |
| 0.9582     | 2.1552          | 42    |
| 0.9497     | 2.1533          | 43    |
| 0.9411     | 2.1621          | 44    |
| 0.9331     | 2.1618          | 45    |
| 0.9248     | 2.1655          | 46    |
| 0.9172     | 2.1755          | 47    |
| 0.9093     | 2.1759          | 48    |
| 0.9014     | 2.1751          | 49    |
| 0.8942     | 2.1813          | 50    |
| 0.8867     | 2.1831          | 51    |
| 0.8795     | 2.1856          | 52    |
| 0.8723     | 2.1909          | 53    |
| 0.8651     | 2.1950          | 54    |
| 0.8581     | 2.1955          | 55    |
| 0.8511     | 2.2007          | 56    |
| 0.8444     | 2.2002          | 57    |
| 0.8380     | 2.2078          | 58    |
| 0.8312     | 2.2077          | 59    |
| 0.8246     | 2.2161          | 60    |
| 0.8186     | 2.2103          | 61    |
| 0.8120     | 2.2180          | 62    |
| 0.8053     | 2.2202          | 63    |
| 0.7994     | 2.2232          | 64    |
| 0.7934     | 2.2290          | 65    |
| 0.7872     | 2.2301          | 66    |
| 0.7816     | 2.2327          | 67    |
| 0.7757     | 2.2369          | 68    |
| 0.7698     | 2.2408          | 69    |
| 0.7640     | 2.2439          | 70    |
| 0.7582     | 2.2451          | 71    |
| 0.7528     | 2.2505          | 72    |
| 0.7475     | 2.2524          | 73    |
| 0.7420     | 2.2520          | 74    |
| 0.7366     | 2.2561          | 75    |
| 0.7313     | 2.2616          | 76    |
| 0.7260     | 2.2628          | 77    |
| 0.7211     | 2.2654          | 78    |
| 0.7158     | 2.2701          | 79    |
| 0.7107     | 2.2704          | 80    |
| 0.7061     | 2.2743          | 81    |
| 0.7008     | 2.2749          | 82    |
| 0.6962     | 2.2769          | 83    |
| 0.6916     | 2.2813          | 84    |
| 0.6869     | 2.2838          | 85    |
| 0.6823     | 2.2853          | 86    |
| 0.6780     | 2.2867          | 87    |
| 0.6737     | 2.2883          | 88    |
| 0.6691     | 2.2921          | 89    |
| 0.6651     | 2.2931          | 90    |
| 0.6608     | 2.2946          | 91    |
| 0.6568     | 2.2957          | 92    |
| 0.6533     | 2.2984          | 93    |
| 0.6494     | 2.2981          | 94    |
| 0.6459     | 2.2994          | 95    |
| 0.6425     | 2.3006          | 96    |
| 0.6395     | 2.3019          | 97    |
| 0.6363     | 2.3026          | 98    |
| 0.6337     | 2.3028          | 99    |


### Framework versions

- Transformers 4.25.1
- TensorFlow 2.9.0
- Datasets 2.8.0
- Tokenizers 0.13.2