File size: 2,616 Bytes
617cef1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a7cd8ae
 
 
617cef1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f50496d
617cef1
 
 
 
 
 
f50496d
a80400c
b5a0bc5
f4d2706
890539e
ec39ee7
3f7e277
950b6fc
4945454
2a7edde
de27863
9e54aae
9a145cd
5e34fa9
7b941fe
e571bae
5b4a4e1
ac6703c
05131a4
4f7c844
48b16e4
f600557
72e37bf
aa93643
5091825
a79fd10
cb859af
7aa96d6
0f539df
a7cd8ae
617cef1
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
---
license: mit
tags:
- generated_from_keras_callback
model-index:
- name: Ashraf-kasem/custom_gpt2_frames_text
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# Ashraf-kasem/custom_gpt2_frames_text

This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 1.3938
- Validation Loss: 2.0834
- Epoch: 29

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 188670, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
- training_precision: mixed_float16

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 5.4252     | 4.4731          | 0     |
| 4.1781     | 3.6928          | 1     |
| 3.5744     | 3.2572          | 2     |
| 3.1856     | 2.9789          | 3     |
| 2.9095     | 2.7887          | 4     |
| 2.6999     | 2.6534          | 5     |
| 2.5334     | 2.5484          | 6     |
| 2.3969     | 2.4706          | 7     |
| 2.2826     | 2.4102          | 8     |
| 2.1842     | 2.3518          | 9     |
| 2.0988     | 2.3096          | 10    |
| 2.0236     | 2.2740          | 11    |
| 1.9569     | 2.2443          | 12    |
| 1.8960     | 2.2214          | 13    |
| 1.8411     | 2.1954          | 14    |
| 1.7913     | 2.1815          | 15    |
| 1.7457     | 2.1652          | 16    |
| 1.7034     | 2.1552          | 17    |
| 1.6648     | 2.1398          | 18    |
| 1.6288     | 2.1289          | 19    |
| 1.5955     | 2.1213          | 20    |
| 1.5643     | 2.1114          | 21    |
| 1.5359     | 2.1071          | 22    |
| 1.5094     | 2.0998          | 23    |
| 1.4846     | 2.0942          | 24    |
| 1.4622     | 2.0911          | 25    |
| 1.4420     | 2.0893          | 26    |
| 1.4233     | 2.0879          | 27    |
| 1.4074     | 2.0838          | 28    |
| 1.3938     | 2.0834          | 29    |


### Framework versions

- Transformers 4.25.1
- TensorFlow 2.9.0
- Datasets 2.8.0
- Tokenizers 0.13.2