File size: 2,654 Bytes
884dcb4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f40284d
 
 
884dcb4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8753975
ceb5efc
1df98e8
ad2eb8e
4cb43c5
7d2204f
4adb025
5f5af08
dbcfbb5
4aef454
0403edd
2ac1784
d93110d
9597daf
9b0055c
8aceab7
c6011d5
10c01a5
96b8f99
3a33096
81ac0a3
8481d3c
f78c083
fc7b1f4
7b5bd6e
4daa16a
74c36f9
afea68b
f40284d
884dcb4
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
---
license: mit
tags:
- generated_from_keras_callback
model-index:
- name: Ashraf-kasem/custom_gpt2_frames_text_original_tokenizer
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# Ashraf-kasem/custom_gpt2_frames_text_original_tokenizer

This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 1.1074
- Validation Loss: 1.6432
- Epoch: 29

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 240780, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
- training_precision: mixed_float16

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 4.3075     | 3.4095          | 0     |
| 3.1973     | 2.8234          | 1     |
| 2.7420     | 2.5057          | 2     |
| 2.4541     | 2.3022          | 3     |
| 2.2507     | 2.1648          | 4     |
| 2.0962     | 2.0612          | 5     |
| 1.9736     | 1.9885          | 6     |
| 1.8729     | 1.9286          | 7     |
| 1.7883     | 1.8823          | 8     |
| 1.7153     | 1.8448          | 9     |
| 1.6517     | 1.8113          | 10    |
| 1.5953     | 1.7864          | 11    |
| 1.5446     | 1.7624          | 12    |
| 1.4994     | 1.7459          | 13    |
| 1.4578     | 1.7294          | 14    |
| 1.4200     | 1.7171          | 15    |
| 1.3851     | 1.7026          | 16    |
| 1.3528     | 1.6958          | 17    |
| 1.3229     | 1.6846          | 18    |
| 1.2950     | 1.6760          | 19    |
| 1.2690     | 1.6704          | 20    |
| 1.2448     | 1.6650          | 21    |
| 1.2223     | 1.6599          | 22    |
| 1.2012     | 1.6539          | 23    |
| 1.1815     | 1.6534          | 24    |
| 1.1635     | 1.6486          | 25    |
| 1.1470     | 1.6457          | 26    |
| 1.1318     | 1.6443          | 27    |
| 1.1185     | 1.6434          | 28    |
| 1.1074     | 1.6432          | 29    |


### Framework versions

- Transformers 4.25.1
- TensorFlow 2.9.0
- Datasets 2.8.0
- Tokenizers 0.13.2