File size: 2,654 Bytes
884dcb4 f40284d 884dcb4 8753975 ceb5efc 1df98e8 ad2eb8e 4cb43c5 7d2204f 4adb025 5f5af08 dbcfbb5 4aef454 0403edd 2ac1784 d93110d 9597daf 9b0055c 8aceab7 c6011d5 10c01a5 96b8f99 3a33096 81ac0a3 8481d3c f78c083 fc7b1f4 7b5bd6e 4daa16a 74c36f9 afea68b f40284d 884dcb4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 |
---
license: mit
tags:
- generated_from_keras_callback
model-index:
- name: Ashraf-kasem/custom_gpt2_frames_text_original_tokenizer
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# Ashraf-kasem/custom_gpt2_frames_text_original_tokenizer
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 1.1074
- Validation Loss: 1.6432
- Epoch: 29
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 240780, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
- training_precision: mixed_float16
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 4.3075 | 3.4095 | 0 |
| 3.1973 | 2.8234 | 1 |
| 2.7420 | 2.5057 | 2 |
| 2.4541 | 2.3022 | 3 |
| 2.2507 | 2.1648 | 4 |
| 2.0962 | 2.0612 | 5 |
| 1.9736 | 1.9885 | 6 |
| 1.8729 | 1.9286 | 7 |
| 1.7883 | 1.8823 | 8 |
| 1.7153 | 1.8448 | 9 |
| 1.6517 | 1.8113 | 10 |
| 1.5953 | 1.7864 | 11 |
| 1.5446 | 1.7624 | 12 |
| 1.4994 | 1.7459 | 13 |
| 1.4578 | 1.7294 | 14 |
| 1.4200 | 1.7171 | 15 |
| 1.3851 | 1.7026 | 16 |
| 1.3528 | 1.6958 | 17 |
| 1.3229 | 1.6846 | 18 |
| 1.2950 | 1.6760 | 19 |
| 1.2690 | 1.6704 | 20 |
| 1.2448 | 1.6650 | 21 |
| 1.2223 | 1.6599 | 22 |
| 1.2012 | 1.6539 | 23 |
| 1.1815 | 1.6534 | 24 |
| 1.1635 | 1.6486 | 25 |
| 1.1470 | 1.6457 | 26 |
| 1.1318 | 1.6443 | 27 |
| 1.1185 | 1.6434 | 28 |
| 1.1074 | 1.6432 | 29 |
### Framework versions
- Transformers 4.25.1
- TensorFlow 2.9.0
- Datasets 2.8.0
- Tokenizers 0.13.2
|