e-hossam96
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -7,16 +7,18 @@ tags:
|
|
7 |
model-index:
|
8 |
- name: arabic-nano-gpt-v2
|
9 |
results: []
|
|
|
|
|
|
|
|
|
10 |
---
|
11 |
|
12 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
-
should probably proofread and complete it, then remove this comment. -->
|
14 |
|
15 |
# arabic-nano-gpt-v2
|
16 |
|
17 |
This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
|
18 |
-
It achieves the following results on the
|
19 |
-
- Loss: 3.
|
20 |
|
21 |
## Model description
|
22 |
|
@@ -46,7 +48,7 @@ The following hyperparameters were used during training:
|
|
46 |
- lr_scheduler_warmup_ratio: 0.01
|
47 |
- num_epochs: 8
|
48 |
|
49 |
-
### Training results
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss |
|
52 |
|:-------------:|:------:|:------:|:---------------:|
|
@@ -76,12 +78,20 @@ The following hyperparameters were used during training:
|
|
76 |
| 3.3786 | 7.0187 | 120000 | 3.2625 |
|
77 |
| 3.3713 | 7.3112 | 125000 | 3.2587 |
|
78 |
| 3.3662 | 7.6036 | 130000 | 3.2553 |
|
79 |
-
| 3.365 | 7.8961 | 135000 | 3.2532 |
|
80 |
|
81 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
82 |
### Framework versions
|
83 |
|
84 |
- Transformers 4.45.2
|
85 |
- Pytorch 2.5.0
|
86 |
- Datasets 3.0.1
|
87 |
-
- Tokenizers 0.20.1
|
|
|
7 |
model-index:
|
8 |
- name: arabic-nano-gpt-v2
|
9 |
results: []
|
10 |
+
datasets:
|
11 |
+
- wikimedia/wikipedia
|
12 |
+
language:
|
13 |
+
- ar
|
14 |
---
|
15 |
|
|
|
|
|
16 |
|
17 |
# arabic-nano-gpt-v2
|
18 |
|
19 |
This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
|
20 |
+
It achieves the following results on the held-out test set:
|
21 |
+
- Loss: 3.25564
|
22 |
|
23 |
## Model description
|
24 |
|
|
|
48 |
- lr_scheduler_warmup_ratio: 0.01
|
49 |
- num_epochs: 8
|
50 |
|
51 |
+
<!-- ### Training results
|
52 |
|
53 |
| Training Loss | Epoch | Step | Validation Loss |
|
54 |
|:-------------:|:------:|:------:|:---------------:|
|
|
|
78 |
| 3.3786 | 7.0187 | 120000 | 3.2625 |
|
79 |
| 3.3713 | 7.3112 | 125000 | 3.2587 |
|
80 |
| 3.3662 | 7.6036 | 130000 | 3.2553 |
|
81 |
+
| 3.365 | 7.8961 | 135000 | 3.2532 | -->
|
82 |
|
83 |
|
84 |
+
### Training Loss
|
85 |
+
|
86 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/Fwe5cHogWPrpkzN-Jp1f3.png)
|
87 |
+
|
88 |
+
### Validation Loss
|
89 |
+
|
90 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/uQ1u25rLcBZJgdrji7TwE.png)
|
91 |
+
|
92 |
### Framework versions
|
93 |
|
94 |
- Transformers 4.45.2
|
95 |
- Pytorch 2.5.0
|
96 |
- Datasets 3.0.1
|
97 |
+
- Tokenizers 0.20.1
|