e-hossam96 commited on
Commit
c80d57a
·
verified ·
1 Parent(s): 82061d3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -7
README.md CHANGED
@@ -7,16 +7,18 @@ tags:
7
  model-index:
8
  - name: arabic-nano-gpt-v2
9
  results: []
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
 
15
  # arabic-nano-gpt-v2
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 3.2532
20
 
21
  ## Model description
22
 
@@ -46,7 +48,7 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_warmup_ratio: 0.01
47
  - num_epochs: 8
48
 
49
- ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:------:|:------:|:---------------:|
@@ -76,12 +78,20 @@ The following hyperparameters were used during training:
76
  | 3.3786 | 7.0187 | 120000 | 3.2625 |
77
  | 3.3713 | 7.3112 | 125000 | 3.2587 |
78
  | 3.3662 | 7.6036 | 130000 | 3.2553 |
79
- | 3.365 | 7.8961 | 135000 | 3.2532 |
80
 
81
 
 
 
 
 
 
 
 
 
82
  ### Framework versions
83
 
84
  - Transformers 4.45.2
85
  - Pytorch 2.5.0
86
  - Datasets 3.0.1
87
- - Tokenizers 0.20.1
 
7
  model-index:
8
  - name: arabic-nano-gpt-v2
9
  results: []
10
+ datasets:
11
+ - wikimedia/wikipedia
12
+ language:
13
+ - ar
14
  ---
15
 
 
 
16
 
17
  # arabic-nano-gpt-v2
18
 
19
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
20
+ It achieves the following results on the held-out test set:
21
+ - Loss: 3.25564
22
 
23
  ## Model description
24
 
 
48
  - lr_scheduler_warmup_ratio: 0.01
49
  - num_epochs: 8
50
 
51
+ <!-- ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:------:|:------:|:---------------:|
 
78
  | 3.3786 | 7.0187 | 120000 | 3.2625 |
79
  | 3.3713 | 7.3112 | 125000 | 3.2587 |
80
  | 3.3662 | 7.6036 | 130000 | 3.2553 |
81
+ | 3.365 | 7.8961 | 135000 | 3.2532 | -->
82
 
83
 
84
+ ### Training Loss
85
+
86
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/Fwe5cHogWPrpkzN-Jp1f3.png)
87
+
88
+ ### Validation Loss
89
+
90
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/uQ1u25rLcBZJgdrji7TwE.png)
91
+
92
  ### Framework versions
93
 
94
  - Transformers 4.45.2
95
  - Pytorch 2.5.0
96
  - Datasets 3.0.1
97
+ - Tokenizers 0.20.1