nlparabic commited on
Commit
e63aa1a
·
verified ·
1 Parent(s): db88370

End of training

Browse files
README.md CHANGED
@@ -3,6 +3,9 @@ license: apache-2.0
3
  base_model: riotu-lab/ArabianGPT-01B
4
  tags:
5
  - generated_from_trainer
 
 
 
6
  model-index:
7
  - name: test-egy_aragpt
8
  results: []
@@ -14,6 +17,12 @@ should probably proofread and complete it, then remove this comment. -->
14
  # test-egy_aragpt
15
 
16
  This model is a fine-tuned version of [riotu-lab/ArabianGPT-01B](https://huggingface.co/riotu-lab/ArabianGPT-01B) on an unknown dataset.
 
 
 
 
 
 
17
 
18
  ## Model description
19
 
 
3
  base_model: riotu-lab/ArabianGPT-01B
4
  tags:
5
  - generated_from_trainer
6
+ metrics:
7
+ - bleu
8
+ - rouge
9
  model-index:
10
  - name: test-egy_aragpt
11
  results: []
 
17
  # test-egy_aragpt
18
 
19
  This model is a fine-tuned version of [riotu-lab/ArabianGPT-01B](https://huggingface.co/riotu-lab/ArabianGPT-01B) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 3.4630
22
+ - Bleu: 0.0984
23
+ - Rouge1: 0.3093
24
+ - Rouge2: 0.0718
25
+ - Rougel: 0.2296
26
 
27
  ## Model description
28
 
all_results.json CHANGED
@@ -1,19 +1,19 @@
1
  {
2
- "epoch": 2.0,
3
- "eval_bleu": 0.11424038411303619,
4
- "eval_loss": 3.128293514251709,
5
- "eval_rouge1": 0.3297614987151056,
6
- "eval_rouge2": 0.08429294540985294,
7
- "eval_rougeL": 0.2561476738686219,
8
- "eval_runtime": 26.3662,
9
  "eval_samples": 847,
10
- "eval_samples_per_second": 32.124,
11
- "eval_steps_per_second": 4.02,
12
- "perplexity": 22.834978708992665,
13
- "total_flos": 2000451796992000.0,
14
- "train_loss": 4.049833530916316,
15
- "train_runtime": 263.3387,
16
  "train_samples": 2552,
17
- "train_samples_per_second": 19.382,
18
- "train_steps_per_second": 2.423
19
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "eval_bleu": 0.09842217214841122,
4
+ "eval_loss": 3.4630110263824463,
5
+ "eval_rouge1": 0.3093411104418377,
6
+ "eval_rouge2": 0.07177511381538038,
7
+ "eval_rougeL": 0.22957881226204224,
8
+ "eval_runtime": 26.8225,
9
  "eval_samples": 847,
10
+ "eval_samples_per_second": 31.578,
11
+ "eval_steps_per_second": 3.952,
12
+ "perplexity": 31.912922646039927,
13
+ "total_flos": 1000225898496000.0,
14
+ "train_loss": 4.835940561324452,
15
+ "train_runtime": 118.408,
16
  "train_samples": 2552,
17
+ "train_samples_per_second": 21.553,
18
+ "train_steps_per_second": 2.694
19
  }
egy_training_log.txt CHANGED
@@ -141,3 +141,5 @@ INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_maren
141
  INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-0637777c38512acf/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-4ddbb6e08bb37d3f.arrow
142
  WARNING:accelerate.utils.other:Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
143
  WARNING:root:Epoch 1.0: No losses recorded yet.
 
 
 
141
  INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-0637777c38512acf/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-4ddbb6e08bb37d3f.arrow
142
  WARNING:accelerate.utils.other:Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
143
  WARNING:root:Epoch 1.0: No losses recorded yet.
144
+ INFO:__main__:*** Evaluate ***
145
+ INFO:absl:Using default tokenizer.
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 2.0,
3
- "eval_bleu": 0.11424038411303619,
4
- "eval_loss": 3.128293514251709,
5
- "eval_rouge1": 0.3297614987151056,
6
- "eval_rouge2": 0.08429294540985294,
7
- "eval_rougeL": 0.2561476738686219,
8
- "eval_runtime": 26.3662,
9
  "eval_samples": 847,
10
- "eval_samples_per_second": 32.124,
11
- "eval_steps_per_second": 4.02,
12
- "perplexity": 22.834978708992665
13
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "eval_bleu": 0.09842217214841122,
4
+ "eval_loss": 3.4630110263824463,
5
+ "eval_rouge1": 0.3093411104418377,
6
+ "eval_rouge2": 0.07177511381538038,
7
+ "eval_rougeL": 0.22957881226204224,
8
+ "eval_runtime": 26.8225,
9
  "eval_samples": 847,
10
+ "eval_samples_per_second": 31.578,
11
+ "eval_steps_per_second": 3.952,
12
+ "perplexity": 31.912922646039927
13
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 2.0,
3
- "total_flos": 2000451796992000.0,
4
- "train_loss": 4.049833530916316,
5
- "train_runtime": 263.3387,
6
  "train_samples": 2552,
7
- "train_samples_per_second": 19.382,
8
- "train_steps_per_second": 2.423
9
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "total_flos": 1000225898496000.0,
4
+ "train_loss": 4.835940561324452,
5
+ "train_runtime": 118.408,
6
  "train_samples": 2552,
7
+ "train_samples_per_second": 21.553,
8
+ "train_steps_per_second": 2.694
9
  }
trainer_state.json CHANGED
@@ -1,46 +1,27 @@
1
  {
2
- "best_metric": 3.128293514251709,
3
- "best_model_checkpoint": "/tmp/test-egy_aragpt/checkpoint-500",
4
- "epoch": 2.0,
5
  "eval_steps": 500,
6
- "global_step": 638,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 1.567398119122257,
13
- "grad_norm": 1.3536401987075806,
14
- "learning_rate": 5e-05,
15
- "loss": 4.3013,
16
- "step": 500
17
- },
18
- {
19
- "epoch": 1.567398119122257,
20
- "eval_bleu": 0.11424038411303619,
21
- "eval_loss": 3.128293514251709,
22
- "eval_rouge1": 0.3297614987151056,
23
- "eval_rouge2": 0.08429294540985294,
24
- "eval_rougeL": 0.2561476738686219,
25
- "eval_runtime": 26.7734,
26
- "eval_samples_per_second": 31.636,
27
- "eval_steps_per_second": 3.959,
28
- "step": 500
29
- },
30
- {
31
- "epoch": 2.0,
32
- "step": 638,
33
- "total_flos": 2000451796992000.0,
34
- "train_loss": 4.049833530916316,
35
- "train_runtime": 263.3387,
36
- "train_samples_per_second": 19.382,
37
- "train_steps_per_second": 2.423
38
  }
39
  ],
40
  "logging_steps": 500,
41
- "max_steps": 638,
42
  "num_input_tokens_seen": 0,
43
- "num_train_epochs": 2,
44
  "save_steps": 500,
45
  "stateful_callbacks": {
46
  "EarlyStoppingCallback": {
@@ -63,7 +44,7 @@
63
  "attributes": {}
64
  }
65
  },
66
- "total_flos": 2000451796992000.0,
67
  "train_batch_size": 8,
68
  "trial_name": null,
69
  "trial_params": null
 
1
  {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 319,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 1.0,
13
+ "step": 319,
14
+ "total_flos": 1000225898496000.0,
15
+ "train_loss": 4.835940561324452,
16
+ "train_runtime": 118.408,
17
+ "train_samples_per_second": 21.553,
18
+ "train_steps_per_second": 2.694
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  }
20
  ],
21
  "logging_steps": 500,
22
+ "max_steps": 319,
23
  "num_input_tokens_seen": 0,
24
+ "num_train_epochs": 1,
25
  "save_steps": 500,
26
  "stateful_callbacks": {
27
  "EarlyStoppingCallback": {
 
44
  "attributes": {}
45
  }
46
  },
47
+ "total_flos": 1000225898496000.0,
48
  "train_batch_size": 8,
49
  "trial_name": null,
50
  "trial_params": null