BounharAbdelaziz commited on
Commit
1affce4
1 Parent(s): c972a93

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -15
README.md CHANGED
@@ -21,19 +21,7 @@ This model is an impovement of the previous version [Terjman-Large](atlasia/Terj
21
 
22
  The finetuning was conducted using a **A100-40GB** and took **17 hours**.
23
 
24
- ## Training hyperparameters
25
-
26
- The following hyperparameters were used during training:
27
- - learning_rate: 2e-04
28
- - train_batch_size: 16
29
- - eval_batch_size: 16
30
- - seed: 42
31
- - gradient_accumulation_steps: 4
32
- - total_train_batch_size: 32
33
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
34
- - lr_scheduler_type: linear
35
- - lr_scheduler_warmup_ratio: 0.03
36
- - num_epochs: 30
37
 
38
  ## Usage
39
 
@@ -67,9 +55,9 @@ print("Translation:", output_text)
67
 
68
  Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
69
 
70
- **Input**: "Hello my friend, how's life in Morocco"
71
 
72
- **Output**: "سالام صاحبي كيف الأحوال فالمغرب"
73
 
74
 
75
  ## Limiations
@@ -82,6 +70,22 @@ We're currently collecting more data with the aim of continous improvements.
82
  We're continuously striving to improve our model's performance and usability and we will be improving it incrementaly.
83
  If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
84
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
  ## Framework versions
86
 
87
  - Transformers 4.39.2
 
21
 
22
  The finetuning was conducted using a **A100-40GB** and took **17 hours**.
23
 
24
+ Try it out on our dedicated [Terjman-Large-v2 Space](https://huggingface.co/spaces/atlasia/Terjman-Large-v2) 🤗
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
  ## Usage
27
 
 
55
 
56
  Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
57
 
58
+ **Input**: "Hi my friend, can you tell me a joke in moroccan darija? I'd be happy to hear that from you!"
59
 
60
+ **Output**: "سلام صاحبتي ممكن تقولي ليا نكتة بالدارجة المغربية؟ نفرح نسمعها منك!"
61
 
62
 
63
  ## Limiations
 
70
  We're continuously striving to improve our model's performance and usability and we will be improving it incrementaly.
71
  If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
72
 
73
+
74
+ ## Training hyperparameters
75
+
76
+ The following hyperparameters were used during training:
77
+ - learning_rate: 2e-04
78
+ - train_batch_size: 16
79
+ - eval_batch_size: 16
80
+ - seed: 42
81
+ - gradient_accumulation_steps: 4
82
+ - total_train_batch_size: 32
83
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
84
+ - lr_scheduler_type: linear
85
+ - lr_scheduler_warmup_ratio: 0.03
86
+ - num_epochs: 30
87
+
88
+
89
  ## Framework versions
90
 
91
  - Transformers 4.39.2