Update README.md
Browse files
README.md
CHANGED
@@ -112,15 +112,15 @@ The model was fine-tuned using the Unsloth and LoRA.
|
|
112 |
|
113 |
#### Training Hyperparameters
|
114 |
|
115 |
-
|
116 |
-
|
117 |
-
|
118 |
-
|
119 |
-
|
120 |
-
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
|
125 |
#### Speeds, Sizes, Times [optional]
|
126 |
|
|
|
112 |
|
113 |
#### Training Hyperparameters
|
114 |
|
115 |
+
- batch size:8,
|
116 |
+
- gradient accumulation steps:4,
|
117 |
+
- optimizer:adamw_torch,
|
118 |
+
- learning rate:2e-5,
|
119 |
+
- warmup_steps:100,
|
120 |
+
- fp16:True,
|
121 |
+
- epochs:2,
|
122 |
+
- weight_decay:0.01,
|
123 |
+
- lr_scheduler_type:linear
|
124 |
|
125 |
#### Speeds, Sizes, Times [optional]
|
126 |
|