The learning rate was not displayed as it should.
#3
by
kapllan
- opened
README.md
CHANGED
@@ -88,7 +88,7 @@ For further details see [Niklaus et al. 2023](https://arxiv.org/abs/2306.02069?u
|
|
88 |
- batche size: 512 samples
|
89 |
- Number of steps: 1M/500K for the base/large model
|
90 |
- Warm-up steps for the first 5\% of the total training steps
|
91 |
-
- Learning rate: (linearly increasing up to)
|
92 |
- Word masking: increased 20/30\% masking rate for base/large models respectively
|
93 |
|
94 |
## Evaluation
|
|
|
88 |
- batche size: 512 samples
|
89 |
- Number of steps: 1M/500K for the base/large model
|
90 |
- Warm-up steps for the first 5\% of the total training steps
|
91 |
+
- Learning rate: (linearly increasing up to) 1e-4
|
92 |
- Word masking: increased 20/30\% masking rate for base/large models respectively
|
93 |
|
94 |
## Evaluation
|