## Fine-tuning run 3 Tried to improve model fine-tuned during run 1. Checkpoint used: checkpoint-12000 * Trained for 6000 steps * Used custom Learning Rate scheduler initialized in: `custom_trainer.Seq2SeqTrainerCustomLinearScheduler`: * `--learning_rate="3e-5"` * `--learning_rate_end="1e-5"` * no warmup was used * no WER improvements compared to checkpoint-12000 of run 1 * using `seed=43` * do not upload checkpoints from that run * uploading src, logs, tensorboard logs, trainer_state ## Advices * I guess, we need to use warmup when resuming training and increasing LR compared to the last LR in previous run * need to set number of steps > 6000. because model improved WER veeery slowly * can use original Mozilla Common Voice dataset instead of a HuggingFace's one.
the reason is that original contains multiple voicings of same sentence - so there is at least twice as more data.
to use this "additional" data, train, validation, test sets need to be enlarged using `validated` set - the one that is absent in HuggingFace's CV11 dataset