AigizK's picture
how to restore messages back
91f33dc
|
raw
history blame
2.11 kB
metadata
language:
  - ba
license: apache-2.0
tags:
  - automatic-speech-recognition
  - mozilla-foundation/common_voice_7_0
  - generated_from_trainer
  - robust-speech-event
datasets:
  - common_voice
model-index:
  - name: wav2vec2-large-xls-r-300m-bashkir-cv7_opt
    results:
      - task:
          name: Automatic Speech Recognition
          type: automatic-speech-recognition
        dataset:
          name: Common Voice 7
          type: mozilla-foundation/common_voice_7_0
          args: ba
        metrics:
          - name: Test WER
            type: wer
            value: 9.46

wav2vec2-large-xls-r-300m-bashkir-cv7_opt

This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - BA dataset. It achieves the following results on the evaluation set:

  • Training Loss: 0.313700
  • Validation Loss: 0.120663
  • Wer: 0.094649

Model description

More information needed

Intended uses & limitations

In order to reduce the number of characters, the following letters have been replaced or removed:

'я' -> 'йа' 'ю' -> 'йу' 'ё' -> 'йо' 'е' -> 'йэ' for first letter 'е' -> 'э' for other cases 'ъ' -> deleted 'ь' -> deleted

Therefore, in order to get the correct text, you need to do the reverse transformation and use the language model.

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 64
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 300
  • num_epochs: 50
  • mixed_precision_training: Native AMP

Framework versions

  • Transformers 4.16.1
  • Pytorch 1.10.0+cu113
  • Datasets 1.18.2
  • Tokenizers 0.10.3