gilkeyio's picture
Update README.md
0daa1eb
metadata
dataset_info:
  features:
    - name: sex
      dtype: string
    - name: subset
      dtype: string
    - name: id
      dtype: string
    - name: audio
      dtype: audio
    - name: transcript
      dtype: string
    - name: words
      list:
        - name: end
          dtype: float64
        - name: start
          dtype: float64
        - name: word
          dtype: string
    - name: phonemes
      list:
        - name: end
          dtype: float64
        - name: phoneme
          dtype: string
        - name: start
          dtype: float64
  splits:
    - name: dev_clean
      num_bytes: 365310608.879
      num_examples: 2703
    - name: dev_other
      num_bytes: 341143993.784
      num_examples: 2864
    - name: test_clean
      num_bytes: 377535532.98
      num_examples: 2620
    - name: test_other
      num_bytes: 351207892.569557
      num_examples: 2938
    - name: train_clean_100
      num_bytes: 6694747231.610863
      num_examples: 28538
    - name: train_clean_360
      num_bytes: 24163659711.787865
      num_examples: 104008
    - name: train_other_500
      num_bytes: 32945085271.89443
      num_examples: 148645
  download_size: 62101682957
  dataset_size: 65238690243.50571
configs:
  - config_name: default
    data_files:
      - split: dev_clean
        path: data/dev_clean-*
      - split: dev_other
        path: data/dev_other-*
      - split: test_clean
        path: data/test_clean-*
      - split: test_other
        path: data/test_other-*
      - split: train_clean_100
        path: data/train_clean_100-*
      - split: train_clean_360
        path: data/train_clean_360-*
      - split: train_other_500
        path: data/train_other_500-*
license: cc-by-4.0
task_categories:
  - automatic-speech-recognition
language:
  - en
pretty_name: Librispeech Alignments
size_categories:
  - 100K<n<1M

Dataset Card for Librispeech Alignments

Librispeech with alignments generated by the Montreal Forced Aligner. The original alignments in TextGrid format can be found here

Dataset Details

Dataset Description

Librispeech is a corpus of read English speech, designed for training and evaluating automatic speech recognition (ASR) systems. The dataset contains 1000 hours of 16kHz read English speech derived from audiobooks.

The Montreal Forced Aligner (MFA) was used to generate word and phoneme level alignments for the Librispeech dataset.

  • Curated by: Vassil Panayotov, Guoguo Chen, Daniel Povey, Sanjeev Khudanpur (for Librispeech)
  • Funded by: DARPA LORELEI
  • Shared by: Loren Lugosch (for Alignments)
  • Language(s) (NLP): English
  • License: Creative Commons Attribution 4.0 International License

Dataset Sources

Uses

Direct Use

The Librispeech dataset can be used to train and evaluate ASR systems. The alignments allow for forced alignment techniques.

Out-of-Scope Use

The dataset only contains read speech, so may not perform as well on spontaneous conversational speech.

Dataset Structure

The dataset contains 1000 hours of segmented read English speech from audiobooks. There are three train subsets: 100 hours (train-clean-100), 360 hours (train-clean-360) and 500 hours (train-other-500).

The alignments connect the audio to the reference text transcripts on word and phoneme level.

Data Fields

  • sex: M for male, F for female

  • subset: dev_clean, dev_other, test_clean, test_other, train_clean_100, train_clean_360, train_other_500

  • id: unique id of the data sample. (speaker id)-(chapter-id)-(utterance-id)

  • audio: the audio, 16kHz

  • transcript: the spoken text of the dataset, normalized and lowercased

  • words: a list of words with fields:

    • word: the text of the word
    • start: the start time in seconds
    • end: the end time in seconds
  • phonemes: a list of phonemes with fields:

    • phoneme: the phoneme spoken
    • start: the start time in seconds
    • end: the end time in seconds

Dataset Creation

Curation Rationale

Librispeech was created to further speech recognition research and to benchmark progress in the field.

Source Data

Data Collection and Processing

The audio and reference texts were sourced from read English audiobooks in the LibriVox project. The data was segmented, filtered and prepared for speech recognition.

Who are the source data producers?

The audiobooks are read by volunteers for the LibriVox project. Information about the readers is available in the LibriVox catalog.

Annotations

Annotation process

The Montreal Forced Aligner was used to create word and phoneme level alignments between the audio and reference texts. The aligner is based on Kaldi. In the process of formatting this into a HuggingFace dataset, words with empty text and phonemes with empty text, silence tokens, or spacing tokens were removed

Who are the annotators?

The alignments were generated automatically by the Montreal Forced Aligner and shared by Loren Lugosch. The TextGrid files were parsed and integrated into this dataset by Kim Gilkey.

Personal and Sensitive Information

The data contains read speech and transcripts. No personal or sensitive information expected.

Bias, Risks, and Limitations

The dataset contains only read speech from published books, not natural conversational speech. Performance on other tasks may be reduced.

Recommendations

Users should understand that the alignments may contain errors and account for this in applications. For example, be wary of tokens.

Citation

Librispeech:

@inproceedings{panayotov2015librispeech,  
  title={Librispeech: an ASR corpus based on public domain audio books},
  author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},  
  booktitle={ICASSP},   
  year={2015},   
  organization={IEEE} 
}

Librispeech Alignments:

Loren Lugosch, Mirco Ravanelli, Patrick Ignoto, Vikrant Singh Tomar, and Yoshua Bengio, "Speech Model Pre-training for End-to-End Spoken Language Understanding", Interspeech 2019.

Montreal Forced Aligner:

Michael McAuliffe, Michaela Socolof, Sarah Mihuc, Michael Wagner, and Morgan Sonderegger. "Montreal Forced Aligner: trainable text-speech alignment using Kaldi", Interspeech 2017.