LICENSE MANIFEST.in README.md pyproject.toml setup.cfg setup.py examples/operators/alignment_train_cpu.cpp examples/operators/alignment_train_cuda.cpp examples/operators/alignment_train_kernel.cu fairseq/__init__.py fairseq/binarizer.py fairseq/checkpoint_utils.py fairseq/file_chunker_utils.py fairseq/file_io.py fairseq/file_utils.py fairseq/hub_utils.py fairseq/incremental_decoding_utils.py fairseq/iterative_refinement_generator.py fairseq/nan_detector.py fairseq/ngram_repeat_block.py fairseq/options.py fairseq/pdb.py fairseq/quantization_utils.py fairseq/registry.py fairseq/search.py fairseq/sequence_generator.py fairseq/sequence_scorer.py fairseq/speech_generator.py fairseq/token_generation_constraints.py fairseq/tokenizer.py fairseq/trainer.py fairseq/utils.py fairseq/version.py fairseq/version.txt fairseq.egg-info/PKG-INFO fairseq.egg-info/SOURCES.txt fairseq.egg-info/dependency_links.txt fairseq.egg-info/entry_points.txt fairseq.egg-info/not-zip-safe fairseq.egg-info/requires.txt fairseq.egg-info/top_level.txt fairseq/benchmark/__init__.py fairseq/benchmark/benchmark_multihead_attention.py fairseq/benchmark/dummy_dataset.py fairseq/benchmark/dummy_lm.py fairseq/benchmark/dummy_masked_lm.py fairseq/benchmark/dummy_model.py fairseq/benchmark/dummy_mt.py fairseq/clib/cuda/ngram_repeat_block_cuda.cpp fairseq/clib/cuda/ngram_repeat_block_cuda_kernel.cu fairseq/clib/libbase/balanced_assignment.cpp fairseq/clib/libbleu/libbleu.cpp fairseq/clib/libbleu/module.cpp fairseq/clib/libnat/edit_dist.cpp fairseq/clib/libnat_cuda/binding.cpp fairseq/clib/libnat_cuda/edit_dist.cu fairseq/config/__init__.py fairseq/config/config.yaml fairseq/config/fb_run_config/slurm.yaml fairseq/config/model/transformer_lm/transformer_lm_baevski_gbw.yaml fairseq/config/model/transformer_lm/transformer_lm_baevski_wiki103.yaml fairseq/config/model/transformer_lm/transformer_lm_big.yaml fairseq/config/model/transformer_lm/transformer_lm_gbw.yaml fairseq/config/model/transformer_lm/transformer_lm_gpt.yaml fairseq/config/model/transformer_lm/transformer_lm_gpt2_big.yaml fairseq/config/model/transformer_lm/transformer_lm_gpt2_medium.yaml fairseq/config/model/transformer_lm/transformer_lm_gpt2_small.yaml fairseq/config/model/transformer_lm/transformer_lm_wiki103.yaml fairseq/config/model/wav2vec/vq_wav2vec_gumbel.yaml fairseq/config/model/wav2vec2/wav2vec2_base.yaml fairseq/config/model/wav2vec2/wav2vec2_large.yaml fairseq/criterions/__init__.py fairseq/criterions/adaptive_loss.py fairseq/criterions/composite_loss.py fairseq/criterions/cross_entropy.py fairseq/criterions/ctc.py fairseq/criterions/fairseq_criterion.py fairseq/criterions/fastspeech2_loss.py fairseq/criterions/hubert_criterion.py fairseq/criterions/label_smoothed_cross_entropy.py fairseq/criterions/label_smoothed_cross_entropy_latency_augmented.py fairseq/criterions/label_smoothed_cross_entropy_with_alignment.py fairseq/criterions/label_smoothed_cross_entropy_with_ctc.py fairseq/criterions/label_smoothed_cross_entropy_with_rdrop.py fairseq/criterions/legacy_masked_lm.py fairseq/criterions/masked_lm.py fairseq/criterions/model_criterion.py fairseq/criterions/nat_loss.py fairseq/criterions/sentence_prediction.py fairseq/criterions/sentence_prediction_adapters.py fairseq/criterions/sentence_ranking.py fairseq/criterions/speech_dlm_criterion.py fairseq/criterions/speech_to_speech_criterion.py fairseq/criterions/speech_ulm_criterion.py fairseq/criterions/tacotron2_loss.py fairseq/criterions/wav2vec_criterion.py fairseq/data/__init__.py fairseq/data/add_class_target_dataset.py fairseq/data/add_target_dataset.py fairseq/data/append_token_dataset.py fairseq/data/backtranslation_dataset.py fairseq/data/base_wrapper_dataset.py fairseq/data/bucket_pad_length_dataset.py fairseq/data/codedataset.py fairseq/data/colorize_dataset.py fairseq/data/concat_dataset.py fairseq/data/concat_sentences_dataset.py fairseq/data/data_utils.py fairseq/data/data_utils_fast.pyx fairseq/data/denoising_dataset.py fairseq/data/dictionary.py fairseq/data/fairseq_dataset.py fairseq/data/fasta_dataset.py fairseq/data/id_dataset.py fairseq/data/indexed_dataset.py fairseq/data/iterators.py fairseq/data/language_pair_dataset.py fairseq/data/list_dataset.py fairseq/data/lm_context_window_dataset.py fairseq/data/lru_cache_dataset.py fairseq/data/mask_tokens_dataset.py fairseq/data/monolingual_dataset.py fairseq/data/multi_corpus_dataset.py fairseq/data/multi_corpus_sampled_dataset.py fairseq/data/nested_dictionary_dataset.py fairseq/data/noising.py fairseq/data/num_samples_dataset.py fairseq/data/numel_dataset.py fairseq/data/offset_tokens_dataset.py fairseq/data/pad_dataset.py fairseq/data/padding_mask_dataset.py fairseq/data/plasma_utils.py fairseq/data/prepend_dataset.py fairseq/data/prepend_token_dataset.py fairseq/data/raw_label_dataset.py fairseq/data/replace_dataset.py fairseq/data/resampling_dataset.py fairseq/data/roll_dataset.py fairseq/data/round_robin_zip_datasets.py fairseq/data/shorten_dataset.py fairseq/data/sort_dataset.py fairseq/data/span_mask_tokens_dataset.py fairseq/data/speech_dlm_dataset.py fairseq/data/strip_token_dataset.py fairseq/data/subsample_dataset.py fairseq/data/text_compressor.py fairseq/data/token_block_dataset.py fairseq/data/token_block_utils_fast.pyx fairseq/data/transform_eos_concat_langpair_dataset.py fairseq/data/transform_eos_dataset.py fairseq/data/transform_eos_lang_pair_dataset.py fairseq/data/audio/__init__.py fairseq/data/audio/audio_utils.py fairseq/data/audio/data_cfg.py fairseq/data/audio/frm_text_to_speech_dataset.py fairseq/data/audio/hubert_dataset.py fairseq/data/audio/multi_modality_dataset.py fairseq/data/audio/raw_audio_dataset.py fairseq/data/audio/speech_to_speech_dataset.py fairseq/data/audio/speech_to_text_dataset.py fairseq/data/audio/speech_to_text_joint_dataset.py fairseq/data/audio/text_to_speech_dataset.py fairseq/data/audio/dataset_transforms/__init__.py fairseq/data/audio/dataset_transforms/concataugment.py fairseq/data/audio/dataset_transforms/noisyoverlapaugment.py fairseq/data/audio/feature_transforms/__init__.py fairseq/data/audio/feature_transforms/delta_deltas.py fairseq/data/audio/feature_transforms/global_cmvn.py fairseq/data/audio/feature_transforms/specaugment.py fairseq/data/audio/feature_transforms/utterance_cmvn.py fairseq/data/audio/waveform_transforms/__init__.py fairseq/data/audio/waveform_transforms/noiseaugment.py fairseq/data/encoders/__init__.py fairseq/data/encoders/byte_bpe.py fairseq/data/encoders/byte_utils.py fairseq/data/encoders/bytes.py fairseq/data/encoders/characters.py fairseq/data/encoders/fastbpe.py fairseq/data/encoders/gpt2_bpe.py fairseq/data/encoders/gpt2_bpe_utils.py fairseq/data/encoders/hf_bert_bpe.py fairseq/data/encoders/hf_byte_bpe.py fairseq/data/encoders/moses_tokenizer.py fairseq/data/encoders/nltk_tokenizer.py fairseq/data/encoders/sentencepiece_bpe.py fairseq/data/encoders/space_tokenizer.py fairseq/data/encoders/subword_nmt_bpe.py fairseq/data/encoders/utils.py fairseq/data/huffman/__init__.py fairseq/data/huffman/huffman_coder.py fairseq/data/huffman/huffman_mmap_indexed_dataset.py fairseq/data/legacy/__init__.py fairseq/data/legacy/block_pair_dataset.py fairseq/data/legacy/masked_lm_dataset.py fairseq/data/legacy/masked_lm_dictionary.py fairseq/data/multilingual/__init__.py fairseq/data/multilingual/multilingual_data_manager.py fairseq/data/multilingual/multilingual_utils.py fairseq/data/multilingual/sampled_multi_dataset.py fairseq/data/multilingual/sampled_multi_epoch_dataset.py fairseq/data/multilingual/sampling_method.py fairseq/dataclass/__init__.py fairseq/dataclass/configs.py fairseq/dataclass/constants.py fairseq/dataclass/initialize.py fairseq/dataclass/utils.py fairseq/distributed/__init__.py fairseq/distributed/distributed_timeout_wrapper.py fairseq/distributed/fully_sharded_data_parallel.py fairseq/distributed/legacy_distributed_data_parallel.py fairseq/distributed/module_proxy_wrapper.py fairseq/distributed/tpu_distributed_data_parallel.py fairseq/distributed/utils.py fairseq/examples/.gitignore fairseq/examples/__init__.py fairseq/examples/MMPT/.gitignore fairseq/examples/MMPT/CONFIG.md fairseq/examples/MMPT/DATASET.md fairseq/examples/MMPT/README.md fairseq/examples/MMPT/endtask.md fairseq/examples/MMPT/locallaunch.py fairseq/examples/MMPT/pretraining.md fairseq/examples/MMPT/setup.py fairseq/examples/MMPT/videoclip.png fairseq/examples/MMPT/vlm.png fairseq/examples/MMPT/mmpt/__init__.py fairseq/examples/MMPT/mmpt/datasets/__init__.py fairseq/examples/MMPT/mmpt/datasets/fairseqmmdataset.py fairseq/examples/MMPT/mmpt/datasets/mmdataset.py fairseq/examples/MMPT/mmpt/evaluators/__init__.py fairseq/examples/MMPT/mmpt/evaluators/evaluator.py fairseq/examples/MMPT/mmpt/evaluators/metric.py fairseq/examples/MMPT/mmpt/evaluators/predictor.py fairseq/examples/MMPT/mmpt/losses/__init__.py fairseq/examples/MMPT/mmpt/losses/fairseqmmloss.py fairseq/examples/MMPT/mmpt/losses/loss.py fairseq/examples/MMPT/mmpt/losses/nce.py fairseq/examples/MMPT/mmpt/models/__init__.py fairseq/examples/MMPT/mmpt/models/fairseqmmmodel.py fairseq/examples/MMPT/mmpt/models/mmfusion.py fairseq/examples/MMPT/mmpt/models/mmfusionnlg.py fairseq/examples/MMPT/mmpt/models/transformermodel.py fairseq/examples/MMPT/mmpt/modules/__init__.py fairseq/examples/MMPT/mmpt/modules/mm.py fairseq/examples/MMPT/mmpt/modules/retri.py fairseq/examples/MMPT/mmpt/modules/vectorpool.py fairseq/examples/MMPT/mmpt/processors/__init__.py fairseq/examples/MMPT/mmpt/processors/dedupprocessor.py fairseq/examples/MMPT/mmpt/processors/dsprocessor.py fairseq/examples/MMPT/mmpt/processors/how2processor.py fairseq/examples/MMPT/mmpt/processors/how2retriprocessor.py fairseq/examples/MMPT/mmpt/processors/processor.py fairseq/examples/MMPT/mmpt/processors/models/s3dg.py fairseq/examples/MMPT/mmpt/tasks/__init__.py fairseq/examples/MMPT/mmpt/tasks/fairseqmmtask.py fairseq/examples/MMPT/mmpt/tasks/milncetask.py fairseq/examples/MMPT/mmpt/tasks/retritask.py fairseq/examples/MMPT/mmpt/tasks/task.py fairseq/examples/MMPT/mmpt/tasks/vlmtask.py fairseq/examples/MMPT/mmpt/utils/__init__.py fairseq/examples/MMPT/mmpt/utils/load_config.py fairseq/examples/MMPT/mmpt/utils/shardedtensor.py fairseq/examples/MMPT/mmpt_cli/localjob.py fairseq/examples/MMPT/mmpt_cli/predict.py fairseq/examples/MMPT/projects/mfmmlm.yaml fairseq/examples/MMPT/projects/mtm/mmfusionmtm.yaml fairseq/examples/MMPT/projects/mtm/vlm.yaml fairseq/examples/MMPT/projects/mtm/vlm/coin.yaml fairseq/examples/MMPT/projects/mtm/vlm/crosstask.yaml fairseq/examples/MMPT/projects/mtm/vlm/how2.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_coin.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_crosstask.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_crosstask_zs.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_vtt.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_vttqa.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_youcook.yaml fairseq/examples/MMPT/projects/mtm/vlm/test_youcookcap.yaml fairseq/examples/MMPT/projects/mtm/vlm/vtt.yaml fairseq/examples/MMPT/projects/mtm/vlm/vttqa.yaml fairseq/examples/MMPT/projects/mtm/vlm/youcook.yaml fairseq/examples/MMPT/projects/mtm/vlm/youcookcap.yaml fairseq/examples/MMPT/projects/retri/videoclip.yaml fairseq/examples/MMPT/projects/retri/videoretri.yaml fairseq/examples/MMPT/projects/retri/videoclip/coin_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/crosstask_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/how2.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_coin_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_coin_zs.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_crosstask_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_crosstask_zs_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_didemo_zs.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_vtt_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_vtt_zs.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_vttqa_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_vttqa_zs.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_youcook_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/test_youcook_zs.yaml fairseq/examples/MMPT/projects/retri/videoclip/vtt_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/vttqa_videoclip.yaml fairseq/examples/MMPT/projects/retri/videoclip/youcook_videoclip.yaml fairseq/examples/MMPT/projects/task/coin.yaml fairseq/examples/MMPT/projects/task/coin_videoclip.yaml fairseq/examples/MMPT/projects/task/crosstask.yaml fairseq/examples/MMPT/projects/task/crosstask_videoclip.yaml fairseq/examples/MMPT/projects/task/default.yaml fairseq/examples/MMPT/projects/task/ft.yaml fairseq/examples/MMPT/projects/task/how2.yaml fairseq/examples/MMPT/projects/task/test.yaml fairseq/examples/MMPT/projects/task/test_coin.yaml fairseq/examples/MMPT/projects/task/test_coin_videoclip.yaml fairseq/examples/MMPT/projects/task/test_coin_zs.yaml fairseq/examples/MMPT/projects/task/test_crosstask.yaml fairseq/examples/MMPT/projects/task/test_crosstask_videoclip.yaml fairseq/examples/MMPT/projects/task/test_crosstask_zs.yaml fairseq/examples/MMPT/projects/task/test_crosstask_zs_videoclip.yaml fairseq/examples/MMPT/projects/task/test_didemo_zs.yaml fairseq/examples/MMPT/projects/task/test_vtt.yaml fairseq/examples/MMPT/projects/task/test_vtt_videoclip.yaml fairseq/examples/MMPT/projects/task/test_vtt_zs.yaml fairseq/examples/MMPT/projects/task/test_vttqa.yaml fairseq/examples/MMPT/projects/task/test_vttqa_videoclip.yaml fairseq/examples/MMPT/projects/task/test_vttqa_zs.yaml fairseq/examples/MMPT/projects/task/test_youcook.yaml fairseq/examples/MMPT/projects/task/test_youcook_videoclip.yaml fairseq/examples/MMPT/projects/task/test_youcook_zs.yaml fairseq/examples/MMPT/projects/task/test_youcookcap.yaml fairseq/examples/MMPT/projects/task/vtt.yaml fairseq/examples/MMPT/projects/task/vtt_videoclip.yaml fairseq/examples/MMPT/projects/task/vttqa.yaml fairseq/examples/MMPT/projects/task/vttqa_videoclip.yaml fairseq/examples/MMPT/projects/task/youcook.yaml fairseq/examples/MMPT/projects/task/youcook_videoclip.yaml fairseq/examples/MMPT/projects/task/youcookcap.yaml fairseq/examples/MMPT/scripts/text_token_extractor/pretokenization.py fairseq/examples/MMPT/scripts/text_token_extractor/configs/bert-base-uncased.yaml fairseq/examples/MMPT/scripts/video_feature_extractor/extract.py fairseq/examples/MMPT/scripts/video_feature_extractor/model.py fairseq/examples/MMPT/scripts/video_feature_extractor/pathbuilder.py fairseq/examples/MMPT/scripts/video_feature_extractor/preprocessing.py fairseq/examples/MMPT/scripts/video_feature_extractor/random_sequence_shuffler.py fairseq/examples/MMPT/scripts/video_feature_extractor/shard_feature.py fairseq/examples/MMPT/scripts/video_feature_extractor/videoreader.py fairseq/examples/MMPT/scripts/video_feature_extractor/how2/s3d.sh fairseq/examples/adaptive_span/README.md fairseq/examples/adaptive_span/__init__.py fairseq/examples/adaptive_span/adagrad_with_grad_clip.py fairseq/examples/adaptive_span/adaptive_span_attention.py fairseq/examples/adaptive_span/adaptive_span_loss.py fairseq/examples/adaptive_span/adaptive_span_model.py fairseq/examples/adaptive_span/adaptive_span_model_wrapper.py fairseq/examples/adaptive_span/truncated_bptt_lm_task.py fairseq/examples/attention_head_selection/README.md fairseq/examples/attention_head_selection/src/__init__.py fairseq/examples/attention_head_selection/src/speech_to_text_head_selection.py fairseq/examples/attention_head_selection/src/data/__init__.py fairseq/examples/attention_head_selection/src/data/speech_to_text_dataset_with_domain.py fairseq/examples/attention_head_selection/src/loss/__init__.py fairseq/examples/attention_head_selection/src/loss/attention_head_selection.py fairseq/examples/attention_head_selection/src/models/__init__.py fairseq/examples/attention_head_selection/src/models/head_selection_s2t_transformer.py fairseq/examples/attention_head_selection/src/models/head_selection_transformer.py fairseq/examples/attention_head_selection/src/modules/__init__.py fairseq/examples/attention_head_selection/src/modules/attn_head_selector.py fairseq/examples/attention_head_selection/src/modules/head_selection_transformer_layer.py fairseq/examples/attention_head_selection/src/modules/multihead_attention_selection.py fairseq/examples/attention_head_selection/src/modules/multihead_functional.py fairseq/examples/audio_nlp/nlu/README.md fairseq/examples/audio_nlp/nlu/create_dict_stop.sh fairseq/examples/audio_nlp/nlu/generate_manifests.py fairseq/examples/audio_nlp/nlu/configs/nlu_finetuning.yaml fairseq/examples/backtranslation/README.md fairseq/examples/backtranslation/deduplicate_lines.py fairseq/examples/backtranslation/extract_bt_data.py fairseq/examples/backtranslation/prepare-de-monolingual.sh fairseq/examples/backtranslation/prepare-wmt18en2de.sh fairseq/examples/backtranslation/sacrebleu.sh fairseq/examples/backtranslation/tokenized_bleu.sh fairseq/examples/bart/README.glue.md fairseq/examples/bart/README.md fairseq/examples/bart/README.summarization.md fairseq/examples/bart/summarize.py fairseq/examples/byte_level_bpe/README.md fairseq/examples/byte_level_bpe/get_bitext.py fairseq/examples/byte_level_bpe/get_data.sh fairseq/examples/byte_level_bpe/gru_transformer.py fairseq/examples/camembert/README.md fairseq/examples/constrained_decoding/README.md fairseq/examples/constrained_decoding/normalize.py fairseq/examples/constrained_decoding/tok.py fairseq/examples/conv_seq2seq/README.md fairseq/examples/criss/README.md fairseq/examples/criss/download_and_preprocess_flores_test.sh fairseq/examples/criss/download_and_preprocess_tatoeba.sh fairseq/examples/criss/save_encoder.py fairseq/examples/criss/mining/mine.py fairseq/examples/criss/mining/mine_example.sh fairseq/examples/criss/sentence_retrieval/encoder_analysis.py fairseq/examples/criss/sentence_retrieval/sentence_retrieval_tatoeba.sh fairseq/examples/criss/unsupervised_mt/eval.sh fairseq/examples/cross_lingual_language_model/README.md fairseq/examples/data2vec/README.md fairseq/examples/data2vec/__init__.py fairseq/examples/data2vec/fb_convert_beit_cp.py fairseq/examples/data2vec/config/audio/classification/base_classification.yaml fairseq/examples/data2vec/config/audio/classification/run_config/slurm_1.yaml fairseq/examples/data2vec/config/audio/classification/run_config/slurm_1g.yaml fairseq/examples/data2vec/config/audio/classification/run_config/slurm_2.yaml fairseq/examples/data2vec/config/audio/pretraining/audioset.yaml fairseq/examples/data2vec/config/audio/pretraining/base_librispeech.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/local.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_1.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_1_aws.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_2.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_2_aws.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_3.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_4.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_4_aws.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_6_aws.yaml fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_8_aws.yaml fairseq/examples/data2vec/config/text/pretraining/base.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/local.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_1_aws.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_2.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_2_aws.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_3.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_4.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_4_aws.yaml fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_8_aws.yaml fairseq/examples/data2vec/config/v2/base_audio_only_task.yaml fairseq/examples/data2vec/config/v2/base_images_only_task.yaml fairseq/examples/data2vec/config/v2/base_text_only_task.yaml fairseq/examples/data2vec/config/v2/huge_images14_only_task.yaml fairseq/examples/data2vec/config/v2/huge_images_only_task.yaml fairseq/examples/data2vec/config/v2/large_audio_only_task.yaml fairseq/examples/data2vec/config/v2/large_images_only_task.yaml fairseq/examples/data2vec/config/v2/large_text_only_task.yaml fairseq/examples/data2vec/config/v2/large_text_only_task_pgrp_1M.yaml fairseq/examples/data2vec/config/v2/run_config/local.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_1.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_1_aws.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_2.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_2_aws.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_3.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_4.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_4_aws.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_6_aws.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_8.yaml fairseq/examples/data2vec/config/v2/run_config/slurm_8_aws.yaml fairseq/examples/data2vec/config/v2/text_finetuning/cola.yaml fairseq/examples/data2vec/config/v2/text_finetuning/mnli.yaml fairseq/examples/data2vec/config/v2/text_finetuning/mrpc.yaml fairseq/examples/data2vec/config/v2/text_finetuning/qnli.yaml fairseq/examples/data2vec/config/v2/text_finetuning/qqp.yaml fairseq/examples/data2vec/config/v2/text_finetuning/rte.yaml fairseq/examples/data2vec/config/v2/text_finetuning/sst_2.yaml fairseq/examples/data2vec/config/v2/text_finetuning/sts_b.yaml fairseq/examples/data2vec/config/v2/text_finetuning/run_config/local.yaml fairseq/examples/data2vec/config/vision/finetuning/imagenet.yaml fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_clean.yaml fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_huge_clean.yaml fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_large_clean.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/local.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_1.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_1_aws.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_2.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_2_aws.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_3.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_4.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_4_aws.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_6_aws.yaml fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_8_aws.yaml fairseq/examples/data2vec/config/vision/pretraining/base_imagenet.yaml fairseq/examples/data2vec/config/vision/pretraining/base_imagenet_d2v1.yaml fairseq/examples/data2vec/config/vision/pretraining/base_mae_imagenet.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/local.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_1.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_1_aws.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_2.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_2_aws.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_3.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_4.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_4_aws.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_6_aws.yaml fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_8_aws.yaml fairseq/examples/data2vec/data/__init__.py fairseq/examples/data2vec/data/add_class_target_dataset.py fairseq/examples/data2vec/data/image_dataset.py fairseq/examples/data2vec/data/mae_finetuning_image_dataset.py fairseq/examples/data2vec/data/mae_image_dataset.py fairseq/examples/data2vec/data/modality.py fairseq/examples/data2vec/data/path_dataset.py fairseq/examples/data2vec/models/__init__.py fairseq/examples/data2vec/models/audio_classification.py fairseq/examples/data2vec/models/data2vec2.py fairseq/examples/data2vec/models/data2vec_audio.py fairseq/examples/data2vec/models/data2vec_image_classification.py fairseq/examples/data2vec/models/data2vec_text.py fairseq/examples/data2vec/models/data2vec_text_classification.py fairseq/examples/data2vec/models/data2vec_vision.py fairseq/examples/data2vec/models/mae.py fairseq/examples/data2vec/models/mae_image_classification.py fairseq/examples/data2vec/models/utils.py fairseq/examples/data2vec/models/modalities/__init__.py fairseq/examples/data2vec/models/modalities/audio.py fairseq/examples/data2vec/models/modalities/base.py fairseq/examples/data2vec/models/modalities/images.py fairseq/examples/data2vec/models/modalities/modules.py fairseq/examples/data2vec/models/modalities/text.py fairseq/examples/data2vec/scripts/convert_audioset_labels.py fairseq/examples/data2vec/scripts/multi/finetune_all_fair_aws_local_lr.sh fairseq/examples/data2vec/scripts/multi/finetune_all_fair_aws_local_lr_nodep.sh fairseq/examples/data2vec/scripts/multi/finetune_all_fair_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_char_fair_aws_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_lr_nopos.sh fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_aws_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_nodep_aws_local_lr.sh fairseq/examples/data2vec/scripts/text/finetune_sst2_qnli_sweep_fair_nodep.sh fairseq/examples/data2vec/scripts/text/glue.py fairseq/examples/data2vec/scripts/text/glue_lr.py fairseq/examples/data2vec/scripts/text/unprocess_data.py fairseq/examples/data2vec/scripts/text/valids.py fairseq/examples/data2vec/tasks/__init__.py fairseq/examples/data2vec/tasks/audio_classification.py fairseq/examples/data2vec/tasks/image_classification.py fairseq/examples/data2vec/tasks/image_pretraining.py fairseq/examples/data2vec/tasks/mae_image_classification.py fairseq/examples/data2vec/tasks/mae_image_pretraining.py fairseq/examples/data2vec/tasks/multimodal.py fairseq/examples/discriminative_reranking_nmt/README.md fairseq/examples/discriminative_reranking_nmt/__init__.py fairseq/examples/discriminative_reranking_nmt/drnmt_rerank.py fairseq/examples/discriminative_reranking_nmt/config/deen.yaml fairseq/examples/discriminative_reranking_nmt/criterions/__init__.py fairseq/examples/discriminative_reranking_nmt/criterions/discriminative_reranking_criterion.py fairseq/examples/discriminative_reranking_nmt/models/__init__.py fairseq/examples/discriminative_reranking_nmt/models/discriminative_reranking_model.py fairseq/examples/discriminative_reranking_nmt/scripts/prep_data.py fairseq/examples/discriminative_reranking_nmt/tasks/__init__.py fairseq/examples/discriminative_reranking_nmt/tasks/discriminative_reranking_task.py fairseq/examples/emotion_conversion/README.md fairseq/examples/emotion_conversion/requirements.txt fairseq/examples/emotion_conversion/synthesize.py fairseq/examples/emotion_conversion/emotion_models/__init__.py fairseq/examples/emotion_conversion/emotion_models/duration_predictor.py fairseq/examples/emotion_conversion/emotion_models/duration_predictor.yaml fairseq/examples/emotion_conversion/emotion_models/pitch_predictor.py fairseq/examples/emotion_conversion/emotion_models/pitch_predictor.yaml fairseq/examples/emotion_conversion/emotion_models/utils.py fairseq/examples/emotion_conversion/fairseq_models/__init__.py fairseq/examples/emotion_conversion/preprocess/__init__.py fairseq/examples/emotion_conversion/preprocess/build_hifigan_manifest.py fairseq/examples/emotion_conversion/preprocess/build_translation_manifests.py fairseq/examples/emotion_conversion/preprocess/create_core_manifest.py fairseq/examples/emotion_conversion/preprocess/extract_f0.py fairseq/examples/emotion_conversion/preprocess/process_km.py fairseq/examples/emotion_conversion/preprocess/split_emov_km_tsv_by_uttid.py fairseq/examples/emotion_conversion/preprocess/split_km.py fairseq/examples/emotion_conversion/preprocess/split_km_tsv.py fairseq/examples/fast_noisy_channel/README.md fairseq/examples/fast_noisy_channel/__init__.py fairseq/examples/fast_noisy_channel/noisy_channel_beam_search.py fairseq/examples/fast_noisy_channel/noisy_channel_sequence_generator.py fairseq/examples/fast_noisy_channel/noisy_channel_translation.py fairseq/examples/flores101/README.md fairseq/examples/flores101/flores_logo.png fairseq/examples/fully_sharded_data_parallel/README.md fairseq/examples/gottbert/README.md fairseq/examples/hubert/README.md fairseq/examples/hubert/measure_teacher_quality.py fairseq/examples/hubert/update_ckpt.py fairseq/examples/hubert/config/decode/infer_fsqlm.yaml fairseq/examples/hubert/config/decode/infer_kenlm.yaml fairseq/examples/hubert/config/decode/infer_viterbi.yaml fairseq/examples/hubert/config/decode/ax_sweep/ngram.yaml fairseq/examples/hubert/config/decode/ax_sweep/transformer.yaml fairseq/examples/hubert/config/decode/run/submitit_slurm.yaml fairseq/examples/hubert/config/decode/run/submitit_slurm_8gpu.yaml fairseq/examples/hubert/config/finetune/base_10h.yaml fairseq/examples/hubert/config/finetune/ckpt/it1.yaml fairseq/examples/hubert/config/finetune/lm/ls_4gram.yaml fairseq/examples/hubert/config/finetune/run/submitit_reg.yaml fairseq/examples/hubert/config/pretrain/hubert_base_librispeech.yaml fairseq/examples/hubert/config/pretrain/hubert_large_librivox.yaml fairseq/examples/hubert/config/pretrain/hubert_xlarge_librivox.yaml fairseq/examples/hubert/config/pretrain/data/iter1.yaml fairseq/examples/hubert/config/pretrain/data/iter2.yaml fairseq/examples/hubert/config/pretrain/run/submitit_reg.yaml fairseq/examples/hubert/simple_kmeans/README.md fairseq/examples/hubert/simple_kmeans/dump_hubert_feature.py fairseq/examples/hubert/simple_kmeans/dump_hubert_feature_s2t.py fairseq/examples/hubert/simple_kmeans/dump_km_label.py fairseq/examples/hubert/simple_kmeans/dump_mfcc_feature.py fairseq/examples/hubert/simple_kmeans/dump_w2v2_feature.py fairseq/examples/hubert/simple_kmeans/feature_utils.py fairseq/examples/hubert/simple_kmeans/learn_kmeans.py fairseq/examples/hubert/tests/6313-76958-0021.flac fairseq/examples/hubert/tests/sample.base.L9.km500.km fairseq/examples/hubert/tests/sample.base.L9.len fairseq/examples/hubert/tests/sample.base.L9.npy fairseq/examples/hubert/tests/sample.large.L20.len fairseq/examples/hubert/tests/sample.large.L20.npy fairseq/examples/hubert/tests/sample.large.hypo.word fairseq/examples/hubert/tests/sample.xlarge.L30.len fairseq/examples/hubert/tests/sample.xlarge.L30.npy fairseq/examples/hubert/tests/sample.xlarge.hypo.word fairseq/examples/hubert/tests/test_feature_and_unit.sh fairseq/examples/hubert/tests/test_finetuned_asr.sh fairseq/examples/joint_alignment_translation/README.md fairseq/examples/joint_alignment_translation/prepare-wmt18en2de_no_norm_no_escape_no_agressive.sh fairseq/examples/language_model/README.adaptive_inputs.md fairseq/examples/language_model/README.conv.md fairseq/examples/language_model/README.md fairseq/examples/language_model/prepare-wikitext-103.sh fairseq/examples/laser/README.md fairseq/examples/laser/laser_src/__init__.py fairseq/examples/laser/laser_src/laser_lstm.py fairseq/examples/laser/laser_src/laser_task.py fairseq/examples/laser/laser_src/laser_transformer.py fairseq/examples/laser/laser_src/multitask_data_utils.py fairseq/examples/latent_depth/README.md fairseq/examples/latent_depth/latent_depth_src/__init__.py fairseq/examples/latent_depth/latent_depth_src/multilingual_translation_latent_depth.py fairseq/examples/latent_depth/latent_depth_src/loss/__init__.py fairseq/examples/latent_depth/latent_depth_src/loss/latent_depth.py fairseq/examples/latent_depth/latent_depth_src/models/__init__.py fairseq/examples/latent_depth/latent_depth_src/models/latent_multilingual_transformer.py fairseq/examples/latent_depth/latent_depth_src/models/latent_transformer.py fairseq/examples/latent_depth/latent_depth_src/modules/__init__.py fairseq/examples/latent_depth/latent_depth_src/modules/latent_layers.py fairseq/examples/layerdrop/README.md fairseq/examples/linformer/README.md fairseq/examples/linformer/linformer_src/__init__.py fairseq/examples/linformer/linformer_src/models/__init__.py fairseq/examples/linformer/linformer_src/models/linformer_roberta.py fairseq/examples/linformer/linformer_src/modules/__init__.py fairseq/examples/linformer/linformer_src/modules/linformer_sentence_encoder.py fairseq/examples/linformer/linformer_src/modules/linformer_sentence_encoder_layer.py fairseq/examples/linformer/linformer_src/modules/multihead_linear_attention.py fairseq/examples/m2m_100/README.md fairseq/examples/m2m_100/install_dependecies.sh fairseq/examples/m2m_100/tok.sh fairseq/examples/m2m_100/process_data/clean_histogram.py fairseq/examples/m2m_100/process_data/dedup_data.py fairseq/examples/m2m_100/process_data/remove_too_much_punc.py fairseq/examples/m2m_100/tokenizers/README.md fairseq/examples/m2m_100/tokenizers/seg_ja.sh fairseq/examples/m2m_100/tokenizers/seg_ko.sh fairseq/examples/m2m_100/tokenizers/tokenize_indic.py fairseq/examples/m2m_100/tokenizers/tokenize_thai.py fairseq/examples/m2m_100/tokenizers/tokenize_zh.py fairseq/examples/m2m_100/tokenizers/tokenizer_ar.sh fairseq/examples/m2m_100/tokenizers/thirdparty/.gitignore fairseq/examples/mbart/README.md fairseq/examples/megatron_11b/README.md fairseq/examples/megatron_11b/detok.py fairseq/examples/mms/MODEL_CARD.md fairseq/examples/mms/README.md fairseq/examples/mms/asr/config/infer_common.yaml fairseq/examples/mms/asr/infer/example_infer_adapter.sh fairseq/examples/mms/asr/infer/mms_infer.py fairseq/examples/mms/asr/tutorial/MMS_ASR_Inference_Colab.ipynb fairseq/examples/mms/data_prep/README.md fairseq/examples/mms/data_prep/align_and_segment.py fairseq/examples/mms/data_prep/align_utils.py fairseq/examples/mms/data_prep/norm_config.py fairseq/examples/mms/data_prep/punctuations.lst fairseq/examples/mms/data_prep/text_normalization.py fairseq/examples/mms/lid/infer.py fairseq/examples/mms/lid/tutorial/MMS_LID_Inference_Colab.ipynb fairseq/examples/mms/lid_rerank/README.md fairseq/examples/mms/lid_rerank/cer_langs.txt fairseq/examples/mms/lid_rerank/requirements.txt fairseq/examples/mms/lid_rerank/mala/infer.py fairseq/examples/mms/lid_rerank/mms/make_parallel_single_runs.py fairseq/examples/mms/lid_rerank/mms/merge_by_lang.py fairseq/examples/mms/lid_rerank/mms/prep_wav_list.py fairseq/examples/mms/lid_rerank/mms/run_single_lang.py fairseq/examples/mms/lid_rerank/mms/split_by_lang.py fairseq/examples/mms/lid_rerank/mms-zs/falign.py fairseq/examples/mms/lid_rerank/mms-zs/lib.py fairseq/examples/mms/lid_rerank/mms-zs/uromanize.py fairseq/examples/mms/lid_rerank/nllb/infer.py fairseq/examples/mms/lid_rerank/rerank/rerank.py fairseq/examples/mms/lid_rerank/rerank/tune_coefficients.py fairseq/examples/mms/lid_rerank/whisper/infer_asr.py fairseq/examples/mms/lid_rerank/whisper/infer_lid.py fairseq/examples/mms/lid_rerank/whisper/lid_mapping.txt fairseq/examples/mms/misc/get_sample_size.py fairseq/examples/mms/tts/infer.py fairseq/examples/mms/tts/tutorial/MMS_TTS_Inference_Colab.ipynb fairseq/examples/mms/zero_shot/README.md fairseq/examples/moe_lm/README.md fairseq/examples/moe_lm/data_card.md fairseq/examples/moe_lm/model_card.md fairseq/examples/mr_hubert/README.md fairseq/examples/mr_hubert/decode.sh fairseq/examples/mr_hubert/finetune.sh fairseq/examples/mr_hubert/train.sh fairseq/examples/mr_hubert/config/decode/infer.yaml fairseq/examples/mr_hubert/config/decode/infer_lm.yaml fairseq/examples/mr_hubert/config/decode/run/submitit_slurm.yaml fairseq/examples/mr_hubert/config/decode/run/submitit_slurm_8gpu.yaml fairseq/examples/mr_hubert/config/finetune/base_100h.yaml fairseq/examples/mr_hubert/config/finetune/base_100h_large.yaml fairseq/examples/mr_hubert/config/finetune/base_10h.yaml fairseq/examples/mr_hubert/config/finetune/base_10h_large.yaml fairseq/examples/mr_hubert/config/finetune/base_1h.yaml fairseq/examples/mr_hubert/config/finetune/base_1h_large.yaml fairseq/examples/mr_hubert/config/pretrain/mrhubert_base_librispeech.yaml fairseq/examples/mr_hubert/config/pretrain/mrhubert_large_librilight.yaml fairseq/examples/mr_hubert/config/pretrain/run/submitit_reg.yaml fairseq/examples/mr_hubert/simple_kmeans/README.md fairseq/examples/mr_hubert/simple_kmeans/dump_hubert_feature.py fairseq/examples/mr_hubert/simple_kmeans/dump_hubert_feature_s2t.py fairseq/examples/mr_hubert/simple_kmeans/dump_km_label.py fairseq/examples/mr_hubert/simple_kmeans/dump_mfcc_feature.py fairseq/examples/mr_hubert/simple_kmeans/dump_w2v2_feature.py fairseq/examples/mr_hubert/simple_kmeans/feature_utils.py fairseq/examples/mr_hubert/simple_kmeans/learn_kmeans.py fairseq/examples/multilingual/ML50_langs.txt fairseq/examples/multilingual/README.md fairseq/examples/multilingual/finetune_multilingual_model.sh fairseq/examples/multilingual/multilingual_fairseq_gen.sh fairseq/examples/multilingual/train_multilingual_model.sh fairseq/examples/multilingual/data_scripts/README.md fairseq/examples/multilingual/data_scripts/binarize.py fairseq/examples/multilingual/data_scripts/check_iswlt_test_data.py fairseq/examples/multilingual/data_scripts/check_self_overlaps.py fairseq/examples/multilingual/data_scripts/check_valid_test_overlaps.py fairseq/examples/multilingual/data_scripts/dedup_all.py fairseq/examples/multilingual/data_scripts/download_ML50_v1.sh fairseq/examples/multilingual/data_scripts/download_af_xh.sh fairseq/examples/multilingual/data_scripts/download_flores_data.sh fairseq/examples/multilingual/data_scripts/download_iitb.sh fairseq/examples/multilingual/data_scripts/download_iwslt_and_extract.sh fairseq/examples/multilingual/data_scripts/download_lotus.sh fairseq/examples/multilingual/data_scripts/download_ted_and_extract.py fairseq/examples/multilingual/data_scripts/download_wat19_my.sh fairseq/examples/multilingual/data_scripts/download_wmt19_and_before.py fairseq/examples/multilingual/data_scripts/download_wmt20.sh fairseq/examples/multilingual/data_scripts/preprocess_ML50_v1.sh fairseq/examples/multilingual/data_scripts/remove_valid_test_in_train.py fairseq/examples/multilingual/data_scripts/requirement.txt fairseq/examples/multilingual/data_scripts/utils/dedup.py fairseq/examples/multilingual/data_scripts/utils/fasttext_multi_filter.py fairseq/examples/multilingual/data_scripts/utils/strip_sgm.sh fairseq/examples/noisychannel/README.md fairseq/examples/noisychannel/__init__.py fairseq/examples/noisychannel/rerank.py fairseq/examples/noisychannel/rerank_generate.py fairseq/examples/noisychannel/rerank_options.py fairseq/examples/noisychannel/rerank_score_bw.py fairseq/examples/noisychannel/rerank_score_lm.py fairseq/examples/noisychannel/rerank_tune.py fairseq/examples/noisychannel/rerank_utils.py fairseq/examples/nonautoregressive_translation/README.md fairseq/examples/nonautoregressive_translation/scripts.md fairseq/examples/normformer/README.md fairseq/examples/normformer/train_lm.sh fairseq/examples/operators/alignment_train_cpu.cpp fairseq/examples/operators/alignment_train_cuda.cpp fairseq/examples/operators/alignment_train_cuda.h fairseq/examples/operators/alignment_train_kernel.cu fairseq/examples/operators/utils.h fairseq/examples/paraphraser/README.md fairseq/examples/paraphraser/paraphrase.py fairseq/examples/pay_less_attention_paper/README.md fairseq/examples/pointer_generator/README.md fairseq/examples/pointer_generator/README.xsum.md fairseq/examples/pointer_generator/postprocess.py fairseq/examples/pointer_generator/preprocess.py fairseq/examples/pointer_generator/pointer_generator_src/__init__.py fairseq/examples/pointer_generator/pointer_generator_src/transformer_pg.py fairseq/examples/quant_noise/README.md fairseq/examples/quant_noise/transformer_quantization_config.yaml fairseq/examples/roberta/README.custom_classification.md fairseq/examples/roberta/README.glue.md fairseq/examples/roberta/README.md fairseq/examples/roberta/README.pretraining.md fairseq/examples/roberta/README.race.md fairseq/examples/roberta/multiprocessing_bpe_encoder.py fairseq/examples/roberta/preprocess_GLUE_tasks.sh fairseq/examples/roberta/preprocess_RACE.py fairseq/examples/roberta/preprocess_RACE.sh fairseq/examples/roberta/commonsense_qa/README.md fairseq/examples/roberta/commonsense_qa/__init__.py fairseq/examples/roberta/commonsense_qa/commonsense_qa_task.py fairseq/examples/roberta/commonsense_qa/download_cqa_data.sh fairseq/examples/roberta/config/finetuning/cola.yaml fairseq/examples/roberta/config/finetuning/mnli.yaml fairseq/examples/roberta/config/finetuning/mrpc.yaml fairseq/examples/roberta/config/finetuning/qnli.yaml fairseq/examples/roberta/config/finetuning/qqp.yaml fairseq/examples/roberta/config/finetuning/rte.yaml fairseq/examples/roberta/config/finetuning/sst_2.yaml fairseq/examples/roberta/config/finetuning/sts_b.yaml fairseq/examples/roberta/config/finetuning/run_config/local.yaml fairseq/examples/roberta/config/finetuning/run_config/slurm_1g.yaml fairseq/examples/roberta/config/finetuning/run_config/slurm_1g_aws.yaml fairseq/examples/roberta/config/pretraining/base.yaml fairseq/examples/roberta/config/pretraining/run_config/local.yaml fairseq/examples/roberta/config/pretraining/run_config/slurm_2.yaml fairseq/examples/roberta/config/pretraining/run_config/slurm_2_aws.yaml fairseq/examples/roberta/config/pretraining/run_config/slurm_3.yaml fairseq/examples/roberta/config/pretraining/run_config/slurm_4.yaml fairseq/examples/roberta/fb_multilingual/README.multilingual.pretraining.md fairseq/examples/roberta/wsc/README.md fairseq/examples/roberta/wsc/__init__.py fairseq/examples/roberta/wsc/wsc_criterion.py fairseq/examples/roberta/wsc/wsc_task.py fairseq/examples/roberta/wsc/wsc_utils.py fairseq/examples/rxf/README.md fairseq/examples/rxf/__init__.py fairseq/examples/rxf/rxf_src/__init__.py fairseq/examples/rxf/rxf_src/label_smoothed_cross_entropy_r3f.py fairseq/examples/rxf/rxf_src/sentence_prediction_r3f.py fairseq/examples/scaling_nmt/README.md fairseq/examples/shuffled_word_order/README.finetuning.md fairseq/examples/shuffled_word_order/README.md fairseq/examples/simultaneous_translation/README.md fairseq/examples/simultaneous_translation/__init__.py fairseq/examples/simultaneous_translation/docs/ende-mma.md fairseq/examples/simultaneous_translation/docs/enja-waitk.md fairseq/examples/simultaneous_translation/eval/agents/simul_t2t_enja.py fairseq/examples/simultaneous_translation/models/__init__.py fairseq/examples/simultaneous_translation/models/convtransformer_simul_trans.py fairseq/examples/simultaneous_translation/models/transformer_monotonic_attention.py fairseq/examples/simultaneous_translation/modules/__init__.py fairseq/examples/simultaneous_translation/modules/fixed_pre_decision.py fairseq/examples/simultaneous_translation/modules/monotonic_multihead_attention.py fairseq/examples/simultaneous_translation/modules/monotonic_transformer_layer.py fairseq/examples/simultaneous_translation/tests/test_alignment_train.py fairseq/examples/simultaneous_translation/tests/test_text_models.py fairseq/examples/simultaneous_translation/utils/__init__.py fairseq/examples/simultaneous_translation/utils/functions.py fairseq/examples/simultaneous_translation/utils/monotonic_attention.py fairseq/examples/simultaneous_translation/utils/p_choose_strategy.py fairseq/examples/speech_recognition/README.md fairseq/examples/speech_recognition/__init__.py fairseq/examples/speech_recognition/infer.py fairseq/examples/speech_recognition/w2l_decoder.py fairseq/examples/speech_recognition/criterions/ASG_loss.py fairseq/examples/speech_recognition/criterions/__init__.py fairseq/examples/speech_recognition/criterions/cross_entropy_acc.py fairseq/examples/speech_recognition/data/__init__.py fairseq/examples/speech_recognition/data/asr_dataset.py fairseq/examples/speech_recognition/data/collaters.py fairseq/examples/speech_recognition/data/data_utils.py fairseq/examples/speech_recognition/data/replabels.py fairseq/examples/speech_recognition/datasets/asr_prep_json.py fairseq/examples/speech_recognition/datasets/prepare-librispeech.sh fairseq/examples/speech_recognition/kaldi/__init__.py fairseq/examples/speech_recognition/kaldi/add-self-loop-simple.cc fairseq/examples/speech_recognition/kaldi/kaldi_decoder.py fairseq/examples/speech_recognition/kaldi/kaldi_initializer.py fairseq/examples/speech_recognition/kaldi/config/kaldi_initializer.yaml fairseq/examples/speech_recognition/models/__init__.py fairseq/examples/speech_recognition/models/vggtransformer.py fairseq/examples/speech_recognition/models/w2l_conv_glu_enc.py fairseq/examples/speech_recognition/new/README.md fairseq/examples/speech_recognition/new/__init__.py fairseq/examples/speech_recognition/new/infer.py fairseq/examples/speech_recognition/new/conf/infer.yaml fairseq/examples/speech_recognition/new/conf/hydra/sweeper/ax.yaml fairseq/examples/speech_recognition/new/conf/hydra/sweeper/ax_sil.yaml fairseq/examples/speech_recognition/new/conf/run_config/fb_slurm_1.yaml fairseq/examples/speech_recognition/new/conf/run_config/fb_slurm_2g.yaml fairseq/examples/speech_recognition/new/decoders/__init__.py fairseq/examples/speech_recognition/new/decoders/base_decoder.py fairseq/examples/speech_recognition/new/decoders/decoder.py fairseq/examples/speech_recognition/new/decoders/decoder_config.py fairseq/examples/speech_recognition/new/decoders/flashlight_decoder.py fairseq/examples/speech_recognition/new/decoders/viterbi_decoder.py fairseq/examples/speech_recognition/tasks/__init__.py fairseq/examples/speech_recognition/tasks/speech_recognition.py fairseq/examples/speech_recognition/utils/wer_utils.py fairseq/examples/speech_synthesis/README.md fairseq/examples/speech_synthesis/__init__.py fairseq/examples/speech_synthesis/data_utils.py fairseq/examples/speech_synthesis/generate_waveform.py fairseq/examples/speech_synthesis/utils.py fairseq/examples/speech_synthesis/docs/common_voice_example.md fairseq/examples/speech_synthesis/docs/ljspeech_example.md fairseq/examples/speech_synthesis/docs/vctk_example.md fairseq/examples/speech_synthesis/evaluation/__init__.py fairseq/examples/speech_synthesis/evaluation/eval_asr.py fairseq/examples/speech_synthesis/evaluation/eval_f0.py fairseq/examples/speech_synthesis/evaluation/eval_sp.py fairseq/examples/speech_synthesis/evaluation/get_eval_manifest.py fairseq/examples/speech_synthesis/preprocessing/__init__.py fairseq/examples/speech_synthesis/preprocessing/denoise_and_vad_audio.py fairseq/examples/speech_synthesis/preprocessing/get_common_voice_audio_manifest.py fairseq/examples/speech_synthesis/preprocessing/get_feature_manifest.py fairseq/examples/speech_synthesis/preprocessing/get_ljspeech_audio_manifest.py fairseq/examples/speech_synthesis/preprocessing/get_speaker_embedding.py fairseq/examples/speech_synthesis/preprocessing/get_vctk_audio_manifest.py fairseq/examples/speech_synthesis/preprocessing/denoiser/__init__.py fairseq/examples/speech_synthesis/preprocessing/denoiser/demucs.py fairseq/examples/speech_synthesis/preprocessing/denoiser/pretrained.py fairseq/examples/speech_synthesis/preprocessing/denoiser/resample.py fairseq/examples/speech_synthesis/preprocessing/denoiser/utils.py fairseq/examples/speech_synthesis/preprocessing/speaker_embedder/__init__.py fairseq/examples/speech_synthesis/preprocessing/vad/__init__.py fairseq/examples/speech_text_joint_to_text/README.md fairseq/examples/speech_text_joint_to_text/__init__.py fairseq/examples/speech_text_joint_to_text/configs/mustc_noise.list fairseq/examples/speech_text_joint_to_text/criterions/__init__.py fairseq/examples/speech_text_joint_to_text/criterions/multi_modality_compound.py fairseq/examples/speech_text_joint_to_text/criterions/multi_modality_cross_entropy.py fairseq/examples/speech_text_joint_to_text/criterions/text_guide_cross_entropy_acc.py fairseq/examples/speech_text_joint_to_text/data/pair_denoising_dataset.py fairseq/examples/speech_text_joint_to_text/docs/ende-mustc.md fairseq/examples/speech_text_joint_to_text/docs/iwslt2021.md fairseq/examples/speech_text_joint_to_text/docs/pre-training.md fairseq/examples/speech_text_joint_to_text/models/__init__.py fairseq/examples/speech_text_joint_to_text/models/joint_speech_text_pretrain_transformer.py fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputtransformer.py fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputwavtransformer.py fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputxmtransformer.py fairseq/examples/speech_text_joint_to_text/scripts/convert_model.py fairseq/examples/speech_text_joint_to_text/scripts/g2p_encode.py fairseq/examples/speech_text_joint_to_text/tasks/__init__.py fairseq/examples/speech_text_joint_to_text/tasks/pair_denoising.py fairseq/examples/speech_text_joint_to_text/tasks/speech_text_denoise_pretrain.py fairseq/examples/speech_text_joint_to_text/tasks/speech_text_joint.py fairseq/examples/speech_to_speech/README.md fairseq/examples/speech_to_speech/__init__.py fairseq/examples/speech_to_speech/generate_waveform_from_code.py fairseq/examples/speech_to_speech/asr_bleu/README.md fairseq/examples/speech_to_speech/asr_bleu/__init__.py fairseq/examples/speech_to_speech/asr_bleu/asr_model_cfgs.json fairseq/examples/speech_to_speech/asr_bleu/compute_asr_bleu.py fairseq/examples/speech_to_speech/asr_bleu/requirements.txt fairseq/examples/speech_to_speech/asr_bleu/utils.py fairseq/examples/speech_to_speech/benchmarking/README.md fairseq/examples/speech_to_speech/benchmarking/core.py fairseq/examples/speech_to_speech/benchmarking/data_utils.py fairseq/examples/speech_to_speech/benchmarking/get_metrics.py fairseq/examples/speech_to_speech/benchmarking/configs/2StageS2ST.yaml fairseq/examples/speech_to_speech/benchmarking/configs/3StageS2ST.yaml fairseq/examples/speech_to_speech/benchmarking/configs/DirectS2U.yaml fairseq/examples/speech_to_speech/benchmarking/configs/S2T.yaml fairseq/examples/speech_to_speech/docs/data_augmentation.md fairseq/examples/speech_to_speech/docs/direct_s2st_discrete_units.md fairseq/examples/speech_to_speech/docs/enhanced_direct_s2st_discrete_units.md fairseq/examples/speech_to_speech/docs/textless_s2st_real_data.md fairseq/examples/speech_to_speech/preprocessing/__init__.py fairseq/examples/speech_to_speech/preprocessing/data_utils.py fairseq/examples/speech_to_speech/preprocessing/prep_s2spect_data.py fairseq/examples/speech_to_speech/preprocessing/prep_s2ut_data.py fairseq/examples/speech_to_speech/preprocessing/prep_sn_data.py fairseq/examples/speech_to_speech/preprocessing/prep_sn_output_data.py fairseq/examples/speech_to_speech/unity/__init__.py fairseq/examples/speech_to_speech/unity/sequence_generator.py fairseq/examples/speech_to_speech/unity/sequence_generator_multi_decoder.py fairseq/examples/speech_to_text/README.md fairseq/examples/speech_to_text/data_utils.py fairseq/examples/speech_to_text/prep_covost_data.py fairseq/examples/speech_to_text/prep_librispeech_data.py fairseq/examples/speech_to_text/prep_mtedx_data.py fairseq/examples/speech_to_text/prep_mustc_data.py fairseq/examples/speech_to_text/seg_mustc_data.py fairseq/examples/speech_to_text/docs/covost_example.md fairseq/examples/speech_to_text/docs/librispeech_example.md fairseq/examples/speech_to_text/docs/mtedx_example.md fairseq/examples/speech_to_text/docs/mustc_example.md fairseq/examples/speech_to_text/docs/simulst_mustc_example.md fairseq/examples/speech_to_text/simultaneous_translation/agents/fairseq_simul_st_agent.py fairseq/examples/stories/README.md fairseq/examples/textless_nlp/dgslm/README.md fairseq/examples/textless_nlp/dgslm/create_code_file.py fairseq/examples/textless_nlp/dgslm/dgslm_utils.py fairseq/examples/textless_nlp/dgslm/sample_speech_dlm.py fairseq/examples/textless_nlp/dgslm/hubert_fisher/README.md fairseq/examples/textless_nlp/dgslm/vocoder_hifigan/README.md fairseq/examples/textless_nlp/dgslm/vocoder_hifigan/generate_stereo_waveform.py fairseq/examples/textless_nlp/gslm/README.md fairseq/examples/textless_nlp/gslm/metrics/README.md fairseq/examples/textless_nlp/gslm/metrics/abx_metrics/README.md fairseq/examples/textless_nlp/gslm/metrics/abx_metrics/dump_abx_feats.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/README.md fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/continuation_eval.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/ppx.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/self_auto_bleu.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/bleu_utils.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/cut_as.py fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/dict.ltr.txt fairseq/examples/textless_nlp/gslm/speech2unit/README.md fairseq/examples/textless_nlp/gslm/speech2unit/__init__.py fairseq/examples/textless_nlp/gslm/speech2unit/clustering/__init__.py fairseq/examples/textless_nlp/gslm/speech2unit/clustering/cluster_kmeans.py fairseq/examples/textless_nlp/gslm/speech2unit/clustering/dump_feats.py fairseq/examples/textless_nlp/gslm/speech2unit/clustering/quantize_with_kmeans.py fairseq/examples/textless_nlp/gslm/speech2unit/clustering/utils.py fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/cpc_feature_reader.py fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/hubert_feature_reader.py fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/logmel_feature_reader.py fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/utils.py fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/w2v2_feature_reader.py fairseq/examples/textless_nlp/gslm/tools/README.md fairseq/examples/textless_nlp/gslm/tools/resynthesize_speech.py fairseq/examples/textless_nlp/gslm/ulm/README.md fairseq/examples/textless_nlp/gslm/ulm/sample.py fairseq/examples/textless_nlp/gslm/unit2speech/README.md fairseq/examples/textless_nlp/gslm/unit2speech/convert_to_16k.py fairseq/examples/textless_nlp/gslm/unit2speech/glow.py fairseq/examples/textless_nlp/gslm/unit2speech/multiproc.py fairseq/examples/textless_nlp/gslm/unit2speech/synthesize_audio_from_units.py fairseq/examples/textless_nlp/gslm/unit2speech/tts_data.py fairseq/examples/textless_nlp/gslm/unit2speech/utils.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/__init__.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/audio_processing.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/cleaners.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/cmudict.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/layers.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/model.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/numbers.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/stft.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/symbols.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/text.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/utils.py fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/waveglow_denoiser.py fairseq/examples/textless_nlp/pgslm/README.md fairseq/examples/textless_nlp/pgslm/data_utils.py fairseq/examples/textless_nlp/pgslm/generate_waveform.py fairseq/examples/textless_nlp/pgslm/inference_dataset.py fairseq/examples/textless_nlp/pgslm/naive_decoder.py fairseq/examples/textless_nlp/pgslm/prepare_dataset.py fairseq/examples/textless_nlp/pgslm/preprocess_f0.py fairseq/examples/textless_nlp/pgslm/quantize_f0.py fairseq/examples/textless_nlp/pgslm/truncated_laplace.py fairseq/examples/textless_nlp/pgslm/eval/__init__.py fairseq/examples/textless_nlp/pgslm/eval/cont_metrics.py fairseq/examples/textless_nlp/pgslm/sample/__init__.py fairseq/examples/textless_nlp/pgslm/sample/sample.py fairseq/examples/textless_nlp/pgslm/scripts/join_units_manifest.py fairseq/examples/textless_nlp/pgslm/scripts/prepare_data.sh fairseq/examples/textless_nlp/pgslm/scripts/prepare_f0_quantization.sh fairseq/examples/textless_nlp/speech-resynth/README.md fairseq/examples/textless_nlp/speech-resynth/img/fig.png fairseq/examples/translation/README.md fairseq/examples/translation/prepare-iwslt14.sh fairseq/examples/translation/prepare-iwslt17-multilingual.sh fairseq/examples/translation/prepare-wmt14en2de.sh fairseq/examples/translation/prepare-wmt14en2fr.sh fairseq/examples/translation_moe/README.md fairseq/examples/translation_moe/score.py fairseq/examples/translation_moe/translation_moe_src/__init__.py fairseq/examples/translation_moe/translation_moe_src/logsumexp_moe.py fairseq/examples/translation_moe/translation_moe_src/mean_pool_gating_network.py fairseq/examples/translation_moe/translation_moe_src/translation_moe.py fairseq/examples/truncated_bptt/README.md fairseq/examples/truncated_bptt/__init__.py fairseq/examples/truncated_bptt/transformer_xl_model.py fairseq/examples/truncated_bptt/truncated_bptt_lm_task.py fairseq/examples/unsupervised_quality_estimation/README.md fairseq/examples/unsupervised_quality_estimation/aggregate_scores.py fairseq/examples/unsupervised_quality_estimation/meteor.py fairseq/examples/unsupervised_quality_estimation/repeat_lines.py fairseq/examples/wav2vec/README.md fairseq/examples/wav2vec/__init__.py fairseq/examples/wav2vec/libri_labels.py fairseq/examples/wav2vec/vq-wav2vec_featurize.py fairseq/examples/wav2vec/wav2vec_featurize.py fairseq/examples/wav2vec/wav2vec_manifest.py fairseq/examples/wav2vec/config/finetuning/base_100h.yaml fairseq/examples/wav2vec/config/finetuning/base_10h.yaml fairseq/examples/wav2vec/config/finetuning/base_10m.yaml fairseq/examples/wav2vec/config/finetuning/base_1h.yaml fairseq/examples/wav2vec/config/finetuning/base_960h.yaml fairseq/examples/wav2vec/config/finetuning/vox_100h.yaml fairseq/examples/wav2vec/config/finetuning/vox_100h_2.yaml fairseq/examples/wav2vec/config/finetuning/vox_100h_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_100h_3.yaml fairseq/examples/wav2vec/config/finetuning/vox_10h.yaml fairseq/examples/wav2vec/config/finetuning/vox_10h_2.yaml fairseq/examples/wav2vec/config/finetuning/vox_10h_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_10h_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_10h_aws_v100.yaml fairseq/examples/wav2vec/config/finetuning/vox_10m.yaml fairseq/examples/wav2vec/config/finetuning/vox_10m_2.yaml fairseq/examples/wav2vec/config/finetuning/vox_10m_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_10m_3.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h_2.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h_3.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h_4.yaml fairseq/examples/wav2vec/config/finetuning/vox_1h_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_960h.yaml fairseq/examples/wav2vec/config/finetuning/vox_960h_2.yaml fairseq/examples/wav2vec/config/finetuning/vox_960h_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/vox_960h_3.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_16.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1_aws.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1_old.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2_aws.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2g.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_3.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_4g.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_4g_aws.yaml fairseq/examples/wav2vec/config/finetuning/run_config/slurm_8.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_base_librispeech.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_conformer_base_librispeech.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_conformer_large_librivox.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu-pod.yaml fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu.yaml fairseq/examples/wav2vec/scripts/binarize_manifest.sh fairseq/examples/wav2vec/unsupervised/README.md fairseq/examples/wav2vec/unsupervised/__init__.py fairseq/examples/wav2vec/unsupervised/w2vu_generate.py fairseq/examples/wav2vec/unsupervised/config/finetuning/w2v_finetune.yaml fairseq/examples/wav2vec/unsupervised/config/gan/w2vu.yaml fairseq/examples/wav2vec/unsupervised/config/gan/w2vu2.yaml fairseq/examples/wav2vec/unsupervised/config/generate/viterbi.yaml fairseq/examples/wav2vec/unsupervised/config/timit_matched/test.uid fairseq/examples/wav2vec/unsupervised/config/timit_matched/train.uid fairseq/examples/wav2vec/unsupervised/config/timit_matched/train_text.uid fairseq/examples/wav2vec/unsupervised/config/timit_matched/valid.uid fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/test.uid fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/train.uid fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/train_text.uid fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/valid.uid fairseq/examples/wav2vec/unsupervised/data/__init__.py fairseq/examples/wav2vec/unsupervised/data/extracted_features_dataset.py fairseq/examples/wav2vec/unsupervised/data/random_input_dataset.py fairseq/examples/wav2vec/unsupervised/kaldi_self_train/README.md fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/cmd.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_phone.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step1.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step2.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/path.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/train.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/copy_aligned_text.py fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/decode.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_data_from_w2v.py fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang_word.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lm.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/score.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/show_wer.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/train_subset_lgbeam.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select.py fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode_word.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_deltas.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_lda_mllt.sh fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_sat.sh fairseq/examples/wav2vec/unsupervised/models/__init__.py fairseq/examples/wav2vec/unsupervised/models/wav2vec_u.py fairseq/examples/wav2vec/unsupervised/scripts/apply_pca.py fairseq/examples/wav2vec/unsupervised/scripts/copy_labels.py fairseq/examples/wav2vec/unsupervised/scripts/filter_lexicon.py fairseq/examples/wav2vec/unsupervised/scripts/filter_tsv.py fairseq/examples/wav2vec/unsupervised/scripts/g2p_wrd_to_phn.py fairseq/examples/wav2vec/unsupervised/scripts/ltr_to_wrd.py fairseq/examples/wav2vec/unsupervised/scripts/mean_pool.py fairseq/examples/wav2vec/unsupervised/scripts/merge_clusters.py fairseq/examples/wav2vec/unsupervised/scripts/normalize_and_filter_text.py fairseq/examples/wav2vec/unsupervised/scripts/normalize_text.py fairseq/examples/wav2vec/unsupervised/scripts/pca.py fairseq/examples/wav2vec/unsupervised/scripts/phonemize_with_sil.py fairseq/examples/wav2vec/unsupervised/scripts/prepare_audio.sh fairseq/examples/wav2vec/unsupervised/scripts/prepare_audio_v2.sh fairseq/examples/wav2vec/unsupervised/scripts/prepare_text.sh fairseq/examples/wav2vec/unsupervised/scripts/prepare_timit.sh fairseq/examples/wav2vec/unsupervised/scripts/remove_silence.py fairseq/examples/wav2vec/unsupervised/scripts/vads.py fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_apply_cluster_faiss.py fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_cluster_faiss.py fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_extract_features.py fairseq/examples/wav2vec/unsupervised/scripts/wer.py fairseq/examples/wav2vec/unsupervised/scripts/wrd_to_ltr.py fairseq/examples/wav2vec/unsupervised/tasks/__init__.py fairseq/examples/wav2vec/unsupervised/tasks/unpaired_audio_text.py fairseq/examples/wav2vec/xlsr/README.md fairseq/examples/wav2vec/xlsr/config/finetune.yaml fairseq/examples/wav2vec/xlsr/scripts/eval_speaker_clf_task.py fairseq/examples/wav2vec/xlsr/scripts/gen_audio_embedding.py fairseq/examples/wmt19/README.md fairseq/examples/wmt20/README.md fairseq/examples/wmt21/README.md fairseq/examples/wmt21/eval.sh fairseq/examples/wmt21/scripts/normalize-punctuation.perl fairseq/examples/wmt21/scripts/replace-unicode-punctuation.perl fairseq/examples/womens_bios/README.md fairseq/examples/womens_bios/query_occupations_from_wikidata.py fairseq/examples/xformers/README.md fairseq/examples/xglm/README.md fairseq/examples/xglm/XStoryCloze.md fairseq/examples/xglm/model_card.md fairseq/examples/xlmr/README.md fairseq/examples/xmod/README.md fairseq/examples/xmod/preprocess_nli.py fairseq/logging/__init__.py fairseq/logging/meters.py fairseq/logging/metrics.py fairseq/logging/progress_bar.py fairseq/model_parallel/__init__.py fairseq/model_parallel/megatron_trainer.py fairseq/model_parallel/criterions/__init__.py fairseq/model_parallel/criterions/vocab_parallel_cross_entropy.py fairseq/model_parallel/models/__init__.py fairseq/model_parallel/models/transformer.py fairseq/model_parallel/models/transformer_lm.py fairseq/model_parallel/models/pipeline_parallel_transformer/__init__.py fairseq/model_parallel/models/pipeline_parallel_transformer/layers.py fairseq/model_parallel/models/pipeline_parallel_transformer/model.py fairseq/model_parallel/models/roberta/__init__.py fairseq/model_parallel/models/roberta/model.py fairseq/model_parallel/modules/__init__.py fairseq/model_parallel/modules/multihead_attention.py fairseq/model_parallel/modules/transformer_layer.py fairseq/models/__init__.py fairseq/models/composite_encoder.py fairseq/models/distributed_fairseq_model.py fairseq/models/fairseq_decoder.py fairseq/models/fairseq_encoder.py fairseq/models/fairseq_incremental_decoder.py fairseq/models/fairseq_model.py fairseq/models/fconv.py fairseq/models/fconv_lm.py fairseq/models/fconv_self_att.py fairseq/models/lightconv.py fairseq/models/lightconv_lm.py fairseq/models/lstm.py fairseq/models/lstm_lm.py fairseq/models/masked_lm.py fairseq/models/model_utils.py fairseq/models/multilingual_transformer.py fairseq/models/transformer_align.py fairseq/models/transformer_from_pretrained_xlm.py fairseq/models/transformer_lm.py fairseq/models/transformer_ulm.py fairseq/models/bart/__init__.py fairseq/models/bart/hub_interface.py fairseq/models/bart/model.py fairseq/models/ema/__init__.py fairseq/models/ema/ema.py fairseq/models/hubert/__init__.py fairseq/models/hubert/hubert.py fairseq/models/hubert/hubert_asr.py fairseq/models/huggingface/__init__.py fairseq/models/huggingface/hf_gpt2.py fairseq/models/multires_hubert/__init__.py fairseq/models/multires_hubert/multires_hubert.py fairseq/models/multires_hubert/multires_hubert_asr.py fairseq/models/nat/__init__.py fairseq/models/nat/cmlm_transformer.py fairseq/models/nat/fairseq_nat_model.py fairseq/models/nat/insertion_transformer.py fairseq/models/nat/iterative_nonautoregressive_transformer.py fairseq/models/nat/levenshtein_transformer.py fairseq/models/nat/levenshtein_utils.py fairseq/models/nat/nat_crf_transformer.py fairseq/models/nat/nonautoregressive_ensembles.py fairseq/models/nat/nonautoregressive_transformer.py fairseq/models/roberta/__init__.py fairseq/models/roberta/alignment_utils.py fairseq/models/roberta/enc_dec.py fairseq/models/roberta/hub_interface.py fairseq/models/roberta/model.py fairseq/models/roberta/model_camembert.py fairseq/models/roberta/model_gottbert.py fairseq/models/roberta/model_xlmr.py fairseq/models/speech_dlm/__init__.py fairseq/models/speech_dlm/hub_interface.py fairseq/models/speech_dlm/speech_dlm.py fairseq/models/speech_dlm/modules/__init__.py fairseq/models/speech_dlm/modules/speech_dlm_decoder.py fairseq/models/speech_dlm/modules/speech_dlm_decoder_layer.py fairseq/models/speech_dlm/sequence_generator/__init__.py fairseq/models/speech_dlm/sequence_generator/multichannel_search.py fairseq/models/speech_dlm/sequence_generator/multichannel_sequence_generator.py fairseq/models/speech_to_speech/__init__.py fairseq/models/speech_to_speech/s2s_conformer.py fairseq/models/speech_to_speech/s2s_conformer_translatotron2.py fairseq/models/speech_to_speech/s2s_conformer_unity.py fairseq/models/speech_to_speech/s2s_transformer.py fairseq/models/speech_to_speech/modules/__init__.py fairseq/models/speech_to_speech/modules/ctc_decoder.py fairseq/models/speech_to_speech/modules/stacked_embedding.py fairseq/models/speech_to_speech/modules/transformer_decoder_aug.py fairseq/models/speech_to_speech/modules/transformer_encoder.py fairseq/models/speech_to_text/__init__.py fairseq/models/speech_to_text/berard.py fairseq/models/speech_to_text/convtransformer.py fairseq/models/speech_to_text/hub_interface.py fairseq/models/speech_to_text/multi_modality_model.py fairseq/models/speech_to_text/s2t_conformer.py fairseq/models/speech_to_text/s2t_transformer.py fairseq/models/speech_to_text/s2t_wav_transformer.py fairseq/models/speech_to_text/utils.py fairseq/models/speech_to_text/xm_transformer.py fairseq/models/speech_to_text/xm_transformer_unity.py fairseq/models/speech_to_text/modules/__init__.py fairseq/models/speech_to_text/modules/augmented_memory_attention.py fairseq/models/speech_to_text/modules/convolution.py fairseq/models/speech_to_text/modules/emformer.py fairseq/models/text_to_speech/__init__.py fairseq/models/text_to_speech/codehifigan.py fairseq/models/text_to_speech/fastspeech2.py fairseq/models/text_to_speech/hifigan.py fairseq/models/text_to_speech/hub_interface.py fairseq/models/text_to_speech/tacotron2.py fairseq/models/text_to_speech/tts_transformer.py fairseq/models/text_to_speech/vocoder.py fairseq/models/transformer/__init__.py fairseq/models/transformer/transformer_base.py fairseq/models/transformer/transformer_config.py fairseq/models/transformer/transformer_decoder.py fairseq/models/transformer/transformer_decoder_aug.py fairseq/models/transformer/transformer_encoder.py fairseq/models/transformer/transformer_legacy.py fairseq/models/wav2vec/__init__.py fairseq/models/wav2vec/utils.py fairseq/models/wav2vec/wav2vec.py fairseq/models/wav2vec/wav2vec2.py fairseq/models/wav2vec/wav2vec2_asr.py fairseq/models/wav2vec/wav2vec2_classification.py fairseq/models/wav2vec/wav2vec2_laser.py fairseq/models/xmod/__init__.py fairseq/models/xmod/hub_interface.py fairseq/models/xmod/model.py fairseq/models/xmod/transformer_layer_xmod.py fairseq/modules/__init__.py fairseq/modules/adaptive_input.py fairseq/modules/adaptive_softmax.py fairseq/modules/base_layer.py fairseq/modules/beamable_mm.py fairseq/modules/character_token_embedder.py fairseq/modules/checkpoint_activations.py fairseq/modules/conformer_layer.py fairseq/modules/conv_tbc.py fairseq/modules/cross_entropy.py fairseq/modules/downsampled_multihead_attention.py fairseq/modules/dynamic_convolution.py fairseq/modules/dynamic_crf_layer.py fairseq/modules/ema_module.py fairseq/modules/espnet_multihead_attention.py fairseq/modules/fairseq_dropout.py fairseq/modules/fp32_batch_norm.py fairseq/modules/fp32_group_norm.py fairseq/modules/fp32_instance_norm.py fairseq/modules/gelu.py fairseq/modules/grad_multiply.py fairseq/modules/gumbel_vector_quantizer.py fairseq/modules/kmeans_attention.py fairseq/modules/kmeans_vector_quantizer.py fairseq/modules/layer_drop.py fairseq/modules/layer_norm.py fairseq/modules/learned_positional_embedding.py fairseq/modules/lightweight_convolution.py fairseq/modules/linearized_convolution.py fairseq/modules/location_attention.py fairseq/modules/lstm_cell_with_zoneout.py fairseq/modules/multihead_attention.py fairseq/modules/positional_embedding.py fairseq/modules/positional_encoding.py fairseq/modules/quant_noise.py fairseq/modules/rotary_positional_embedding.py fairseq/modules/same_pad.py fairseq/modules/scalar_bias.py fairseq/modules/sinusoidal_positional_embedding.py fairseq/modules/sparse_multihead_attention.py fairseq/modules/sparse_transformer_sentence_encoder.py fairseq/modules/sparse_transformer_sentence_encoder_layer.py fairseq/modules/transformer_layer.py fairseq/modules/transformer_layer_aug.py fairseq/modules/transformer_sentence_encoder.py fairseq/modules/transformer_sentence_encoder_layer.py fairseq/modules/transpose_last.py fairseq/modules/unfold.py fairseq/modules/vggblock.py fairseq/modules/dynamicconv_layer/__init__.py fairseq/modules/dynamicconv_layer/cuda_function_gen.py fairseq/modules/dynamicconv_layer/dynamicconv_layer.py fairseq/modules/dynamicconv_layer/setup.py fairseq/modules/lightconv_layer/__init__.py fairseq/modules/lightconv_layer/cuda_function_gen.py fairseq/modules/lightconv_layer/lightconv_layer.py fairseq/modules/lightconv_layer/setup.py fairseq/modules/quantization/__init__.py fairseq/modules/quantization/quantization_options.py fairseq/modules/quantization/pq/__init__.py fairseq/modules/quantization/pq/em.py fairseq/modules/quantization/pq/pq.py fairseq/modules/quantization/pq/utils.py fairseq/modules/quantization/pq/modules/__init__.py fairseq/modules/quantization/pq/modules/qconv.py fairseq/modules/quantization/pq/modules/qemb.py fairseq/modules/quantization/pq/modules/qlinear.py fairseq/modules/quantization/scalar/__init__.py fairseq/modules/quantization/scalar/ops.py fairseq/modules/quantization/scalar/utils.py fairseq/modules/quantization/scalar/modules/__init__.py fairseq/modules/quantization/scalar/modules/qact.py fairseq/modules/quantization/scalar/modules/qconv.py fairseq/modules/quantization/scalar/modules/qemb.py fairseq/modules/quantization/scalar/modules/qlinear.py fairseq/optim/__init__.py fairseq/optim/adadelta.py fairseq/optim/adafactor.py fairseq/optim/adagrad.py fairseq/optim/adam.py fairseq/optim/adamax.py fairseq/optim/amp_optimizer.py fairseq/optim/bmuf.py fairseq/optim/composite.py fairseq/optim/cpu_adam.py fairseq/optim/dynamic_loss_scaler.py fairseq/optim/fairseq_optimizer.py fairseq/optim/fp16_optimizer.py fairseq/optim/fused_adam.py fairseq/optim/fused_lamb.py fairseq/optim/nag.py fairseq/optim/sgd.py fairseq/optim/shard.py fairseq/optim/lr_scheduler/__init__.py fairseq/optim/lr_scheduler/cosine_lr_scheduler.py fairseq/optim/lr_scheduler/fairseq_lr_scheduler.py fairseq/optim/lr_scheduler/fixed_schedule.py fairseq/optim/lr_scheduler/inverse_square_root_schedule.py fairseq/optim/lr_scheduler/manual_lr_scheduler.py fairseq/optim/lr_scheduler/pass_through.py fairseq/optim/lr_scheduler/polynomial_decay_schedule.py fairseq/optim/lr_scheduler/reduce_lr_on_plateau.py fairseq/optim/lr_scheduler/step_lr_scheduler.py fairseq/optim/lr_scheduler/tri_stage_lr_scheduler.py fairseq/optim/lr_scheduler/triangular_lr_scheduler.py fairseq/scoring/__init__.py fairseq/scoring/bertscore.py fairseq/scoring/bleu.py fairseq/scoring/chrf.py fairseq/scoring/meteor.py fairseq/scoring/tokenizer.py fairseq/scoring/wer.py fairseq/tasks/__init__.py fairseq/tasks/audio_classification.py fairseq/tasks/audio_finetuning.py fairseq/tasks/audio_pretraining.py fairseq/tasks/cross_lingual_lm.py fairseq/tasks/denoising.py fairseq/tasks/fairseq_task.py fairseq/tasks/frm_text_to_speech.py fairseq/tasks/hubert_pretraining.py fairseq/tasks/language_modeling.py fairseq/tasks/legacy_masked_lm.py fairseq/tasks/masked_lm.py fairseq/tasks/multilingual_denoising.py fairseq/tasks/multilingual_language_modeling.py fairseq/tasks/multilingual_masked_lm.py fairseq/tasks/multilingual_translation.py fairseq/tasks/multires_hubert_pretraining.py fairseq/tasks/nlu_finetuning.py fairseq/tasks/online_backtranslation.py fairseq/tasks/semisupervised_translation.py fairseq/tasks/sentence_prediction.py fairseq/tasks/sentence_prediction_adapters.py fairseq/tasks/sentence_ranking.py fairseq/tasks/simultaneous_translation.py fairseq/tasks/span_masked_lm.py fairseq/tasks/speech_dlm_task.py fairseq/tasks/speech_to_speech.py fairseq/tasks/speech_to_text.py fairseq/tasks/speech_ulm_task.py fairseq/tasks/text_to_speech.py fairseq/tasks/translation.py fairseq/tasks/translation_from_pretrained_bart.py fairseq/tasks/translation_from_pretrained_xlm.py fairseq/tasks/translation_lev.py fairseq/tasks/translation_multi_simple_epoch.py fairseq_cli/__init__.py fairseq_cli/eval_lm.py fairseq_cli/generate.py fairseq_cli/hydra_train.py fairseq_cli/hydra_validate.py fairseq_cli/interactive.py fairseq_cli/preprocess.py fairseq_cli/score.py fairseq_cli/train.py fairseq_cli/validate.py tests/test_activation_checkpointing.py tests/test_amp_optimizer.py tests/test_average_checkpoints.py tests/test_backtranslation_dataset.py tests/test_binaries.py tests/test_binarizer.py tests/test_character_token_embedder.py tests/test_checkpoint_utils.py tests/test_checkpoint_utils_for_task_level_attributes.py tests/test_concat_dataset.py tests/test_constraints.py tests/test_convtbc.py tests/test_data_utils.py tests/test_dataclass_utils.py tests/test_dataset.py tests/test_dictionary.py tests/test_ema.py tests/test_espnet_multihead_attention.py tests/test_export.py tests/test_file_chunker_utils.py tests/test_file_io.py tests/test_fp16_optimizer.py tests/test_hf_hub.py tests/test_huffman.py tests/test_inference_dropout.py tests/test_iopath.py tests/test_iterators.py tests/test_label_smoothing.py tests/test_lm_context_window.py tests/test_lstm_jitable.py tests/test_memory_efficient_fp16.py tests/test_metrics.py tests/test_multi_corpus_dataset.py tests/test_multi_corpus_sampled_dataset.py tests/test_multihead_attention.py tests/test_noising.py tests/test_online_backtranslation.py tests/test_plasma_utils.py tests/test_positional_encoding.py tests/test_reproducibility.py tests/test_resampling_dataset.py tests/test_roberta.py tests/test_rotary_positional_embedding.py tests/test_sequence_generator.py tests/test_sequence_scorer.py tests/test_sparse_multihead_attention.py tests/test_token_block_dataset.py tests/test_train.py tests/test_transformer.py tests/test_utils.py tests/test_valid_subset_checks.py