|
LICENSE |
|
MANIFEST.in |
|
README.md |
|
pyproject.toml |
|
setup.cfg |
|
setup.py |
|
examples/operators/alignment_train_cpu.cpp |
|
examples/operators/alignment_train_cuda.cpp |
|
examples/operators/alignment_train_kernel.cu |
|
fairseq/__init__.py |
|
fairseq/binarizer.py |
|
fairseq/checkpoint_utils.py |
|
fairseq/file_chunker_utils.py |
|
fairseq/file_io.py |
|
fairseq/file_utils.py |
|
fairseq/hub_utils.py |
|
fairseq/incremental_decoding_utils.py |
|
fairseq/iterative_refinement_generator.py |
|
fairseq/nan_detector.py |
|
fairseq/ngram_repeat_block.py |
|
fairseq/options.py |
|
fairseq/pdb.py |
|
fairseq/quantization_utils.py |
|
fairseq/registry.py |
|
fairseq/search.py |
|
fairseq/sequence_generator.py |
|
fairseq/sequence_scorer.py |
|
fairseq/speech_generator.py |
|
fairseq/token_generation_constraints.py |
|
fairseq/tokenizer.py |
|
fairseq/trainer.py |
|
fairseq/utils.py |
|
fairseq/version.py |
|
fairseq/version.txt |
|
fairseq.egg-info/PKG-INFO |
|
fairseq.egg-info/SOURCES.txt |
|
fairseq.egg-info/dependency_links.txt |
|
fairseq.egg-info/entry_points.txt |
|
fairseq.egg-info/not-zip-safe |
|
fairseq.egg-info/requires.txt |
|
fairseq.egg-info/top_level.txt |
|
fairseq/benchmark/__init__.py |
|
fairseq/benchmark/benchmark_multihead_attention.py |
|
fairseq/benchmark/dummy_dataset.py |
|
fairseq/benchmark/dummy_lm.py |
|
fairseq/benchmark/dummy_masked_lm.py |
|
fairseq/benchmark/dummy_model.py |
|
fairseq/benchmark/dummy_mt.py |
|
fairseq/clib/cuda/ngram_repeat_block_cuda.cpp |
|
fairseq/clib/cuda/ngram_repeat_block_cuda_kernel.cu |
|
fairseq/clib/libbase/balanced_assignment.cpp |
|
fairseq/clib/libbleu/libbleu.cpp |
|
fairseq/clib/libbleu/module.cpp |
|
fairseq/clib/libnat/edit_dist.cpp |
|
fairseq/clib/libnat_cuda/binding.cpp |
|
fairseq/clib/libnat_cuda/edit_dist.cu |
|
fairseq/config/__init__.py |
|
fairseq/config/config.yaml |
|
fairseq/config/fb_run_config/slurm.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_baevski_gbw.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_baevski_wiki103.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_big.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_gbw.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_gpt.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_gpt2_big.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_gpt2_medium.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_gpt2_small.yaml |
|
fairseq/config/model/transformer_lm/transformer_lm_wiki103.yaml |
|
fairseq/config/model/wav2vec/vq_wav2vec_gumbel.yaml |
|
fairseq/config/model/wav2vec2/wav2vec2_base.yaml |
|
fairseq/config/model/wav2vec2/wav2vec2_large.yaml |
|
fairseq/criterions/__init__.py |
|
fairseq/criterions/adaptive_loss.py |
|
fairseq/criterions/composite_loss.py |
|
fairseq/criterions/cross_entropy.py |
|
fairseq/criterions/ctc.py |
|
fairseq/criterions/fairseq_criterion.py |
|
fairseq/criterions/fastspeech2_loss.py |
|
fairseq/criterions/hubert_criterion.py |
|
fairseq/criterions/label_smoothed_cross_entropy.py |
|
fairseq/criterions/label_smoothed_cross_entropy_latency_augmented.py |
|
fairseq/criterions/label_smoothed_cross_entropy_with_alignment.py |
|
fairseq/criterions/label_smoothed_cross_entropy_with_ctc.py |
|
fairseq/criterions/label_smoothed_cross_entropy_with_rdrop.py |
|
fairseq/criterions/legacy_masked_lm.py |
|
fairseq/criterions/masked_lm.py |
|
fairseq/criterions/model_criterion.py |
|
fairseq/criterions/nat_loss.py |
|
fairseq/criterions/sentence_prediction.py |
|
fairseq/criterions/sentence_prediction_adapters.py |
|
fairseq/criterions/sentence_ranking.py |
|
fairseq/criterions/speech_dlm_criterion.py |
|
fairseq/criterions/speech_to_speech_criterion.py |
|
fairseq/criterions/speech_ulm_criterion.py |
|
fairseq/criterions/tacotron2_loss.py |
|
fairseq/criterions/wav2vec_criterion.py |
|
fairseq/data/__init__.py |
|
fairseq/data/add_class_target_dataset.py |
|
fairseq/data/add_target_dataset.py |
|
fairseq/data/append_token_dataset.py |
|
fairseq/data/backtranslation_dataset.py |
|
fairseq/data/base_wrapper_dataset.py |
|
fairseq/data/bucket_pad_length_dataset.py |
|
fairseq/data/codedataset.py |
|
fairseq/data/colorize_dataset.py |
|
fairseq/data/concat_dataset.py |
|
fairseq/data/concat_sentences_dataset.py |
|
fairseq/data/data_utils.py |
|
fairseq/data/data_utils_fast.pyx |
|
fairseq/data/denoising_dataset.py |
|
fairseq/data/dictionary.py |
|
fairseq/data/fairseq_dataset.py |
|
fairseq/data/fasta_dataset.py |
|
fairseq/data/id_dataset.py |
|
fairseq/data/indexed_dataset.py |
|
fairseq/data/iterators.py |
|
fairseq/data/language_pair_dataset.py |
|
fairseq/data/list_dataset.py |
|
fairseq/data/lm_context_window_dataset.py |
|
fairseq/data/lru_cache_dataset.py |
|
fairseq/data/mask_tokens_dataset.py |
|
fairseq/data/monolingual_dataset.py |
|
fairseq/data/multi_corpus_dataset.py |
|
fairseq/data/multi_corpus_sampled_dataset.py |
|
fairseq/data/nested_dictionary_dataset.py |
|
fairseq/data/noising.py |
|
fairseq/data/num_samples_dataset.py |
|
fairseq/data/numel_dataset.py |
|
fairseq/data/offset_tokens_dataset.py |
|
fairseq/data/pad_dataset.py |
|
fairseq/data/padding_mask_dataset.py |
|
fairseq/data/plasma_utils.py |
|
fairseq/data/prepend_dataset.py |
|
fairseq/data/prepend_token_dataset.py |
|
fairseq/data/raw_label_dataset.py |
|
fairseq/data/replace_dataset.py |
|
fairseq/data/resampling_dataset.py |
|
fairseq/data/roll_dataset.py |
|
fairseq/data/round_robin_zip_datasets.py |
|
fairseq/data/shorten_dataset.py |
|
fairseq/data/sort_dataset.py |
|
fairseq/data/span_mask_tokens_dataset.py |
|
fairseq/data/speech_dlm_dataset.py |
|
fairseq/data/strip_token_dataset.py |
|
fairseq/data/subsample_dataset.py |
|
fairseq/data/text_compressor.py |
|
fairseq/data/token_block_dataset.py |
|
fairseq/data/token_block_utils_fast.pyx |
|
fairseq/data/transform_eos_concat_langpair_dataset.py |
|
fairseq/data/transform_eos_dataset.py |
|
fairseq/data/transform_eos_lang_pair_dataset.py |
|
fairseq/data/audio/__init__.py |
|
fairseq/data/audio/audio_utils.py |
|
fairseq/data/audio/data_cfg.py |
|
fairseq/data/audio/frm_text_to_speech_dataset.py |
|
fairseq/data/audio/hubert_dataset.py |
|
fairseq/data/audio/multi_modality_dataset.py |
|
fairseq/data/audio/raw_audio_dataset.py |
|
fairseq/data/audio/speech_to_speech_dataset.py |
|
fairseq/data/audio/speech_to_text_dataset.py |
|
fairseq/data/audio/speech_to_text_joint_dataset.py |
|
fairseq/data/audio/text_to_speech_dataset.py |
|
fairseq/data/audio/dataset_transforms/__init__.py |
|
fairseq/data/audio/dataset_transforms/concataugment.py |
|
fairseq/data/audio/dataset_transforms/noisyoverlapaugment.py |
|
fairseq/data/audio/feature_transforms/__init__.py |
|
fairseq/data/audio/feature_transforms/delta_deltas.py |
|
fairseq/data/audio/feature_transforms/global_cmvn.py |
|
fairseq/data/audio/feature_transforms/specaugment.py |
|
fairseq/data/audio/feature_transforms/utterance_cmvn.py |
|
fairseq/data/audio/waveform_transforms/__init__.py |
|
fairseq/data/audio/waveform_transforms/noiseaugment.py |
|
fairseq/data/encoders/__init__.py |
|
fairseq/data/encoders/byte_bpe.py |
|
fairseq/data/encoders/byte_utils.py |
|
fairseq/data/encoders/bytes.py |
|
fairseq/data/encoders/characters.py |
|
fairseq/data/encoders/fastbpe.py |
|
fairseq/data/encoders/gpt2_bpe.py |
|
fairseq/data/encoders/gpt2_bpe_utils.py |
|
fairseq/data/encoders/hf_bert_bpe.py |
|
fairseq/data/encoders/hf_byte_bpe.py |
|
fairseq/data/encoders/moses_tokenizer.py |
|
fairseq/data/encoders/nltk_tokenizer.py |
|
fairseq/data/encoders/sentencepiece_bpe.py |
|
fairseq/data/encoders/space_tokenizer.py |
|
fairseq/data/encoders/subword_nmt_bpe.py |
|
fairseq/data/encoders/utils.py |
|
fairseq/data/huffman/__init__.py |
|
fairseq/data/huffman/huffman_coder.py |
|
fairseq/data/huffman/huffman_mmap_indexed_dataset.py |
|
fairseq/data/legacy/__init__.py |
|
fairseq/data/legacy/block_pair_dataset.py |
|
fairseq/data/legacy/masked_lm_dataset.py |
|
fairseq/data/legacy/masked_lm_dictionary.py |
|
fairseq/data/multilingual/__init__.py |
|
fairseq/data/multilingual/multilingual_data_manager.py |
|
fairseq/data/multilingual/multilingual_utils.py |
|
fairseq/data/multilingual/sampled_multi_dataset.py |
|
fairseq/data/multilingual/sampled_multi_epoch_dataset.py |
|
fairseq/data/multilingual/sampling_method.py |
|
fairseq/dataclass/__init__.py |
|
fairseq/dataclass/configs.py |
|
fairseq/dataclass/constants.py |
|
fairseq/dataclass/initialize.py |
|
fairseq/dataclass/utils.py |
|
fairseq/distributed/__init__.py |
|
fairseq/distributed/distributed_timeout_wrapper.py |
|
fairseq/distributed/fully_sharded_data_parallel.py |
|
fairseq/distributed/legacy_distributed_data_parallel.py |
|
fairseq/distributed/module_proxy_wrapper.py |
|
fairseq/distributed/tpu_distributed_data_parallel.py |
|
fairseq/distributed/utils.py |
|
fairseq/examples/.gitignore |
|
fairseq/examples/__init__.py |
|
fairseq/examples/MMPT/.gitignore |
|
fairseq/examples/MMPT/CONFIG.md |
|
fairseq/examples/MMPT/DATASET.md |
|
fairseq/examples/MMPT/README.md |
|
fairseq/examples/MMPT/endtask.md |
|
fairseq/examples/MMPT/locallaunch.py |
|
fairseq/examples/MMPT/pretraining.md |
|
fairseq/examples/MMPT/setup.py |
|
fairseq/examples/MMPT/videoclip.png |
|
fairseq/examples/MMPT/vlm.png |
|
fairseq/examples/MMPT/mmpt/__init__.py |
|
fairseq/examples/MMPT/mmpt/datasets/__init__.py |
|
fairseq/examples/MMPT/mmpt/datasets/fairseqmmdataset.py |
|
fairseq/examples/MMPT/mmpt/datasets/mmdataset.py |
|
fairseq/examples/MMPT/mmpt/evaluators/__init__.py |
|
fairseq/examples/MMPT/mmpt/evaluators/evaluator.py |
|
fairseq/examples/MMPT/mmpt/evaluators/metric.py |
|
fairseq/examples/MMPT/mmpt/evaluators/predictor.py |
|
fairseq/examples/MMPT/mmpt/losses/__init__.py |
|
fairseq/examples/MMPT/mmpt/losses/fairseqmmloss.py |
|
fairseq/examples/MMPT/mmpt/losses/loss.py |
|
fairseq/examples/MMPT/mmpt/losses/nce.py |
|
fairseq/examples/MMPT/mmpt/models/__init__.py |
|
fairseq/examples/MMPT/mmpt/models/fairseqmmmodel.py |
|
fairseq/examples/MMPT/mmpt/models/mmfusion.py |
|
fairseq/examples/MMPT/mmpt/models/mmfusionnlg.py |
|
fairseq/examples/MMPT/mmpt/models/transformermodel.py |
|
fairseq/examples/MMPT/mmpt/modules/__init__.py |
|
fairseq/examples/MMPT/mmpt/modules/mm.py |
|
fairseq/examples/MMPT/mmpt/modules/retri.py |
|
fairseq/examples/MMPT/mmpt/modules/vectorpool.py |
|
fairseq/examples/MMPT/mmpt/processors/__init__.py |
|
fairseq/examples/MMPT/mmpt/processors/dedupprocessor.py |
|
fairseq/examples/MMPT/mmpt/processors/dsprocessor.py |
|
fairseq/examples/MMPT/mmpt/processors/how2processor.py |
|
fairseq/examples/MMPT/mmpt/processors/how2retriprocessor.py |
|
fairseq/examples/MMPT/mmpt/processors/processor.py |
|
fairseq/examples/MMPT/mmpt/processors/models/s3dg.py |
|
fairseq/examples/MMPT/mmpt/tasks/__init__.py |
|
fairseq/examples/MMPT/mmpt/tasks/fairseqmmtask.py |
|
fairseq/examples/MMPT/mmpt/tasks/milncetask.py |
|
fairseq/examples/MMPT/mmpt/tasks/retritask.py |
|
fairseq/examples/MMPT/mmpt/tasks/task.py |
|
fairseq/examples/MMPT/mmpt/tasks/vlmtask.py |
|
fairseq/examples/MMPT/mmpt/utils/__init__.py |
|
fairseq/examples/MMPT/mmpt/utils/load_config.py |
|
fairseq/examples/MMPT/mmpt/utils/shardedtensor.py |
|
fairseq/examples/MMPT/mmpt_cli/localjob.py |
|
fairseq/examples/MMPT/mmpt_cli/predict.py |
|
fairseq/examples/MMPT/projects/mfmmlm.yaml |
|
fairseq/examples/MMPT/projects/mtm/mmfusionmtm.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/coin.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/crosstask.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/how2.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_coin.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_crosstask.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_crosstask_zs.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_vtt.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_vttqa.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_youcook.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/test_youcookcap.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/vtt.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/vttqa.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/youcook.yaml |
|
fairseq/examples/MMPT/projects/mtm/vlm/youcookcap.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoretri.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/coin_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/crosstask_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/how2.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_coin_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_coin_zs.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_crosstask_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_crosstask_zs_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_didemo_zs.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_vtt_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_vtt_zs.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_vttqa_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_vttqa_zs.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_youcook_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/test_youcook_zs.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/vtt_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/vttqa_videoclip.yaml |
|
fairseq/examples/MMPT/projects/retri/videoclip/youcook_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/coin.yaml |
|
fairseq/examples/MMPT/projects/task/coin_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/crosstask.yaml |
|
fairseq/examples/MMPT/projects/task/crosstask_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/default.yaml |
|
fairseq/examples/MMPT/projects/task/ft.yaml |
|
fairseq/examples/MMPT/projects/task/how2.yaml |
|
fairseq/examples/MMPT/projects/task/test.yaml |
|
fairseq/examples/MMPT/projects/task/test_coin.yaml |
|
fairseq/examples/MMPT/projects/task/test_coin_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_coin_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_crosstask.yaml |
|
fairseq/examples/MMPT/projects/task/test_crosstask_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_crosstask_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_crosstask_zs_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_didemo_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_vtt.yaml |
|
fairseq/examples/MMPT/projects/task/test_vtt_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_vtt_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_vttqa.yaml |
|
fairseq/examples/MMPT/projects/task/test_vttqa_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_vttqa_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_youcook.yaml |
|
fairseq/examples/MMPT/projects/task/test_youcook_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/test_youcook_zs.yaml |
|
fairseq/examples/MMPT/projects/task/test_youcookcap.yaml |
|
fairseq/examples/MMPT/projects/task/vtt.yaml |
|
fairseq/examples/MMPT/projects/task/vtt_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/vttqa.yaml |
|
fairseq/examples/MMPT/projects/task/vttqa_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/youcook.yaml |
|
fairseq/examples/MMPT/projects/task/youcook_videoclip.yaml |
|
fairseq/examples/MMPT/projects/task/youcookcap.yaml |
|
fairseq/examples/MMPT/scripts/text_token_extractor/pretokenization.py |
|
fairseq/examples/MMPT/scripts/text_token_extractor/configs/bert-base-uncased.yaml |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/extract.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/model.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/pathbuilder.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/preprocessing.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/random_sequence_shuffler.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/shard_feature.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/videoreader.py |
|
fairseq/examples/MMPT/scripts/video_feature_extractor/how2/s3d.sh |
|
fairseq/examples/adaptive_span/README.md |
|
fairseq/examples/adaptive_span/__init__.py |
|
fairseq/examples/adaptive_span/adagrad_with_grad_clip.py |
|
fairseq/examples/adaptive_span/adaptive_span_attention.py |
|
fairseq/examples/adaptive_span/adaptive_span_loss.py |
|
fairseq/examples/adaptive_span/adaptive_span_model.py |
|
fairseq/examples/adaptive_span/adaptive_span_model_wrapper.py |
|
fairseq/examples/adaptive_span/truncated_bptt_lm_task.py |
|
fairseq/examples/attention_head_selection/README.md |
|
fairseq/examples/attention_head_selection/src/__init__.py |
|
fairseq/examples/attention_head_selection/src/speech_to_text_head_selection.py |
|
fairseq/examples/attention_head_selection/src/data/__init__.py |
|
fairseq/examples/attention_head_selection/src/data/speech_to_text_dataset_with_domain.py |
|
fairseq/examples/attention_head_selection/src/loss/__init__.py |
|
fairseq/examples/attention_head_selection/src/loss/attention_head_selection.py |
|
fairseq/examples/attention_head_selection/src/models/__init__.py |
|
fairseq/examples/attention_head_selection/src/models/head_selection_s2t_transformer.py |
|
fairseq/examples/attention_head_selection/src/models/head_selection_transformer.py |
|
fairseq/examples/attention_head_selection/src/modules/__init__.py |
|
fairseq/examples/attention_head_selection/src/modules/attn_head_selector.py |
|
fairseq/examples/attention_head_selection/src/modules/head_selection_transformer_layer.py |
|
fairseq/examples/attention_head_selection/src/modules/multihead_attention_selection.py |
|
fairseq/examples/attention_head_selection/src/modules/multihead_functional.py |
|
fairseq/examples/audio_nlp/nlu/README.md |
|
fairseq/examples/audio_nlp/nlu/create_dict_stop.sh |
|
fairseq/examples/audio_nlp/nlu/generate_manifests.py |
|
fairseq/examples/audio_nlp/nlu/configs/nlu_finetuning.yaml |
|
fairseq/examples/backtranslation/README.md |
|
fairseq/examples/backtranslation/deduplicate_lines.py |
|
fairseq/examples/backtranslation/extract_bt_data.py |
|
fairseq/examples/backtranslation/prepare-de-monolingual.sh |
|
fairseq/examples/backtranslation/prepare-wmt18en2de.sh |
|
fairseq/examples/backtranslation/sacrebleu.sh |
|
fairseq/examples/backtranslation/tokenized_bleu.sh |
|
fairseq/examples/bart/README.glue.md |
|
fairseq/examples/bart/README.md |
|
fairseq/examples/bart/README.summarization.md |
|
fairseq/examples/bart/summarize.py |
|
fairseq/examples/byte_level_bpe/README.md |
|
fairseq/examples/byte_level_bpe/get_bitext.py |
|
fairseq/examples/byte_level_bpe/get_data.sh |
|
fairseq/examples/byte_level_bpe/gru_transformer.py |
|
fairseq/examples/camembert/README.md |
|
fairseq/examples/constrained_decoding/README.md |
|
fairseq/examples/constrained_decoding/normalize.py |
|
fairseq/examples/constrained_decoding/tok.py |
|
fairseq/examples/conv_seq2seq/README.md |
|
fairseq/examples/criss/README.md |
|
fairseq/examples/criss/download_and_preprocess_flores_test.sh |
|
fairseq/examples/criss/download_and_preprocess_tatoeba.sh |
|
fairseq/examples/criss/save_encoder.py |
|
fairseq/examples/criss/mining/mine.py |
|
fairseq/examples/criss/mining/mine_example.sh |
|
fairseq/examples/criss/sentence_retrieval/encoder_analysis.py |
|
fairseq/examples/criss/sentence_retrieval/sentence_retrieval_tatoeba.sh |
|
fairseq/examples/criss/unsupervised_mt/eval.sh |
|
fairseq/examples/cross_lingual_language_model/README.md |
|
fairseq/examples/data2vec/README.md |
|
fairseq/examples/data2vec/__init__.py |
|
fairseq/examples/data2vec/fb_convert_beit_cp.py |
|
fairseq/examples/data2vec/config/audio/classification/base_classification.yaml |
|
fairseq/examples/data2vec/config/audio/classification/run_config/slurm_1.yaml |
|
fairseq/examples/data2vec/config/audio/classification/run_config/slurm_1g.yaml |
|
fairseq/examples/data2vec/config/audio/classification/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/audioset.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/base_librispeech.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/local.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_1.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_1_aws.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_2_aws.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_3.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_4.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_4_aws.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_6_aws.yaml |
|
fairseq/examples/data2vec/config/audio/pretraining/run_config/slurm_8_aws.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/base.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/local.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_1_aws.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_2_aws.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_3.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_4.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_4_aws.yaml |
|
fairseq/examples/data2vec/config/text/pretraining/run_config/slurm_8_aws.yaml |
|
fairseq/examples/data2vec/config/v2/base_audio_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/base_images_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/base_text_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/huge_images14_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/huge_images_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/large_audio_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/large_images_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/large_text_only_task.yaml |
|
fairseq/examples/data2vec/config/v2/large_text_only_task_pgrp_1M.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/local.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_1.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_1_aws.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_2_aws.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_3.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_4.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_4_aws.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_6_aws.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_8.yaml |
|
fairseq/examples/data2vec/config/v2/run_config/slurm_8_aws.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/cola.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/mnli.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/mrpc.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/qnli.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/qqp.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/rte.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/sst_2.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/sts_b.yaml |
|
fairseq/examples/data2vec/config/v2/text_finetuning/run_config/local.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/imagenet.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_clean.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_huge_clean.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/mae_imagenet_large_clean.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/local.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_1.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_1_aws.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_2_aws.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_3.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_4.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_4_aws.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_6_aws.yaml |
|
fairseq/examples/data2vec/config/vision/finetuning/run_config/slurm_8_aws.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/base_imagenet.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/base_imagenet_d2v1.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/base_mae_imagenet.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/local.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_1.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_1_aws.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_2.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_2_aws.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_3.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_4.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_4_aws.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_6_aws.yaml |
|
fairseq/examples/data2vec/config/vision/pretraining/run_config/slurm_8_aws.yaml |
|
fairseq/examples/data2vec/data/__init__.py |
|
fairseq/examples/data2vec/data/add_class_target_dataset.py |
|
fairseq/examples/data2vec/data/image_dataset.py |
|
fairseq/examples/data2vec/data/mae_finetuning_image_dataset.py |
|
fairseq/examples/data2vec/data/mae_image_dataset.py |
|
fairseq/examples/data2vec/data/modality.py |
|
fairseq/examples/data2vec/data/path_dataset.py |
|
fairseq/examples/data2vec/models/__init__.py |
|
fairseq/examples/data2vec/models/audio_classification.py |
|
fairseq/examples/data2vec/models/data2vec2.py |
|
fairseq/examples/data2vec/models/data2vec_audio.py |
|
fairseq/examples/data2vec/models/data2vec_image_classification.py |
|
fairseq/examples/data2vec/models/data2vec_text.py |
|
fairseq/examples/data2vec/models/data2vec_text_classification.py |
|
fairseq/examples/data2vec/models/data2vec_vision.py |
|
fairseq/examples/data2vec/models/mae.py |
|
fairseq/examples/data2vec/models/mae_image_classification.py |
|
fairseq/examples/data2vec/models/utils.py |
|
fairseq/examples/data2vec/models/modalities/__init__.py |
|
fairseq/examples/data2vec/models/modalities/audio.py |
|
fairseq/examples/data2vec/models/modalities/base.py |
|
fairseq/examples/data2vec/models/modalities/images.py |
|
fairseq/examples/data2vec/models/modalities/modules.py |
|
fairseq/examples/data2vec/models/modalities/text.py |
|
fairseq/examples/data2vec/scripts/convert_audioset_labels.py |
|
fairseq/examples/data2vec/scripts/multi/finetune_all_fair_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/multi/finetune_all_fair_aws_local_lr_nodep.sh |
|
fairseq/examples/data2vec/scripts/multi/finetune_all_fair_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_char_fair_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_aws_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_fair_nodep_aws_lr_nopos.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_all_large_fair_nodep_aws_local_lr.sh |
|
fairseq/examples/data2vec/scripts/text/finetune_sst2_qnli_sweep_fair_nodep.sh |
|
fairseq/examples/data2vec/scripts/text/glue.py |
|
fairseq/examples/data2vec/scripts/text/glue_lr.py |
|
fairseq/examples/data2vec/scripts/text/unprocess_data.py |
|
fairseq/examples/data2vec/scripts/text/valids.py |
|
fairseq/examples/data2vec/tasks/__init__.py |
|
fairseq/examples/data2vec/tasks/audio_classification.py |
|
fairseq/examples/data2vec/tasks/image_classification.py |
|
fairseq/examples/data2vec/tasks/image_pretraining.py |
|
fairseq/examples/data2vec/tasks/mae_image_classification.py |
|
fairseq/examples/data2vec/tasks/mae_image_pretraining.py |
|
fairseq/examples/data2vec/tasks/multimodal.py |
|
fairseq/examples/discriminative_reranking_nmt/README.md |
|
fairseq/examples/discriminative_reranking_nmt/__init__.py |
|
fairseq/examples/discriminative_reranking_nmt/drnmt_rerank.py |
|
fairseq/examples/discriminative_reranking_nmt/config/deen.yaml |
|
fairseq/examples/discriminative_reranking_nmt/criterions/__init__.py |
|
fairseq/examples/discriminative_reranking_nmt/criterions/discriminative_reranking_criterion.py |
|
fairseq/examples/discriminative_reranking_nmt/models/__init__.py |
|
fairseq/examples/discriminative_reranking_nmt/models/discriminative_reranking_model.py |
|
fairseq/examples/discriminative_reranking_nmt/scripts/prep_data.py |
|
fairseq/examples/discriminative_reranking_nmt/tasks/__init__.py |
|
fairseq/examples/discriminative_reranking_nmt/tasks/discriminative_reranking_task.py |
|
fairseq/examples/emotion_conversion/README.md |
|
fairseq/examples/emotion_conversion/requirements.txt |
|
fairseq/examples/emotion_conversion/synthesize.py |
|
fairseq/examples/emotion_conversion/emotion_models/__init__.py |
|
fairseq/examples/emotion_conversion/emotion_models/duration_predictor.py |
|
fairseq/examples/emotion_conversion/emotion_models/duration_predictor.yaml |
|
fairseq/examples/emotion_conversion/emotion_models/pitch_predictor.py |
|
fairseq/examples/emotion_conversion/emotion_models/pitch_predictor.yaml |
|
fairseq/examples/emotion_conversion/emotion_models/utils.py |
|
fairseq/examples/emotion_conversion/fairseq_models/__init__.py |
|
fairseq/examples/emotion_conversion/preprocess/__init__.py |
|
fairseq/examples/emotion_conversion/preprocess/build_hifigan_manifest.py |
|
fairseq/examples/emotion_conversion/preprocess/build_translation_manifests.py |
|
fairseq/examples/emotion_conversion/preprocess/create_core_manifest.py |
|
fairseq/examples/emotion_conversion/preprocess/extract_f0.py |
|
fairseq/examples/emotion_conversion/preprocess/process_km.py |
|
fairseq/examples/emotion_conversion/preprocess/split_emov_km_tsv_by_uttid.py |
|
fairseq/examples/emotion_conversion/preprocess/split_km.py |
|
fairseq/examples/emotion_conversion/preprocess/split_km_tsv.py |
|
fairseq/examples/fast_noisy_channel/README.md |
|
fairseq/examples/fast_noisy_channel/__init__.py |
|
fairseq/examples/fast_noisy_channel/noisy_channel_beam_search.py |
|
fairseq/examples/fast_noisy_channel/noisy_channel_sequence_generator.py |
|
fairseq/examples/fast_noisy_channel/noisy_channel_translation.py |
|
fairseq/examples/flores101/README.md |
|
fairseq/examples/flores101/flores_logo.png |
|
fairseq/examples/fully_sharded_data_parallel/README.md |
|
fairseq/examples/gottbert/README.md |
|
fairseq/examples/hubert/README.md |
|
fairseq/examples/hubert/measure_teacher_quality.py |
|
fairseq/examples/hubert/update_ckpt.py |
|
fairseq/examples/hubert/config/decode/infer_fsqlm.yaml |
|
fairseq/examples/hubert/config/decode/infer_kenlm.yaml |
|
fairseq/examples/hubert/config/decode/infer_viterbi.yaml |
|
fairseq/examples/hubert/config/decode/ax_sweep/ngram.yaml |
|
fairseq/examples/hubert/config/decode/ax_sweep/transformer.yaml |
|
fairseq/examples/hubert/config/decode/run/submitit_slurm.yaml |
|
fairseq/examples/hubert/config/decode/run/submitit_slurm_8gpu.yaml |
|
fairseq/examples/hubert/config/finetune/base_10h.yaml |
|
fairseq/examples/hubert/config/finetune/ckpt/it1.yaml |
|
fairseq/examples/hubert/config/finetune/lm/ls_4gram.yaml |
|
fairseq/examples/hubert/config/finetune/run/submitit_reg.yaml |
|
fairseq/examples/hubert/config/pretrain/hubert_base_librispeech.yaml |
|
fairseq/examples/hubert/config/pretrain/hubert_large_librivox.yaml |
|
fairseq/examples/hubert/config/pretrain/hubert_xlarge_librivox.yaml |
|
fairseq/examples/hubert/config/pretrain/data/iter1.yaml |
|
fairseq/examples/hubert/config/pretrain/data/iter2.yaml |
|
fairseq/examples/hubert/config/pretrain/run/submitit_reg.yaml |
|
fairseq/examples/hubert/simple_kmeans/README.md |
|
fairseq/examples/hubert/simple_kmeans/dump_hubert_feature.py |
|
fairseq/examples/hubert/simple_kmeans/dump_hubert_feature_s2t.py |
|
fairseq/examples/hubert/simple_kmeans/dump_km_label.py |
|
fairseq/examples/hubert/simple_kmeans/dump_mfcc_feature.py |
|
fairseq/examples/hubert/simple_kmeans/dump_w2v2_feature.py |
|
fairseq/examples/hubert/simple_kmeans/feature_utils.py |
|
fairseq/examples/hubert/simple_kmeans/learn_kmeans.py |
|
fairseq/examples/hubert/tests/6313-76958-0021.flac |
|
fairseq/examples/hubert/tests/sample.base.L9.km500.km |
|
fairseq/examples/hubert/tests/sample.base.L9.len |
|
fairseq/examples/hubert/tests/sample.base.L9.npy |
|
fairseq/examples/hubert/tests/sample.large.L20.len |
|
fairseq/examples/hubert/tests/sample.large.L20.npy |
|
fairseq/examples/hubert/tests/sample.large.hypo.word |
|
fairseq/examples/hubert/tests/sample.xlarge.L30.len |
|
fairseq/examples/hubert/tests/sample.xlarge.L30.npy |
|
fairseq/examples/hubert/tests/sample.xlarge.hypo.word |
|
fairseq/examples/hubert/tests/test_feature_and_unit.sh |
|
fairseq/examples/hubert/tests/test_finetuned_asr.sh |
|
fairseq/examples/joint_alignment_translation/README.md |
|
fairseq/examples/joint_alignment_translation/prepare-wmt18en2de_no_norm_no_escape_no_agressive.sh |
|
fairseq/examples/language_model/README.adaptive_inputs.md |
|
fairseq/examples/language_model/README.conv.md |
|
fairseq/examples/language_model/README.md |
|
fairseq/examples/language_model/prepare-wikitext-103.sh |
|
fairseq/examples/laser/README.md |
|
fairseq/examples/laser/laser_src/__init__.py |
|
fairseq/examples/laser/laser_src/laser_lstm.py |
|
fairseq/examples/laser/laser_src/laser_task.py |
|
fairseq/examples/laser/laser_src/laser_transformer.py |
|
fairseq/examples/laser/laser_src/multitask_data_utils.py |
|
fairseq/examples/latent_depth/README.md |
|
fairseq/examples/latent_depth/latent_depth_src/__init__.py |
|
fairseq/examples/latent_depth/latent_depth_src/multilingual_translation_latent_depth.py |
|
fairseq/examples/latent_depth/latent_depth_src/loss/__init__.py |
|
fairseq/examples/latent_depth/latent_depth_src/loss/latent_depth.py |
|
fairseq/examples/latent_depth/latent_depth_src/models/__init__.py |
|
fairseq/examples/latent_depth/latent_depth_src/models/latent_multilingual_transformer.py |
|
fairseq/examples/latent_depth/latent_depth_src/models/latent_transformer.py |
|
fairseq/examples/latent_depth/latent_depth_src/modules/__init__.py |
|
fairseq/examples/latent_depth/latent_depth_src/modules/latent_layers.py |
|
fairseq/examples/layerdrop/README.md |
|
fairseq/examples/linformer/README.md |
|
fairseq/examples/linformer/linformer_src/__init__.py |
|
fairseq/examples/linformer/linformer_src/models/__init__.py |
|
fairseq/examples/linformer/linformer_src/models/linformer_roberta.py |
|
fairseq/examples/linformer/linformer_src/modules/__init__.py |
|
fairseq/examples/linformer/linformer_src/modules/linformer_sentence_encoder.py |
|
fairseq/examples/linformer/linformer_src/modules/linformer_sentence_encoder_layer.py |
|
fairseq/examples/linformer/linformer_src/modules/multihead_linear_attention.py |
|
fairseq/examples/m2m_100/README.md |
|
fairseq/examples/m2m_100/install_dependecies.sh |
|
fairseq/examples/m2m_100/tok.sh |
|
fairseq/examples/m2m_100/process_data/clean_histogram.py |
|
fairseq/examples/m2m_100/process_data/dedup_data.py |
|
fairseq/examples/m2m_100/process_data/remove_too_much_punc.py |
|
fairseq/examples/m2m_100/tokenizers/README.md |
|
fairseq/examples/m2m_100/tokenizers/seg_ja.sh |
|
fairseq/examples/m2m_100/tokenizers/seg_ko.sh |
|
fairseq/examples/m2m_100/tokenizers/tokenize_indic.py |
|
fairseq/examples/m2m_100/tokenizers/tokenize_thai.py |
|
fairseq/examples/m2m_100/tokenizers/tokenize_zh.py |
|
fairseq/examples/m2m_100/tokenizers/tokenizer_ar.sh |
|
fairseq/examples/m2m_100/tokenizers/thirdparty/.gitignore |
|
fairseq/examples/mbart/README.md |
|
fairseq/examples/megatron_11b/README.md |
|
fairseq/examples/megatron_11b/detok.py |
|
fairseq/examples/mms/MODEL_CARD.md |
|
fairseq/examples/mms/README.md |
|
fairseq/examples/mms/asr/config/infer_common.yaml |
|
fairseq/examples/mms/asr/infer/example_infer_adapter.sh |
|
fairseq/examples/mms/asr/infer/mms_infer.py |
|
fairseq/examples/mms/asr/tutorial/MMS_ASR_Inference_Colab.ipynb |
|
fairseq/examples/mms/data_prep/README.md |
|
fairseq/examples/mms/data_prep/align_and_segment.py |
|
fairseq/examples/mms/data_prep/align_utils.py |
|
fairseq/examples/mms/data_prep/norm_config.py |
|
fairseq/examples/mms/data_prep/punctuations.lst |
|
fairseq/examples/mms/data_prep/text_normalization.py |
|
fairseq/examples/mms/lid/infer.py |
|
fairseq/examples/mms/lid/tutorial/MMS_LID_Inference_Colab.ipynb |
|
fairseq/examples/mms/lid_rerank/README.md |
|
fairseq/examples/mms/lid_rerank/cer_langs.txt |
|
fairseq/examples/mms/lid_rerank/requirements.txt |
|
fairseq/examples/mms/lid_rerank/mala/infer.py |
|
fairseq/examples/mms/lid_rerank/mms/make_parallel_single_runs.py |
|
fairseq/examples/mms/lid_rerank/mms/merge_by_lang.py |
|
fairseq/examples/mms/lid_rerank/mms/prep_wav_list.py |
|
fairseq/examples/mms/lid_rerank/mms/run_single_lang.py |
|
fairseq/examples/mms/lid_rerank/mms/split_by_lang.py |
|
fairseq/examples/mms/lid_rerank/mms-zs/falign.py |
|
fairseq/examples/mms/lid_rerank/mms-zs/lib.py |
|
fairseq/examples/mms/lid_rerank/mms-zs/uromanize.py |
|
fairseq/examples/mms/lid_rerank/nllb/infer.py |
|
fairseq/examples/mms/lid_rerank/rerank/rerank.py |
|
fairseq/examples/mms/lid_rerank/rerank/tune_coefficients.py |
|
fairseq/examples/mms/lid_rerank/whisper/infer_asr.py |
|
fairseq/examples/mms/lid_rerank/whisper/infer_lid.py |
|
fairseq/examples/mms/lid_rerank/whisper/lid_mapping.txt |
|
fairseq/examples/mms/misc/get_sample_size.py |
|
fairseq/examples/mms/tts/infer.py |
|
fairseq/examples/mms/tts/tutorial/MMS_TTS_Inference_Colab.ipynb |
|
fairseq/examples/mms/zero_shot/README.md |
|
fairseq/examples/moe_lm/README.md |
|
fairseq/examples/moe_lm/data_card.md |
|
fairseq/examples/moe_lm/model_card.md |
|
fairseq/examples/mr_hubert/README.md |
|
fairseq/examples/mr_hubert/decode.sh |
|
fairseq/examples/mr_hubert/finetune.sh |
|
fairseq/examples/mr_hubert/train.sh |
|
fairseq/examples/mr_hubert/config/decode/infer.yaml |
|
fairseq/examples/mr_hubert/config/decode/infer_lm.yaml |
|
fairseq/examples/mr_hubert/config/decode/run/submitit_slurm.yaml |
|
fairseq/examples/mr_hubert/config/decode/run/submitit_slurm_8gpu.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_100h.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_100h_large.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_10h.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_10h_large.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_1h.yaml |
|
fairseq/examples/mr_hubert/config/finetune/base_1h_large.yaml |
|
fairseq/examples/mr_hubert/config/pretrain/mrhubert_base_librispeech.yaml |
|
fairseq/examples/mr_hubert/config/pretrain/mrhubert_large_librilight.yaml |
|
fairseq/examples/mr_hubert/config/pretrain/run/submitit_reg.yaml |
|
fairseq/examples/mr_hubert/simple_kmeans/README.md |
|
fairseq/examples/mr_hubert/simple_kmeans/dump_hubert_feature.py |
|
fairseq/examples/mr_hubert/simple_kmeans/dump_hubert_feature_s2t.py |
|
fairseq/examples/mr_hubert/simple_kmeans/dump_km_label.py |
|
fairseq/examples/mr_hubert/simple_kmeans/dump_mfcc_feature.py |
|
fairseq/examples/mr_hubert/simple_kmeans/dump_w2v2_feature.py |
|
fairseq/examples/mr_hubert/simple_kmeans/feature_utils.py |
|
fairseq/examples/mr_hubert/simple_kmeans/learn_kmeans.py |
|
fairseq/examples/multilingual/ML50_langs.txt |
|
fairseq/examples/multilingual/README.md |
|
fairseq/examples/multilingual/finetune_multilingual_model.sh |
|
fairseq/examples/multilingual/multilingual_fairseq_gen.sh |
|
fairseq/examples/multilingual/train_multilingual_model.sh |
|
fairseq/examples/multilingual/data_scripts/README.md |
|
fairseq/examples/multilingual/data_scripts/binarize.py |
|
fairseq/examples/multilingual/data_scripts/check_iswlt_test_data.py |
|
fairseq/examples/multilingual/data_scripts/check_self_overlaps.py |
|
fairseq/examples/multilingual/data_scripts/check_valid_test_overlaps.py |
|
fairseq/examples/multilingual/data_scripts/dedup_all.py |
|
fairseq/examples/multilingual/data_scripts/download_ML50_v1.sh |
|
fairseq/examples/multilingual/data_scripts/download_af_xh.sh |
|
fairseq/examples/multilingual/data_scripts/download_flores_data.sh |
|
fairseq/examples/multilingual/data_scripts/download_iitb.sh |
|
fairseq/examples/multilingual/data_scripts/download_iwslt_and_extract.sh |
|
fairseq/examples/multilingual/data_scripts/download_lotus.sh |
|
fairseq/examples/multilingual/data_scripts/download_ted_and_extract.py |
|
fairseq/examples/multilingual/data_scripts/download_wat19_my.sh |
|
fairseq/examples/multilingual/data_scripts/download_wmt19_and_before.py |
|
fairseq/examples/multilingual/data_scripts/download_wmt20.sh |
|
fairseq/examples/multilingual/data_scripts/preprocess_ML50_v1.sh |
|
fairseq/examples/multilingual/data_scripts/remove_valid_test_in_train.py |
|
fairseq/examples/multilingual/data_scripts/requirement.txt |
|
fairseq/examples/multilingual/data_scripts/utils/dedup.py |
|
fairseq/examples/multilingual/data_scripts/utils/fasttext_multi_filter.py |
|
fairseq/examples/multilingual/data_scripts/utils/strip_sgm.sh |
|
fairseq/examples/noisychannel/README.md |
|
fairseq/examples/noisychannel/__init__.py |
|
fairseq/examples/noisychannel/rerank.py |
|
fairseq/examples/noisychannel/rerank_generate.py |
|
fairseq/examples/noisychannel/rerank_options.py |
|
fairseq/examples/noisychannel/rerank_score_bw.py |
|
fairseq/examples/noisychannel/rerank_score_lm.py |
|
fairseq/examples/noisychannel/rerank_tune.py |
|
fairseq/examples/noisychannel/rerank_utils.py |
|
fairseq/examples/nonautoregressive_translation/README.md |
|
fairseq/examples/nonautoregressive_translation/scripts.md |
|
fairseq/examples/normformer/README.md |
|
fairseq/examples/normformer/train_lm.sh |
|
fairseq/examples/operators/alignment_train_cpu.cpp |
|
fairseq/examples/operators/alignment_train_cuda.cpp |
|
fairseq/examples/operators/alignment_train_cuda.h |
|
fairseq/examples/operators/alignment_train_kernel.cu |
|
fairseq/examples/operators/utils.h |
|
fairseq/examples/paraphraser/README.md |
|
fairseq/examples/paraphraser/paraphrase.py |
|
fairseq/examples/pay_less_attention_paper/README.md |
|
fairseq/examples/pointer_generator/README.md |
|
fairseq/examples/pointer_generator/README.xsum.md |
|
fairseq/examples/pointer_generator/postprocess.py |
|
fairseq/examples/pointer_generator/preprocess.py |
|
fairseq/examples/pointer_generator/pointer_generator_src/__init__.py |
|
fairseq/examples/pointer_generator/pointer_generator_src/transformer_pg.py |
|
fairseq/examples/quant_noise/README.md |
|
fairseq/examples/quant_noise/transformer_quantization_config.yaml |
|
fairseq/examples/roberta/README.custom_classification.md |
|
fairseq/examples/roberta/README.glue.md |
|
fairseq/examples/roberta/README.md |
|
fairseq/examples/roberta/README.pretraining.md |
|
fairseq/examples/roberta/README.race.md |
|
fairseq/examples/roberta/multiprocessing_bpe_encoder.py |
|
fairseq/examples/roberta/preprocess_GLUE_tasks.sh |
|
fairseq/examples/roberta/preprocess_RACE.py |
|
fairseq/examples/roberta/preprocess_RACE.sh |
|
fairseq/examples/roberta/commonsense_qa/README.md |
|
fairseq/examples/roberta/commonsense_qa/__init__.py |
|
fairseq/examples/roberta/commonsense_qa/commonsense_qa_task.py |
|
fairseq/examples/roberta/commonsense_qa/download_cqa_data.sh |
|
fairseq/examples/roberta/config/finetuning/cola.yaml |
|
fairseq/examples/roberta/config/finetuning/mnli.yaml |
|
fairseq/examples/roberta/config/finetuning/mrpc.yaml |
|
fairseq/examples/roberta/config/finetuning/qnli.yaml |
|
fairseq/examples/roberta/config/finetuning/qqp.yaml |
|
fairseq/examples/roberta/config/finetuning/rte.yaml |
|
fairseq/examples/roberta/config/finetuning/sst_2.yaml |
|
fairseq/examples/roberta/config/finetuning/sts_b.yaml |
|
fairseq/examples/roberta/config/finetuning/run_config/local.yaml |
|
fairseq/examples/roberta/config/finetuning/run_config/slurm_1g.yaml |
|
fairseq/examples/roberta/config/finetuning/run_config/slurm_1g_aws.yaml |
|
fairseq/examples/roberta/config/pretraining/base.yaml |
|
fairseq/examples/roberta/config/pretraining/run_config/local.yaml |
|
fairseq/examples/roberta/config/pretraining/run_config/slurm_2.yaml |
|
fairseq/examples/roberta/config/pretraining/run_config/slurm_2_aws.yaml |
|
fairseq/examples/roberta/config/pretraining/run_config/slurm_3.yaml |
|
fairseq/examples/roberta/config/pretraining/run_config/slurm_4.yaml |
|
fairseq/examples/roberta/fb_multilingual/README.multilingual.pretraining.md |
|
fairseq/examples/roberta/wsc/README.md |
|
fairseq/examples/roberta/wsc/__init__.py |
|
fairseq/examples/roberta/wsc/wsc_criterion.py |
|
fairseq/examples/roberta/wsc/wsc_task.py |
|
fairseq/examples/roberta/wsc/wsc_utils.py |
|
fairseq/examples/rxf/README.md |
|
fairseq/examples/rxf/__init__.py |
|
fairseq/examples/rxf/rxf_src/__init__.py |
|
fairseq/examples/rxf/rxf_src/label_smoothed_cross_entropy_r3f.py |
|
fairseq/examples/rxf/rxf_src/sentence_prediction_r3f.py |
|
fairseq/examples/scaling_nmt/README.md |
|
fairseq/examples/shuffled_word_order/README.finetuning.md |
|
fairseq/examples/shuffled_word_order/README.md |
|
fairseq/examples/simultaneous_translation/README.md |
|
fairseq/examples/simultaneous_translation/__init__.py |
|
fairseq/examples/simultaneous_translation/docs/ende-mma.md |
|
fairseq/examples/simultaneous_translation/docs/enja-waitk.md |
|
fairseq/examples/simultaneous_translation/eval/agents/simul_t2t_enja.py |
|
fairseq/examples/simultaneous_translation/models/__init__.py |
|
fairseq/examples/simultaneous_translation/models/convtransformer_simul_trans.py |
|
fairseq/examples/simultaneous_translation/models/transformer_monotonic_attention.py |
|
fairseq/examples/simultaneous_translation/modules/__init__.py |
|
fairseq/examples/simultaneous_translation/modules/fixed_pre_decision.py |
|
fairseq/examples/simultaneous_translation/modules/monotonic_multihead_attention.py |
|
fairseq/examples/simultaneous_translation/modules/monotonic_transformer_layer.py |
|
fairseq/examples/simultaneous_translation/tests/test_alignment_train.py |
|
fairseq/examples/simultaneous_translation/tests/test_text_models.py |
|
fairseq/examples/simultaneous_translation/utils/__init__.py |
|
fairseq/examples/simultaneous_translation/utils/functions.py |
|
fairseq/examples/simultaneous_translation/utils/monotonic_attention.py |
|
fairseq/examples/simultaneous_translation/utils/p_choose_strategy.py |
|
fairseq/examples/speech_recognition/README.md |
|
fairseq/examples/speech_recognition/__init__.py |
|
fairseq/examples/speech_recognition/infer.py |
|
fairseq/examples/speech_recognition/w2l_decoder.py |
|
fairseq/examples/speech_recognition/criterions/ASG_loss.py |
|
fairseq/examples/speech_recognition/criterions/__init__.py |
|
fairseq/examples/speech_recognition/criterions/cross_entropy_acc.py |
|
fairseq/examples/speech_recognition/data/__init__.py |
|
fairseq/examples/speech_recognition/data/asr_dataset.py |
|
fairseq/examples/speech_recognition/data/collaters.py |
|
fairseq/examples/speech_recognition/data/data_utils.py |
|
fairseq/examples/speech_recognition/data/replabels.py |
|
fairseq/examples/speech_recognition/datasets/asr_prep_json.py |
|
fairseq/examples/speech_recognition/datasets/prepare-librispeech.sh |
|
fairseq/examples/speech_recognition/kaldi/__init__.py |
|
fairseq/examples/speech_recognition/kaldi/add-self-loop-simple.cc |
|
fairseq/examples/speech_recognition/kaldi/kaldi_decoder.py |
|
fairseq/examples/speech_recognition/kaldi/kaldi_initializer.py |
|
fairseq/examples/speech_recognition/kaldi/config/kaldi_initializer.yaml |
|
fairseq/examples/speech_recognition/models/__init__.py |
|
fairseq/examples/speech_recognition/models/vggtransformer.py |
|
fairseq/examples/speech_recognition/models/w2l_conv_glu_enc.py |
|
fairseq/examples/speech_recognition/new/README.md |
|
fairseq/examples/speech_recognition/new/__init__.py |
|
fairseq/examples/speech_recognition/new/infer.py |
|
fairseq/examples/speech_recognition/new/conf/infer.yaml |
|
fairseq/examples/speech_recognition/new/conf/hydra/sweeper/ax.yaml |
|
fairseq/examples/speech_recognition/new/conf/hydra/sweeper/ax_sil.yaml |
|
fairseq/examples/speech_recognition/new/conf/run_config/fb_slurm_1.yaml |
|
fairseq/examples/speech_recognition/new/conf/run_config/fb_slurm_2g.yaml |
|
fairseq/examples/speech_recognition/new/decoders/__init__.py |
|
fairseq/examples/speech_recognition/new/decoders/base_decoder.py |
|
fairseq/examples/speech_recognition/new/decoders/decoder.py |
|
fairseq/examples/speech_recognition/new/decoders/decoder_config.py |
|
fairseq/examples/speech_recognition/new/decoders/flashlight_decoder.py |
|
fairseq/examples/speech_recognition/new/decoders/viterbi_decoder.py |
|
fairseq/examples/speech_recognition/tasks/__init__.py |
|
fairseq/examples/speech_recognition/tasks/speech_recognition.py |
|
fairseq/examples/speech_recognition/utils/wer_utils.py |
|
fairseq/examples/speech_synthesis/README.md |
|
fairseq/examples/speech_synthesis/__init__.py |
|
fairseq/examples/speech_synthesis/data_utils.py |
|
fairseq/examples/speech_synthesis/generate_waveform.py |
|
fairseq/examples/speech_synthesis/utils.py |
|
fairseq/examples/speech_synthesis/docs/common_voice_example.md |
|
fairseq/examples/speech_synthesis/docs/ljspeech_example.md |
|
fairseq/examples/speech_synthesis/docs/vctk_example.md |
|
fairseq/examples/speech_synthesis/evaluation/__init__.py |
|
fairseq/examples/speech_synthesis/evaluation/eval_asr.py |
|
fairseq/examples/speech_synthesis/evaluation/eval_f0.py |
|
fairseq/examples/speech_synthesis/evaluation/eval_sp.py |
|
fairseq/examples/speech_synthesis/evaluation/get_eval_manifest.py |
|
fairseq/examples/speech_synthesis/preprocessing/__init__.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoise_and_vad_audio.py |
|
fairseq/examples/speech_synthesis/preprocessing/get_common_voice_audio_manifest.py |
|
fairseq/examples/speech_synthesis/preprocessing/get_feature_manifest.py |
|
fairseq/examples/speech_synthesis/preprocessing/get_ljspeech_audio_manifest.py |
|
fairseq/examples/speech_synthesis/preprocessing/get_speaker_embedding.py |
|
fairseq/examples/speech_synthesis/preprocessing/get_vctk_audio_manifest.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoiser/__init__.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoiser/demucs.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoiser/pretrained.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoiser/resample.py |
|
fairseq/examples/speech_synthesis/preprocessing/denoiser/utils.py |
|
fairseq/examples/speech_synthesis/preprocessing/speaker_embedder/__init__.py |
|
fairseq/examples/speech_synthesis/preprocessing/vad/__init__.py |
|
fairseq/examples/speech_text_joint_to_text/README.md |
|
fairseq/examples/speech_text_joint_to_text/__init__.py |
|
fairseq/examples/speech_text_joint_to_text/configs/mustc_noise.list |
|
fairseq/examples/speech_text_joint_to_text/criterions/__init__.py |
|
fairseq/examples/speech_text_joint_to_text/criterions/multi_modality_compound.py |
|
fairseq/examples/speech_text_joint_to_text/criterions/multi_modality_cross_entropy.py |
|
fairseq/examples/speech_text_joint_to_text/criterions/text_guide_cross_entropy_acc.py |
|
fairseq/examples/speech_text_joint_to_text/data/pair_denoising_dataset.py |
|
fairseq/examples/speech_text_joint_to_text/docs/ende-mustc.md |
|
fairseq/examples/speech_text_joint_to_text/docs/iwslt2021.md |
|
fairseq/examples/speech_text_joint_to_text/docs/pre-training.md |
|
fairseq/examples/speech_text_joint_to_text/models/__init__.py |
|
fairseq/examples/speech_text_joint_to_text/models/joint_speech_text_pretrain_transformer.py |
|
fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputtransformer.py |
|
fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputwavtransformer.py |
|
fairseq/examples/speech_text_joint_to_text/models/s2t_dualinputxmtransformer.py |
|
fairseq/examples/speech_text_joint_to_text/scripts/convert_model.py |
|
fairseq/examples/speech_text_joint_to_text/scripts/g2p_encode.py |
|
fairseq/examples/speech_text_joint_to_text/tasks/__init__.py |
|
fairseq/examples/speech_text_joint_to_text/tasks/pair_denoising.py |
|
fairseq/examples/speech_text_joint_to_text/tasks/speech_text_denoise_pretrain.py |
|
fairseq/examples/speech_text_joint_to_text/tasks/speech_text_joint.py |
|
fairseq/examples/speech_to_speech/README.md |
|
fairseq/examples/speech_to_speech/__init__.py |
|
fairseq/examples/speech_to_speech/generate_waveform_from_code.py |
|
fairseq/examples/speech_to_speech/asr_bleu/README.md |
|
fairseq/examples/speech_to_speech/asr_bleu/__init__.py |
|
fairseq/examples/speech_to_speech/asr_bleu/asr_model_cfgs.json |
|
fairseq/examples/speech_to_speech/asr_bleu/compute_asr_bleu.py |
|
fairseq/examples/speech_to_speech/asr_bleu/requirements.txt |
|
fairseq/examples/speech_to_speech/asr_bleu/utils.py |
|
fairseq/examples/speech_to_speech/benchmarking/README.md |
|
fairseq/examples/speech_to_speech/benchmarking/core.py |
|
fairseq/examples/speech_to_speech/benchmarking/data_utils.py |
|
fairseq/examples/speech_to_speech/benchmarking/get_metrics.py |
|
fairseq/examples/speech_to_speech/benchmarking/configs/2StageS2ST.yaml |
|
fairseq/examples/speech_to_speech/benchmarking/configs/3StageS2ST.yaml |
|
fairseq/examples/speech_to_speech/benchmarking/configs/DirectS2U.yaml |
|
fairseq/examples/speech_to_speech/benchmarking/configs/S2T.yaml |
|
fairseq/examples/speech_to_speech/docs/data_augmentation.md |
|
fairseq/examples/speech_to_speech/docs/direct_s2st_discrete_units.md |
|
fairseq/examples/speech_to_speech/docs/enhanced_direct_s2st_discrete_units.md |
|
fairseq/examples/speech_to_speech/docs/textless_s2st_real_data.md |
|
fairseq/examples/speech_to_speech/preprocessing/__init__.py |
|
fairseq/examples/speech_to_speech/preprocessing/data_utils.py |
|
fairseq/examples/speech_to_speech/preprocessing/prep_s2spect_data.py |
|
fairseq/examples/speech_to_speech/preprocessing/prep_s2ut_data.py |
|
fairseq/examples/speech_to_speech/preprocessing/prep_sn_data.py |
|
fairseq/examples/speech_to_speech/preprocessing/prep_sn_output_data.py |
|
fairseq/examples/speech_to_speech/unity/__init__.py |
|
fairseq/examples/speech_to_speech/unity/sequence_generator.py |
|
fairseq/examples/speech_to_speech/unity/sequence_generator_multi_decoder.py |
|
fairseq/examples/speech_to_text/README.md |
|
fairseq/examples/speech_to_text/data_utils.py |
|
fairseq/examples/speech_to_text/prep_covost_data.py |
|
fairseq/examples/speech_to_text/prep_librispeech_data.py |
|
fairseq/examples/speech_to_text/prep_mtedx_data.py |
|
fairseq/examples/speech_to_text/prep_mustc_data.py |
|
fairseq/examples/speech_to_text/seg_mustc_data.py |
|
fairseq/examples/speech_to_text/docs/covost_example.md |
|
fairseq/examples/speech_to_text/docs/librispeech_example.md |
|
fairseq/examples/speech_to_text/docs/mtedx_example.md |
|
fairseq/examples/speech_to_text/docs/mustc_example.md |
|
fairseq/examples/speech_to_text/docs/simulst_mustc_example.md |
|
fairseq/examples/speech_to_text/simultaneous_translation/agents/fairseq_simul_st_agent.py |
|
fairseq/examples/stories/README.md |
|
fairseq/examples/textless_nlp/dgslm/README.md |
|
fairseq/examples/textless_nlp/dgslm/create_code_file.py |
|
fairseq/examples/textless_nlp/dgslm/dgslm_utils.py |
|
fairseq/examples/textless_nlp/dgslm/sample_speech_dlm.py |
|
fairseq/examples/textless_nlp/dgslm/hubert_fisher/README.md |
|
fairseq/examples/textless_nlp/dgslm/vocoder_hifigan/README.md |
|
fairseq/examples/textless_nlp/dgslm/vocoder_hifigan/generate_stereo_waveform.py |
|
fairseq/examples/textless_nlp/gslm/README.md |
|
fairseq/examples/textless_nlp/gslm/metrics/README.md |
|
fairseq/examples/textless_nlp/gslm/metrics/abx_metrics/README.md |
|
fairseq/examples/textless_nlp/gslm/metrics/abx_metrics/dump_abx_feats.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/README.md |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/continuation_eval.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/ppx.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/self_auto_bleu.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/bleu_utils.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/cut_as.py |
|
fairseq/examples/textless_nlp/gslm/metrics/asr_metrics/misc/dict.ltr.txt |
|
fairseq/examples/textless_nlp/gslm/speech2unit/README.md |
|
fairseq/examples/textless_nlp/gslm/speech2unit/__init__.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/clustering/__init__.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/clustering/cluster_kmeans.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/clustering/dump_feats.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/clustering/quantize_with_kmeans.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/clustering/utils.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/cpc_feature_reader.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/hubert_feature_reader.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/logmel_feature_reader.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/utils.py |
|
fairseq/examples/textless_nlp/gslm/speech2unit/pretrained/w2v2_feature_reader.py |
|
fairseq/examples/textless_nlp/gslm/tools/README.md |
|
fairseq/examples/textless_nlp/gslm/tools/resynthesize_speech.py |
|
fairseq/examples/textless_nlp/gslm/ulm/README.md |
|
fairseq/examples/textless_nlp/gslm/ulm/sample.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/README.md |
|
fairseq/examples/textless_nlp/gslm/unit2speech/convert_to_16k.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/glow.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/multiproc.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/synthesize_audio_from_units.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tts_data.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/utils.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/__init__.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/audio_processing.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/cleaners.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/cmudict.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/layers.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/model.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/numbers.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/stft.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/symbols.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/text.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/utils.py |
|
fairseq/examples/textless_nlp/gslm/unit2speech/tacotron2/waveglow_denoiser.py |
|
fairseq/examples/textless_nlp/pgslm/README.md |
|
fairseq/examples/textless_nlp/pgslm/data_utils.py |
|
fairseq/examples/textless_nlp/pgslm/generate_waveform.py |
|
fairseq/examples/textless_nlp/pgslm/inference_dataset.py |
|
fairseq/examples/textless_nlp/pgslm/naive_decoder.py |
|
fairseq/examples/textless_nlp/pgslm/prepare_dataset.py |
|
fairseq/examples/textless_nlp/pgslm/preprocess_f0.py |
|
fairseq/examples/textless_nlp/pgslm/quantize_f0.py |
|
fairseq/examples/textless_nlp/pgslm/truncated_laplace.py |
|
fairseq/examples/textless_nlp/pgslm/eval/__init__.py |
|
fairseq/examples/textless_nlp/pgslm/eval/cont_metrics.py |
|
fairseq/examples/textless_nlp/pgslm/sample/__init__.py |
|
fairseq/examples/textless_nlp/pgslm/sample/sample.py |
|
fairseq/examples/textless_nlp/pgslm/scripts/join_units_manifest.py |
|
fairseq/examples/textless_nlp/pgslm/scripts/prepare_data.sh |
|
fairseq/examples/textless_nlp/pgslm/scripts/prepare_f0_quantization.sh |
|
fairseq/examples/textless_nlp/speech-resynth/README.md |
|
fairseq/examples/textless_nlp/speech-resynth/img/fig.png |
|
fairseq/examples/translation/README.md |
|
fairseq/examples/translation/prepare-iwslt14.sh |
|
fairseq/examples/translation/prepare-iwslt17-multilingual.sh |
|
fairseq/examples/translation/prepare-wmt14en2de.sh |
|
fairseq/examples/translation/prepare-wmt14en2fr.sh |
|
fairseq/examples/translation_moe/README.md |
|
fairseq/examples/translation_moe/score.py |
|
fairseq/examples/translation_moe/translation_moe_src/__init__.py |
|
fairseq/examples/translation_moe/translation_moe_src/logsumexp_moe.py |
|
fairseq/examples/translation_moe/translation_moe_src/mean_pool_gating_network.py |
|
fairseq/examples/translation_moe/translation_moe_src/translation_moe.py |
|
fairseq/examples/truncated_bptt/README.md |
|
fairseq/examples/truncated_bptt/__init__.py |
|
fairseq/examples/truncated_bptt/transformer_xl_model.py |
|
fairseq/examples/truncated_bptt/truncated_bptt_lm_task.py |
|
fairseq/examples/unsupervised_quality_estimation/README.md |
|
fairseq/examples/unsupervised_quality_estimation/aggregate_scores.py |
|
fairseq/examples/unsupervised_quality_estimation/meteor.py |
|
fairseq/examples/unsupervised_quality_estimation/repeat_lines.py |
|
fairseq/examples/wav2vec/README.md |
|
fairseq/examples/wav2vec/__init__.py |
|
fairseq/examples/wav2vec/libri_labels.py |
|
fairseq/examples/wav2vec/vq-wav2vec_featurize.py |
|
fairseq/examples/wav2vec/wav2vec_featurize.py |
|
fairseq/examples/wav2vec/wav2vec_manifest.py |
|
fairseq/examples/wav2vec/config/finetuning/base_100h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/base_10h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/base_10m.yaml |
|
fairseq/examples/wav2vec/config/finetuning/base_1h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/base_960h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_100h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_100h_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_100h_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_100h_3.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10h_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10h_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10h_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10h_aws_v100.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10m.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10m_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10m_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_10m_3.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h_3.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h_4.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_1h_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_960h.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_960h_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_960h_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/vox_960h_3.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_16.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_1_old.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_2g.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_3.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_4g.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_4g_aws.yaml |
|
fairseq/examples/wav2vec/config/finetuning/run_config/slurm_8.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_base_librispeech.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_conformer_base_librispeech.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_conformer_large_librivox.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu-pod.yaml |
|
fairseq/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu.yaml |
|
fairseq/examples/wav2vec/scripts/binarize_manifest.sh |
|
fairseq/examples/wav2vec/unsupervised/README.md |
|
fairseq/examples/wav2vec/unsupervised/__init__.py |
|
fairseq/examples/wav2vec/unsupervised/w2vu_generate.py |
|
fairseq/examples/wav2vec/unsupervised/config/finetuning/w2v_finetune.yaml |
|
fairseq/examples/wav2vec/unsupervised/config/gan/w2vu.yaml |
|
fairseq/examples/wav2vec/unsupervised/config/gan/w2vu2.yaml |
|
fairseq/examples/wav2vec/unsupervised/config/generate/viterbi.yaml |
|
fairseq/examples/wav2vec/unsupervised/config/timit_matched/test.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_matched/train.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_matched/train_text.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_matched/valid.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/test.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/train.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/train_text.uid |
|
fairseq/examples/wav2vec/unsupervised/config/timit_unmatched/valid.uid |
|
fairseq/examples/wav2vec/unsupervised/data/__init__.py |
|
fairseq/examples/wav2vec/unsupervised/data/extracted_features_dataset.py |
|
fairseq/examples/wav2vec/unsupervised/data/random_input_dataset.py |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/README.md |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/cmd.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_phone.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step1.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step2.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/path.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/train.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/copy_aligned_text.py |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/decode.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_data_from_w2v.py |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang_word.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lm.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/score.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/show_wer.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/train_subset_lgbeam.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select.py |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode_word.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_deltas.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_lda_mllt.sh |
|
fairseq/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_sat.sh |
|
fairseq/examples/wav2vec/unsupervised/models/__init__.py |
|
fairseq/examples/wav2vec/unsupervised/models/wav2vec_u.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/apply_pca.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/copy_labels.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/filter_lexicon.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/filter_tsv.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/g2p_wrd_to_phn.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/ltr_to_wrd.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/mean_pool.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/merge_clusters.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/normalize_and_filter_text.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/normalize_text.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/pca.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/phonemize_with_sil.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/prepare_audio.sh |
|
fairseq/examples/wav2vec/unsupervised/scripts/prepare_audio_v2.sh |
|
fairseq/examples/wav2vec/unsupervised/scripts/prepare_text.sh |
|
fairseq/examples/wav2vec/unsupervised/scripts/prepare_timit.sh |
|
fairseq/examples/wav2vec/unsupervised/scripts/remove_silence.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/vads.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_apply_cluster_faiss.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_cluster_faiss.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/wav2vec_extract_features.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/wer.py |
|
fairseq/examples/wav2vec/unsupervised/scripts/wrd_to_ltr.py |
|
fairseq/examples/wav2vec/unsupervised/tasks/__init__.py |
|
fairseq/examples/wav2vec/unsupervised/tasks/unpaired_audio_text.py |
|
fairseq/examples/wav2vec/xlsr/README.md |
|
fairseq/examples/wav2vec/xlsr/config/finetune.yaml |
|
fairseq/examples/wav2vec/xlsr/scripts/eval_speaker_clf_task.py |
|
fairseq/examples/wav2vec/xlsr/scripts/gen_audio_embedding.py |
|
fairseq/examples/wmt19/README.md |
|
fairseq/examples/wmt20/README.md |
|
fairseq/examples/wmt21/README.md |
|
fairseq/examples/wmt21/eval.sh |
|
fairseq/examples/wmt21/scripts/normalize-punctuation.perl |
|
fairseq/examples/wmt21/scripts/replace-unicode-punctuation.perl |
|
fairseq/examples/womens_bios/README.md |
|
fairseq/examples/womens_bios/query_occupations_from_wikidata.py |
|
fairseq/examples/xformers/README.md |
|
fairseq/examples/xglm/README.md |
|
fairseq/examples/xglm/XStoryCloze.md |
|
fairseq/examples/xglm/model_card.md |
|
fairseq/examples/xlmr/README.md |
|
fairseq/examples/xmod/README.md |
|
fairseq/examples/xmod/preprocess_nli.py |
|
fairseq/logging/__init__.py |
|
fairseq/logging/meters.py |
|
fairseq/logging/metrics.py |
|
fairseq/logging/progress_bar.py |
|
fairseq/model_parallel/__init__.py |
|
fairseq/model_parallel/megatron_trainer.py |
|
fairseq/model_parallel/criterions/__init__.py |
|
fairseq/model_parallel/criterions/vocab_parallel_cross_entropy.py |
|
fairseq/model_parallel/models/__init__.py |
|
fairseq/model_parallel/models/transformer.py |
|
fairseq/model_parallel/models/transformer_lm.py |
|
fairseq/model_parallel/models/pipeline_parallel_transformer/__init__.py |
|
fairseq/model_parallel/models/pipeline_parallel_transformer/layers.py |
|
fairseq/model_parallel/models/pipeline_parallel_transformer/model.py |
|
fairseq/model_parallel/models/roberta/__init__.py |
|
fairseq/model_parallel/models/roberta/model.py |
|
fairseq/model_parallel/modules/__init__.py |
|
fairseq/model_parallel/modules/multihead_attention.py |
|
fairseq/model_parallel/modules/transformer_layer.py |
|
fairseq/models/__init__.py |
|
fairseq/models/composite_encoder.py |
|
fairseq/models/distributed_fairseq_model.py |
|
fairseq/models/fairseq_decoder.py |
|
fairseq/models/fairseq_encoder.py |
|
fairseq/models/fairseq_incremental_decoder.py |
|
fairseq/models/fairseq_model.py |
|
fairseq/models/fconv.py |
|
fairseq/models/fconv_lm.py |
|
fairseq/models/fconv_self_att.py |
|
fairseq/models/lightconv.py |
|
fairseq/models/lightconv_lm.py |
|
fairseq/models/lstm.py |
|
fairseq/models/lstm_lm.py |
|
fairseq/models/masked_lm.py |
|
fairseq/models/model_utils.py |
|
fairseq/models/multilingual_transformer.py |
|
fairseq/models/transformer_align.py |
|
fairseq/models/transformer_from_pretrained_xlm.py |
|
fairseq/models/transformer_lm.py |
|
fairseq/models/transformer_ulm.py |
|
fairseq/models/bart/__init__.py |
|
fairseq/models/bart/hub_interface.py |
|
fairseq/models/bart/model.py |
|
fairseq/models/ema/__init__.py |
|
fairseq/models/ema/ema.py |
|
fairseq/models/hubert/__init__.py |
|
fairseq/models/hubert/hubert.py |
|
fairseq/models/hubert/hubert_asr.py |
|
fairseq/models/huggingface/__init__.py |
|
fairseq/models/huggingface/hf_gpt2.py |
|
fairseq/models/multires_hubert/__init__.py |
|
fairseq/models/multires_hubert/multires_hubert.py |
|
fairseq/models/multires_hubert/multires_hubert_asr.py |
|
fairseq/models/nat/__init__.py |
|
fairseq/models/nat/cmlm_transformer.py |
|
fairseq/models/nat/fairseq_nat_model.py |
|
fairseq/models/nat/insertion_transformer.py |
|
fairseq/models/nat/iterative_nonautoregressive_transformer.py |
|
fairseq/models/nat/levenshtein_transformer.py |
|
fairseq/models/nat/levenshtein_utils.py |
|
fairseq/models/nat/nat_crf_transformer.py |
|
fairseq/models/nat/nonautoregressive_ensembles.py |
|
fairseq/models/nat/nonautoregressive_transformer.py |
|
fairseq/models/roberta/__init__.py |
|
fairseq/models/roberta/alignment_utils.py |
|
fairseq/models/roberta/enc_dec.py |
|
fairseq/models/roberta/hub_interface.py |
|
fairseq/models/roberta/model.py |
|
fairseq/models/roberta/model_camembert.py |
|
fairseq/models/roberta/model_gottbert.py |
|
fairseq/models/roberta/model_xlmr.py |
|
fairseq/models/speech_dlm/__init__.py |
|
fairseq/models/speech_dlm/hub_interface.py |
|
fairseq/models/speech_dlm/speech_dlm.py |
|
fairseq/models/speech_dlm/modules/__init__.py |
|
fairseq/models/speech_dlm/modules/speech_dlm_decoder.py |
|
fairseq/models/speech_dlm/modules/speech_dlm_decoder_layer.py |
|
fairseq/models/speech_dlm/sequence_generator/__init__.py |
|
fairseq/models/speech_dlm/sequence_generator/multichannel_search.py |
|
fairseq/models/speech_dlm/sequence_generator/multichannel_sequence_generator.py |
|
fairseq/models/speech_to_speech/__init__.py |
|
fairseq/models/speech_to_speech/s2s_conformer.py |
|
fairseq/models/speech_to_speech/s2s_conformer_translatotron2.py |
|
fairseq/models/speech_to_speech/s2s_conformer_unity.py |
|
fairseq/models/speech_to_speech/s2s_transformer.py |
|
fairseq/models/speech_to_speech/modules/__init__.py |
|
fairseq/models/speech_to_speech/modules/ctc_decoder.py |
|
fairseq/models/speech_to_speech/modules/stacked_embedding.py |
|
fairseq/models/speech_to_speech/modules/transformer_decoder_aug.py |
|
fairseq/models/speech_to_speech/modules/transformer_encoder.py |
|
fairseq/models/speech_to_text/__init__.py |
|
fairseq/models/speech_to_text/berard.py |
|
fairseq/models/speech_to_text/convtransformer.py |
|
fairseq/models/speech_to_text/hub_interface.py |
|
fairseq/models/speech_to_text/multi_modality_model.py |
|
fairseq/models/speech_to_text/s2t_conformer.py |
|
fairseq/models/speech_to_text/s2t_transformer.py |
|
fairseq/models/speech_to_text/s2t_wav_transformer.py |
|
fairseq/models/speech_to_text/utils.py |
|
fairseq/models/speech_to_text/xm_transformer.py |
|
fairseq/models/speech_to_text/xm_transformer_unity.py |
|
fairseq/models/speech_to_text/modules/__init__.py |
|
fairseq/models/speech_to_text/modules/augmented_memory_attention.py |
|
fairseq/models/speech_to_text/modules/convolution.py |
|
fairseq/models/speech_to_text/modules/emformer.py |
|
fairseq/models/text_to_speech/__init__.py |
|
fairseq/models/text_to_speech/codehifigan.py |
|
fairseq/models/text_to_speech/fastspeech2.py |
|
fairseq/models/text_to_speech/hifigan.py |
|
fairseq/models/text_to_speech/hub_interface.py |
|
fairseq/models/text_to_speech/tacotron2.py |
|
fairseq/models/text_to_speech/tts_transformer.py |
|
fairseq/models/text_to_speech/vocoder.py |
|
fairseq/models/transformer/__init__.py |
|
fairseq/models/transformer/transformer_base.py |
|
fairseq/models/transformer/transformer_config.py |
|
fairseq/models/transformer/transformer_decoder.py |
|
fairseq/models/transformer/transformer_decoder_aug.py |
|
fairseq/models/transformer/transformer_encoder.py |
|
fairseq/models/transformer/transformer_legacy.py |
|
fairseq/models/wav2vec/__init__.py |
|
fairseq/models/wav2vec/utils.py |
|
fairseq/models/wav2vec/wav2vec.py |
|
fairseq/models/wav2vec/wav2vec2.py |
|
fairseq/models/wav2vec/wav2vec2_asr.py |
|
fairseq/models/wav2vec/wav2vec2_classification.py |
|
fairseq/models/wav2vec/wav2vec2_laser.py |
|
fairseq/models/xmod/__init__.py |
|
fairseq/models/xmod/hub_interface.py |
|
fairseq/models/xmod/model.py |
|
fairseq/models/xmod/transformer_layer_xmod.py |
|
fairseq/modules/__init__.py |
|
fairseq/modules/adaptive_input.py |
|
fairseq/modules/adaptive_softmax.py |
|
fairseq/modules/base_layer.py |
|
fairseq/modules/beamable_mm.py |
|
fairseq/modules/character_token_embedder.py |
|
fairseq/modules/checkpoint_activations.py |
|
fairseq/modules/conformer_layer.py |
|
fairseq/modules/conv_tbc.py |
|
fairseq/modules/cross_entropy.py |
|
fairseq/modules/downsampled_multihead_attention.py |
|
fairseq/modules/dynamic_convolution.py |
|
fairseq/modules/dynamic_crf_layer.py |
|
fairseq/modules/ema_module.py |
|
fairseq/modules/espnet_multihead_attention.py |
|
fairseq/modules/fairseq_dropout.py |
|
fairseq/modules/fp32_batch_norm.py |
|
fairseq/modules/fp32_group_norm.py |
|
fairseq/modules/fp32_instance_norm.py |
|
fairseq/modules/gelu.py |
|
fairseq/modules/grad_multiply.py |
|
fairseq/modules/gumbel_vector_quantizer.py |
|
fairseq/modules/kmeans_attention.py |
|
fairseq/modules/kmeans_vector_quantizer.py |
|
fairseq/modules/layer_drop.py |
|
fairseq/modules/layer_norm.py |
|
fairseq/modules/learned_positional_embedding.py |
|
fairseq/modules/lightweight_convolution.py |
|
fairseq/modules/linearized_convolution.py |
|
fairseq/modules/location_attention.py |
|
fairseq/modules/lstm_cell_with_zoneout.py |
|
fairseq/modules/multihead_attention.py |
|
fairseq/modules/positional_embedding.py |
|
fairseq/modules/positional_encoding.py |
|
fairseq/modules/quant_noise.py |
|
fairseq/modules/rotary_positional_embedding.py |
|
fairseq/modules/same_pad.py |
|
fairseq/modules/scalar_bias.py |
|
fairseq/modules/sinusoidal_positional_embedding.py |
|
fairseq/modules/sparse_multihead_attention.py |
|
fairseq/modules/sparse_transformer_sentence_encoder.py |
|
fairseq/modules/sparse_transformer_sentence_encoder_layer.py |
|
fairseq/modules/transformer_layer.py |
|
fairseq/modules/transformer_layer_aug.py |
|
fairseq/modules/transformer_sentence_encoder.py |
|
fairseq/modules/transformer_sentence_encoder_layer.py |
|
fairseq/modules/transpose_last.py |
|
fairseq/modules/unfold.py |
|
fairseq/modules/vggblock.py |
|
fairseq/modules/dynamicconv_layer/__init__.py |
|
fairseq/modules/dynamicconv_layer/cuda_function_gen.py |
|
fairseq/modules/dynamicconv_layer/dynamicconv_layer.py |
|
fairseq/modules/dynamicconv_layer/setup.py |
|
fairseq/modules/lightconv_layer/__init__.py |
|
fairseq/modules/lightconv_layer/cuda_function_gen.py |
|
fairseq/modules/lightconv_layer/lightconv_layer.py |
|
fairseq/modules/lightconv_layer/setup.py |
|
fairseq/modules/quantization/__init__.py |
|
fairseq/modules/quantization/quantization_options.py |
|
fairseq/modules/quantization/pq/__init__.py |
|
fairseq/modules/quantization/pq/em.py |
|
fairseq/modules/quantization/pq/pq.py |
|
fairseq/modules/quantization/pq/utils.py |
|
fairseq/modules/quantization/pq/modules/__init__.py |
|
fairseq/modules/quantization/pq/modules/qconv.py |
|
fairseq/modules/quantization/pq/modules/qemb.py |
|
fairseq/modules/quantization/pq/modules/qlinear.py |
|
fairseq/modules/quantization/scalar/__init__.py |
|
fairseq/modules/quantization/scalar/ops.py |
|
fairseq/modules/quantization/scalar/utils.py |
|
fairseq/modules/quantization/scalar/modules/__init__.py |
|
fairseq/modules/quantization/scalar/modules/qact.py |
|
fairseq/modules/quantization/scalar/modules/qconv.py |
|
fairseq/modules/quantization/scalar/modules/qemb.py |
|
fairseq/modules/quantization/scalar/modules/qlinear.py |
|
fairseq/optim/__init__.py |
|
fairseq/optim/adadelta.py |
|
fairseq/optim/adafactor.py |
|
fairseq/optim/adagrad.py |
|
fairseq/optim/adam.py |
|
fairseq/optim/adamax.py |
|
fairseq/optim/amp_optimizer.py |
|
fairseq/optim/bmuf.py |
|
fairseq/optim/composite.py |
|
fairseq/optim/cpu_adam.py |
|
fairseq/optim/dynamic_loss_scaler.py |
|
fairseq/optim/fairseq_optimizer.py |
|
fairseq/optim/fp16_optimizer.py |
|
fairseq/optim/fused_adam.py |
|
fairseq/optim/fused_lamb.py |
|
fairseq/optim/nag.py |
|
fairseq/optim/sgd.py |
|
fairseq/optim/shard.py |
|
fairseq/optim/lr_scheduler/__init__.py |
|
fairseq/optim/lr_scheduler/cosine_lr_scheduler.py |
|
fairseq/optim/lr_scheduler/fairseq_lr_scheduler.py |
|
fairseq/optim/lr_scheduler/fixed_schedule.py |
|
fairseq/optim/lr_scheduler/inverse_square_root_schedule.py |
|
fairseq/optim/lr_scheduler/manual_lr_scheduler.py |
|
fairseq/optim/lr_scheduler/pass_through.py |
|
fairseq/optim/lr_scheduler/polynomial_decay_schedule.py |
|
fairseq/optim/lr_scheduler/reduce_lr_on_plateau.py |
|
fairseq/optim/lr_scheduler/step_lr_scheduler.py |
|
fairseq/optim/lr_scheduler/tri_stage_lr_scheduler.py |
|
fairseq/optim/lr_scheduler/triangular_lr_scheduler.py |
|
fairseq/scoring/__init__.py |
|
fairseq/scoring/bertscore.py |
|
fairseq/scoring/bleu.py |
|
fairseq/scoring/chrf.py |
|
fairseq/scoring/meteor.py |
|
fairseq/scoring/tokenizer.py |
|
fairseq/scoring/wer.py |
|
fairseq/tasks/__init__.py |
|
fairseq/tasks/audio_classification.py |
|
fairseq/tasks/audio_finetuning.py |
|
fairseq/tasks/audio_pretraining.py |
|
fairseq/tasks/cross_lingual_lm.py |
|
fairseq/tasks/denoising.py |
|
fairseq/tasks/fairseq_task.py |
|
fairseq/tasks/frm_text_to_speech.py |
|
fairseq/tasks/hubert_pretraining.py |
|
fairseq/tasks/language_modeling.py |
|
fairseq/tasks/legacy_masked_lm.py |
|
fairseq/tasks/masked_lm.py |
|
fairseq/tasks/multilingual_denoising.py |
|
fairseq/tasks/multilingual_language_modeling.py |
|
fairseq/tasks/multilingual_masked_lm.py |
|
fairseq/tasks/multilingual_translation.py |
|
fairseq/tasks/multires_hubert_pretraining.py |
|
fairseq/tasks/nlu_finetuning.py |
|
fairseq/tasks/online_backtranslation.py |
|
fairseq/tasks/semisupervised_translation.py |
|
fairseq/tasks/sentence_prediction.py |
|
fairseq/tasks/sentence_prediction_adapters.py |
|
fairseq/tasks/sentence_ranking.py |
|
fairseq/tasks/simultaneous_translation.py |
|
fairseq/tasks/span_masked_lm.py |
|
fairseq/tasks/speech_dlm_task.py |
|
fairseq/tasks/speech_to_speech.py |
|
fairseq/tasks/speech_to_text.py |
|
fairseq/tasks/speech_ulm_task.py |
|
fairseq/tasks/text_to_speech.py |
|
fairseq/tasks/translation.py |
|
fairseq/tasks/translation_from_pretrained_bart.py |
|
fairseq/tasks/translation_from_pretrained_xlm.py |
|
fairseq/tasks/translation_lev.py |
|
fairseq/tasks/translation_multi_simple_epoch.py |
|
fairseq_cli/__init__.py |
|
fairseq_cli/eval_lm.py |
|
fairseq_cli/generate.py |
|
fairseq_cli/hydra_train.py |
|
fairseq_cli/hydra_validate.py |
|
fairseq_cli/interactive.py |
|
fairseq_cli/preprocess.py |
|
fairseq_cli/score.py |
|
fairseq_cli/train.py |
|
fairseq_cli/validate.py |
|
tests/test_activation_checkpointing.py |
|
tests/test_amp_optimizer.py |
|
tests/test_average_checkpoints.py |
|
tests/test_backtranslation_dataset.py |
|
tests/test_binaries.py |
|
tests/test_binarizer.py |
|
tests/test_character_token_embedder.py |
|
tests/test_checkpoint_utils.py |
|
tests/test_checkpoint_utils_for_task_level_attributes.py |
|
tests/test_concat_dataset.py |
|
tests/test_constraints.py |
|
tests/test_convtbc.py |
|
tests/test_data_utils.py |
|
tests/test_dataclass_utils.py |
|
tests/test_dataset.py |
|
tests/test_dictionary.py |
|
tests/test_ema.py |
|
tests/test_espnet_multihead_attention.py |
|
tests/test_export.py |
|
tests/test_file_chunker_utils.py |
|
tests/test_file_io.py |
|
tests/test_fp16_optimizer.py |
|
tests/test_hf_hub.py |
|
tests/test_huffman.py |
|
tests/test_inference_dropout.py |
|
tests/test_iopath.py |
|
tests/test_iterators.py |
|
tests/test_label_smoothing.py |
|
tests/test_lm_context_window.py |
|
tests/test_lstm_jitable.py |
|
tests/test_memory_efficient_fp16.py |
|
tests/test_metrics.py |
|
tests/test_multi_corpus_dataset.py |
|
tests/test_multi_corpus_sampled_dataset.py |
|
tests/test_multihead_attention.py |
|
tests/test_noising.py |
|
tests/test_online_backtranslation.py |
|
tests/test_plasma_utils.py |
|
tests/test_positional_encoding.py |
|
tests/test_reproducibility.py |
|
tests/test_resampling_dataset.py |
|
tests/test_roberta.py |
|
tests/test_rotary_positional_embedding.py |
|
tests/test_sequence_generator.py |
|
tests/test_sequence_scorer.py |
|
tests/test_sparse_multihead_attention.py |
|
tests/test_token_block_dataset.py |
|
tests/test_train.py |
|
tests/test_transformer.py |
|
tests/test_utils.py |
|
tests/test_valid_subset_checks.py |