|
|
|
|
|
"""Lexicap: Lex Friedman Podcast Whisper Captions.""" |
|
|
|
|
|
import csv |
|
|
|
import datasets |
|
|
|
|
|
_CITATION = """\ |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
Lexicap contains the captions for every Lex Friedman Podcast episode. It it created by [Dr. Andrej Karpathy](https://twitter.com/karpathy). |
|
There are 430 caption files available. There are 2 types of files: |
|
- large |
|
- small |
|
Each file name follows the format `episode_{episode_number}_{file_type}.vtt`. |
|
""" |
|
|
|
|
|
|
|
class LexicapConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for Lexicap.""" |
|
|
|
def __init__(self, **kwargs): |
|
"""Constructs a LexicapConfig. |
|
Args: |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super(LexicapConfig, self).__init__(version=datasets.Version("0.1.0", ""), **kwargs), |
|
|
|
|
|
class Lexicap(datasets.GeneratorBasedBuilder): |
|
"""Lexicap dataset.""" |
|
|
|
BUILDER_CONFIGS = [ |
|
LexicapConfig( |
|
description=( |
|
f"A dataset consisting of captions for every Lex Friedman Podcast episode, generated using OpenAI Whisper. This dataset is created by [Dr. Andrej Karpathy](https://twitter.com/karpathy). |
|
), |
|
) |
|
] |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
# There is no predefined train/val/test split for this dataset. |
|
return [ |
|
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"file_path": 'vtt'}), |
|
] |
|
|