import os

import pyarrow as pa
import pyarrow.parquet as pq
import datasets

_REPO_NAME = 'Fsoft-AIC/the-vault'

_LANG_TO_TEXT = {
    "python": "python",
    "c": "c",
    # "c#": "c_sharp",
    # "c++": "cpp",
    # "go": "go",
    # "Java": "java",
    # "javascript": "javascript",
    # "php": "php",
    # "ruby": "ruby",
    # "rust": "rust",
}


_DESCRIPTION = """The Vault"""

_HOMEPAGE = "https://huggingface.co/Fsoft-AIC"


_TEXT_TO_LANG = {}
for lang in _LANG_TO_TEXT:
    _TEXT_TO_LANG[_LANG_TO_TEXT[lang]] = lang


        
_LANG_CONFIGS = ["all"] + list(_TEXT_TO_LANG.keys())

num_shard_split = {
    'train/small/python': 1,
    'train/medium/python': 1,
    'train/small/c': 1,
    'train/medium/c': 1
}  
_SPLIT_CONFIGS = ["all", "train", "train/small", "train/medium"]

class TheVaultFunctionConfig(datasets.BuilderConfig):
    """BuilderConfig for The Vault dataset."""

    def __init__(self, *args, languages=["all"], split_set= ["all"], **kwargs):
        """BuilderConfig for the GitHub Code dataset.
        Args:
            split_set (:obj:`List[str]`): List of split set to load.
            languages (:obj:`List[str]`): List of languages to load.
            **kwargs: keyword arguments forwarded to super.
        """
        super().__init__(
            *args,
            name= "+".join([split.replace("/", "_") for split in split_set]) + "-" + "+".join(languages),
            **kwargs,
        )
        
        languages = set([lang.lower() for lang in languages])
        split_set = set([split.lower() for split in split_set]) 
        
        assert all([language in _LANG_CONFIGS for language in languages]), f"languages {languages} contains language not in {_LANG_CONFIGS}."
        assert all([split in _SPLIT_CONFIGS for split in split_set]), f"split_set {split_set} contains element not in {_SPLIT_CONFIGS}."
        
        if "all" in split_set:
            assert len(split_set)==1, f"Passed 'all' together with other split sets. {split_set}"
        elif "train" in split_set:
            for split in split_set:
                if "train" in split and split != "train":
                    raise ValueError(f"Split set 'train' already contains '{split}'. Please only include one.")

        if "all" in languages:
            assert len(languages)==1, f"Passed 'all' together with other languages. {languages}"
        #     self.filter_languages = False
        # else:
        #     self.filter_languages = True
        
        self.languages = set(languages)
        self.split_set= split_set


class TheVaultFunction(datasets.GeneratorBasedBuilder):
    """The Vault dataset."""

    VERSION = datasets.Version("1.0.0")
    
    BUILDER_CONFIG_CLASS = TheVaultFunctionConfig
    BUILDER_CONFIGS = [TheVaultFunctionConfig(languages=[lang], split_set=[spl]) for lang in _LANG_CONFIGS for spl in _SPLIT_CONFIGS]
    DEFAULT_CONFIG_NAME = "all-all"

    
    def _info(self):
        return datasets.DatasetInfo(
            description=_DESCRIPTION,
            features=datasets.Features({
                                        "repo": datasets.Value("string"),
                                        "path": datasets.Value("string"), 
                                        "license": datasets.Value("string"),
                                        "language": datasets.Value("string"),
                                        "identifier": datasets.Value("string"),
                                        "return_type": datasets.Value("string"),
                                        # "original_string": datasets.Value("string"),
                                        "original_docstring": datasets.Value("string"),
                                        "docstring": datasets.Value("string"),
                                        "code": datasets.Value("string"),
                                        "code_tokens": datasets.Value("string"),
                                        "docstring_tokens": datasets.Value("string"),
                                        "short_docstring": datasets.Value("string"),
                                        "short_docstring_tokens": datasets.Value("string"),
                                        "comment": datasets.Value("string"),
                                        }),
            supervised_keys=None,
            homepage=_HOMEPAGE,
            
        )

    def _split_generators(self, dl_manager):
        generators = []
        split_set = list(self.config.split_set)
        languages = list(self.config.languages)

        if "all" in split_set:
            split_set = _SPLIT_CONFIGS[1:]

        if "train" in split_set:
            split_set.remove('train')
            split_set.extend(["train/small", "train/medium"])
        
        if "all" in languages:
            languages = _LANG_CONFIGS[1:]

        for split in split_set:
            split_files = []
            for language in languages:
                num_shards = num_shard_split[f"{split}/{language}"]
                data_files = [
                    f"data/{split}/{language}-{_index:05d}-of-{num_shards:05d}.parquet"
                    for _index in range(num_shards)
                ]
                files = dl_manager.download(data_files)
                split_files.extend(files)

            generators.append(
                datasets.SplitGenerator(
                    name=split.replace("/", "_"),
                    gen_kwargs={
                        "files": split_files,
                    },
                ),
            )
        return generators

    def _generate_examples(self, files):
        key = 0
        for file_idx, file in enumerate(files):
            with open(file, "rb") as f:
                parquet_file = pq.ParquetFile(f)
                for batch_idx, record_batch in enumerate(parquet_file.iter_batches(batch_size=10_000)):
                    pa_table = pa.Table.from_batches([record_batch])
                    for row_index in range(pa_table.num_rows):
                        row = pa_table.slice(row_index, 1).to_pydict()
                        
                        # lang = row['language'][0]
                        
                        # if self.config.filter_languages and not lang in self.config.languages:
                        #     continue
                        
                        yield key, {
                                        "repo": row['repo'][0],
                                        "path": row['path'][0], 
                                        "license": row['license'][0], 
                                        "language": row['language'][0],
                                        "identifier": row['identifier'][0],
                                        "return_type": row['return_type'][0],
                                        # "original_string": row['original_string'][0],
                                        "original_docstring": row['original_docstring'][0],
                                        "docstring": row['docstring'][0],
                                        "docstring_tokens": row['docstring_tokens'][0],
                                        "code": row['code'][0],
                                        "code_tokens": row['code_tokens'][0],
                                        "short_docstring": row['short_docstring'][0],
                                        "short_docstring_tokens": row['short_docstring_tokens'][0],
                                        "comment": row['comment'][0]
                                    } 
                        key += 1