--- license: mit language: - en --- Submission to the BabyLM challenge 2024 trained on [Baby-cosmo-fine-100M](https://huggingface.co/datasets/ltg/babylm-2024-baby-cosmo-fine-100m). ```latex @misc{charpentier2024gptbertboth, title={GPT or BERT: why not both?}, author={Lucas Georges Gabriel Charpentier and David Samuel}, year={2024}, eprint={2410.24159}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2410.24159}, } ```