|
--- |
|
license: mit |
|
language: |
|
- en |
|
--- |
|
|
|
Submission to the BabyLM challenge 2024 trained on [Baby-cosmo-fine-100M](https://huggingface.co/datasets/ltg/babylm-2024-baby-cosmo-fine-100m). |
|
|
|
The training scripts are published here: https://github.com/ltgoslo/gpt-bert |
|
|
|
```latex |
|
@misc{charpentier2024gptbertboth, |
|
title={GPT or BERT: why not both?}, |
|
author={Lucas Georges Gabriel Charpentier and David Samuel}, |
|
year={2024}, |
|
eprint={2410.24159}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL}, |
|
url={https://arxiv.org/abs/2410.24159}, |
|
} |
|
``` |