ape = Algorithmic Pattern Emulation

A finetune of Mistral Nemo Instruct 2407 using my fullfictions-85kmax dataset. The prompts are fairly simple. My goal is to train a model that can write long ficiton that make sense. The training data contains the full text of public domain short stories and novels. 85k of context is about the limit I've been able to train without getting OOM errors using rented GPUs.

Thanks to unsloth brothers, gutenberg volunteers, the Mistral Nemo team, and the folks in various discord servers who have helped me out.

Downloads last month
6
Safetensors
Model size
12.2B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for leftyfeep/ape-fiction-full-instruct

Quantizations
1 model