This is a finetune of https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T trained on a small 50mb subset of the Kobble Dataset. Training was done in under 2 hours on a single Nvidia RTX 2060 Mobile GPU with qLora (LR 1.5e-4, rank 8, alpha 16, batch size 2, gradient acc. 4, 2048 ctx).
You can obtain the GGUF quantization of this model here: https://huggingface.co/concedo/KobbleTinyV2-1.1B-GGUF
Update: KobbleTiny has been upgraded to V2! The old V1 is still available at this link.
Try it live now: https://concedo-koboldcpp-kobbletiny.hf.space/
Dataset and Objectives
The Kobble Dataset is a semi-private aggregated dataset made from multiple online sources and web scrapes. It contains content chosen and formatted specifically to work with KoboldAI software and Kobold Lite.
Dataset Categories:
- Instruct: Single turn instruct examples presented in the Alpaca format, with an emphasis on uncensored and unrestricted responses.
- Chat: Two participant roleplay conversation logs in a multi-turn raw chat format that KoboldAI uses.
- Story: Unstructured fiction excerpts, including literature containing various erotic and provocative content.
Prompt template: Alpaca
### Instruction:
{prompt}
### Response:
Note: No assurances will be provided about the origins, safety, or copyright status of this model, or of any content within the Kobble dataset.
If you belong to a country or organization that has strict AI laws or restrictions against unlabelled or unrestricted content, you are advised not to use this model.
- Downloads last month
- 40