---
license: apache-2.0
language:
- en
---
# KobbleTinyV2-1.1B
This is a finetune of https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T trained on a small 50mb subset of the Kobble Dataset.
Training was done in under 2 hours on a single Nvidia RTX 2060 Mobile GPU with qLora (LR 1.5e-4, rank 8, alpha 16, batch size 2, gradient acc. 4, 2048 ctx).
You can obtain the GGUF quantization of this model here: https://huggingface.co/concedo/KobbleTinyV2-1.1B-GGUF
Update: KobbleTiny has been upgraded to V2! The old V1 is [still available at this link](https://huggingface.co/concedo/KobbleTiny/tree/eb0c96864bfecfd6ac9ece1a42c4654b4997eb72).
Try it live now: https://concedo-koboldcpp-kobbletiny.hf.space/
## Dataset and Objectives
The Kobble Dataset is a semi-private aggregated dataset made from multiple online sources and web scrapes.
It contains content chosen and formatted specifically to work with KoboldAI software and Kobold Lite.
#### Dataset Categories:
- Instruct: Single turn instruct examples presented in the Alpaca format, with an emphasis on uncensored and unrestricted responses.
- Chat: Two participant roleplay conversation logs in a multi-turn raw chat format that KoboldAI uses.
- Story: Unstructured fiction excerpts, including literature containing various erotic and provocative content.
## Prompt template: Alpaca
```
### Instruction:
{prompt}
### Response:
```
**Note:** *No assurances will be provided about the **origins, safety, or copyright status** of this model, or of **any content** within the Kobble dataset.*
*If you belong to a country or organization that has strict AI laws or restrictions against unlabelled or unrestricted content, you are advised not to use this model.*