KobbleTinyV2-1.1B

This is a finetune of https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T trained on a small 50mb subset of the Kobble Dataset. Training was done in under 2 hours on a single Nvidia RTX 2060 Mobile GPU with qLora (LR 1.5e-4, rank 8, alpha 16, batch size 2, gradient acc. 4, 2048 ctx).

You can obtain the GGUF quantization of this model here: https://huggingface.co/concedo/KobbleTinyV2-1.1B-GGUF

Update: KobbleTiny has been upgraded to V2! The old V1 is still available at this link.

Try it live now: https://concedo-koboldcpp-kobbletiny.hf.space/

Dataset and Objectives

The Kobble Dataset is a semi-private aggregated dataset made from multiple online sources and web scrapes. It contains content chosen and formatted specifically to work with KoboldAI software and Kobold Lite.

Dataset Categories:

  • Instruct: Single turn instruct examples presented in the Alpaca format, with an emphasis on uncensored and unrestricted responses.
  • Chat: Two participant roleplay conversation logs in a multi-turn raw chat format that KoboldAI uses.
  • Story: Unstructured fiction excerpts, including literature containing various erotic and provocative content.

Prompt template: Alpaca

### Instruction:
{prompt}

### Response:

Note: No assurances will be provided about the origins, safety, or copyright status of this model, or of any content within the Kobble dataset.
If you belong to a country or organization that has strict AI laws or restrictions against unlabelled or unrestricted content, you are advised not to use this model.

Downloads last month
40
Safetensors
Model size
1.1B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for concedo/KobbleTinyV2-1.1B

Adapters
1 model
Finetunes
3 models
Quantizations
2 models