---


Zicklein: A german finetuned instructions following LLaMA

This is a ggml conversion of Zicklein 7B.

Zicklein itself is a LLaMA finetuned model with a cleaned and german translated Alpaca dataset.

Currently I have only converted it into new k-quant method Q5_K_M. I will gladly make more versions on request.

Other possible quantizations include: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q5_K_M, q6_K

A f-16 version could be found here: nikuya3/alpaca-lora-7b-german-base-51k-ggml

Compatible with llama.cpp, but also with:

  • text-generation-webui
  • KoboldCpp
  • ParisNeo/GPT4All-UI
  • llama-cpp-python
  • ctransformers

Prompt format

Since this model is based on alpaca dataset, the right prompt formatting should look like this:

Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
{instruction}

### Input:
{input}

### Response:

Or without addiotional input:


Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{instruction}

### Response:

That's it!

If you have any further questions, feel free to contact me or start a discussion

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train phi0112358/Zicklein-7B-german_Alpaca-ggml