Clarified quantization type
Browse filesAlso added original model card
README.md
CHANGED
@@ -3,6 +3,26 @@ license: llama2
|
|
3 |
language:
|
4 |
- en
|
5 |
---
|
6 |
-
This is an
|
7 |
|
8 |
-
Alpaca and ChatML prompt templates seem to work fine with this model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
---
|
6 |
+
This is an ExLlamaV2 quantized model in 3bpw of [jebcarter/Psyfighter-13B](https://huggingface.co/jebcarter/Psyfighter-13B) using [PIPPA](https://huggingface.co/datasets/jasonkstevens/pippa-llama2-chat/blob/refs%2Fconvert%2Fparquet/default/train/0000.parquet) as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)
|
7 |
|
8 |
+
Alpaca and ChatML prompt templates seem to work fine with this model.
|
9 |
+
|
10 |
+
# Original Model card
|
11 |
+
|
12 |
+
```
|
13 |
+
merge_method: task_arithmaetic
|
14 |
+
base_model: TheBloke/Llama-2-13B-fp16
|
15 |
+
models:
|
16 |
+
- model: TheBloke/Llama-2-13B-fp16
|
17 |
+
- model: KoboldAI/LLaMA2-13B-Tiefighter
|
18 |
+
parameters:
|
19 |
+
weight: 1.0
|
20 |
+
- model: chaoyi-wu/MedLLaMA_13B
|
21 |
+
parameters:
|
22 |
+
weight: 0.01
|
23 |
+
- model: Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged
|
24 |
+
parameters:
|
25 |
+
weight: 0.02
|
26 |
+
dtype: float16
|
27 |
+
```
|
28 |
+
This model was made possible thanks to the Compute provided by the KoboldAI community.
|