Triangle104
commited on
Commit
•
e49e708
1
Parent(s):
0a5bbb9
Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,29 @@ tags:
|
|
11 |
This model was converted to GGUF format from [`concedo/Beepo-22B`](https://huggingface.co/concedo/Beepo-22B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
12 |
Refer to the [original model card](https://huggingface.co/concedo/Beepo-22B) for more details on the model.
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
## Use with llama.cpp
|
15 |
Install llama.cpp through brew (works on Mac and Linux)
|
16 |
|
|
|
11 |
This model was converted to GGUF format from [`concedo/Beepo-22B`](https://huggingface.co/concedo/Beepo-22B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
12 |
Refer to the [original model card](https://huggingface.co/concedo/Beepo-22B) for more details on the model.
|
13 |
|
14 |
+
---
|
15 |
+
Model details:
|
16 |
+
-
|
17 |
+
This is a finetune done on top of https://huggingface.co/mistralai/Mistral-Small-Instruct-2409 making it less censored in general, while attempting to maintain excellent instruct capabilities.
|
18 |
+
|
19 |
+
Key Features:
|
20 |
+
-
|
21 |
+
Retains Intelligence - LR was kept low and dataset heavily pruned to avoid losing too much of the original model's intelligence.
|
22 |
+
Instruct prompt format supports Alpaca - Honestly, I don't know why more models don't use it. If you are an Alpaca format lover like me, this should help. The original Mistral instruct format can still be used, but is not recommended.
|
23 |
+
Instruct Decensoring Applied - You should not need a jailbreak for a model to obey the user. The model should always do what you tell it to. No need for weird "Sure, I will" or kitten-murdering-threat tricks. No abliteration was done, only finetuning. This model is not evil. It does not judge or moralize. Like a good tool, it simply obeys.
|
24 |
+
|
25 |
+
You can obtain the GGUF quantization of this model here: https://huggingface.co/concedo/Beepo-22B-GGUF
|
26 |
+
|
27 |
+
Prompt template: Alpaca
|
28 |
+
|
29 |
+
### Instruction:
|
30 |
+
{prompt}
|
31 |
+
|
32 |
+
### Response:
|
33 |
+
|
34 |
+
Please leave any feedback or issues that you may have.
|
35 |
+
|
36 |
+
---
|
37 |
## Use with llama.cpp
|
38 |
Install llama.cpp through brew (works on Mac and Linux)
|
39 |
|