File size: 3,550 Bytes
7ba68cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
08019b1
7ba68cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8cc638c
 
7ba68cc
 
 
cb01a6f
 
7ba68cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
---
license: cc-by-nc-4.0
base_model_relation: quantized
quantized_by: Quant-Cartel
base_model: InferenceIllusionist/SorcererLM-22B

---
```
  e88 88e                               d8     
 d888 888b  8888 8888  ,"Y88b 888 8e   d88     
C8888 8888D 8888 8888 "8" 888 888 88b d88888   
 Y888 888P  Y888 888P ,ee 888 888 888  888     
  "88 88"    "88 88"  "88 888 888 888  888     
      b                                        
      8b,                                      
 
  e88'Y88                  d8           888    
 d888  'Y  ,"Y88b 888,8,  d88    ,e e,  888    
C8888     "8" 888 888 "  d88888 d88 88b 888    
 Y888  ,d ,ee 888 888     888   888   , 888    
  "88,d88 "88 888 888     888    "YeeP" 888    
                                               
PROUDLY PRESENTS         
```
# SorcererLM-22B-exl2-longcal

Quantized using 115 rows of 8192 tokens from the default ExLlamav2-calibration dataset.

Branches:
- `main` -- `measurement.json`
- 8.0b8h -- 8.0bpw, 8bit lm_head
- 6.0b6h -- 6.0bpw, 6bit lm_head
- 5.0b6h -- 5.0bpw, 6bit lm_head
- 4.0b6h -- 4.0bpw, 6bit lm_head
- 3.5b6h -- 3.5bpw, 6bit lm_head
- 2.25b6h -- 2.25bpw, 6bit lm_head

Original model link: [InferenceIllusionist/SorcererLM-22B](https://huggingface.co/InferenceIllusionist/SorcererLM-22B)

Original model README below.

-----
## SorcererLM-22B

<img src="https://files.catbox.moe/ya4zca.png" width="500"/>
<i>Because good things always come in threes!</i>


**SorcererLM-22B** is here, rounding out the trinity of Mistral-Small-Instruct tunes from the [Quant Cartel](https://huggingface.co/Quant-Cartel).


## Prompt Format

* Prompt Template: Mistral V2 & V3 Context / Instruct Templates 
* Samplers / Advanced Instruct Template: See [Quant-Cartel/Recommended-Settings/SorcererLM-22B](https://huggingface.co/Quant-Cartel/Recommended-Settings/tree/main/SorcererLM-22B)

## Quantized Versions

* [exl2 longcal](https://huggingface.co/Quant-Cartel/SorcererLM-22B-exl2-longcal)
* [iMat GGUF](https://huggingface.co/Quant-Cartel/SorcererLM-22B-iMat-GGUF)

## Training 

For starters this is a LORA tune on top of Mistral-Small-Instruct-2409 and **not** a pruned version of [SorcererLM-8x22b](https://huggingface.co/rAIfle/SorcererLM-8x22b-bf16).

Trained with a whole lot of love on 1 epoch of cleaned and deduped c2 logs. This model is 100% 'born-local', the result of roughly 27 hours and a little bit of patience on a single RTX 4080 SUPER.

As hyperparameters and dataset intentionally mirror ones used in the original Sorcerer 8x22b tune, this is considered its 'lite' counterpart aiming to provide the same bespoke conversational experience relative to its size and reduced hardware requirements.

While all three share the same Mistral-Small-Instruct base, in contrast to its sisters [Mistral-Small-NovusKyver](https://huggingface.co/Envoid/Mistral-Small-NovusKyver) and [Acolyte-22B](https://huggingface.co/rAIfle/Acolyte-22B) this release did not SLERP the resulting model with the original in a 50/50 ratio post-training. Instead, alpha was dropped when the lora was merged with full precision weights in the final step.

## Acknowledgments

* First and foremost a huge thank you my brilliant teammates [envoid](https://huggingface.co/envoid/) and [rAIfle](https://huggingface.co/rAIfle/). Special shout-out to rAIfle for critical last minute advice that got this one through the finish line
* Props to unsloth as well for helping make this local tune possible
* And of course, none of this would matter without users like you. Thank you :)

## Safety
...