--- base_model: - unsloth/Mistral-Small-Instruct-2409 - TheDrummer/Cydonia-22B-v1.2 - Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small - anthracite-org/magnum-v4-22b - ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1 - spow12/ChatWaifu_v2.0_22B - rAIfle/Acolyte-22B - Envoid/Mistral-Small-NovusKyver - InferenceIllusionist/SorcererLM-22B - allura-org/MS-Meadowlark-22B - crestf411/MS-sunfall-v0.7.0 library_name: transformers tags: - mergekit - merge license: other language: - en --- *** ## Schisandra Many thanks to the authors of the models used! [RPMax v1.1](https://huggingface.co/ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1) | [Pantheon-RP](https://huggingface.co/Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small) | [UnslopSmall-v1](https://huggingface.co/TheDrummer/UnslopSmall-22B-v1) | [Magnum V4](https://huggingface.co/anthracite-org/magnum-v4-22b) | [ChatWaifu v2.0](https://huggingface.co/spow12/ChatWaifu_v2.0_22B) | [SorcererLM](https://huggingface.co/InferenceIllusionist/SorcererLM-22B) | [Acolyte](https://huggingface.co/rAIfle/Acolyte-22B) | [NovusKyver](https://huggingface.co/Envoid/Mistral-Small-NovusKyver) | (Meadowlark)[https://huggingface.co/allura-org/MS-Meadowlark-22B] | (Sunfall)[https://huggingface.co/crestf411/MS-sunfall-v0.7.0] *** ### Overview Main uses: RP, Storywriting An intelligent model that is attentive to details and has a low-slop writing style. This time with a stable tokenizer. Oh, and it now contains 10 finetunes! Not sure if some of them actually contribute to the output, but it's nice to see the numbers growing. *** ### Quants GGUF: [Static](https://huggingface.co/mradermacher/MS-Schisandra-22B-v0.2-GGUF) | [Imatrix](https://huggingface.co/mradermacher/MS-Schisandra-22B-v0.2-i1-GGUF) exl2: [4.65bpw](https://huggingface.co/waldie/MS-Schisandra-22B-v0.2-4.65bpw-h6-exl2) [5.5bpw](https://huggingface.co/waldie/MS-Schisandra-22B-v0.2-5.5bpw-h6-exl2) [6.5bpw](https://huggingface.co/waldie/MS-Schisandra-22B-v0.2-6.5bpw-h6-exl2) *** ### Settings Prompt format: Mistral-V3 or (this)[https://huggingface.co/ToastyPigeon/ST-Presets-Mistral-Small/resolve/main/ST-formatting-Mistral-Small-Oneshot.json] Samplers: [These](https://qu.ax/BCRKS.json) or [These](https://qu.ax/OusTx.json) *** ## Merge Details ### Merging steps ## Step1 (Config partially taken from (here)[https://huggingface.co/Casual-Autopsy/L3-Super-Nova-RP-8B]) ```yaml base_model: spow12/ChatWaifu_v2.0_22B parameters: int8_mask: true rescale: true normalize: false dtype: bfloat16 tokenizer_source: base merge_method: della models: - model: Envoid/Mistral-Small-NovusKyver parameters: density: [0.35, 0.65, 0.5, 0.65, 0.35] epsilon: [0.1, 0.1, 0.25, 0.1, 0.1] lambda: 0.85 weight: [-0.01891, 0.01554, -0.01325, 0.01791, -0.01458] - model: rAIfle/Acolyte-22B parameters: density: [0.6, 0.4, 0.5, 0.4, 0.6] epsilon: [0.1, 0.1, 0.25, 0.1, 0.1] lambda: 0.85 weight: [0.01847, -0.01468, 0.01503, -0.01822, 0.01459] ``` ## Step2 (Config partially taken from (here)[https://huggingface.co/Casual-Autopsy/L3-Super-Nova-RP-8B]) ```yaml base_model: InferenceIllusionist/SorcererLM-22B parameters: int8_mask: true rescale: true normalize: false dtype: bfloat16 tokenizer_source: base merge_method: della models: - model: crestf411/MS-sunfall-v0.7.0 parameters: density: [0.35, 0.65, 0.5, 0.65, 0.35] epsilon: [0.1, 0.1, 0.25, 0.1, 0.1] lambda: 0.85 weight: [-0.01891, 0.01554, -0.01325, 0.01791, -0.01458] - model: anthracite-org/magnum-v4-22b parameters: density: [0.6, 0.4, 0.5, 0.4, 0.6] epsilon: [0.1, 0.1, 0.25, 0.1, 0.1] lambda: 0.85 weight: [0.01847, -0.01468, 0.01503, -0.01822, 0.01459] ``` ## SchisandraVA2 (Config taken from (here)[https://huggingface.co/HiroseKoichi/Llama-3-8B-Stroganoff-4.0]) ```yaml merge_method: della_linear dtype: bfloat16 parameters: normalize: true int8_mask: true tokenizer_source: base base_model: TheDrummer/UnslopSmall-22B-v1 models: - model: ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1 parameters: density: 0.55 weight: 1 - model: Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small parameters: density: 0.55 weight: 1 - model: Step1 parameters: density: 0.55 weight: 1 - model: allura-org/MS-Meadowlark-22B parameters: density: 0.55 weight: 1 - model: Step2 parameters: density: 0.55 weight: 1 ``` ## Schisandra-v0.2 ```yaml dtype: bfloat16 tokenizer_source: base merge_method: della_linear parameters: density: 0.5 base_model: SchisandraVA2 models: - model: unsloth/Mistral-Small-Instruct-2409 parameters: weight: - filter: v_proj value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0] - filter: o_proj value: [1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1] - filter: up_proj value: [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] - filter: gate_proj value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0] - filter: down_proj value: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] - value: 0 - model: SchisandraVA2 parameters: weight: - filter: v_proj value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1] - filter: o_proj value: [0, 1, 0, 1, 1, 1, 1, 1, 0, 0, 0] - filter: up_proj value: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] - filter: gate_proj value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1] - filter: down_proj value: [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] - value: 1 ```