QMB15's picture
Upload 8 files
b3be29f
|
raw
history blame
2.75 kB
metadata
license: llama2
language:
  - en

An experimental merging of Several Models using two various methods, Ties-Merge and BlockMerge_Gradient

I plan for this to be the base of my Model with my own [Stheno: ERP-Based LORA] merged in, some time in the future.

Stheno:
Gradient Merge of Stheno-P1 & Stheno-P2.

SISTER MODEL HERE: Stheno-Inverted-L2-13B

Quants courtesy of TheBloke!
GPTQ
GGUF
GGML

Test Checklist:
Censorship - Fairly Uncensored
Writing - Good Prose, Fairly Descriptive
NSFW - Yes
IQ Level - Pretty Smart
Formatting - Proper Formatting with Examples

Stheno-P1 [Ties-Merge]
-----elinas/chronos-13b-v2
-----jondurbin/airoboros-l2-13b-2.1
-----NousResearch/Nous-Hermes-Llama2-13b+nRuaif/Kimiko-v2 LORA

Stheno-P2 [Ties-Merge]
-----CalderaAI/13B-Legerdemain-L2+lemonilia/limarp-llama2-v2 LORA
-----ehartford/WizardLM-1.0-Uncensored-Llama2-13b
-----Henk717/spring-dragon

Most formats could work, but my tests have all been done in Alpaca format and it works well.

### Instruction:
Your instruction or question here.
For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.

### Response:

Below is the Illustration for the Final Merge:

ILLUSTRATION

Once Again, thanks to Chargoddard for his amazing and simple ties-merge script, and Gryphe for their great BlockMerge_Gradient script. Thanks to the original model creators too!

Art by wada_kazu / わだかず (pixiv page private?)