--- tags: - roleplay - creative-writing - merge - mergekit base_model: - Delta-Vector/Archaeo-12B pipeline_tag: text-generation library_name: transformers --- These are static GGUF quants. --- ``` __~a~_ ~~; ~_ _ ~ ~_ _ '_\;__._._._._._._] ~_._._._._._.__;/_` '(/'/'/'/'|'|'|'| ( )|'|'|'|'\'\'\'\)' (/ / / /, | | | |(/ \) | | | ,\ \ \ \) (/ / / / / | | | ~(/ \) ~ | | \ \ \ \ \) (/ / / / / ~ ~ ~ (/ \) ~ ~ \ \ \ \ \) (/ / / / ~ / (||)| ~ \ \ \ \) ~ / / ~ M /||\M ~ \ \ ~ ~ ~ /||\ ~ ~ //||\\ //||\\ //||\\ '/||\' "Archaeopteryx" ```

A series of Merges made for Roleplaying & Creative Writing, This model uses Rei-12B and Francois-Huali-12B and Slerp to merge the 2 models.

ChatML formatting

"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""
        

MergeKit Configuration

models:
  - model: Delta-Vector/Francois-Huali-12B
  - model: Delta-Vector/Rei-12B
merge_method: slerp
base_model: Delta-Vector/Rei-12B
parameters:
  t:
   - value: 0.2
dtype: bfloat16
tokenizer_source: base
        

Quants:

Credits

Thank you to: Kubernetes-bad, LucyKnada, Intervitens & The rest of Anthracite