BlackMagic-7B / README.md
ToastyPigeon's picture
Update README.md
42cdcc9 verified
|
raw
history blame
1.98 kB
metadata
base_model: []
library_name: transformers
tags:
  - mergekit
  - merge

merge

This is a merge of pre-trained language models created using mergekit.

Thank you to MatchaSpeed for picking out the models to go into this merge. You're awesome! 🤗

The models in this merge were chosen for their attention to detail and/or their creativity.

Based on each model's individual strength, weights were applied to prioritize the model's intelligence (self-attention weights) or style/creativity (multi-layer perceptron weights).

Merge Details

Merge Method

This model was merged using the task arithmetic merge method using Mistral-7B-v0.1 as a base.

Models Merged

The following models were included in the merge:

  • fearlessdots/WizardLM-2-7B-abliterated
  • Gryphe/MythoMist-7b
  • Sao10K/Frostwind-v2.1-m7
  • senseable/Westlake-7b-v2
  • maywell/PiVoT-0.1-Evil-a
  • Undi95/Toppy-M-7B

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: fearlessdots/WizardLM-2-7B-abliterated
    parameters:
      weight: 1.0
  - model: Undi95/Toppy-M-7B
    parameters:
      weight: 
        - filter: self_attn
          value: 0.8
        - value: 0.5
  - model: senseable/Westlake-7b-v2
    parameters:
      weight:
        - filter: self_attn
          value: 0.6
        - value: 0.4
  - model: maywell/PiVoT-0.1-Evil-a
    parameters:
      weight: 
        - filter: mlp
          value: 0.2
        - value: 0.0
  - model: Sao10K/Frostwind-v2.1-m7
    parameters:
      weight: 
        - filter: self_attn
          value: 0.2
        - filter: mlp
          value: 0.8
        - value: 0.5
  - model: Gryphe/MythoMist-7b
    parameters:
      weight: 
        - filter: mlp
          value: 0.6
        - value: 0.0
base_model: TeeZee/Mistral-7B-v0.1-fp32
merge_method: task_arithmetic
dtype: float32