jeffmeloy's picture
Update README.md
dc0d2aa verified
metadata
license: apache-2.0
base_model:
  - Qwen/Qwen2.5-7B
pipeline_tag: text-generation
tags:
  - not-for-all-audiences
language:
  - en
library_name: transformers

Model Description

Model created by analyzing and selecting the optimal layers from other Qwen2.5-7B models based on their dimensional utilization efficiency, measured by the Normalized Effective Rank (NER). Computed like:

  • Input: Weight matrix for each model layer
  • Compute singular values σᵢ where σᵢ ≥ 0 # σᵢ represents the importance of each dimension
  • Filter values above numerical threshold (>1e-12)
  • Sum all singular values: S = Σσᵢ # S acts as normalization factor
  • Create probability distribution: pᵢ = σᵢ/S # converts singular values to probabilities summing to 1
  • Compute Shannon entropy: H = -Σ(pᵢ * log₂(pᵢ)) # measures information content
  • Calculate maximum possible entropy: H_max = log₂(n)
  • Final NER score = H/H_max # normalizes score to [0,1] range
  • Results in value between 0 and 1 for each model layer

Creating Composite Model

Code here: https://huggingface.co/jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0/blob/main/ner_merge.py

Code functions:

  • Download selected models from Hugging Face Hub
  • Calculate Normalized Effective Rank (NER) for each layer within each model
  • Define model and layer name pairs that have highest NER for each layer based on their NER scores
  • Incrementally build a composite model using layer with highest NER from model pool
  • Save merge reports documenting layer sources
  • Copy config and tokenizer files from base model
  • Save the composite model with complete weights # model ready to use

Configfile:

base_model: "Qwen/Qwen2.5-7B"

fine_tuned_models: # uncomment the models you want to merge

#- "Qwen/Qwen2.5-7B"

#- "Qwen/Qwen2.5-7B-Instruct"

#- "EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1"

#- "FourOhFour/Vapor_v2_7B"

#- "Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2"

#- "happzy2633/qwen2.5-7b-ins-v3"

#- "huihui-ai/Qwen2.5-7B-Instruct-abliterated-v2"

#- "HumanLLMs/Humanish-Qwen2.5-7B-Instruct"

#- "Orion-zhen/Qwen2.5-7B-Instruct-Uncensored"

#- "Orion-zhen/Meissa-Qwen2.5-7B-Instruct"

#- "jeffmeloy/Qwen2.5-7B-nerd-uncensored-v0.9"

#- "jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0"

#- "jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.1"

#- "jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.2"

#- "AmberYifan/Qwen2.5-7B-dpo-2k"

#- "sethuiyer/Qwen2.5-7B-Anvita"

#- "rombodawg/Rombos-LLM-V2.5-Qwen-7b"

#- "Cran-May/T.E-8.1"

#- "beomi/Qwen2.5-7B-Instruct-kowiki-qa"

#- "Orion-zhen/Qwen2.5-7B-Gutenberg-KTO"

#- "fblgit/cybertron-v4-qw7B-MGS"

#- "nguyentd/FinancialAdvice-Qwen2.5-7B"

#- "WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B"

#- "edgerunner-ai/EdgeRunner-Command-Nested"

#- "katanemo/Arch-Function-7B"

#- "DeepGlint-AI/llava-mlcd-qwen2.5-7b"

#- "mergekit-community/mergekit-slerp-aflqaqy"

#- "mergekit-community/mergekit-ties-inxwsfo"

#- "Qwen/Qwen2.5-Coder-7B-Instruct"

#- "Qwen/Qwen2.5-Math-7B-Instruct"

#- "Qwen/Qwen2.5-Coder-7B"

#- "Qwen/Qwen2.5-Math-7B"

#- "thomas-yanxin/XinYuan-Qwen2.5-7B-0917"

#- "jbjeong91/Qwen2.5_7B_IST_StoryGen_vanilla"

#- "AmberYifan/Qwen2.5-7B-dpo-2k-hhrlhf"

#- "jbjeong91/Qwen2.5_7B_IST_StoryGen_test2"

models_dir: "./input_models/"

output_dir: "./merged_model/"

metric_dir: "./metrics/"