flam-kit
Building off of the success of my last merge, I identified some areas of weakness that I perceived in the resulting model and this model is an attempt to address them.
I created a wild model with a della merge (schonsense/Flamlama_70B_della), using models that all had strengths I was hoping to incorporate into my existing model SLERP. To create this new model (schonsense/flam-kit), I SLERP merged my previous model (schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP) as a base and gently brought in the wild flavors of the della merged model. The intent being to keep the instruct following and proper model function without having to stomp on it with excessive sampling parameters, while changing the voice and capabilities of the model. I believe this model to finally be a success, after a number of failures. I find (schonsense/flam-kit) to be superior in most respects to (schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP), requiring only the most modest of sampling parameters to function well.
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
- schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
- schonsense/Flamlama_70B_della
Configuration
The following YAML configuration was used to produce this model:
models:
- model: schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
- model: schonsense/Flamlama_70B_della
merge_method: slerp
base_model: schonsense/Llama-3.3-70B-Inst-Ablit-Flammades-SLERP
dtype: bfloat16
parameters:
t: [
0, 0.0, 0.0, 0.055, 0.109, 0.127, 0.145, 0.164, 0.182, 0.2,
0.218, 0.236, 0.255, 0.273, 0.291, 0.309, 0.327, 0.345, 0.364, 0.382, 0.4,
0.418, 0.436, 0.455, 0.473, 0.491, 0.509, 0.527, 0.545, 0.564, 0.582, 0.6,
0.588, 0.576, 0.564, 0.552, 0.54, 0.527, 0.515, 0.503, 0.491, 0.479, 0.467,
0.455, 0.442, 0.43, 0.418, 0.406, 0.394, 0.382, 0.369, 0.357, 0.345, 0.333,
0.321, 0.309, 0.297, 0.285, 0.273, 0.26, 0.248, 0.236, 0.224, 0.212, 0.2,
0.188, 0.176, 0.164, 0.151, 0.139, 0.127, 0.115, 0.103, 0.091, 0.079, 0.067,
0.055, 0, 0, 0
]
- Downloads last month
- 43