base_model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 dtype: bfloat16 merge_method: dare_ties parameters: int8_mask: 1.0 normalize: 1.0 slices: - sources: - layer_range: [0, 4] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 0.9567125290278002 weight: 0.7329399429419414 - layer_range: [0, 4] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 1.0 weight: 0.6175016127199866 - sources: - layer_range: [4, 8] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 0.7351530312710608 weight: 0.9361918263111237 - layer_range: [4, 8] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 1.0 weight: 0.6261500333536962 - sources: - layer_range: [8, 12] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 1.0 weight: 0.38219531855733224 - layer_range: [8, 12] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 0.9560023967707558 weight: 0.4847363738604221 - sources: - layer_range: [12, 16] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 1.0 weight: 0.9760238855152437 - layer_range: [12, 16] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 0.9962392996911643 weight: 0.6535045223316338 - sources: - layer_range: [16, 20] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 1.0 weight: 0.32930274558082606 - layer_range: [16, 20] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 0.9095162947498548 weight: 0.7439598517576353 - sources: - layer_range: [20, 24] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 0.756559714204041 weight: 0.44719009636986334 - layer_range: [20, 24] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 1.0 weight: 0.348213220068222 - sources: - layer_range: [24, 28] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 1.0 weight: 0.5663144522369852 - layer_range: [24, 28] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 1.0 weight: 0.5221351804388025 - sources: - layer_range: [28, 32] model: ./evol_merge_storage_ja_e57b_tydija/input_models/e5-mistral-7b-instruct_2385958088 parameters: density: 1.0 weight: 0.3974171827818004 - layer_range: [28, 32] model: ./evol_merge_storage_ja_e57b_tydija/input_models/japanese-stablelm-base-gamma-7b_545310900 parameters: density: 0.9839795858964644 weight: 0.2576111697863762