dtype: bfloat16 merge_method: linear parameters: int8_mask: 1.0 normalize: 1.0 slices: - sources: - layer_range: [0, 4] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.6235769265047518 - layer_range: [0, 4] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.7274442555681364 - sources: - layer_range: [4, 8] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.5271398694239577 - layer_range: [4, 8] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.3489250438855029 - sources: - layer_range: [8, 12] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.15496421762028023 - layer_range: [8, 12] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.541330668871115 - sources: - layer_range: [12, 16] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.5267269624685371 - layer_range: [12, 16] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.8265113027826562 - sources: - layer_range: [16, 20] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.6599861585345389 - layer_range: [16, 20] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: -0.249060520039947 - sources: - layer_range: [20, 24] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.7761318532349375 - layer_range: [20, 24] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.7040995904551324 - sources: - layer_range: [24, 28] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: 0.40152017541360374 - layer_range: [24, 28] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.767141768059921 - sources: - layer_range: [28, 32] model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 parameters: weight: -0.004536646708608122 - layer_range: [28, 32] model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct parameters: weight: 0.8295357241419378