|
dtype: bfloat16 |
|
merge_method: linear |
|
parameters: |
|
int8_mask: 1.0 |
|
normalize: 1.0 |
|
slices: |
|
- sources: |
|
- layer_range: [0, 4] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.6235769265047518 |
|
- layer_range: [0, 4] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.7274442555681364 |
|
- sources: |
|
- layer_range: [4, 8] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.5271398694239577 |
|
- layer_range: [4, 8] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.3489250438855029 |
|
- sources: |
|
- layer_range: [8, 12] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.15496421762028023 |
|
- layer_range: [8, 12] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.541330668871115 |
|
- sources: |
|
- layer_range: [12, 16] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.5267269624685371 |
|
- layer_range: [12, 16] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.8265113027826562 |
|
- sources: |
|
- layer_range: [16, 20] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.6599861585345389 |
|
- layer_range: [16, 20] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: -0.249060520039947 |
|
- sources: |
|
- layer_range: [20, 24] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.7761318532349375 |
|
- layer_range: [20, 24] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.7040995904551324 |
|
- sources: |
|
- layer_range: [24, 28] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: 0.40152017541360374 |
|
- layer_range: [24, 28] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.767141768059921 |
|
- sources: |
|
- layer_range: [28, 32] |
|
model: Nisk36/finetuned-lmsys_vicuna-7b-v1.5 |
|
parameters: |
|
weight: -0.004536646708608122 |
|
- layer_range: [28, 32] |
|
model: Nisk36/FT_elyza_ELYZA-japanese-Llama-2-7b-instruct |
|
parameters: |
|
weight: 0.8295357241419378 |