--- base_model: - Aratako/Antler-7B-RP-v2 - NTQAI/chatntq-ja-7b-v1.0 - TFMC/Japanese-Starling-ChatV-7B library_name: transformers tags: - mergekit - merge --- # japanese-chat-umievo-itr004-7b This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method. ### Models Merged The following models were included in the merge: * [Aratako/Antler-7B-RP-v2](https://huggingface.co/Aratako/Antler-7B-RP-v2) * [NTQAI/chatntq-ja-7b-v1.0](https://huggingface.co/NTQAI/chatntq-ja-7b-v1.0) * [TFMC/Japanese-Starling-ChatV-7B](https://huggingface.co/TFMC/Japanese-Starling-ChatV-7B) ### Configuration The following YAML configuration was used to produce this model: ```yaml dtype: bfloat16 merge_method: linear parameters: int8_mask: 1.0 normalize: 1.0 slices: - sources: - layer_range: [0, 4] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.3117821548077149 - layer_range: [0, 4] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.06622915649259134 - layer_range: [0, 4] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.1456932001643398 - sources: - layer_range: [4, 8] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.235386369865962 - layer_range: [4, 8] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.4897808973741158 - layer_range: [4, 8] model: Aratako/Antler-7B-RP-v2 parameters: weight: -0.13610409063514917 - sources: - layer_range: [8, 12] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: -0.02904316547026939 - layer_range: [8, 12] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.27984511027305725 - layer_range: [8, 12] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.17613114842003957 - sources: - layer_range: [12, 16] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.30708851106133633 - layer_range: [12, 16] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.633483361515513 - layer_range: [12, 16] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.23184742233331504 - sources: - layer_range: [16, 20] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.4261576863911635 - layer_range: [16, 20] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.448790034700236 - layer_range: [16, 20] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.908083073839566 - sources: - layer_range: [20, 24] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.49909163132229717 - layer_range: [20, 24] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.46200102252131314 - layer_range: [20, 24] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.3537781762598641 - sources: - layer_range: [24, 28] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.3053835052195379 - layer_range: [24, 28] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: 0.554162215544036 - layer_range: [24, 28] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.3526529317468468 - sources: - layer_range: [28, 32] model: NTQAI/chatntq-ja-7b-v1.0 parameters: weight: 0.458122927784649 - layer_range: [28, 32] model: TFMC/Japanese-Starling-ChatV-7B parameters: weight: -0.20730388433466773 - layer_range: [28, 32] model: Aratako/Antler-7B-RP-v2 parameters: weight: 0.5077362589077673 ```