base_model: [intervitens/internlm2-base-20b-llama] | |
language: | |
- en | |
pipeline_tag: text-generation | |
tags: | |
- mergekit | |
- merge | |
license: other | |
datasets: | |
- lemonilia/LimaRP | |
Experimental model, LimaRP LoRA trained on top of internlm2-base-20b with 8192 context length and merged with internlm2-chat-20b. | |
Prompt format is ChatML. | |
*** | |
# internlm2-06limarp-1chat-TASK_ARITHM-20b-v0.03 | |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). | |
## Merge Details | |
### Merge Method | |
This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using intervitens/internlm2-base-20b-llama as a base. | |
### Models Merged | |
The following models were included in the merge: | |
* ./internlm2-limarp-20b-v0.03 | |
* ./internlm2-chat-20b-llama | |
### Configuration | |
The following YAML configuration was used to produce this model: | |
```yaml | |
models: | |
- model: ./internlm2-chat-20b-llama | |
parameters: | |
weight: 1.0 | |
- model: ./internlm2-limarp-20b-v0.03 | |
parameters: | |
weight: 0.6 | |
merge_method: task_arithmetic | |
base_model: ./internlm2-base-20b-llama | |
parameters: | |
#normalize: false | |
#int8_mask: true | |
dtype: bfloat16 | |
``` | |