File size: 2,614 Bytes
9c8b90f
 
 
 
 
 
 
 
 
 
 
 
458d7bf
9c8b90f
f77b4a0
 
 
 
 
 
 
 
9b2c19e
 
 
 
 
 
f77b4a0
 
9c8b90f
 
 
 
6a59887
 
 
9c8b90f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
---
base_model:
- EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
- allura-org/Teleut-7b
- FourOhFour/Vapor_v2_7B
library_name: transformers
tags:
- mergekit
- merge

---

![](https://i.imgur.com/3rVKAcZ.jpeg)

## EDIT: MAY NOT WORK FOR GGUFs

I don't know if its an issue with me or the model, but I can't seem to make quants of this model. I consistently get 
`llama_model_quantize: failed to quantize: tensor 'blk.24.attn_norm.weight' has invalid data`. My whole setup has so 
many stds and idiosynchasies that it may just be my system, but I tried redoing the whole thing and same stuff happened.
At this point it may be an issue with the NuSLERP method or one of the models I'm using in the merge. Not sure gang. I will
try swapping out a model or two in the merge and trying again to upload as a v2.0.

## EDIT (again): DO NOT USE THIS MODEL

I tried it four more times, swapping out models, swapping base models and models entirely, swapping params, 
`git pull`ing llamacpp and mergekit, nothing. Errors when making quants, every one. I'm declaring this a lost cause. 
I'm leaving this up in case someone gets it working.

# info

Merge using the brand new NuSLERP method. Fresh out of the oven. Performance not guaranteed.

Uses the slightly-unstable EVA and two other finetunes I found. I also turned on both the NuSLERP exclusive mergekit options for fun.

Named after the **nemesia**, a temperate shrubby flower. I tried to pick a flower that sounded kind of like NuSLERP. 
It doesn't, but the name still has the '''essence''' of NuSLERP I guess? (it doesn't.) Very pretty flower nonetheless 

# mergekit

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
### Merge Method

This model was merged using the NuSLERP merge method using [EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) as a base.

### Models Merged

The following models were included in the merge:
* [allura-org/Teleut-7b](https://huggingface.co/allura-org/Teleut-7b)
* [FourOhFour/Vapor_v2_7B](https://huggingface.co/FourOhFour/Vapor_v2_7B)

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: allura-org/Teleut-7b
    parameters:
      weight: 0.6
  - model: FourOhFour/Vapor_v2_7B
    parameters:
      weight: 0.2
  - model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
    parameters:
      weight: 1.0
merge_method: nuslerp
base_model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
parameters:
  normalize: true
  int8_mask: true
  nuslerp_flatten: false
  nuslerp_row_wise: true
dtype: float16
```