twosmoothslateslabs commited on
Commit
9c8b90f
·
verified ·
1 Parent(s): 37194c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -50
README.md CHANGED
@@ -1,50 +1,57 @@
1
- ---
2
- base_model:
3
- - EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
4
- - allura-org/Teleut-7b
5
- - FourOhFour/Vapor_v2_7B
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
-
11
- ---
12
- # temp
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- ## Merge Details
17
- ### Merge Method
18
-
19
- This model was merged using the NuSLERP merge method using [EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) as a base.
20
-
21
- ### Models Merged
22
-
23
- The following models were included in the merge:
24
- * [allura-org/Teleut-7b](https://huggingface.co/allura-org/Teleut-7b)
25
- * [FourOhFour/Vapor_v2_7B](https://huggingface.co/FourOhFour/Vapor_v2_7B)
26
-
27
- ### Configuration
28
-
29
- The following YAML configuration was used to produce this model:
30
-
31
- ```yaml
32
- models:
33
- - model: allura-org/Teleut-7b
34
- parameters:
35
- weight: 0.6
36
- - model: FourOhFour/Vapor_v2_7B
37
- parameters:
38
- weight: 0.2
39
- - model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
40
- parameters:
41
- weight: 1.0
42
- merge_method: nuslerp
43
- base_model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
44
- parameters:
45
- normalize: true
46
- int8_mask: true
47
- nuslerp_flatten: false
48
- nuslerp_row_wise: true
49
- dtype: float16
50
- ```
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
4
+ - allura-org/Teleut-7b
5
+ - FourOhFour/Vapor_v2_7B
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+
11
+ ---
12
+
13
+ ![](https://huggingface.co/twosmoothslateslabs/Nemesia-Qwen-2.5-v1.0/blob/main/Nemesia.jpg)
14
+
15
+ Merge using the brand new NuSLERP method. Fresh out of the oven. Performance not guaranteed.
16
+
17
+ Uses the slightly-unstable EVA and two other finetunes I found. I also turned on both the NuSLERP exclusive mergekit options for fun.
18
+
19
+ # mergekit
20
+
21
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
22
+
23
+ ## Merge Details
24
+ ### Merge Method
25
+
26
+ This model was merged using the NuSLERP merge method using [EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) as a base.
27
+
28
+ ### Models Merged
29
+
30
+ The following models were included in the merge:
31
+ * [allura-org/Teleut-7b](https://huggingface.co/allura-org/Teleut-7b)
32
+ * [FourOhFour/Vapor_v2_7B](https://huggingface.co/FourOhFour/Vapor_v2_7B)
33
+
34
+ ### Configuration
35
+
36
+ The following YAML configuration was used to produce this model:
37
+
38
+ ```yaml
39
+ models:
40
+ - model: allura-org/Teleut-7b
41
+ parameters:
42
+ weight: 0.6
43
+ - model: FourOhFour/Vapor_v2_7B
44
+ parameters:
45
+ weight: 0.2
46
+ - model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
47
+ parameters:
48
+ weight: 1.0
49
+ merge_method: nuslerp
50
+ base_model: EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1
51
+ parameters:
52
+ normalize: true
53
+ int8_mask: true
54
+ nuslerp_flatten: false
55
+ nuslerp_row_wise: true
56
+ dtype: float16
57
+ ```