giraffe176 commited on
Commit
f0d7b41
·
verified ·
1 Parent(s): 3eb05d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -5
README.md CHANGED
@@ -18,8 +18,8 @@ This model was merged using the SLERP merge method.
18
  ### Models Merged
19
 
20
  The following models were included in the merge:
21
- * /home/giraffe176/mergekit/workspace2
22
- * /home/giraffe176/mergekit/workspace1
23
 
24
  ### Configuration
25
 
@@ -27,12 +27,12 @@ The following YAML configuration was used to produce this model:
27
 
28
  ```yaml
29
  models:
30
- - model: /home/giraffe176/mergekit/workspace1
31
  layer_range: [0, 32]
32
- - model: /home/giraffe176/mergekit/workspace2
33
  layer_range: [0, 32]
34
  merge_method: slerp
35
- base_model: /home/giraffe176/mergekit/workspace1
36
  parameters:
37
  t:
38
  - filter: self_attn
@@ -41,6 +41,43 @@ parameters:
41
  value: [1, 0.5, 0.7, 0.3, 0]
42
  - value: 0.5
43
  dtype: bfloat16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
 
45
 
46
 
 
18
  ### Models Merged
19
 
20
  The following models were included in the merge:
21
+ * workspace2
22
+ * workspace1
23
 
24
  ### Configuration
25
 
 
27
 
28
  ```yaml
29
  models:
30
+ - model: cognitivecomputations/samantha-1.1-westlake-7b
31
  layer_range: [0, 32]
32
+ - model: NeverSleep/Noromaid-7B-0.4-DPO
33
  layer_range: [0, 32]
34
  merge_method: slerp
35
+ base_model: NeverSleep/Noromaid-7B-0.4-DPO
36
  parameters:
37
  t:
38
  - filter: self_attn
 
41
  value: [1, 0.5, 0.7, 0.3, 0]
42
  - value: 0.5
43
  dtype: bfloat16
44
+ name: workspace1
45
+ ---
46
+ models:
47
+ - model: teknium/OpenHermes-2.5-Mistral-7B
48
+ layer_range: [0, 32]
49
+ - model: Open-Orca/Mistral-7B-OpenOrca
50
+ layer_range: [0, 32]
51
+ merge_method: slerp
52
+ base_model: teknium/OpenHermes-2.5-Mistral-7B
53
+ parameters:
54
+ t:
55
+ - filter: self_attn
56
+ value: [0, 0.5, 0.3, 0.7, 1]
57
+ - filter: mlp
58
+ value: [1, 0.5, 0.7, 0.3, 0]
59
+ - value: 0.5
60
+ dtype: bfloat16
61
+ name: workspace2
62
+ ---
63
+ models:
64
+ - model: workspace1
65
+ layer_range: [0, 32]
66
+ - model: workspace2
67
+ layer_range: [0, 32]
68
+ merge_method: slerp
69
+ base_model: workspace1
70
+ parameters:
71
+ t:
72
+ - filter: self_attn
73
+ value: [0, 0.5, 0.3, 0.7, 1]
74
+ - filter: mlp
75
+ value: [1, 0.5, 0.7, 0.3, 0]
76
+ - value: 0.5
77
+ dtype: bfloat16
78
+
79
+
80
+
81
 
82
 
83