InferenceIllusionist commited on
Commit
9aba670
·
verified ·
1 Parent(s): de6157c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +97 -0
README.md ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ tags:
4
+ - gguf
5
+ ---
6
+
7
+ # Sloppy-Wingman-8x7B-GGUF
8
+
9
+ Quantized from fp16 with love.
10
+
11
+ Uploading Q5_K_M for starters, other sizes available upon request.
12
+
13
+ See original model card details below.
14
+
15
+
16
+ ---
17
+ # Sloppy-Wingman-8x7B-hf
18
+ ![Sloppy Wingman](https://files.catbox.moe/7ay3me.png)
19
+
20
+ Big slop, good model.
21
+ Running better at slightly higher temp (1.1-ish) than usual, along with 0.05 MinP and 0.28 snoot.
22
+ Bog-standard ChatML works best imo, but Alpaca and Mixtral formats work (to some degree) too.
23
+
24
+ Parts:
25
+ ```yaml
26
+ models:
27
+ - model: mistralai/Mixtral-8x7B-v0.1+retrieval-bar/Mixtral-8x7B-v0.1_case-briefs
28
+ parameters:
29
+ weight: 0.33
30
+ - model: mistralai/Mixtral-8x7B-v0.1+wandb/Mixtral-8x7b-Remixtral
31
+ parameters:
32
+ weight: 0.33
33
+ merge_method: task_arithmetic
34
+ base_model: mistralai/Mixtral-8x7B-v0.1
35
+ dtype: float16
36
+ ```
37
+ and
38
+ ```yaml
39
+ models:
40
+ - model: mistralai/Mixtral-8x7B-Instruct-v0.1+/ai/LLM/tmp/pefts/daybreak-peft/mixtral-8x7b
41
+ parameters:
42
+ weight: 0.85
43
+ - model: notstoic/Nous-Hermes-2-Mixtruct-v0.1-8x7B-DPO-DARE_TIES
44
+ parameters:
45
+ weight: 0.25
46
+ - model: ycros/BagelWorldTour-8x7B
47
+ parameters:
48
+ weight: 0.1
49
+ merge_method: task_arithmetic
50
+ base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
51
+ dtype: float16
52
+ ```
53
+ SLERP:ed together as per below.
54
+
55
+ ---
56
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
57
+
58
+ ## Merge Details
59
+ ### Merge Method
60
+
61
+ This model was merged using the SLERP merge method.
62
+
63
+ ### Models Merged
64
+
65
+ The following models were included in the merge:
66
+ * ./02-friend2-instruct
67
+ * ./01-friend2-base
68
+
69
+ ### Configuration
70
+
71
+ The following YAML configuration was used to produce this model:
72
+
73
+ ```yaml
74
+ models:
75
+ - model: ./01-friend2-base
76
+ - model: ./02-friend2-instruct
77
+ merge_method: slerp
78
+ base_model: ./01-friend2-base
79
+ parameters:
80
+ t:
81
+ - value: 0.5
82
+ dtype: float16
83
+
84
+ ```
85
+
86
+ ```yaml
87
+ models:
88
+ - model: ./temp-output-base
89
+ - model: ./temp-output-instruct
90
+ merge_method: slerp
91
+ base_model: ./temp-output-base
92
+ parameters:
93
+ t:
94
+ - value: 0.5
95
+ dtype: float16
96
+
97
+ ```