Nohobby commited on
Commit
47132bd
·
verified ·
1 Parent(s): 4cb71a1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +88 -13
README.md CHANGED
@@ -1,31 +1,107 @@
1
  ---
2
  base_model:
3
- - Nohobby/L3.3-Prikol-70B-v0.1a
 
 
 
4
  - Sao10K/70B-L3.3-Cirrus-x1
 
5
  library_name: transformers
6
  tags:
7
  - mergekit
8
  - merge
9
 
10
  ---
11
- # prikol
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
 
15
  ## Merge Details
16
- ### Merge Method
17
 
18
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using AbominationSnowPig as a base.
19
 
20
- ### Models Merged
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
 
22
- The following models were included in the merge:
23
- * [Nohobby/L3.3-Prikol-70B-v0.1a](https://huggingface.co/Nohobby/L3.3-Prikol-70B-v0.1a)
24
- * [Sao10K/70B-L3.3-Cirrus-x1](https://huggingface.co/Sao10K/70B-L3.3-Cirrus-x1)
25
 
26
- ### Configuration
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
- The following YAML configuration was used to produce this model:
29
 
30
  ```yaml
31
  base_model: AbominationSnowPig
@@ -34,5 +110,4 @@ dtype: bfloat16
34
  models:
35
  - model: Sao10K/70B-L3.3-Cirrus-x1
36
  - model: Nohobby/L3.3-Prikol-70B-v0.1a
37
-
38
- ```
 
1
  ---
2
  base_model:
3
+ - unsloth/Llama-3.3-70B-Instruct
4
+ - pankajmathur/orca_mini_v9_3_70B
5
+ - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
6
+ - Undi95/Sushi-v1.4
7
  - Sao10K/70B-L3.3-Cirrus-x1
8
+ - Nohobby/L3.3-Prikol-70B-v0.1a
9
  library_name: transformers
10
  tags:
11
  - mergekit
12
  - merge
13
 
14
  ---
15
+ # Prikol
16
+
17
+ > I don't even know anymore
18
+
19
+ ![Меня нужно изолировать от общества](https://files.catbox.moe/x9t3zo.png)
20
+
21
+ ### Overview
22
+
23
+ A merge of some Llama 3.3 models because um uh yeah
24
+
25
+ Went extra schizo on the recepie hoping for some extra fun result and... Well, I guess it's an overall improvement over the previous revision. It's a tiny bit smarter, has even more distinct swipes and nicer writing thanks to Cirrus, but for some reason it's damn sloppy.
26
+
27
+ I've published the second step of this merge as a separate model, and I'd say the results are more interesting, but not as usable as this one. https://huggingface.co/Nohobby/AbominationSnowPig
28
+
29
+ Propmt format: Llama3
30
+
31
+ Samplers: [This kinda works but I'm weird](https://files.catbox.moe/olsiei.json)
32
+
33
+ ### Quants
34
 
35
+ [Static](https://huggingface.co/mradermacher/L3.3-Prikol-70B-v0.2-GGUF)
36
 
37
  ## Merge Details
38
+ ### Merging Steps
39
 
40
+ ### Step1
41
 
42
+ ```yaml
43
+ models:
44
+ - model: pankajmathur/orca_mini_v9_3_70B
45
+ - model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
46
+ parameters:
47
+ weight: 1
48
+ density: 0.55
49
+ gamma: 0.03
50
+ - model: Undi95/Sushi-v1.4
51
+ parameters:
52
+ weight: 0.069
53
+ gamma: 0.001
54
+ density: 0.911
55
+ merge_method: breadcrumbs
56
+ base_model: pankajmathur/orca_mini_v9_3_70B
57
+ parameters:
58
+ int8_mask: true
59
+ rescale: true
60
+ normalize: true
61
+ dtype: bfloat16
62
+ tokenizer_source: base
63
+ ```
64
 
65
+ ### Step2 [(AbominationSnowPig)](https://huggingface.co/Nohobby/AbominationSnowPig)
 
 
66
 
67
+ ```yaml
68
+ dtype: bfloat16
69
+ tokenizer_source: base
70
+ merge_method: nuslerp
71
+ parameters:
72
+ nuslerp_row_wise: true
73
+ models:
74
+ - model: unsloth/Llama-3.3-70B-Instruct
75
+ parameters:
76
+ weight:
77
+ - filter: v_proj
78
+ value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0]
79
+ - filter: o_proj
80
+ value: [1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1]
81
+ - filter: up_proj
82
+ value: [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
83
+ - filter: gate_proj
84
+ value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0]
85
+ - filter: down_proj
86
+ value: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
87
+ - value: 0
88
+ - model: Step1
89
+ parameters:
90
+ weight:
91
+ - filter: v_proj
92
+ value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1]
93
+ - filter: o_proj
94
+ value: [0, 1, 0, 1, 1, 1, 1, 1, 0, 0, 0]
95
+ - filter: up_proj
96
+ value: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
97
+ - filter: gate_proj
98
+ value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1]
99
+ - filter: down_proj
100
+ value: [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
101
+ - value: 1
102
+ ```
103
 
104
+ ### Step3
105
 
106
  ```yaml
107
  base_model: AbominationSnowPig
 
110
  models:
111
  - model: Sao10K/70B-L3.3-Cirrus-x1
112
  - model: Nohobby/L3.3-Prikol-70B-v0.1a
113
+ ```