aashish1904 commited on
Commit
9a27eab
·
verified ·
1 Parent(s): 362a2ee

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +104 -0
README.md ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - akjindal53244/Llama-3.1-Storm-8B
6
+ - Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
7
+ library_name: transformers
8
+ tags:
9
+ - merge
10
+ - llama
11
+ - not-for-all-audiences
12
+
13
+ ---
14
+
15
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
16
+
17
+
18
+ # QuantFactory/L3-Umbral-Storm-8B-t0.0001-GGUF
19
+ This is quantized version of [v000000/L3-Umbral-Storm-8B-t0.0001](https://huggingface.co/v000000/L3-Umbral-Storm-8B-t0.0001) created using llama.cpp
20
+
21
+ # Original Model Card
22
+
23
+
24
+ # Llama-3-Umbral-Storm-8B (8K)
25
+
26
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/79tIjC6Ykm4rlwOHa9uzZ.png)
27
+
28
+ RP model, "L3-Umbral-Mind-v2.0" as a base, nearswapped with one of the smartest L3.1 models "Storm".
29
+
30
+ * Warning: Based on Mopey-Mule so it should be negative, don't use this model for any truthful information or advice.
31
+
32
+ * <b>----></b>[ GGUF Q8 static](https://huggingface.co/v000000/L3-Umbral-Storm-8B-t0.0001-Q8_0-GGUF)
33
+
34
+ # Thank you mradermacher for the quants!
35
+
36
+ * [GGUFs](https://huggingface.co/mradermacher/L3-Umbral-Storm-8B-t0.0001-GGUF)
37
+ * [GGUFs imatrix](https://huggingface.co/mradermacher/L3-Umbral-Storm-8B-t0.0001-i1-GGUF)
38
+
39
+ -------------------------------------------------------------------------------
40
+
41
+ ## merge
42
+
43
+ This is a merge of pre-trained language models.
44
+
45
+ ## Merge Details
46
+
47
+ This model is on the Llama-3 arch with Llama-3.1 merged in, so it has 8k context length. But could possibly be extended slightly with RoPE due to the L3.1 layers.
48
+
49
+ ### Merge Method
50
+
51
+ This model was merged using the <b>NEARSWAP t0.0001</b> merge algorithm.
52
+
53
+ ### Models Merged
54
+
55
+ The following models were included in the merge:
56
+ * Base Model: [Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B)
57
+ * [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B)
58
+
59
+ ### Configuration
60
+
61
+ ```yaml
62
+ slices:
63
+ - sources:
64
+ - model: Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
65
+ layer_range: [0, 32]
66
+ - model: akjindal53244/Llama-3.1-Storm-8B
67
+ layer_range: [0, 32]
68
+ merge_method: nearswap
69
+ base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
70
+ parameters:
71
+ t:
72
+ - value: 0.0001
73
+ dtype: bfloat16
74
+ ```
75
+
76
+ # Prompt Template:
77
+ ```bash
78
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
79
+
80
+ {system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
81
+
82
+ {input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
83
+
84
+ {output}<|eot_id|>
85
+
86
+ ```
87
+
88
+ Credit to Alchemonaut:
89
+
90
+ ```python
91
+ def lerp(a, b, t):
92
+ return a * (1 - t) + b * t
93
+
94
+ def nearswap(v0, v1, t):
95
+ lweight = np.abs(v0 - v1)
96
+ with np.errstate(divide='ignore', invalid='ignore'):
97
+ lweight = np.where(lweight != 0, t / lweight, 1.0)
98
+ lweight = np.nan_to_num(lweight, nan=1.0, posinf=1.0, neginf=1.0)
99
+ np.clip(lweight, a_min=0.0, a_max=1.0, out=lweight)
100
+ return lerp(v0, v1, lweight)
101
+ ```
102
+
103
+ Credit to Numbra for idea.
104
+