Xin Liu commited on
Commit
145dcde
1 Parent(s): bcecae4

Add Q2 model

Browse files

Signed-off-by: Xin Liu <[email protected]>

Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +54 -1
  3. Yi-34Bx2-MoE-60B-Q2_K.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,56 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: cloudyu/Yi-34Bx2-MoE-60B
3
+ license: cc-by-nc-4.0
4
+ model_creator: cloudyu
5
+ model_name: Yi 34Bx2 MoE 60B
6
+ model_type: mistral
7
+ quantized_by: Second State Inc.
8
+ tags:
9
+ - moe
10
  ---
11
+
12
+ <!-- header start -->
13
+ <!-- 200823 -->
14
+ <div style="width: auto; margin-left: auto; margin-right: auto">
15
+ <img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
16
+ </div>
17
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
18
+ <!-- header end -->
19
+
20
+ # Yi-34Bx2-MoE-60B-GGUF
21
+
22
+ ## Original Model
23
+
24
+ [cloudyu/Yi-34Bx2-MoE-60B](https://huggingface.co/cloudyu/Yi-34Bx2-MoE-60B)
25
+
26
+ ## Run with LlamaEdge
27
+
28
+ - LlamaEdge version: [v0.2.4](https://github.com/second-state/LlamaEdge/releases/tag/0.2.4)
29
+
30
+ - Prompt template
31
+
32
+ - Prompt type: `chatml`
33
+
34
+ - Prompt string
35
+
36
+ ```text
37
+ <|im_start|>system
38
+ {system_message}<|im_end|>
39
+ <|im_start|>user
40
+ {prompt}<|im_end|>
41
+ <|im_start|>assistant
42
+ ```
43
+
44
+ - Reverse prompt: `<|im_end|>`
45
+
46
+ - Run as LlamaEdge service
47
+
48
+ ```bash
49
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34B-Chat-ggml-model-q4_0.gguf llama-api-server.wasm -p chatml -r '<|im_end|>'
50
+ ```
51
+
52
+ - Run as LlamaEdge command app
53
+
54
+ ```bash
55
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34B-Chat-ggml-model-q4_0.gguf llama-chat.wasm -p chatml -r '<|im_end|>'
56
+ ```
Yi-34Bx2-MoE-60B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a51505254c1b7072b61e8f2b3e7ce380237d7da6e10fddbada956d6ca7503ced
3
+ size 22391949280