apepkuss79 commited on
Commit
e93a836
1 Parent(s): cfe0078

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
  <!-- header start -->
13
  <!-- 200823 -->
14
  <div style="width: auto; margin-left: auto; margin-right: auto">
15
- <img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
16
  </div>
17
  <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
18
  <!-- header end -->
@@ -25,7 +25,7 @@ tags:
25
 
26
  ## Run with LlamaEdge
27
 
28
- - LlamaEdge version: [v0.2.8](https://github.com/second-state/LlamaEdge/releases/tag/0.2.8) and above
29
 
30
  - Prompt template
31
 
@@ -46,13 +46,13 @@ tags:
46
  - Run as LlamaEdge service
47
 
48
  ```bash
49
- wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34B-Chat-ggml-model-q4_0.gguf llama-api-server.wasm -p chatml -r '<|im_end|>'
50
  ```
51
 
52
  - Run as LlamaEdge command app
53
 
54
  ```bash
55
- wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34B-Chat-ggml-model-q4_0.gguf llama-chat.wasm -p chatml -r '<|im_end|>'
56
  ```
57
 
58
  ## Quantized GGUF Models
 
12
  <!-- header start -->
13
  <!-- 200823 -->
14
  <div style="width: auto; margin-left: auto; margin-right: auto">
15
+ <img src="https://github.com/LlamaEdge/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
16
  </div>
17
  <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
18
  <!-- header end -->
 
25
 
26
  ## Run with LlamaEdge
27
 
28
+ - LlamaEdge version: [v0.2.8](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.2.8) and above
29
 
30
  - Prompt template
31
 
 
46
  - Run as LlamaEdge service
47
 
48
  ```bash
49
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34Bx2-MoE-60B-Q5_K_M.gguf llama-api-server.wasm -p chatml -r '<|im_end|>'
50
  ```
51
 
52
  - Run as LlamaEdge command app
53
 
54
  ```bash
55
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:Yi-34Bx2-MoE-60B-Q5_K_M.gguf llama-chat.wasm -p chatml -r '<|im_end|>'
56
  ```
57
 
58
  ## Quantized GGUF Models