newsletter commited on
Commit
d8a6fd9
1 Parent(s): 5575b36

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +25 -15
README.md CHANGED
@@ -13,39 +13,49 @@ tags:
13
  - malteos/hermeo-7b
14
  - llama-cpp
15
  - gguf-my-repo
16
- base_model:
17
- - DiscoResearch/DiscoLM_German_7b_v1
18
- - DRXD1000/Phoenix
19
- - VAGOsolutions/SauerkrautLM-7b-v1-mistral
20
- - malteos/hermeo-7b
21
  ---
22
 
23
  # newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF
24
  This model was converted to GGUF format from [`mayflowergmbh/Wiedervereinigung-7b-dpo`](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
25
  Refer to the [original model card](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo) for more details on the model.
26
- ## Use with llama.cpp
27
 
28
- Install llama.cpp through brew.
 
29
 
30
  ```bash
31
- brew install ggerganov/ggerganov/llama.cpp
 
32
  ```
33
  Invoke the llama.cpp server or the CLI.
34
 
35
- CLI:
36
-
37
  ```bash
38
- llama-cli --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --model wiedervereinigung-7b-dpo.Q6_K.gguf -p "The meaning to life and the universe is"
39
  ```
40
 
41
- Server:
42
-
43
  ```bash
44
- llama-server --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --model wiedervereinigung-7b-dpo.Q6_K.gguf -c 2048
45
  ```
46
 
47
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
48
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  ```
50
- git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m wiedervereinigung-7b-dpo.Q6_K.gguf -n 128
51
  ```
 
13
  - malteos/hermeo-7b
14
  - llama-cpp
15
  - gguf-my-repo
16
+ base_model: mayflowergmbh/Wiedervereinigung-7b-dpo
 
 
 
 
17
  ---
18
 
19
  # newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF
20
  This model was converted to GGUF format from [`mayflowergmbh/Wiedervereinigung-7b-dpo`](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
21
  Refer to the [original model card](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo) for more details on the model.
 
22
 
23
+ ## Use with llama.cpp
24
+ Install llama.cpp through brew (works on Mac and Linux)
25
 
26
  ```bash
27
+ brew install llama.cpp
28
+
29
  ```
30
  Invoke the llama.cpp server or the CLI.
31
 
32
+ ### CLI:
 
33
  ```bash
34
+ llama --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --hf-file wiedervereinigung-7b-dpo-q6_k.gguf -p "The meaning to life and the universe is"
35
  ```
36
 
37
+ ### Server:
 
38
  ```bash
39
+ llama-server --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --hf-file wiedervereinigung-7b-dpo-q6_k.gguf -c 2048
40
  ```
41
 
42
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
43
 
44
+ Step 1: Clone llama.cpp from GitHub.
45
+ ```
46
+ git clone https://github.com/ggerganov/llama.cpp
47
+ ```
48
+
49
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
50
+ ```
51
+ cd llama.cpp && LLAMA_CURL=1 make
52
+ ```
53
+
54
+ Step 3: Run inference through the main binary.
55
+ ```
56
+ ./main --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --hf-file wiedervereinigung-7b-dpo-q6_k.gguf -p "The meaning to life and the universe is"
57
+ ```
58
+ or
59
  ```
60
+ ./server --hf-repo newsletter/Wiedervereinigung-7b-dpo-Q6_K-GGUF --hf-file wiedervereinigung-7b-dpo-q6_k.gguf -c 2048
61
  ```