codegood commited on
Commit
3dea944
1 Parent(s): 93b1128

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: HPAI-BSC/Llama3-Aloe-8B-Alpha
3
+ datasets:
4
+ - argilla/dpo-mix-7k
5
+ - nvidia/HelpSteer
6
+ - jondurbin/airoboros-3.2
7
+ - hkust-nlp/deita-10k-v0
8
+ - LDJnr/Capybara
9
+ - HPAI-BSC/CareQA
10
+ - GBaker/MedQA-USMLE-4-options
11
+ - lukaemon/mmlu
12
+ - bigbio/pubmed_qa
13
+ - openlifescienceai/medmcqa
14
+ - bigbio/med_qa
15
+ - HPAI-BSC/better-safe-than-sorry
16
+ - HPAI-BSC/pubmedqa-cot
17
+ - HPAI-BSC/medmcqa-cot
18
+ - HPAI-BSC/medqa-cot
19
+ language:
20
+ - en
21
+ library_name: transformers
22
+ license: cc-by-nc-4.0
23
+ pipeline_tag: question-answering
24
+ tags:
25
+ - biology
26
+ - medical
27
+ - llama-cpp
28
+ - gguf-my-repo
29
+ ---
30
+
31
+ # codegood/Llama3-Aloe-8B-Alpha-Q4_K_M-GGUF
32
+ This model was converted to GGUF format from [`HPAI-BSC/Llama3-Aloe-8B-Alpha`](https://huggingface.co/HPAI-BSC/Llama3-Aloe-8B-Alpha) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
33
+ Refer to the [original model card](https://huggingface.co/HPAI-BSC/Llama3-Aloe-8B-Alpha) for more details on the model.
34
+
35
+ ## Use with llama.cpp
36
+ Install llama.cpp through brew (works on Mac and Linux)
37
+
38
+ ```bash
39
+ brew install llama.cpp
40
+
41
+ ```
42
+ Invoke the llama.cpp server or the CLI.
43
+
44
+ ### CLI:
45
+ ```bash
46
+ llama-cli --hf-repo codegood/Llama3-Aloe-8B-Alpha-Q4_K_M-GGUF --hf-file llama3-aloe-8b-alpha-q4_k_m.gguf -p "The meaning to life and the universe is"
47
+ ```
48
+
49
+ ### Server:
50
+ ```bash
51
+ llama-server --hf-repo codegood/Llama3-Aloe-8B-Alpha-Q4_K_M-GGUF --hf-file llama3-aloe-8b-alpha-q4_k_m.gguf -c 2048
52
+ ```
53
+
54
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
55
+
56
+ Step 1: Clone llama.cpp from GitHub.
57
+ ```
58
+ git clone https://github.com/ggerganov/llama.cpp
59
+ ```
60
+
61
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
62
+ ```
63
+ cd llama.cpp && LLAMA_CURL=1 make
64
+ ```
65
+
66
+ Step 3: Run inference through the main binary.
67
+ ```
68
+ ./llama-cli --hf-repo codegood/Llama3-Aloe-8B-Alpha-Q4_K_M-GGUF --hf-file llama3-aloe-8b-alpha-q4_k_m.gguf -p "The meaning to life and the universe is"
69
+ ```
70
+ or
71
+ ```
72
+ ./llama-server --hf-repo codegood/Llama3-Aloe-8B-Alpha-Q4_K_M-GGUF --hf-file llama3-aloe-8b-alpha-q4_k_m.gguf -c 2048
73
+ ```