Transformers
GGUF
mergekit
Merge
Not-For-All-Audiences
llama-cpp
gguf-my-repo
Eval Results
Inference Endpoints
conversational
bamec66557 commited on
Commit
56bd45e
1 Parent(s): 853669e

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +164 -0
README.md ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - fr
5
+ - de
6
+ - es
7
+ - it
8
+ - pt
9
+ - ru
10
+ - zh
11
+ - ja
12
+ - ko
13
+ license: apache-2.0
14
+ library_name: transformers
15
+ tags:
16
+ - mergekit
17
+ - merge
18
+ - not-for-all-audiences
19
+ - llama-cpp
20
+ - gguf-my-repo
21
+ base_model: bamec66557/VICIOUS_MESH-12B-OMEGA
22
+ datasets:
23
+ - open-llm-leaderboard/bamec66557__VICIOUS_MESH-12B-BETA-details
24
+ - open-llm-leaderboard/bamec66557__VICIOUS_MESH-12B-OMEGA-details
25
+ model-index:
26
+ - name: VICIOUS_MESH-12B-OMEGA
27
+ results:
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: IFEval (0-Shot)
33
+ type: HuggingFaceH4/ifeval
34
+ args:
35
+ num_few_shot: 0
36
+ metrics:
37
+ - type: inst_level_strict_acc and prompt_level_strict_acc
38
+ value: 67
39
+ name: strict accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: BBH (3-Shot)
48
+ type: BBH
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 31.52
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: hendrycks/competition_math
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 12.08
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ args:
80
+ num_few_shot: 0
81
+ metrics:
82
+ - type: acc_norm
83
+ value: 8.72
84
+ name: acc_norm
85
+ source:
86
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MuSR (0-shot)
93
+ type: TAUR-Lab/MuSR
94
+ args:
95
+ num_few_shot: 0
96
+ metrics:
97
+ - type: acc_norm
98
+ value: 14.54
99
+ name: acc_norm
100
+ source:
101
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: MMLU-PRO (5-shot)
108
+ type: TIGER-Lab/MMLU-Pro
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 29.74
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bamec66557/VICIOUS_MESH-12B-OMEGA
119
+ name: Open LLM Leaderboard
120
+ ---
121
+
122
+ # bamec66557/VICIOUS_MESH-12B-OMEGA-Q4_K_M-GGUF
123
+ This model was converted to GGUF format from [`bamec66557/VICIOUS_MESH-12B-OMEGA`](https://huggingface.co/bamec66557/VICIOUS_MESH-12B-OMEGA) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
124
+ Refer to the [original model card](https://huggingface.co/bamec66557/VICIOUS_MESH-12B-OMEGA) for more details on the model.
125
+
126
+ ## Use with llama.cpp
127
+ Install llama.cpp through brew (works on Mac and Linux)
128
+
129
+ ```bash
130
+ brew install llama.cpp
131
+
132
+ ```
133
+ Invoke the llama.cpp server or the CLI.
134
+
135
+ ### CLI:
136
+ ```bash
137
+ llama-cli --hf-repo bamec66557/VICIOUS_MESH-12B-OMEGA-Q4_K_M-GGUF --hf-file vicious_mesh-12b-omega-q4_k_m.gguf -p "The meaning to life and the universe is"
138
+ ```
139
+
140
+ ### Server:
141
+ ```bash
142
+ llama-server --hf-repo bamec66557/VICIOUS_MESH-12B-OMEGA-Q4_K_M-GGUF --hf-file vicious_mesh-12b-omega-q4_k_m.gguf -c 2048
143
+ ```
144
+
145
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
146
+
147
+ Step 1: Clone llama.cpp from GitHub.
148
+ ```
149
+ git clone https://github.com/ggerganov/llama.cpp
150
+ ```
151
+
152
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
153
+ ```
154
+ cd llama.cpp && LLAMA_CURL=1 make
155
+ ```
156
+
157
+ Step 3: Run inference through the main binary.
158
+ ```
159
+ ./llama-cli --hf-repo bamec66557/VICIOUS_MESH-12B-OMEGA-Q4_K_M-GGUF --hf-file vicious_mesh-12b-omega-q4_k_m.gguf -p "The meaning to life and the universe is"
160
+ ```
161
+ or
162
+ ```
163
+ ./llama-server --hf-repo bamec66557/VICIOUS_MESH-12B-OMEGA-Q4_K_M-GGUF --hf-file vicious_mesh-12b-omega-q4_k_m.gguf -c 2048
164
+ ```