ParmyJack commited on
Commit
88a488f
·
verified ·
1 Parent(s): 5f5d924

Add model card

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md CHANGED
@@ -137,3 +137,85 @@ quantized_by: Suparious
137
  ---
138
  # Locutusque/Hyperion-1.5-Mistral-7B AWQ
139
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
137
  ---
138
  # Locutusque/Hyperion-1.5-Mistral-7B AWQ
139
 
140
+ **UPLOAD IN PROGRESS**
141
+
142
+ - Model creator: [Locutusque](https://huggingface.co/Locutusque)
143
+ - Original model: [Hyperion-1.5-Mistral-7B](https://huggingface.co/Locutusque/Hyperion-1.5-Mistral-7B)
144
+
145
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6437292ecd93f4c9a34b0d47/1lL97kzuxqykXGUT6F593.png)
146
+
147
+ ## Model Summary
148
+
149
+ `Locutusque/Hyperion-1.5-Mistral-7B` is a state-of-the-art language model fine-tuned on the Hyperion dataset for advanced reasoning across scientific domains. This model is designed to handle complex inquiries and instructions, leveraging the diverse and rich information contained in the Hyperion dataset. Its primary use cases include but are not limited to complex question answering, conversational understanding, code generation, medical text comprehension, mathematical reasoning, and logical reasoning.
150
+
151
+ ## How to use
152
+
153
+ ### Install the necessary packages
154
+
155
+ ```bash
156
+ pip install --upgrade autoawq autoawq-kernels
157
+ ```
158
+
159
+ ### Example Python code
160
+
161
+ ```python
162
+ from awq import AutoAWQForCausalLM
163
+ from transformers import AutoTokenizer, TextStreamer
164
+
165
+ model_path = "solidrust/Hyperion-1.5-Mistral-7B-AWQ"
166
+ system_message = "You are Hyperion, incarnated as a powerful AI."
167
+
168
+ # Load model
169
+ model = AutoAWQForCausalLM.from_quantized(model_path,
170
+ fuse_layers=True)
171
+ tokenizer = AutoTokenizer.from_pretrained(model_path,
172
+ trust_remote_code=True)
173
+ streamer = TextStreamer(tokenizer,
174
+ skip_prompt=True,
175
+ skip_special_tokens=True)
176
+
177
+ # Convert prompt to tokens
178
+ prompt_template = """\
179
+ <|im_start|>system
180
+ {system_message}<|im_end|>
181
+ <|im_start|>user
182
+ {prompt}<|im_end|>
183
+ <|im_start|>assistant"""
184
+
185
+ prompt = "You're standing on the surface of the Earth. "\
186
+ "You walk one mile south, one mile west and one mile north. "\
187
+ "You end up exactly where you started. Where are you?"
188
+
189
+ tokens = tokenizer(prompt_template.format(system_message=system_message,prompt=prompt),
190
+ return_tensors='pt').input_ids.cuda()
191
+
192
+ # Generate output
193
+ generation_output = model.generate(tokens,
194
+ streamer=streamer,
195
+ max_new_tokens=512)
196
+
197
+ ```
198
+
199
+ ### About AWQ
200
+
201
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
202
+
203
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
204
+
205
+ It is supported by:
206
+
207
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
208
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
209
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
210
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
211
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
212
+
213
+ ## Prompt template: ChatML
214
+
215
+ ```plaintext
216
+ <|im_start|>system
217
+ {system_message}<|im_end|>
218
+ <|im_start|>user
219
+ {prompt}<|im_end|>
220
+ <|im_start|>assistant
221
+ ```