Elkhayyat17
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -24,8 +24,7 @@ quantized_by: Elkhayyat
|
|
24 |
|
25 |
|
26 |
|
27 |
-
#
|
28 |
-
- Model creator: [Meta](https://huggingface.co/meta-llama)
|
29 |
- Original model: [meta](meta-llama/Llama-2-7b-chat-hf)
|
30 |
|
31 |
<!-- description start -->
|
@@ -107,7 +106,7 @@ The following clients/libraries will automatically download models for you, prov
|
|
107 |
|
108 |
### In `text-generation-webui`
|
109 |
|
110 |
-
Under Download Model, you can enter the model repo: Elkhayyat17/llama2-Med-gguf and below it, a specific filename to download, such as:
|
111 |
|
112 |
Then click Download.
|
113 |
|
@@ -158,7 +157,7 @@ Windows CLI users: Use `set HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1` before running
|
|
158 |
Make sure you are using `llama.cpp` from commit [d0cee0d36d5be95a0d9088b674dbb27354107221](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
159 |
|
160 |
```shell
|
161 |
-
./main -ngl 32 -m
|
162 |
```
|
163 |
|
164 |
Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
|
@@ -231,7 +230,7 @@ from transformers import AutoTokenizer
|
|
231 |
import transformers
|
232 |
import torch
|
233 |
|
234 |
-
model = "
|
235 |
|
236 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
237 |
pipeline = transformers.pipeline(
|
@@ -256,25 +255,20 @@ for seq in sequences:
|
|
256 |
```
|
257 |
|
258 |
## Model Details
|
259 |
-
*Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs).
|
260 |
|
261 |
**Model Developers** Meta
|
262 |
|
263 |
**Variations**
|
264 |
## Training Data
|
265 |
|
266 |
-
All experiments reported here and the released models have been trained and fine-tuned using the same data as Llama 2 with different weights (see Section 2 and Table 1 in the [research paper](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/) for details).
|
267 |
|
268 |
## Evaluation Results
|
269 |
|
270 |
-
See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper.
|
271 |
|
272 |
|
273 |
## Ethical Considerations and Limitations
|
274 |
|
275 |
-
Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model.
|
276 |
|
277 |
-
Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-user-guide](https://ai.meta.com/llama/responsible-user-guide).
|
278 |
|
279 |
<!-- original-model-card end -->
|
280 |
---
|
|
|
24 |
|
25 |
|
26 |
|
27 |
+
# Llama2-Med - GGUF
|
|
|
28 |
- Original model: [meta](meta-llama/Llama-2-7b-chat-hf)
|
29 |
|
30 |
<!-- description start -->
|
|
|
106 |
|
107 |
### In `text-generation-webui`
|
108 |
|
109 |
+
Under Download Model, you can enter the model repo: Elkhayyat17/llama2-Med-gguf and below it, a specific filename to download, such as: ggml-model-Q4_K_M.gguf
|
110 |
|
111 |
Then click Download.
|
112 |
|
|
|
157 |
Make sure you are using `llama.cpp` from commit [d0cee0d36d5be95a0d9088b674dbb27354107221](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
158 |
|
159 |
```shell
|
160 |
+
./main -ngl 32 -m c ggml-model-Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "{prompt}"
|
161 |
```
|
162 |
|
163 |
Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
|
|
|
230 |
import transformers
|
231 |
import torch
|
232 |
|
233 |
+
model = "meta-llama/Llama-2-7b-chat-hf"
|
234 |
|
235 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
236 |
pipeline = transformers.pipeline(
|
|
|
255 |
```
|
256 |
|
257 |
## Model Details
|
|
|
258 |
|
259 |
**Model Developers** Meta
|
260 |
|
261 |
**Variations**
|
262 |
## Training Data
|
263 |
|
|
|
264 |
|
265 |
## Evaluation Results
|
266 |
|
|
|
267 |
|
268 |
|
269 |
## Ethical Considerations and Limitations
|
270 |
|
|
|
271 |
|
|
|
272 |
|
273 |
<!-- original-model-card end -->
|
274 |
---
|