File size: 222 Bytes
261a189 |
1 2 3 4 5 6 7 8 |
---
license: apache-2.0
---
## Introduce
Quantizing the [shibing624/llama-3-8b-instruct-262k-chinese](https://huggingface.co/shibing624/llama-3-8b-instruct-262k-chinese) to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp.
|