license: apache-2.0 | |
## Introduce | |
Quantizing the [shibing624/llama-3-8b-instruct-262k-chinese](https://huggingface.co/shibing624/llama-3-8b-instruct-262k-chinese) to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp. | |
license: apache-2.0 | |
## Introduce | |
Quantizing the [shibing624/llama-3-8b-instruct-262k-chinese](https://huggingface.co/shibing624/llama-3-8b-instruct-262k-chinese) to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp. | |