license: apache-2.0 | |
## Introduce | |
Quantizing the [UnicomLLM/Unichat-llama3-Chinese-8B-28K](https://huggingface.co/UnicomLLM/Unichat-llama3-Chinese-8B-28K) to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp. | |
license: apache-2.0 | |
## Introduce | |
Quantizing the [UnicomLLM/Unichat-llama3-Chinese-8B-28K](https://huggingface.co/UnicomLLM/Unichat-llama3-Chinese-8B-28K) to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp. | |