Original Model: https://huggingface.co/aetherwiing/MN-12B-Starcannon-v2

made with https://huggingface.co/FantasiaFoundry/GGUF-Quantization-Script

Models Q2_K_L, Q4_K_L, Q5_K_L, Q6_K_L, are using Q_8 output tensors and token embeddings

using bartowski's imatrix dataset

(happened to already have made these before mradermacher, probably use theirs instead)

Downloads last month
16
GGUF
Model size
12.2B params
Architecture
llama

4-bit

6-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for Reiterate3680/MN-12B-Starcannon-v2-GGUF

Quantized
(5)
this model