prsyahmi's picture
readme: Revert accidental changes
63774cb verified
|
raw
history blame
3.41 kB
---
base_model: mesolitica/malaysian-mistral-7b-32k-instructions-v2
inference: false
model_creator: mesolitica
model_name: Malaysian Mistral 7B 32k Instructions v2
model_type: mistral
pipeline_tag: text-generation
prompt_template: '<s>[INST] {prompt} [/INST]
'
quantized_by: prsyahmi
tags:
- finetuned
---
<!-- markdownlint-disable MD041 -->
<!-- header start --><!-- header end -->
# Malaysian Mistral 7B 32k Instructions - GGUF
- Model creator: [mesolotica](https://huggingface.co/mesolitica)
- Original model: [Malaysian Mistral 7B 32k Instructions v2](https://huggingface.co/mesolitica/malaysian-mistral-7b-32k-instructions-v2)
<!-- description start -->
## Pengenalan
Repo ini mengandungi model berformat GGUF untuk [mesolitica's Malaysian Mistral 7B 32k Instructions v2](https://huggingface.co/mesolitica/malaysian-mistral-7b-32k-instructions-v2).
GGUF adalah format kepada llama.cpp yang dibangunkan menggunakan C/C++ dimana pergantungan dengan software/library lain kurang menjadikan aplikasi ini ringan berbanding rata-rata aplikasi python.
<!-- description end -->
<!-- prompt-template start -->
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
<!-- prompt-template end -->
<!-- README_GGUF.md-provided-files start -->
## Fail yang diberikan
| Nama | Kaedah Quant | Saiz Fail |
| ---- | ---- | ---- |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q2_K.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q2_K.gguf) | Q2_K | 2.86 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q3_K_M.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q3_K_M.gguf) | Q3_K_M | 3.27 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q4_K_S.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q4_K_S.gguf) | Q4_K_S | 3.86 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q4_K_M.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q4_K_M.gguf) | Q4_K_M | 4.06 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q5_K_M.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q5_K_M.gguf) | Q5_K_M | 4.77 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q6_K.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.Q6_K.gguf) | Q6_K | 5.53 GB |
| [malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.fp16.gguf](https://huggingface.co/prsyahmi/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700-GGUF/blob/main/malaysian-mistral-7b-32k-instructions-v2-ckpt-1700.fp16.gguf) | FP16 | 13.5 GB |
<!-- README_GGUF.md-provided-files end -->
## Penghargaan
Terima kasih kepada Husein Zolkepli dan keseluruhan team [mesolotica](https://huggingface.co/mesolitica)!
Atas jasa mereka, kita dapat menggunakan atau mencuba AI peringkat tempatan.
<!-- footer end -->
-------
<!-- original-model-card start -->
<!-- original-model-card end -->