This is a more recent quantized GGUF version of OpenOrca-Platypus-13B quantized to 4_K-M respecting the new format from (https://github.com/ggerganov/llama.cpp).
(link to the original model : https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B)
Instruction template :
### Instruction:
<prompt> (without the <>)
### Response:
- Downloads last month
- 18
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
HF Inference deployability: The model has no library tag.