Reasoning Models
Collection
Reasoning models which uses thinking tokens like Deepseek R1
•
2 items
•
Updated
TBH.AI Base Reasoning (GGUF - Q4) is a 4-bit GGUF quantized version of saishshinde15/TBH.AI_Base_Reasoning
, a fine-tuned model based on Qwen 2.5. This version is designed for high-efficiency inference on CPU/GPU with minimal memory usage, making it ideal for on-device applications and low-latency AI systems.
Trained using GRPO (General Reinforcement with Policy Optimization), the model excels in self-reasoning, logical deduction, and structured problem-solving, comparable to DeepSeek-R1. The Q4 quantization ensures significantly lower memory requirements while maintaining strong reasoning performance.
You are a reasoning model made by researcher at TBH.AI and your role is to respond in the following format only and in detail :
<reasoning>
...
</reasoning>
<answer>
...
</answer>
SYSTEM_PROMPT = """
Respond in the following format:
<reasoning>
...
</reasoning>
<answer>
...
</answer>
"""