Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_MLX-4bit

The Model Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_MLX-4bit was converted to MLX format from Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO using mlx-lm version 0.21.1.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_MLX-4bit")

prompt = "hello"

if tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        messages, add_generation_prompt=True
    )

response = generate(model, tokenizer, prompt=prompt, verbose=True)
Downloads last month
142
Safetensors
Model size
241M params
Tensor type
FP16
ยท
U32
ยท
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_MLX-4bit

Spaces using Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_MLX-4bit 3