LoRA Fine-tuned Qwen3-32B

This model is a LoRA fine-tuned version of Qwen2.5-32B-Instruct.

Training Details

  • Training runtime: 13,587 seconds (3h 46m)
  • Training samples per second: 2.208
  • Final training loss: 0.664
  • Epochs: 1.0
  • Training steps: 1,161

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-32B-Instruct")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-32B-Instruct")
model = PeftModel.from_pretrained(base_model, "abhayesian/lora-qwen3-32b-docs")
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for abhayesian/lora-qwen3-32b-docs

Base model

Qwen/Qwen2.5-32B
Adapter
(44)
this model