Velma-9b

Model Overview

Velma-9b is a fine-tuned version of google/gemma-2-9b-it, optimized to improve reasoning capabilities. The model has been trained using the GSM8K dataset, a benchmark dataset designed to enhance mathematical and logical reasoning skills in language models.

This fine-tuning process allows Velma-9b to excel at structured problem-solving, step-by-step reasoning, and logical inference, making it an ideal choice for tasks requiring in-depth analytical thinking.

Features

  • Fine-Tuned on GSM8K: Enhanced for mathematical reasoning and step-by-step logical problem-solving.
  • Transformer-Based: Built on the powerful gemma-2-9b-it architecture.
  • Optimized for SFT (Supervised Fine-Tuning): Fine-tuned using TRL (Transformer Reinforcement Learning) for improved inference and structured output generation.
  • Efficient Deployment: Compatible with transformers and supports GPU acceleration for fast inference.

Quick Start

Use the following code to generate text using Velma-9b:

from transformers import pipeline

# Initialize the pipeline
generator = pipeline("text-generation", model="AmirMohseni/Velma-9b", device="cuda")

# Example prompt
question = "If you had a time machine but could only go to the past or the future once and never return, which would you choose and why?"

# Generate output
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])

Training Procedure

Velma-9b was fine-tuned using the Supervised Fine-Tuning (SFT) approach with the GSM8K dataset. This dataset contains high-quality mathematical and logical reasoning problems that help models develop structured thinking and problem-solving skills.

Training Details

  • Base Model: google/gemma-2-9b-it
  • Dataset Used: GSM8K
  • Fine-Tuning Method: SFT(STaR) using TRL
  • Optimization Objective: Supervised fine-tuning to enhance structured reasoning

Framework Versions

  • TRL: 0.12.1
  • Transformers: 4.46.3
  • PyTorch: 2.1.1
  • Datasets: 3.1.0
  • Tokenizers: 0.20.3

Use Cases

Velma-9b is best suited for tasks requiring structured reasoning and logical inference:

  • Mathematical & Logical Reasoning Tasks: Providing step-by-step explanations and structured problem-solving.
  • AI Research & Experimentation: Evaluating fine-tuning strategies for reasoning-focused language models.
Downloads last month
12
Safetensors
Model size
9.24B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for AmirMohseni/Velma-9b

Base model

google/gemma-2-9b
Finetuned
(144)
this model

Dataset used to train AmirMohseni/Velma-9b