metadata
language:
- en
base_model:
- arcee-ai/Virtuoso-Lite
datasets:
- Open-Orca/OpenOrca
pipeline_tag: text-generation
license: other
license_name: falcon-llm-license
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
library_name: transformers
tags:
- unsloth
- trl
- sft
Maestro-10B
![Model banner](/suayptalha/Maestro-10B/resolve/main/Maestro-Logo.png)
Created by
suayptalha
→
Model Information
Maestro-10B
Base Model
Maestro-10B is a 10 billion parameter model fine-tuned from Virtuoso-Lite,
a next-generation language model developed by arcee-ai. Virtuoso-Lite itself
is based on the Llama-3 architecture, distilled from Deepseek-v3 using
approximately 1.1 billion tokens/logits. This distillation process allows
Virtuoso-Lite to achieve robust performance with a smaller parameter count,
excelling in reasoning, code generation, and mathematical problem-solving.
Maestro-10B inherits these strengths from its base model, Virtuoso-Lite, and
further enhances them through fine-tuning on the OpenOrca dataset. This combination
of a distilled base model and targeted fine-tuning makes Maestro-10B a powerful and
efficient language model.
Loss Graph
![Model banner](/suayptalha/Maestro-10B/resolve/main/loss.png)