suayptalha commited on
Commit
1e2b618
·
verified ·
1 Parent(s): 22cdac2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -17
README.md CHANGED
@@ -845,7 +845,7 @@ a:hover .link-arrow {
845
  <div class="creator-badge">
846
  <span class="creator-label">Created by</span>
847
  <a href="https://huggingface.co/suayptalha" target="_blank" class="creator-link">
848
- <span class="creator-name">SteelSkull</span>
849
  <span class="creator-arrow">→</span>
850
  </a>
851
  </div>
@@ -856,30 +856,29 @@ a:hover .link-arrow {
856
  <div class="info-header">
857
  <h3>L3.3-Nevoria-R1-70b</h3>
858
  <div class="model-tags">
859
- <span class="model-tag">L3.3 = Llama 3.3</span>
860
- <span class="model-tag">R1 = DeepSeek-R1</span>
 
861
  <span class="model-tag">10b Parameters</span>
862
  </div>
863
  </div>
864
  <div class="model-composition">
865
- <h4>Model Composition</h4>
866
  <ul class="composition-list">
867
- <li><span class="model-component"><a href="https://huggingface.co/EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1" target="_blank">EVA-LLAMA-0.1</a></span> Storytelling capabilities</li>
868
- <li><span class="model-component"><a href="https://huggingface.co/Sao10K/L3.3-70B-Euryale-v2.3" target="_blank">EURYALE-v2.3</a></span> Detailed scene descriptions</li>
869
- <li><span class="model-component"><a href="https://huggingface.co/TheDrummer/Anubis-70B-v1" target="_blank">Anubis-v1</a></span> Enhanced prose details</li>
870
- <li><span class="model-component"><a href="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B" target="_blank">Negative_LLAMA</a></span> Reduced positive bias</li>
871
- <li><span class="model-component"><a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B" target="_blank">DeepSeek-R1-Distill-Llama-70B</a></span> Increased Intelligence / Dialog / Awareness</li>
872
- <li><span class="model-component base-model"><a href="https://huggingface.co/nbeerbower/Llama-3.1-Nemotron-lorablated-70B" target="_blank">Nemotron-lorablated</a></span> Base model</li>
873
  </ul>
874
  </div>
875
  <div class="model-description">
876
- Maestro-10B is a 10 billion parameter model built upon the Virtuoso-Lite architecture,
877
- enhanced by fine-tuning with 10,000 carefully selected data points from the OpenOrca dataset.
878
- This process imbues Maestro with improved capabilities in following instructions and generating
879
- creative, contextually relevant responses. While Virtuoso-Lite provides a strong foundation,
880
- the OpenOrca fine-tuning specifically targets enhancing the model's ability to understand and
881
- respond to complex prompts. This training regimen was chosen to optimize the model's performance
882
- in tasks requiring both comprehension and imaginative text generation.
 
 
 
883
  </div>
884
  </div>
885
  <div class="support-section">
 
845
  <div class="creator-badge">
846
  <span class="creator-label">Created by</span>
847
  <a href="https://huggingface.co/suayptalha" target="_blank" class="creator-link">
848
+ <span class="creator-name">suayptalha</span>
849
  <span class="creator-arrow">→</span>
850
  </a>
851
  </div>
 
856
  <div class="info-header">
857
  <h3>L3.3-Nevoria-R1-70b</h3>
858
  <div class="model-tags">
859
+ <span class="model-tag">suayptalha/Maestro-10B</span>
860
+ <span class="model-tag">arcee-ai/Virtuoso-Lite</span>
861
+ <span class="model-tag">DeepSeek-V3</span>
862
  <span class="model-tag">10b Parameters</span>
863
  </div>
864
  </div>
865
  <div class="model-composition">
866
+ <h4>Base Model</h4>
867
  <ul class="composition-list">
868
+ <li><span class="model-component"><a href="arcee-ai/Virtuoso-Lite" target="_blank">EVA-LLAMA-0.1</a></span></li>
 
 
 
 
 
869
  </ul>
870
  </div>
871
  <div class="model-description">
872
+ Maestro-10B is a 10 billion parameter model fine-tuned from Virtuoso-Lite,
873
+ a next-generation language model developed by arcee-ai. Virtuoso-Lite itself
874
+ is based on the Llama-3 architecture, distilled from Deepseek-v3 using
875
+ approximately 1.1 billion tokens/logits. This distillation process allows
876
+ Virtuoso-Lite to achieve robust performance with a smaller parameter count,
877
+ excelling in reasoning, code generation, and mathematical problem-solving.
878
+ Maestro-10B inherits these strengths from its base model, Virtuoso-Lite, and
879
+ further enhances them through fine-tuning on the OpenOrca dataset. This combination
880
+ of a distilled base model and targeted fine-tuning makes Maestro-10B a powerful and
881
+ efficient language model.
882
  </div>
883
  </div>
884
  <div class="support-section">