Update README.md
Browse files
README.md
CHANGED
@@ -845,7 +845,7 @@ a:hover .link-arrow {
|
|
845 |
<div class="creator-badge">
|
846 |
<span class="creator-label">Created by</span>
|
847 |
<a href="https://huggingface.co/suayptalha" target="_blank" class="creator-link">
|
848 |
-
<span class="creator-name">
|
849 |
<span class="creator-arrow">→</span>
|
850 |
</a>
|
851 |
</div>
|
@@ -856,30 +856,29 @@ a:hover .link-arrow {
|
|
856 |
<div class="info-header">
|
857 |
<h3>L3.3-Nevoria-R1-70b</h3>
|
858 |
<div class="model-tags">
|
859 |
-
<span class="model-tag">
|
860 |
-
<span class="model-tag">
|
|
|
861 |
<span class="model-tag">10b Parameters</span>
|
862 |
</div>
|
863 |
</div>
|
864 |
<div class="model-composition">
|
865 |
-
<h4>Model
|
866 |
<ul class="composition-list">
|
867 |
-
<li><span class="model-component"><a href="
|
868 |
-
<li><span class="model-component"><a href="https://huggingface.co/Sao10K/L3.3-70B-Euryale-v2.3" target="_blank">EURYALE-v2.3</a></span> Detailed scene descriptions</li>
|
869 |
-
<li><span class="model-component"><a href="https://huggingface.co/TheDrummer/Anubis-70B-v1" target="_blank">Anubis-v1</a></span> Enhanced prose details</li>
|
870 |
-
<li><span class="model-component"><a href="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B" target="_blank">Negative_LLAMA</a></span> Reduced positive bias</li>
|
871 |
-
<li><span class="model-component"><a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B" target="_blank">DeepSeek-R1-Distill-Llama-70B</a></span> Increased Intelligence / Dialog / Awareness</li>
|
872 |
-
<li><span class="model-component base-model"><a href="https://huggingface.co/nbeerbower/Llama-3.1-Nemotron-lorablated-70B" target="_blank">Nemotron-lorablated</a></span> Base model</li>
|
873 |
</ul>
|
874 |
</div>
|
875 |
<div class="model-description">
|
876 |
-
Maestro-10B is a 10 billion parameter model
|
877 |
-
|
878 |
-
|
879 |
-
|
880 |
-
|
881 |
-
|
882 |
-
|
|
|
|
|
|
|
883 |
</div>
|
884 |
</div>
|
885 |
<div class="support-section">
|
|
|
845 |
<div class="creator-badge">
|
846 |
<span class="creator-label">Created by</span>
|
847 |
<a href="https://huggingface.co/suayptalha" target="_blank" class="creator-link">
|
848 |
+
<span class="creator-name">suayptalha</span>
|
849 |
<span class="creator-arrow">→</span>
|
850 |
</a>
|
851 |
</div>
|
|
|
856 |
<div class="info-header">
|
857 |
<h3>L3.3-Nevoria-R1-70b</h3>
|
858 |
<div class="model-tags">
|
859 |
+
<span class="model-tag">suayptalha/Maestro-10B</span>
|
860 |
+
<span class="model-tag">arcee-ai/Virtuoso-Lite</span>
|
861 |
+
<span class="model-tag">DeepSeek-V3</span>
|
862 |
<span class="model-tag">10b Parameters</span>
|
863 |
</div>
|
864 |
</div>
|
865 |
<div class="model-composition">
|
866 |
+
<h4>Base Model</h4>
|
867 |
<ul class="composition-list">
|
868 |
+
<li><span class="model-component"><a href="arcee-ai/Virtuoso-Lite" target="_blank">EVA-LLAMA-0.1</a></span></li>
|
|
|
|
|
|
|
|
|
|
|
869 |
</ul>
|
870 |
</div>
|
871 |
<div class="model-description">
|
872 |
+
Maestro-10B is a 10 billion parameter model fine-tuned from Virtuoso-Lite,
|
873 |
+
a next-generation language model developed by arcee-ai. Virtuoso-Lite itself
|
874 |
+
is based on the Llama-3 architecture, distilled from Deepseek-v3 using
|
875 |
+
approximately 1.1 billion tokens/logits. This distillation process allows
|
876 |
+
Virtuoso-Lite to achieve robust performance with a smaller parameter count,
|
877 |
+
excelling in reasoning, code generation, and mathematical problem-solving.
|
878 |
+
Maestro-10B inherits these strengths from its base model, Virtuoso-Lite, and
|
879 |
+
further enhances them through fine-tuning on the OpenOrca dataset. This combination
|
880 |
+
of a distilled base model and targeted fine-tuning makes Maestro-10B a powerful and
|
881 |
+
efficient language model.
|
882 |
</div>
|
883 |
</div>
|
884 |
<div class="support-section">
|