File size: 925 Bytes
74253ab a3ccdcc 74253ab a266222 74253ab |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 |
---
license: llama3.1
---
# <a href="https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Nemotron-70b">Rombos-LLM-V2.6-Nemotron-70b</a> by Rombodawg
---
<p><h2>ExLlamaV2 Quantization</h2></p>
<p>Quantized with <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.2.2">ExLlamaV2 v0.2.3</a></p>
[2.2 Bits Per Weight](https://huggingface.co/UnstableLlama/Rombos-LLM-V2.6-Nemotron-70b-exl2/tree/2_2)
[4.65 Bits Per Weight](https://huggingface.co/UnstableLlama/Rombos-LLM-V2.6-Nemotron-70b-exl2/tree/4_65)
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/642cc1c253e76b4c2286c58e/lLAQ7FGTYDz0xT8rs9-ti.jpeg)
I applied the last step of my continuous finetuning method to the Nemotron-70b model from Nvidia. More details bellow:
- https://docs.google.com/document/d/1OjbjU5AOz4Ftn9xHQrX3oFQGhQ6RDUuXQipnQ9gn6tU/edit?usp=sharing
Quants: (Coming Soon)
Open-LLM-Leaderboard scores: (Coming soon) |