--- license: llama3.1 --- # Rombos-LLM-V2.6-Nemotron-70b by Rombodawg ---

ExLlamaV2 Quantization

Quantized with ExLlamaV2 v0.2.3

[2.2 Bits Per Weight](https://huggingface.co/UnstableLlama/Rombos-LLM-V2.6-Nemotron-70b-exl2/tree/2_2) [4.65 Bits Per Weight](https://huggingface.co/UnstableLlama/Rombos-LLM-V2.6-Nemotron-70b-exl2/tree/4_65) --- ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/642cc1c253e76b4c2286c58e/lLAQ7FGTYDz0xT8rs9-ti.jpeg) I applied the last step of my continuous finetuning method to the Nemotron-70b model from Nvidia. More details bellow: - https://docs.google.com/document/d/1OjbjU5AOz4Ftn9xHQrX3oFQGhQ6RDUuXQipnQ9gn6tU/edit?usp=sharing Quants: (Coming Soon) Open-LLM-Leaderboard scores: (Coming soon)