exllamav2 quant for fixed version of mattshumer/Reflection-Llama-3.1-70B
Runs smoothly on 2x3090 with 48GB VRAM

All comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel: Buy Me A Coffee

Downloads last month
8
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including TeeZee/Reflection-Llama-3.1-70B-bpw4.0-h8-exl2