Llama-3.1-Nemotron-70B-Reward-HF-INT4

This consumes about 52 GB VRAM.

Please refer to the original license and documentation:
https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Reward-HF

Downloads last month
5
Safetensors
Model size
37.4B params
Tensor type
F32
·
BF16
·
U8
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including mikeriess/Llama-3.1-Nemotron-70B-Reward-HF-INT4