Adding `safetensors` variant of this model
#2 opened about 2 months ago
by
SFconvertbot

what about an quantized version so we can load in Exlama with large context size?
#1 opened over 1 year ago
by
DQ83