OSError with Llama3.2-3B-Instruct-QLORA_INT4_EO8
#2
by
steveq
- opened
When trying to run Llama3.2-3B-Instruct-QLORA_INT4_EO8
, I'm getting the error:
OSError: meta-llama/Llama-3.2-3B-Instruct-QLORA_INT4_EO8 does not appear to have a file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt or flax_model.msgpack.
I've tried using transformers pipeline
and also AutoModelForCausalLM
to pull the model but get the error in both cases.
The weights were uploaded in their "original" (meta) format, and they need to be translated to the HuggingFace format to be used with the pipelines. I'm sure they will upload the reformatted version soon.
Following this. (I thought I was missing something esoteric and GGUF related at first :D)