Spaces:
Running
on
A10G
Running
on
A10G
Error converting Video-LLaVA-7B
#71
by
nahidalam
- opened
I tried to use https://huggingface.co/spaces/ggml-org/gguf-my-repo for creating GGUF for this repo https://huggingface.co/LanguageBind/Video-LLaVA-7B-hf/tree/main but got this error
Error: Error converting to fp16: b'INFO:hf-to-gguf:Loading model: Video-LLaVA-7B-hf\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2774, in \n main()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2752, in main\n model_class = Model.from_model_architecture(hparams["architectures"][0])\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 370, in from_model_architecture\n raise NotImplementedError(f'Architecture {arch!r} not supported!') from None\nNotImplementedError: Architecture 'VideoLlavaForConditionalGeneration' not supported!\n'
Is this likely due to https://huggingface.co/spaces/ggml-org/gguf-my-repo having older version of transformer?
Hi @nahidalam - I don't think this is a transformers issue - for context - we rebuild from source every couple of hours.
The model is not supported by the quantisation script (as you can see in the error stack). I'm not sure if it is even supported by llama.cpp :/
reach-vb
changed discussion status to
closed