runtime error
Exit code: 1. Reason: ��████▊ | 3.42G/3.86G [00:07<00:00, 535MB/s][A model-00004-of-00005.safetensors: 100%|█████████▉| 3.86G/3.86G [00:08<00:00, 467MB/s] model-00005-of-00005.safetensors: 0%| | 0.00/1.09G [00:00<?, ?B/s][A model-00005-of-00005.safetensors: 3%|▎ | 31.5M/1.09G [00:01<00:36, 28.8MB/s][A model-00005-of-00005.safetensors: 40%|████ | 440M/1.09G [00:02<00:02, 237MB/s] [A model-00005-of-00005.safetensors: 100%|█████████▉| 1.09G/1.09G [00:02<00:00, 365MB/s] Loading checkpoint shards: 0%| | 0/5 [00:00<?, ?it/s][A Loading checkpoint shards: 0%| | 0/5 [00:03<?, ?it/s] Traceback (most recent call last): File "/home/user/app/app.py", line 3, in <module> from chatbot import model_inference, EXAMPLES, chatbot File "/home/user/app/chatbot.py", line 30, in <module> model = Qwen2VLForConditionalGeneration.from_pretrained(MODEL_ID, trust_remote_code=True, torch_dtype=torch.float16).to("cuda").eval() File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 273, in _wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4486, in from_pretrained ) = cls._load_pretrained_model( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4936, in _load_pretrained_model model_to_load.load_state_dict(state_dict, strict=False, assign=assign_params) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2581, in load_state_dict raise RuntimeError( RuntimeError: Error(s) in loading state_dict for Qwen2VLForConditionalGeneration: size mismatch for visual.merger.mlp.2.weight: copying a param with shape torch.Size([3584, 5120]) from checkpoint, the shape in current model is torch.Size([1280, 5120]). size mismatch for visual.merger.mlp.2.bias: copying a param with shape torch.Size([3584]) from checkpoint, the shape in current model is torch.Size([1280]).
Container logs:
Fetching error logs...