runtime error
Exit code: 1. Reason: ] model-00002-of-00002.safetensors: 0%| | 0.00/1.05G [00:00<?, ?B/s][A model-00002-of-00002.safetensors: 1%| | 12.6M/1.05G [00:01<01:37, 10.6MB/s][A model-00002-of-00002.safetensors: 4%|▍ | 44.0M/1.05G [00:02<00:56, 17.8MB/s][A model-00002-of-00002.safetensors: 50%|█████ | 526M/1.05G [00:03<00:02, 198MB/s] [A model-00002-of-00002.safetensors: 100%|█████████▉| 1.05G/1.05G [00:04<00:00, 259MB/s] [2025-08-25 16:15:22] INFO modeling.py:1592: Based on the current allocation process, no modules could be assigned to the following devices due to insufficient memory: - cpu: 1293967360 bytes required These minimum requirements are specific to this allocation attempt and may vary. Consider increasing the available memory for these devices to at least the specified minimum, or adjusting the model config. Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s][A Loading checkpoint shards: 100%|██████████| 2/2 [00:00<00:00, 1378.35it/s] generation_config.json: 0%| | 0.00/184 [00:00<?, ?B/s][A generation_config.json: 100%|██████████| 184/184 [00:00<00:00, 1.12MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 16, in <module> model = AutoModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 571, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 309, in _wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4668, in from_pretrained dispatch_model(model, **device_map_kwargs) File "/usr/local/lib/python3.10/site-packages/accelerate/big_modeling.py", line 504, in dispatch_model raise ValueError( ValueError: You are trying to offload the whole model to the disk. Please use the `disk_offload` function instead.
Container logs:
Fetching error logs...