runtime error

Exit code: 1. Reason: tokenizer_config.json: 0%| | 0.00/578 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 578/578 [00:00<00:00, 2.42MB/s] tokenizer.json: 0%| | 0.00/1.33M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 1.33M/1.33M [00:00<00:00, 18.4MB/s] special_tokens_map.json: 0%| | 0.00/439 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 439/439 [00:00<00:00, 4.73MB/s] config.json: 0%| | 0.00/1.27k [00:00<?, ?B/s] config.json: 100%|██████████| 1.27k/1.27k [00:00<00:00, 12.1MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 9, in <module> model = AutoModelForCausalLM.from_pretrained(model_id) File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 531, in from_pretrained config, kwargs = AutoConfig.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 1143, in from_pretrained return config_class.from_dict(config_dict, **unused_kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/configuration_utils.py", line 737, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/site-packages/transformers/models/mpt/configuration_mpt.py", line 209, in __init__ self.attn_config = MptAttentionConfig(**attn_config) File "/usr/local/lib/python3.10/site-packages/transformers/models/mpt/configuration_mpt.py", line 98, in __init__ raise ValueError( ValueError: `attn_type` has to be either `multihead_attention` or `multiquery_attention`. Received: grouped_query_attention

Container logs:

Fetching error logs...