Spaces:
Sleeping
Sleeping
===== Application Startup at 2024-07-13 21:48:51 ===== | |
Running on local URL: http://0.0.0.0:7860 | |
To create a public link, set `share=True` in `launch()`. | |
Traceback (most recent call last): | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 304, in hf_raise_for_status | |
response.raise_for_status() | |
File "/usr/local/lib/python3.10/site-packages/requests/models.py", line 1024, in raise_for_status | |
raise HTTPError(http_error_msg, response=self) | |
requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://api-inference.huggingface.co/models/mistralai/Mixtral-8x7B-Instruct-v0.1 | |
The above exception was the direct cause of the following exception: | |
Traceback (most recent call last): | |
File "/usr/local/lib/python3.10/site-packages/gradio/queueing.py", line 541, in process_events | |
response = await route_utils.call_process_api( | |
File "/usr/local/lib/python3.10/site-packages/gradio/route_utils.py", line 276, in call_process_api | |
output = await app.get_blocks().process_api( | |
File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 1928, in process_api | |
result = await self.call_function( | |
File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 1514, in call_function | |
prediction = await anyio.to_thread.run_sync( | |
File "/usr/local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync | |
return await get_async_backend().run_sync_in_worker_thread( | |
File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2177, in run_sync_in_worker_thread | |
return await future | |
File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 859, in run | |
result = context.run(func, *args) | |
File "/usr/local/lib/python3.10/site-packages/gradio/utils.py", line 833, in wrapper | |
response = f(*args, **kwargs) | |
File "/home/user/app/app.py", line 276, in chat | |
history, history_output = agent_interaction( | |
File "/home/user/app/app.py", line 106, in agent_interaction | |
response = run_llm( | |
File "/home/user/app/app.py", line 77, in run_llm | |
resp = client.text_generation( | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2061, in text_generation | |
raise_text_generation_error(e) | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_common.py", line 460, in raise_text_generation_error | |
raise http_error | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2032, in text_generation | |
bytes_output = self.post(json=payload, model=model, task="text-generation", stream=stream) # type: ignore | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 273, in post | |
hf_raise_for_status(response) | |
File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 371, in hf_raise_for_status | |
raise HfHubHTTPError(str(e), response=response) from e | |
huggingface_hub.utils._errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://api-inference.huggingface.co/models/mistralai/Mixtral-8x7B-Instruct-v0.1 (Request ID: 41TZN5nlKL2We1-SnVOGU) |