runtime error

Exit code: 1. Reason: async for x in self.fn(*args): # noqa: B007 # type: ignore File "/usr/local/lib/python3.10/site-packages/gradio/chat_interface.py", line 633, in _examples_stream_fn async for response in generator: File "/usr/local/lib/python3.10/site-packages/gradio/utils.py", line 649, in __anext__ return await anyio.to_thread.run_sync( File "/usr/local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2461, in run_sync_in_worker_thread return await future File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 962, in run result = context.run(func, *args) File "/usr/local/lib/python3.10/site-packages/gradio/utils.py", line 632, in run_sync_iterator_async return next(iterator) File "/home/user/app/app.py", line 72, in generate stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2383, in text_generation raise_text_generation_error(e) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_common.py", line 426, in raise_text_generation_error raise http_error File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2353, in text_generation bytes_output = self._inner_post(request_parameters, stream=stream) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 327, in _inner_post hf_raise_for_status(response) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 477, in hf_raise_for_status raise _format(HfHubHTTPError, str(e), response) from e huggingface_hub.errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://api-inference.huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.3

Container logs:

Fetching error logs...