THe notebook does not work with llama or the other endpoint
#76
by
GozaloMoreno
- opened
Hi: I am getting this:
HfHubHTTPError: 503 Server Error: Service Temporarily Unavailable for url: https://router.huggingface.co/hf-inference/models/meta-llama/Llama-3.2-3B-Instruct/v1/chat/completions
When I try with the other endpoint I am getting:
if the outputs for next cells are wrong, the free model may be overloaded. You can also use this public endpoint that contains Llama-3.2-3B-Instruct
client = InferenceClient("https://jc26mwg228mkj8dw.us-east-1.aws.endpoints.huggingface.cloud")
BadRequestError: (Request ID: XfTaeR)
Bad request:
Bad Request: Invalid state
This issue is still present. Is there a good alternative model to use that has an active inference server?
The same thing is happening to me
same here
Same here
same here
Same here.
Same
I've changed it to another model and it works,client = InferenceClient("meta-llama/Llama-3.1-8B-Instruct")