THe notebook does not work with llama or the other endpoint
#76
by
GozaloMoreno
- opened
Hi: I am getting this:
HfHubHTTPError: 503 Server Error: Service Temporarily Unavailable for url: https://router.huggingface.co/hf-inference/models/meta-llama/Llama-3.2-3B-Instruct/v1/chat/completions
When I try with the other endpoint I am getting:
if the outputs for next cells are wrong, the free model may be overloaded. You can also use this public endpoint that contains Llama-3.2-3B-Instruct
client = InferenceClient("https://jc26mwg228mkj8dw.us-east-1.aws.endpoints.huggingface.cloud")
BadRequestError: (Request ID: XfTaeR)
Bad request:
Bad Request: Invalid state
This issue is still present. Is there a good alternative model to use that has an active inference server?
The same thing is happening to me
same here