THe notebook does not work with llama or the other endpoint

#76
by GozaloMoreno - opened

Hi: I am getting this:

HfHubHTTPError: 503 Server Error: Service Temporarily Unavailable for url: https://router.huggingface.co/hf-inference/models/meta-llama/Llama-3.2-3B-Instruct/v1/chat/completions

When I try with the other endpoint I am getting:

if the outputs for next cells are wrong, the free model may be overloaded. You can also use this public endpoint that contains Llama-3.2-3B-Instruct

client = InferenceClient("https://jc26mwg228mkj8dw.us-east-1.aws.endpoints.huggingface.cloud")

BadRequestError: (Request ID: XfTaeR)

Bad request:
Bad Request: Invalid state

This issue is still present. Is there a good alternative model to use that has an active inference server?

The same thing is happening to me

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment