Cannot download the model with huggingface-cli
Hello,
I have trouble downloading this model using the CLI
huggingface-cli download meta-llama/Meta-Llama-3.2-3B-Instruct
Gives me the following error :
Traceback (most recent call last):
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/utils/_errors.py", line 304, in hf_raise_for_status
response.raise_for_status()
File "/home/vscode/py-env/lib/python3.11/site-packages/requests/models.py", line 1024, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/api/models/meta-llama/Meta-Llama-3.2-3B-Instruct/revision/main
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/vscode/py-env/bin/huggingface-cli", line 8, in <module>
sys.exit(main())
^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/commands/huggingface_cli.py", line 51, in main
service.run()
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/commands/download.py", line 146, in run
print(self._download()) # Print path to downloaded files
^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/commands/download.py", line 180, in _download
return snapshot_download(
^^^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/_snapshot_download.py", line 233, in snapshot_download
raise api_call_error
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/_snapshot_download.py", line 164, in snapshot_download
repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/hf_api.py", line 2491, in repo_info
return method(
^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/hf_api.py", line 2301, in model_info
hf_raise_for_status(r)
File "/home/vscode/py-env/lib/python3.11/site-packages/huggingface_hub/utils/_errors.py", line 352, in hf_raise_for_status
raise RepositoryNotFoundError(message, response) from e
huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66f54b74-57e3a2477a203b7c3805d63f;0b7ae875-1169-4a3b-9c8d-467c511a1230)
Repository Not Found for url: https://huggingface.co/api/models/meta-llama/Meta-Llama-3.2-3B-Instruct/revision/main.
Please make sure you specified the correct `repo_id` and `repo_type`.
If you are trying to access a private or gated repo, make sure you are authenticated.
Are you authenticated with a User Access Token?
@supportend Yes, I can download other models with no issues
Llama-Models are special, because you have "to agree to share your contact information" and use a User Access Token, to verify, you have done it - to access the model files.
You saved the tiken in a envionment variable? Because i don't see options like login or login --token in your input.
See chapter huggingface-cli login here:
https://huggingface.co/docs/huggingface_hub/guides/cli#huggingface-cli-login
What I've done is
huggingface-cli login
Then the terminal asks for my token and I just copy paste it in the console. By the way I made sure having accepted the license for the 3.2 family as well.
I also previously accepted the license for Llama3.1 models and have no issue downloading them.
i have same error too , first time i loged huggingface-cli and paste my token ı downloaded the llama 3.2 3b after ı tried ı got [error 404 cannto ](ValueError: Error raised by inference API HTTP code: 404, {"error":"model "meta-llama/Llama-3.2-1B" not found, try pulling it first"}).