|
# This file was autogenerated by uv via the following command: |
|
# uv pip compile pyproject.toml -o requirements.txt |
|
accelerate==1.0.0 |
|
# via llama-2-7b-chat (pyproject.toml) |
|
aiofiles==23.2.1 |
|
# via gradio |
|
annotated-types==0.7.0 |
|
# via pydantic |
|
anyio==4.6.0 |
|
# via |
|
# gradio |
|
# httpx |
|
# starlette |
|
bitsandbytes==0.44.1 |
|
# via llama-2-7b-chat (pyproject.toml) |
|
certifi==2024.8.30 |
|
# via |
|
# httpcore |
|
# httpx |
|
# requests |
|
charset-normalizer==3.3.2 |
|
# via requests |
|
click==8.1.7 |
|
# via |
|
# typer |
|
# uvicorn |
|
exceptiongroup==1.2.2 |
|
# via anyio |
|
fastapi==0.115.0 |
|
# via gradio |
|
ffmpy==0.4.0 |
|
# via gradio |
|
filelock==3.16.1 |
|
# via |
|
# huggingface-hub |
|
# torch |
|
# transformers |
|
# triton |
|
fsspec==2024.9.0 |
|
# via |
|
# gradio-client |
|
# huggingface-hub |
|
# torch |
|
gradio==5.0.1 |
|
# via |
|
# llama-2-7b-chat (pyproject.toml) |
|
# spaces |
|
gradio-client==1.4.0 |
|
# via gradio |
|
h11==0.14.0 |
|
# via |
|
# httpcore |
|
# uvicorn |
|
hf-transfer==0.1.8 |
|
# via llama-2-7b-chat (pyproject.toml) |
|
httpcore==1.0.5 |
|
# via httpx |
|
httpx==0.27.2 |
|
# via |
|
# gradio |
|
# gradio-client |
|
# spaces |
|
huggingface-hub==0.25.1 |
|
# via |
|
# accelerate |
|
# gradio |
|
# gradio-client |
|
# tokenizers |
|
# transformers |
|
idna==3.10 |
|
# via |
|
# anyio |
|
# httpx |
|
# requests |
|
jinja2==3.1.4 |
|
# via |
|
# gradio |
|
# torch |
|
markdown-it-py==3.0.0 |
|
# via rich |
|
markupsafe==2.1.5 |
|
# via |
|
# gradio |
|
# jinja2 |
|
mdurl==0.1.2 |
|
# via markdown-it-py |
|
mpmath==1.3.0 |
|
# via sympy |
|
networkx==3.3 |
|
# via torch |
|
numpy==2.1.1 |
|
# via |
|
# accelerate |
|
# bitsandbytes |
|
# gradio |
|
# pandas |
|
# transformers |
|
nvidia-cublas-cu12==12.1.3.1 |
|
# via |
|
# nvidia-cudnn-cu12 |
|
# nvidia-cusolver-cu12 |
|
# torch |
|
nvidia-cuda-cupti-cu12==12.1.105 |
|
# via torch |
|
nvidia-cuda-nvrtc-cu12==12.1.105 |
|
# via torch |
|
nvidia-cuda-runtime-cu12==12.1.105 |
|
# via torch |
|
nvidia-cudnn-cu12==9.1.0.70 |
|
# via torch |
|
nvidia-cufft-cu12==11.0.2.54 |
|
# via torch |
|
nvidia-curand-cu12==10.3.2.106 |
|
# via torch |
|
nvidia-cusolver-cu12==11.4.5.107 |
|
# via torch |
|
nvidia-cusparse-cu12==12.1.0.106 |
|
# via |
|
# nvidia-cusolver-cu12 |
|
# torch |
|
nvidia-nccl-cu12==2.20.5 |
|
# via torch |
|
nvidia-nvjitlink-cu12==12.6.68 |
|
# via |
|
# nvidia-cusolver-cu12 |
|
# nvidia-cusparse-cu12 |
|
nvidia-nvtx-cu12==12.1.105 |
|
# via torch |
|
orjson==3.10.7 |
|
# via gradio |
|
packaging==24.1 |
|
# via |
|
# accelerate |
|
# gradio |
|
# gradio-client |
|
# huggingface-hub |
|
# spaces |
|
# transformers |
|
pandas==2.2.3 |
|
# via gradio |
|
pillow==10.4.0 |
|
# via gradio |
|
psutil==5.9.8 |
|
# via |
|
# accelerate |
|
# spaces |
|
pydantic==2.9.2 |
|
# via |
|
# fastapi |
|
# gradio |
|
# spaces |
|
pydantic-core==2.23.4 |
|
# via pydantic |
|
pydub==0.25.1 |
|
# via gradio |
|
pygments==2.18.0 |
|
# via rich |
|
python-dateutil==2.9.0.post0 |
|
# via pandas |
|
python-multipart==0.0.12 |
|
# via gradio |
|
pytz==2024.2 |
|
# via pandas |
|
pyyaml==6.0.2 |
|
# via |
|
# accelerate |
|
# gradio |
|
# huggingface-hub |
|
# transformers |
|
regex==2024.9.11 |
|
# via transformers |
|
requests==2.32.3 |
|
# via |
|
# huggingface-hub |
|
# spaces |
|
# transformers |
|
rich==13.8.1 |
|
# via typer |
|
ruff==0.6.8 |
|
# via gradio |
|
safetensors==0.4.5 |
|
# via |
|
# accelerate |
|
# transformers |
|
semantic-version==2.10.0 |
|
# via gradio |
|
shellingham==1.5.4 |
|
# via typer |
|
six==1.16.0 |
|
# via python-dateutil |
|
sniffio==1.3.1 |
|
# via |
|
# anyio |
|
# httpx |
|
spaces==0.30.3 |
|
# via llama-2-7b-chat (pyproject.toml) |
|
starlette==0.38.6 |
|
# via fastapi |
|
sympy==1.13.3 |
|
# via torch |
|
tokenizers==0.20.0 |
|
# via transformers |
|
tomlkit==0.12.0 |
|
# via gradio |
|
torch==2.4.0 |
|
# via |
|
# llama-2-7b-chat (pyproject.toml) |
|
# accelerate |
|
# bitsandbytes |
|
tqdm==4.66.5 |
|
# via |
|
# huggingface-hub |
|
# transformers |
|
transformers==4.45.2 |
|
# via llama-2-7b-chat (pyproject.toml) |
|
triton==3.0.0 |
|
# via torch |
|
typer==0.12.5 |
|
# via gradio |
|
typing-extensions==4.12.2 |
|
# via |
|
# anyio |
|
# fastapi |
|
# gradio |
|
# gradio-client |
|
# huggingface-hub |
|
# pydantic |
|
# pydantic-core |
|
# spaces |
|
# torch |
|
# typer |
|
# uvicorn |
|
tzdata==2024.2 |
|
# via pandas |
|
urllib3==2.2.3 |
|
# via requests |
|
uvicorn==0.31.0 |
|
# via gradio |
|
websockets==12.0 |
|
# via gradio-client |
|
|
|
optimum |
|
auto-gptq |
|
tiktoken |
|
|