Spaces:
Runtime error
Runtime error
davanstrien
HF Staff
chore: Add requirements for vllm, outlines, llama_index, datasets, and rich
4dc4bf5
| # This file was autogenerated by uv via the following command: | |
| # uv pip compile requirements.in -o requirements.txt | |
| aiohttp==3.9.5 | |
| # via | |
| # datasets | |
| # fsspec | |
| # llama-index-core | |
| # llama-index-legacy | |
| # vllm | |
| aiosignal==1.3.1 | |
| # via | |
| # aiohttp | |
| # ray | |
| annotated-types==0.7.0 | |
| # via pydantic | |
| anyio==4.4.0 | |
| # via | |
| # httpx | |
| # openai | |
| # starlette | |
| # watchfiles | |
| async-timeout==4.0.3 | |
| # via aiohttp | |
| attrs==23.2.0 | |
| # via | |
| # aiohttp | |
| # jsonschema | |
| # referencing | |
| beautifulsoup4==4.12.3 | |
| # via llama-index-readers-file | |
| certifi==2024.6.2 | |
| # via | |
| # httpcore | |
| # httpx | |
| # requests | |
| charset-normalizer==3.3.2 | |
| # via requests | |
| click==8.1.7 | |
| # via | |
| # nltk | |
| # ray | |
| # typer | |
| # uvicorn | |
| cloudpickle==3.0.0 | |
| # via outlines | |
| cmake==3.29.5.1 | |
| # via vllm | |
| dataclasses-json==0.6.7 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| datasets==2.19.2 | |
| # via | |
| # -r requirements.in | |
| # outlines | |
| deprecated==1.2.14 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| dill==0.3.7 | |
| # via | |
| # datasets | |
| # multiprocess | |
| dirtyjson==1.0.8 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| diskcache==5.6.3 | |
| # via outlines | |
| distro==1.9.0 | |
| # via openai | |
| dnspython==2.6.1 | |
| # via email-validator | |
| email-validator==2.1.1 | |
| # via fastapi | |
| exceptiongroup==1.2.1 | |
| # via anyio | |
| fastapi==0.111.0 | |
| # via vllm | |
| fastapi-cli==0.0.4 | |
| # via fastapi | |
| filelock==3.15.1 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # ray | |
| # torch | |
| # transformers | |
| # triton | |
| # vllm | |
| frozenlist==1.4.1 | |
| # via | |
| # aiohttp | |
| # aiosignal | |
| # ray | |
| fsspec==2024.3.1 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # llama-index-core | |
| # llama-index-legacy | |
| # torch | |
| greenlet==3.0.3 | |
| # via sqlalchemy | |
| h11==0.14.0 | |
| # via | |
| # httpcore | |
| # uvicorn | |
| httpcore==1.0.5 | |
| # via httpx | |
| httptools==0.6.1 | |
| # via uvicorn | |
| httpx==0.27.0 | |
| # via | |
| # fastapi | |
| # llama-index-core | |
| # llama-index-legacy | |
| # llamaindex-py-client | |
| # openai | |
| huggingface-hub==0.23.3 | |
| # via | |
| # datasets | |
| # tokenizers | |
| # transformers | |
| idna==3.7 | |
| # via | |
| # anyio | |
| # email-validator | |
| # httpx | |
| # requests | |
| # yarl | |
| interegular==0.3.3 | |
| # via | |
| # lm-format-enforcer | |
| # outlines | |
| jinja2==3.1.4 | |
| # via | |
| # fastapi | |
| # outlines | |
| # torch | |
| joblib==1.4.2 | |
| # via nltk | |
| jsonschema==4.22.0 | |
| # via | |
| # outlines | |
| # ray | |
| jsonschema-specifications==2023.12.1 | |
| # via jsonschema | |
| lark==1.1.9 | |
| # via outlines | |
| llama-index==0.10.44 | |
| # via -r requirements.in | |
| llama-index-agent-openai==0.2.7 | |
| # via | |
| # llama-index | |
| # llama-index-program-openai | |
| llama-index-cli==0.1.12 | |
| # via llama-index | |
| llama-index-core==0.10.44 | |
| # via | |
| # llama-index | |
| # llama-index-agent-openai | |
| # llama-index-cli | |
| # llama-index-embeddings-openai | |
| # llama-index-indices-managed-llama-cloud | |
| # llama-index-llms-openai | |
| # llama-index-multi-modal-llms-openai | |
| # llama-index-program-openai | |
| # llama-index-question-gen-openai | |
| # llama-index-readers-file | |
| # llama-index-readers-llama-parse | |
| # llama-parse | |
| llama-index-embeddings-openai==0.1.10 | |
| # via | |
| # llama-index | |
| # llama-index-cli | |
| llama-index-indices-managed-llama-cloud==0.1.6 | |
| # via llama-index | |
| llama-index-legacy==0.9.48 | |
| # via llama-index | |
| llama-index-llms-openai==0.1.22 | |
| # via | |
| # llama-index | |
| # llama-index-agent-openai | |
| # llama-index-cli | |
| # llama-index-multi-modal-llms-openai | |
| # llama-index-program-openai | |
| # llama-index-question-gen-openai | |
| llama-index-multi-modal-llms-openai==0.1.6 | |
| # via llama-index | |
| llama-index-program-openai==0.1.6 | |
| # via | |
| # llama-index | |
| # llama-index-question-gen-openai | |
| llama-index-question-gen-openai==0.1.3 | |
| # via llama-index | |
| llama-index-readers-file==0.1.25 | |
| # via llama-index | |
| llama-index-readers-llama-parse==0.1.4 | |
| # via llama-index | |
| llama-parse==0.4.4 | |
| # via llama-index-readers-llama-parse | |
| llamaindex-py-client==0.1.19 | |
| # via | |
| # llama-index-core | |
| # llama-index-indices-managed-llama-cloud | |
| llvmlite==0.42.0 | |
| # via numba | |
| lm-format-enforcer==0.10.1 | |
| # via vllm | |
| markdown-it-py==3.0.0 | |
| # via rich | |
| markupsafe==2.1.5 | |
| # via jinja2 | |
| marshmallow==3.21.3 | |
| # via dataclasses-json | |
| mdurl==0.1.2 | |
| # via markdown-it-py | |
| mpmath==1.3.0 | |
| # via sympy | |
| msgpack==1.0.8 | |
| # via ray | |
| multidict==6.0.5 | |
| # via | |
| # aiohttp | |
| # yarl | |
| multiprocess==0.70.15 | |
| # via datasets | |
| mypy-extensions==1.0.0 | |
| # via typing-inspect | |
| nest-asyncio==1.6.0 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| # outlines | |
| networkx==3.2.1 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| # torch | |
| ninja==1.11.1.1 | |
| # via vllm | |
| nltk==3.8.1 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| numba==0.59.1 | |
| # via outlines | |
| numpy==1.26.4 | |
| # via | |
| # datasets | |
| # llama-index-core | |
| # llama-index-legacy | |
| # numba | |
| # outlines | |
| # pandas | |
| # pyarrow | |
| # transformers | |
| # vllm | |
| # xformers | |
| nvidia-cublas-cu12==12.1.3.1 | |
| # via | |
| # nvidia-cudnn-cu12 | |
| # nvidia-cusolver-cu12 | |
| # torch | |
| nvidia-cuda-cupti-cu12==12.1.105 | |
| # via torch | |
| nvidia-cuda-nvrtc-cu12==12.1.105 | |
| # via torch | |
| nvidia-cuda-runtime-cu12==12.1.105 | |
| # via torch | |
| nvidia-cudnn-cu12==8.9.2.26 | |
| # via torch | |
| nvidia-cufft-cu12==11.0.2.54 | |
| # via torch | |
| nvidia-curand-cu12==10.3.2.106 | |
| # via torch | |
| nvidia-cusolver-cu12==11.4.5.107 | |
| # via torch | |
| nvidia-cusparse-cu12==12.1.0.106 | |
| # via | |
| # nvidia-cusolver-cu12 | |
| # torch | |
| nvidia-ml-py==12.555.43 | |
| # via vllm | |
| nvidia-nccl-cu12==2.20.5 | |
| # via torch | |
| nvidia-nvjitlink-cu12==12.5.40 | |
| # via | |
| # nvidia-cusolver-cu12 | |
| # nvidia-cusparse-cu12 | |
| nvidia-nvtx-cu12==12.1.105 | |
| # via torch | |
| openai==1.34.0 | |
| # via | |
| # llama-index-agent-openai | |
| # llama-index-core | |
| # llama-index-legacy | |
| # vllm | |
| orjson==3.10.4 | |
| # via fastapi | |
| outlines==0.0.43 | |
| # via | |
| # -r requirements.in | |
| # vllm | |
| packaging==24.1 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # lm-format-enforcer | |
| # marshmallow | |
| # ray | |
| # transformers | |
| pandas==2.2.2 | |
| # via | |
| # datasets | |
| # llama-index-core | |
| # llama-index-legacy | |
| pillow==10.3.0 | |
| # via | |
| # llama-index-core | |
| # vllm | |
| prometheus-client==0.20.0 | |
| # via | |
| # prometheus-fastapi-instrumentator | |
| # vllm | |
| prometheus-fastapi-instrumentator==7.0.0 | |
| # via vllm | |
| protobuf==5.27.1 | |
| # via ray | |
| psutil==5.9.8 | |
| # via vllm | |
| py-cpuinfo==9.0.0 | |
| # via vllm | |
| pyairports==2.1.1 | |
| # via outlines | |
| pyarrow==16.1.0 | |
| # via datasets | |
| pyarrow-hotfix==0.6 | |
| # via datasets | |
| pycountry==24.6.1 | |
| # via outlines | |
| pydantic==2.7.4 | |
| # via | |
| # fastapi | |
| # llamaindex-py-client | |
| # lm-format-enforcer | |
| # openai | |
| # outlines | |
| # vllm | |
| pydantic-core==2.18.4 | |
| # via pydantic | |
| pygments==2.18.0 | |
| # via rich | |
| pypdf==4.2.0 | |
| # via llama-index-readers-file | |
| python-dateutil==2.9.0.post0 | |
| # via pandas | |
| python-dotenv==1.0.1 | |
| # via uvicorn | |
| python-multipart==0.0.9 | |
| # via fastapi | |
| pytz==2024.1 | |
| # via pandas | |
| pyyaml==6.0.1 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # llama-index-core | |
| # lm-format-enforcer | |
| # ray | |
| # transformers | |
| # uvicorn | |
| ray==2.24.0 | |
| # via vllm | |
| referencing==0.35.1 | |
| # via | |
| # jsonschema | |
| # jsonschema-specifications | |
| # outlines | |
| regex==2024.5.15 | |
| # via | |
| # nltk | |
| # tiktoken | |
| # transformers | |
| requests==2.32.3 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # llama-index-core | |
| # llama-index-legacy | |
| # outlines | |
| # ray | |
| # tiktoken | |
| # transformers | |
| # vllm | |
| rich==13.7.1 | |
| # via | |
| # -r requirements.in | |
| # typer | |
| rpds-py==0.18.1 | |
| # via | |
| # jsonschema | |
| # referencing | |
| safetensors==0.4.3 | |
| # via transformers | |
| sentencepiece==0.2.0 | |
| # via vllm | |
| shellingham==1.5.4 | |
| # via typer | |
| six==1.16.0 | |
| # via python-dateutil | |
| sniffio==1.3.1 | |
| # via | |
| # anyio | |
| # httpx | |
| # openai | |
| soupsieve==2.5 | |
| # via beautifulsoup4 | |
| sqlalchemy==2.0.30 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| starlette==0.37.2 | |
| # via | |
| # fastapi | |
| # prometheus-fastapi-instrumentator | |
| striprtf==0.0.26 | |
| # via llama-index-readers-file | |
| sympy==1.12.1 | |
| # via torch | |
| tenacity==8.3.0 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| tiktoken==0.7.0 | |
| # via | |
| # llama-index-core | |
| # llama-index-legacy | |
| # vllm | |
| tokenizers==0.19.1 | |
| # via | |
| # transformers | |
| # vllm | |
| torch==2.3.0 | |
| # via | |
| # vllm | |
| # vllm-flash-attn | |
| # xformers | |
| tqdm==4.66.4 | |
| # via | |
| # datasets | |
| # huggingface-hub | |
| # llama-index-core | |
| # nltk | |
| # openai | |
| # outlines | |
| # transformers | |
| transformers==4.41.2 | |
| # via vllm | |
| triton==2.3.0 | |
| # via torch | |
| typer==0.12.3 | |
| # via fastapi-cli | |
| typing-extensions==4.12.2 | |
| # via | |
| # anyio | |
| # fastapi | |
| # huggingface-hub | |
| # llama-index-core | |
| # llama-index-legacy | |
| # openai | |
| # pydantic | |
| # pydantic-core | |
| # pypdf | |
| # sqlalchemy | |
| # starlette | |
| # torch | |
| # typer | |
| # typing-inspect | |
| # uvicorn | |
| # vllm | |
| typing-inspect==0.9.0 | |
| # via | |
| # dataclasses-json | |
| # llama-index-core | |
| # llama-index-legacy | |
| tzdata==2024.1 | |
| # via pandas | |
| ujson==5.10.0 | |
| # via fastapi | |
| urllib3==2.2.1 | |
| # via requests | |
| uvicorn==0.30.1 | |
| # via | |
| # fastapi | |
| # vllm | |
| uvloop==0.19.0 | |
| # via uvicorn | |
| vllm==0.5.0 | |
| # via -r requirements.in | |
| vllm-flash-attn==2.5.9 | |
| # via vllm | |
| watchfiles==0.22.0 | |
| # via uvicorn | |
| websockets==12.0 | |
| # via uvicorn | |
| wrapt==1.16.0 | |
| # via | |
| # deprecated | |
| # llama-index-core | |
| xformers==0.0.26.post1 | |
| # via vllm | |
| xxhash==3.4.1 | |
| # via datasets | |
| yarl==1.9.4 | |
| # via aiohttp | |