| # LITELLM PROXY DEPENDENCIES # | |
| anyio==4.2.0 # openai + http req. | |
| openai>=1.0.0 # openai req. | |
| fastapi # server dep | |
| pydantic>=2.5 # openai req. | |
| backoff==2.2.1 # server dep | |
| pyyaml==6.0 # server dep | |
| uvicorn==0.22.0 # server dep | |
| gunicorn==21.2.0 # server dep | |
| boto3==1.28.58 # aws bedrock/sagemaker calls | |
| redis==4.6.0 # caching | |
| prisma==0.11.0 # for db | |
| mangum==0.17.0 # for aws lambda functions | |
| google-generativeai==0.1.0 # for vertex ai calls | |
| async_generator==1.10.0 # for async ollama calls | |
| traceloop-sdk==0.5.3 # for open telemetry logging | |
| langfuse>=2.0.0 # for langfuse self-hosted logging | |
| orjson==3.9.7 # fast /embedding responses | |
| ### LITELLM PACKAGE DEPENDENCIES | |
| python-dotenv>=0.2.0 # for env | |
| tiktoken>=0.4.0 # for calculating usage | |
| importlib-metadata>=6.8.0 # for random utils | |
| tokenizers==0.14.0 # for calculating usage | |
| click==8.1.7 # for proxy cli | |
| jinja2==3.1.2 # for prompt templates | |
| certifi>=2023.7.22 # [TODO] clean up | |
| aiohttp==3.9.0 # for network calls | |
| #### |