Spaces:
Running
Running
# LITELLM PROXY DEPENDENCIES # | |
anyio==4.2.0 # openai + http req. | |
openai>=1.0.0 # openai req. | |
fastapi # server dep | |
pydantic>=2.5 # openai req. | |
backoff==2.2.1 # server dep | |
pyyaml==6.0 # server dep | |
uvicorn==0.22.0 # server dep | |
gunicorn==21.2.0 # server dep | |
boto3==1.28.58 # aws bedrock/sagemaker calls | |
redis==4.6.0 # caching | |
prisma==0.11.0 # for db | |
mangum==0.17.0 # for aws lambda functions | |
google-generativeai==0.1.0 # for vertex ai calls | |
async_generator==1.10.0 # for async ollama calls | |
traceloop-sdk==0.5.3 # for open telemetry logging | |
langfuse>=2.0.0 # for langfuse self-hosted logging | |
orjson==3.9.7 # fast /embedding responses | |
### LITELLM PACKAGE DEPENDENCIES | |
python-dotenv>=0.2.0 # for env | |
tiktoken>=0.4.0 # for calculating usage | |
importlib-metadata>=6.8.0 # for random utils | |
tokenizers==0.14.0 # for calculating usage | |
click==8.1.7 # for proxy cli | |
jinja2==3.1.2 # for prompt templates | |
certifi>=2023.7.22 # [TODO] clean up | |
aiohttp==3.9.0 # for network calls | |
#### |