inferencing-llm / app.py
Shyamnath's picture
Update to use LiteLLM proxy server configuration
7d17b48
raw
history blame
703 Bytes
from fastapi import FastAPI
from litellm.proxy.proxy_server import ProxyServer
from litellm.proxy.config import ProxyConfig
app = FastAPI(
title="LiteLLM Proxy",
description="LiteLLM OpenAI-compatible proxy",
version="1.0",
docs_url="/proxy/docs", # Swagger UI
redoc_url="/proxy/redoc", # Optional: ReDoc UI
openapi_url="/proxy/openapi.json"
)
# Load LiteLLM Proxy
proxy_config = ProxyConfig()
proxy_server = ProxyServer(config=proxy_config)
proxy_server.add_routes(app)
@app.get("/")
async def root():
return {"message": "LiteLLM is running. Visit /proxy/docs"}
if __name__ == "__main__":
import uvicorn
uvicorn.run(app, host="0.0.0.0", port=7860)