[project] name = "llm-guard-api" description = "LLM Guard API is a deployment of LLM Guard as an API." authors = [ { name = "Protect AI", email = "community@protectai.com"} ] readme = "README.md" dynamic = ["version"] classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", ] requires-python = ">=3.9" dependencies = [ "asyncio==3.4.3", "fastapi==0.110.0", "llm-guard==0.3.10", "pydantic==1.10.14", "pyyaml==6.0.1", "uvicorn[standard]==0.29.0", "structlog>=24", "slowapi==0.1.9", "opentelemetry-instrumentation-fastapi==0.44b0", "opentelemetry-api==1.23.0", "opentelemetry-sdk==1.23.0", "opentelemetry-exporter-otlp-proto-http==1.23.0", "opentelemetry-exporter-prometheus==0.44b0", "opentelemetry-sdk-extension-aws==2.0.1", "opentelemetry-propagator-aws-xray==1.0.1" ] [project.optional-dependencies] cpu = [ "llm-guard[onnxruntime]==0.3.10", ] gpu = [ "llm-guard[onnxruntime-gpu]==0.3.10", ] [tool.setuptools] packages = ["app"] [tool.setuptools.dynamic] version = {attr = "app.version.__version__"} [build-system] requires = ["setuptools", "wheel"] build-backend = "setuptools.build_meta" [project.scripts] llm_guard_api = "app.app:run_app"