Spaces:
Running
Running
File size: 1,413 Bytes
020aaa1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
from api.config import SETTINGS
from api.models import app, EMBEDDED_MODEL, GENERATE_ENGINE
prefix = SETTINGS.api_prefix
if EMBEDDED_MODEL is not None:
from api.routes.embedding import embedding_router
app.include_router(embedding_router, prefix=prefix, tags=["Embedding"])
if GENERATE_ENGINE is not None:
from api.routes import model_router
app.include_router(model_router, prefix=prefix, tags=["Model"])
if SETTINGS.engine == "vllm":
from api.vllm_routes import chat_router as chat_router
from api.vllm_routes import completion_router as completion_router
elif SETTINGS.engine == "llama.cpp":
from api.llama_cpp_routes import chat_router as chat_router
from api.llama_cpp_routes import completion_router as completion_router
elif SETTINGS.engine == "tgi":
from api.tgi_routes import chat_router as chat_router
from api.tgi_routes.completion import completion_router as completion_router
else:
from api.routes.chat import chat_router as chat_router
from api.routes.completion import completion_router as completion_router
app.include_router(chat_router, prefix=prefix, tags=["Chat Completion"])
app.include_router(completion_router, prefix=prefix, tags=["Completion"])
if __name__ == '__main__':
import uvicorn
uvicorn.run(app, host=SETTINGS.host, port=SETTINGS.port, log_level="info")
|