--- title: Backend emoji: 🐢 colorFrom: pink colorTo: blue sdk: docker pinned: false license: mit app_port: 8000 --- This is a [LlamaIndex](https://www.llamaindex.ai/) project using [FastAPI](https://fastapi.tiangolo.com/) bootstrapped with [`create-llama`](https://github.com/run-llama/LlamaIndexTS/tree/main/packages/create-llama). ## Getting Started First, setup the environment with poetry: > **_Note:_** This step is not needed if you are using the dev-container. ``` poetry install poetry shell ``` Then check the parameters that have been pre-configured in the `.env` file in this directory. (E.g. you might need to configure an `OPENAI_API_KEY` if you're using OpenAI as model provider). If you are using any tools or data sources, you can update their config files in the `config` folder. Second, generate the embeddings of the documents in the `./data` directory (if this folder exists - otherwise, skip this step): ``` poetry run generate ``` Third, run the development server: ``` python main.py ``` The example provides two different API endpoints: 1. `/api/chat` - a streaming chat endpoint 2. `/api/chat/request` - a non-streaming chat endpoint You can test the streaming endpoint with the following curl request: ``` curl --location 'localhost:8000/api/chat' \ --header 'Content-Type: application/json' \ --data '{ "messages": [{ "role": "user", "content": "Hello" }] }' ``` And for the non-streaming endpoint run: ``` curl --location 'localhost:8000/api/chat/request' \ --header 'Content-Type: application/json' \ --data '{ "messages": [{ "role": "user", "content": "Hello" }] }' ``` You can start editing the API endpoints by modifying `app/api/routers/chat.py`. The endpoints auto-update as you save the file. You can delete the endpoint you're not using. Open [http://localhost:8000/docs](http://localhost:8000/docs) with your browser to see the Swagger UI of the API. The API allows CORS for all origins to simplify development. You can change this behavior by setting the `ENVIRONMENT` environment variable to `prod`: ``` ENVIRONMENT=prod python main.py ``` ## Local Postgres database setup To setup a local postgres database, run: 1. Build the docker image: ```bash make build-postgres ``` 2. Start the docker container: ```bash make run-postgres ``` ## Running Migrations To generate new migrations, run: ```bash make generate-migrations migration_title="" ``` To locally verify your changes, run: ```bash make run-migrations ``` ## Using Docker 1. Build an image for the FastAPI app: ``` docker build -t . ``` 2. Generate embeddings: Parse the data and generate the vector embeddings if the `./data` folder exists - otherwise, skip this step: ``` docker run \ --rm \ -v $(pwd)/.env:/app/.env \ # Use ENV variables and configuration from your file-system -v $(pwd)/config:/app/config \ -v $(pwd)/data:/app/data \ # Use your local folder to read the data -v $(pwd)/storage:/app/storage \ # Use your file system to store the vector database \ poetry run generate ``` 3. Start the API: ``` docker run \ -v $(pwd)/.env:/app/.env \ # Use ENV variables and configuration from your file-system -v $(pwd)/config:/app/config \ -v $(pwd)/storage:/app/storage \ # Use your file system to store gea vector database -p 8000:8000 \ ``` ## Learn More To learn more about LlamaIndex, take a look at the following resources: - [LlamaIndex Documentation](https://docs.llamaindex.ai) - learn about LlamaIndex. You can check out [the LlamaIndex GitHub repository](https://github.com/run-llama/llama_index) - your feedback and contributions are welcome!