--- # https://huggingface.co/docs/hub/en/spaces-config-reference title: API Inference short_description: Inference on API endpoints emoji: ⚡ colorFrom: blue colorTo: green sdk: streamlit sdk_version: 1.37.1 python_version: 3.11.9 app_file: 0_🏠_Home.py full_width: true pinned: true header: mini license: apache-2.0 models: - stabilityai/stable-diffusion-xl-base-1.0 - black-forest-labs/FLUX.1-schnell - black-forest-labs/FLUX.1-dev - codellama/CodeLlama-34b-Instruct-hf - meta-llama/Llama-2-13b-chat-hf - meta-llama/Meta-Llama-3.1-405B-Instruct-FP8 - mistralai/Mistral-7B-Instruct-v0.2 - NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO --- # ⚡ API Inference [Streamlit](https://streamlit.io) app for running inference on generative AI endpoints. ## Configuration Streamlit [configuration](https://docs.streamlit.io/develop/concepts/configuration/options) has the following priority order: 1. `streamlit run` flags (highest priority) 2. environment variables 3. local config file 4. global config file So you can deploy the local config and override with environment variables as necessary: ```bash STREAMLIT_LOGGER_LEVEL=error STREAMLIT_CLIENT_SHOW_ERROR_DETAILS=false STREAMLIT_CLIENT_TOOLBAR_MODE=viewer STREAMLIT_SERVER_FILE_WATCHER_TYPE=none STREAMLIT_SERVER_RUN_ON_SAVE=false STREAMLIT_BROWSER_SERVER_ADDRESS=adamelliotfields-api-inference.hf.space ``` ## Secrets ```bash FAL_KEY=... HF_TOKEN=... PERPLEXITY_API_KEY=... ``` ## Installation ```sh # clone git clone https://huggingface.co/spaces/adamelliotfields/api-inference.git cd api-inference git remote set-url origin https://adamelliotfields:$HF_TOKEN@huggingface.co/spaces/adamelliotfields/api-inference # install python -m venv .venv source .venv/bin/activate pip install -r requirements.txt # run python app.py ```