metadata
title: API Inference
short_description: Inference on API endpoints
emoji: ⚡
colorFrom: blue
colorTo: green
sdk: streamlit
sdk_version: 1.37.1
python_version: 3.11.9
app_file: 0_🏠_Home.py
full_width: true
pinned: true
header: mini
license: apache-2.0
models:
- stabilityai/stable-diffusion-xl-base-1.0
- black-forest-labs/FLUX.1-schnell
- black-forest-labs/FLUX.1-dev
- codellama/CodeLlama-34b-Instruct-hf
- meta-llama/Llama-2-13b-chat-hf
- meta-llama/Meta-Llama-3.1-405B-Instruct-FP8
- mistralai/Mistral-7B-Instruct-v0.2
- NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
⚡ API Inference
Streamlit app for running inference on generative AI endpoints.
Configuration
Streamlit configuration has the following priority order:
streamlit run
flags (highest priority)- environment variables
- local config file
- global config file
So you can deploy the local config and override with environment variables as necessary:
STREAMLIT_LOGGER_LEVEL=error
STREAMLIT_CLIENT_SHOW_ERROR_DETAILS=false
STREAMLIT_CLIENT_TOOLBAR_MODE=viewer
STREAMLIT_SERVER_FILE_WATCHER_TYPE=none
STREAMLIT_SERVER_RUN_ON_SAVE=false
STREAMLIT_BROWSER_SERVER_ADDRESS=adamelliotfields-api-inference.hf.space
Installation
# clone
git clone https://huggingface.co/spaces/adamelliotfields/api-inference.git
cd api-inference
git remote set-url origin https://adamelliotfields:$HF_TOKEN@huggingface.co/spaces/adamelliotfields/api-inference
# install
python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt
# run
python app.py