playground-template / README.md
adamelliotfields's picture
Update readme
11aa54f verified
|
raw
history blame
1.74 kB
metadata
title: API Inference
short_description: Inference on API endpoints
emoji: 
colorFrom: blue
colorTo: green
sdk: streamlit
sdk_version: 1.37.1
python_version: 3.11.9
app_file: 0_🏠_Home.py
full_width: true
pinned: true
header: mini
license: apache-2.0
models:
  - stabilityai/stable-diffusion-xl-base-1.0
  - black-forest-labs/FLUX.1-schnell
  - black-forest-labs/FLUX.1-dev
  - codellama/CodeLlama-34b-Instruct-hf
  - meta-llama/Llama-2-13b-chat-hf
  - meta-llama/Meta-Llama-3.1-405B-Instruct-FP8
  - mistralai/Mistral-7B-Instruct-v0.2
  - NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO

⚡ API Inference

Streamlit app for running inference on generative AI endpoints.

Configuration

Streamlit configuration has the following priority order:

  1. streamlit run flags (highest priority)
  2. environment variables
  3. local config file
  4. global config file

So you can deploy the local config and override with environment variables as necessary:

STREAMLIT_LOGGER_LEVEL=error
STREAMLIT_CLIENT_SHOW_ERROR_DETAILS=false
STREAMLIT_CLIENT_TOOLBAR_MODE=viewer
STREAMLIT_SERVER_FILE_WATCHER_TYPE=none
STREAMLIT_SERVER_RUN_ON_SAVE=false
STREAMLIT_BROWSER_SERVER_ADDRESS=adamelliotfields-api-inference.hf.space

Installation

# clone
git clone https://huggingface.co/spaces/adamelliotfields/api-inference.git
cd api-inference
git remote set-url origin https://adamelliotfields:$HF_TOKEN@huggingface.co/spaces/adamelliotfields/api-inference

# install
python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt

# run
python app.py