Spaces:
Running
on
Zero
Running
on
Zero
A newer version of the Gradio SDK is available:
5.21.0
metadata
title: Chat
short_description: Serverless small language model inference
emoji: 🤖
colorFrom: blue
colorTo: yellow
sdk: gradio
sdk_version: 4.44.1
python_version: 3.11.9
app_file: app.py
fullWidth: false
pinned: true
header: mini
license: apache-2.0
preload_from_hub:
- >-
01-ai/Yi-Coder-1.5B-Chat
config.json,generation_config.json,model.safetensors,special_tokens_map.json,tokenizer.model,tokenizer_config.json
- >-
HuggingFaceTB/SmolLM2-135M-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
HuggingFaceTB/SmolLM2-360M-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
HuggingFaceTB/SmolLM2-1.7B-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
ibm-granite/granite-3.0-2b-instruct
added_tokens.json,config.json,merges.txt,model-00001-of-00002.safetensors,model-00002-of-00002.safetensors,model.safetensors.index.json,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
Qwen/Qwen2.5-0.5B-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
Qwen/Qwen2.5-1.5B-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
Qwen/Qwen2.5-Coder-1.5B-Instruct
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
stabilityai/stablelm-2-zephyr-1_6b
config.json,generation_config.json,merges.txt,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json,vocab.json
- >-
THUDM/glm-edge-1.5b-chat
config.json,generation_config.json,model.safetensors,special_tokens_map.json,tokenizer.json,tokenizer_config.json
chat
Serverless small language model inference.
Models
Ungated models under 2B parameters:
- 01-ai/Yi-Coder-1.5B-Chat
- HuggingFaceTB/SmolLM2-135M-Instruct
- HuggingFaceTB/SmolLM2-360M-Instruct
- HuggingFaceTB/SmolLM2-1.7B-Instruct
- ibm-granite/granite-3.0-2b-instruct
- Qwen/Qwen2.5-0.5B-Instruct
- Qwen/Qwen2.5-1.5B-Instruct
- Qwen/Qwen2.5-Coder-1.5B-Instruct
- stabilityai/stablelm-2-zephyr-1_6b
- THUDM/glm-edge-1.5b-chat
Installation
# clone
git clone https://huggingface.co/spaces/adamelliotfields/chat.git
cd chat
# install
uv venv
uv pip install -r requirements.txt
# gradio
source .venv/bin/activate
gradio app.py
Development
Auth
Use existing HF_TOKEN
:
git remote set-url origin https://adamelliotfields:$HF_TOKEN@huggingface.co/spaces/adamelliotfields/chat
PRs
See pull requests and discussions.
git fetch origin refs/pr/42:pr/42
git checkout pr/42
# ...
git add .
git commit -m "Commit message"
git push origin pr/42:refs/pr/42