Spaces:
Running
Running
docs(README): update model support and HF-Inferoxy details
Browse files- [docs] Update introductory text for universal model support and HF-Inferoxy (README.md:78)
- [docs] Replace "Chat Models" section with universal support and new table (README.md:80-87)
- [docs] Replace "Image Models" section with universal support and new table (README.md:89-95)
- [docs] Add "Supported Providers" section listing 11 HF inference providers (README.md:97-110)
- [docs] Add "How It Works" section detailing HF-Inferoxy token management and routing (README.md:112-123)
README.md
CHANGED
|
@@ -78,22 +78,56 @@ The app requires:
|
|
| 78 |
|
| 79 |
## π Supported Models & Providers
|
| 80 |
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
| 85 |
-
|
| 86 |
-
|
| 87 |
-
|
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
|
| 96 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 97 |
|
| 98 |
## π¨ Usage Examples
|
| 99 |
|
|
|
|
| 78 |
|
| 79 |
## π Supported Models & Providers
|
| 80 |
|
| 81 |
+
This application supports **any and all models** that are compatible with Hugging Face's inference providers. The platform uses HF-Inferoxy for intelligent token management and can automatically route requests to the best available provider.
|
| 82 |
+
|
| 83 |
+
### π€ Chat Models
|
| 84 |
+
|
| 85 |
+
**Universal Support**: Any text generation model available through HF inference providers, including:
|
| 86 |
+
|
| 87 |
+
| Model Category | Examples | Providers |
|
| 88 |
+
|----------------|----------|-----------|
|
| 89 |
+
| **Open Source LLMs** | `openai/gpt-oss-20b`, `meta-llama/Llama-2-7b-chat-hf`, `microsoft/DialoGPT-medium` | Auto, Fireworks AI, Cerebras, Groq |
|
| 90 |
+
| **Instruction Models** | `google/flan-t5-base`, `mistralai/Mistral-7B-Instruct-v0.2` | Auto, Together, Cohere |
|
| 91 |
+
| **Multilingual Models** | `CohereLabs/c4ai-command-r-plus`, `bigscience/bloomz` | Cohere, Auto, Together |
|
| 92 |
+
| **Specialized Models** | Any HF-hosted model with chat completion API | All supported providers |
|
| 93 |
+
|
| 94 |
+
### π¨ Image Models
|
| 95 |
+
|
| 96 |
+
**Universal Support**: Any text-to-image model available through HF inference providers, including:
|
| 97 |
+
|
| 98 |
+
| Model Category | Examples | Providers |
|
| 99 |
+
|----------------|----------|-----------|
|
| 100 |
+
| **Diffusion Models** | `stabilityai/stable-diffusion-xl-base-1.0`, `runwayml/stable-diffusion-v1-5` | HF Inference, NScale, Together |
|
| 101 |
+
| **Advanced Generators** | `Qwen/Qwen-Image`, `black-forest-labs/FLUX.1-dev` | Fal.ai, Replicate, Nebius |
|
| 102 |
+
| **Specialized Models** | Any HF-hosted image generation model | All supported providers |
|
| 103 |
+
|
| 104 |
+
### π Supported Providers
|
| 105 |
+
|
| 106 |
+
The application automatically works with all Hugging Face inference providers:
|
| 107 |
+
|
| 108 |
+
- **Auto** - Intelligent provider selection (default)
|
| 109 |
+
- **HF Inference** - Core Hugging Face API
|
| 110 |
+
- **Fireworks AI** - Fast and reliable inference
|
| 111 |
+
- **Cerebras** - High-performance computing
|
| 112 |
+
- **Groq** - Ultra-fast inference
|
| 113 |
+
- **Together** - Collaborative AI hosting
|
| 114 |
+
- **Cohere** - Advanced language models
|
| 115 |
+
- **Fal.ai** - High-quality image generation
|
| 116 |
+
- **Replicate** - Collaborative AI hosting
|
| 117 |
+
- **Nebius** - Cloud-native services
|
| 118 |
+
- **NScale** - Optimized inference performance
|
| 119 |
+
|
| 120 |
+
### π‘ How It Works
|
| 121 |
+
|
| 122 |
+
1. **Model Format**: Use `model_name` or `model_name:provider` format
|
| 123 |
+
2. **Auto Provider**: When no provider is specified, HF-Inferoxy automatically selects the best available provider
|
| 124 |
+
3. **Fallback System**: If one provider fails, the system automatically tries alternatives
|
| 125 |
+
4. **Token Management**: HF-Inferoxy handles token rotation and quota management automatically
|
| 126 |
+
|
| 127 |
+
**Examples:**
|
| 128 |
+
- `openai/gpt-oss-20b` (auto provider selection)
|
| 129 |
+
- `openai/gpt-oss-20b:fireworks-ai` (specific provider)
|
| 130 |
+
- `Qwen/Qwen-Image:fal-ai` (image model with specific provider)
|
| 131 |
|
| 132 |
## π¨ Usage Examples
|
| 133 |
|