nazdridoy commited on
Commit
2d515ec
Β·
verified Β·
1 Parent(s): 7d06e0a

docs(README): update model support and HF-Inferoxy details

Browse files

- [docs] Update introductory text for universal model support and HF-Inferoxy (README.md:78)
- [docs] Replace "Chat Models" section with universal support and new table (README.md:80-87)
- [docs] Replace "Image Models" section with universal support and new table (README.md:89-95)
- [docs] Add "Supported Providers" section listing 11 HF inference providers (README.md:97-110)
- [docs] Add "How It Works" section detailing HF-Inferoxy token management and routing (README.md:112-123)

Files changed (1) hide show
  1. README.md +50 -16
README.md CHANGED
@@ -78,22 +78,56 @@ The app requires:
78
 
79
  ## 🌟 Supported Models & Providers
80
 
81
- ### Chat Models
82
-
83
- | Model | Provider | Description |
84
- |-------|----------|-------------|
85
- | `openai/gpt-oss-20b` | Fireworks AI, Cerebras, Groq | Fast general purpose model |
86
- | `meta-llama/Llama-2-7b-chat-hf` | HF Inference | Chat-optimized model |
87
- | `mistralai/Mistral-7B-Instruct-v0.2` | Featherless AI | Instruction following |
88
- | `CohereLabs/c4ai-command-r-plus` | Cohere | Advanced language model |
89
-
90
- ### Image Models
91
-
92
- | Model | Provider | Description |
93
- |-------|----------|-------------|
94
- | `Qwen/Qwen-Image` | Fal.ai, Replicate | Advanced image generation (default) |
95
- | `black-forest-labs/FLUX.1-dev` | Nebius, Together | State-of-the-art image model |
96
- | `stabilityai/stable-diffusion-xl-base-1.0` | HF Inference, NScale | High-quality SDXL model |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
 
98
  ## 🎨 Usage Examples
99
 
 
78
 
79
  ## 🌟 Supported Models & Providers
80
 
81
+ This application supports **any and all models** that are compatible with Hugging Face's inference providers. The platform uses HF-Inferoxy for intelligent token management and can automatically route requests to the best available provider.
82
+
83
+ ### πŸ€– Chat Models
84
+
85
+ **Universal Support**: Any text generation model available through HF inference providers, including:
86
+
87
+ | Model Category | Examples | Providers |
88
+ |----------------|----------|-----------|
89
+ | **Open Source LLMs** | `openai/gpt-oss-20b`, `meta-llama/Llama-2-7b-chat-hf`, `microsoft/DialoGPT-medium` | Auto, Fireworks AI, Cerebras, Groq |
90
+ | **Instruction Models** | `google/flan-t5-base`, `mistralai/Mistral-7B-Instruct-v0.2` | Auto, Together, Cohere |
91
+ | **Multilingual Models** | `CohereLabs/c4ai-command-r-plus`, `bigscience/bloomz` | Cohere, Auto, Together |
92
+ | **Specialized Models** | Any HF-hosted model with chat completion API | All supported providers |
93
+
94
+ ### 🎨 Image Models
95
+
96
+ **Universal Support**: Any text-to-image model available through HF inference providers, including:
97
+
98
+ | Model Category | Examples | Providers |
99
+ |----------------|----------|-----------|
100
+ | **Diffusion Models** | `stabilityai/stable-diffusion-xl-base-1.0`, `runwayml/stable-diffusion-v1-5` | HF Inference, NScale, Together |
101
+ | **Advanced Generators** | `Qwen/Qwen-Image`, `black-forest-labs/FLUX.1-dev` | Fal.ai, Replicate, Nebius |
102
+ | **Specialized Models** | Any HF-hosted image generation model | All supported providers |
103
+
104
+ ### 🌐 Supported Providers
105
+
106
+ The application automatically works with all Hugging Face inference providers:
107
+
108
+ - **Auto** - Intelligent provider selection (default)
109
+ - **HF Inference** - Core Hugging Face API
110
+ - **Fireworks AI** - Fast and reliable inference
111
+ - **Cerebras** - High-performance computing
112
+ - **Groq** - Ultra-fast inference
113
+ - **Together** - Collaborative AI hosting
114
+ - **Cohere** - Advanced language models
115
+ - **Fal.ai** - High-quality image generation
116
+ - **Replicate** - Collaborative AI hosting
117
+ - **Nebius** - Cloud-native services
118
+ - **NScale** - Optimized inference performance
119
+
120
+ ### πŸ’‘ How It Works
121
+
122
+ 1. **Model Format**: Use `model_name` or `model_name:provider` format
123
+ 2. **Auto Provider**: When no provider is specified, HF-Inferoxy automatically selects the best available provider
124
+ 3. **Fallback System**: If one provider fails, the system automatically tries alternatives
125
+ 4. **Token Management**: HF-Inferoxy handles token rotation and quota management automatically
126
+
127
+ **Examples:**
128
+ - `openai/gpt-oss-20b` (auto provider selection)
129
+ - `openai/gpt-oss-20b:fireworks-ai` (specific provider)
130
+ - `Qwen/Qwen-Image:fal-ai` (image model with specific provider)
131
 
132
  ## 🎨 Usage Examples
133