Spaces:
Runtime error
Runtime error
Update config.yaml
Browse files- config.yaml +5 -1
config.yaml
CHANGED
|
@@ -2,6 +2,9 @@
|
|
| 2 |
# Optimisée pour Hugging Face Space gratuit (2GB RAM, CPU)
|
| 3 |
# Version: 1.0.0
|
| 4 |
|
|
|
|
|
|
|
|
|
|
| 5 |
# ===== CONFIGURATION MODÈLES =====
|
| 6 |
models:
|
| 7 |
# LLM principal (compatible LlamaIndex v0.11.29 du projet)
|
|
@@ -16,7 +19,7 @@ models:
|
|
| 16 |
model_kwargs:
|
| 17 |
n_gpu_layers: 0
|
| 18 |
torch_dtype: float32
|
| 19 |
-
cache_dir:
|
| 20 |
|
| 21 |
|
| 22 |
# Embedding local (pas d'API externe)
|
|
@@ -27,6 +30,7 @@ models:
|
|
| 27 |
max_length: 512
|
| 28 |
normalize: true
|
| 29 |
batch_size: 32 # Optimisé pour Space gratuit
|
|
|
|
| 30 |
alternatives:
|
| 31 |
- "sentence-transformers/all-MiniLM-L6-v2" # Plus léger
|
| 32 |
- "jinaai/jina-embeddings-v2-small-en" # Alternative
|
|
|
|
| 2 |
# Optimisée pour Hugging Face Space gratuit (2GB RAM, CPU)
|
| 3 |
# Version: 1.0.0
|
| 4 |
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
|
| 8 |
# ===== CONFIGURATION MODÈLES =====
|
| 9 |
models:
|
| 10 |
# LLM principal (compatible LlamaIndex v0.11.29 du projet)
|
|
|
|
| 19 |
model_kwargs:
|
| 20 |
n_gpu_layers: 0
|
| 21 |
torch_dtype: float32
|
| 22 |
+
cache_dir: /tmp/llm # ✅ CORRIGE ICI
|
| 23 |
|
| 24 |
|
| 25 |
# Embedding local (pas d'API externe)
|
|
|
|
| 30 |
max_length: 512
|
| 31 |
normalize: true
|
| 32 |
batch_size: 32 # Optimisé pour Space gratuit
|
| 33 |
+
cache_dir: /tmp/embeddings # ✅ Écrivable sur HF
|
| 34 |
alternatives:
|
| 35 |
- "sentence-transformers/all-MiniLM-L6-v2" # Plus léger
|
| 36 |
- "jinaai/jina-embeddings-v2-small-en" # Alternative
|