|
""" |
|
Central configuration for models and providers |
|
|
|
This file centralizes all configurations related to models and providers used in the application. |
|
""" |
|
|
|
|
|
|
|
PREFERRED_PROVIDERS = ["fireworks-ai", "sambanova", "novita"] |
|
|
|
|
|
DEFAULT_EVALUATION_MODELS = [ |
|
"Qwen/QwQ-32B", |
|
"Qwen/Qwen2.5-72B-Instruct", |
|
"Qwen/Qwen2.5-32B-Instruct", |
|
"meta-llama/Llama-3.1-8B-Instruct", |
|
"meta-llama/Llama-3.3-70B-Instruct", |
|
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B", |
|
"mistralai/Mistral-Small-24B-Instruct-2501", |
|
] |
|
|
|
|
|
ALTERNATIVE_BENCHMARK_MODELS = [ |
|
"meta-llama/Llama-3.3-70B-Instruct", |
|
"meta-llama/Llama-3.1-8B-Instruct", |
|
"Qwen/Qwen2.5-72B-Instruct", |
|
"mistralai/Mistral-Small-24B-Instruct-2501", |
|
|
|
"HuggingFaceH4/zephyr-7b-beta", |
|
"TinyLlama/TinyLlama-1.1B-Chat-v1.0", |
|
"microsoft/phi-2", |
|
] |
|
|
|
|
|
DEFAULT_BENCHMARK_MODEL = "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" |
|
|
|
|
|
|
|
BENCHMARK_MODEL_ROLES = { |
|
"ingestion": [DEFAULT_BENCHMARK_MODEL], |
|
"summarization": [DEFAULT_BENCHMARK_MODEL], |
|
"chunking": ["intfloat/multilingual-e5-large-instruct"], |
|
"single_shot_question_generation": [DEFAULT_BENCHMARK_MODEL], |
|
"multi_hop_question_generation": [DEFAULT_BENCHMARK_MODEL], |
|
} |
|
|
|
|
|
DEFAULT_EVALUATION_TIMEOUT = 60.0 |
|
|
|
|
|
DEFAULT_BENCHMARK_TIMEOUT = 300.0 |