Spaces:
Paused
Paused
# | |
# Copyright 2024 The InfiniFlow Authors. All Rights Reserved. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
# | |
import os | |
from enum import IntEnum, Enum | |
from api.utils.file_utils import get_project_base_directory | |
from api.utils.log_utils import LoggerFactory, getLogger | |
# Logger | |
LoggerFactory.set_directory( | |
os.path.join( | |
get_project_base_directory(), | |
"logs", | |
"api")) | |
# {CRITICAL: 50, FATAL:50, ERROR:40, WARNING:30, WARN:30, INFO:20, DEBUG:10, NOTSET:0} | |
LoggerFactory.LEVEL = 30 | |
stat_logger = getLogger("stat") | |
access_logger = getLogger("access") | |
database_logger = getLogger("database") | |
chat_logger = getLogger("chat") | |
from rag.utils.es_conn import ELASTICSEARCH | |
from rag.nlp import search | |
from graphrag import search as kg_search | |
from api.utils import get_base_config, decrypt_database_config | |
API_VERSION = "v1" | |
RAG_FLOW_SERVICE_NAME = "ragflow" | |
SERVER_MODULE = "rag_flow_server.py" | |
TEMP_DIRECTORY = os.path.join(get_project_base_directory(), "temp") | |
RAG_FLOW_CONF_PATH = os.path.join(get_project_base_directory(), "conf") | |
SUBPROCESS_STD_LOG_NAME = "std.log" | |
ERROR_REPORT = True | |
ERROR_REPORT_WITH_PATH = False | |
MAX_TIMESTAMP_INTERVAL = 60 | |
SESSION_VALID_PERIOD = 7 * 24 * 60 * 60 | |
REQUEST_TRY_TIMES = 3 | |
REQUEST_WAIT_SEC = 2 | |
REQUEST_MAX_WAIT_SEC = 300 | |
USE_REGISTRY = get_base_config("use_registry") | |
default_llm = { | |
"Tongyi-Qianwen": { | |
"chat_model": "qwen-plus", | |
"embedding_model": "text-embedding-v2", | |
"image2text_model": "qwen-vl-max", | |
"asr_model": "paraformer-realtime-8k-v1", | |
}, | |
"OpenAI": { | |
"chat_model": "gpt-3.5-turbo", | |
"embedding_model": "text-embedding-ada-002", | |
"image2text_model": "gpt-4-vision-preview", | |
"asr_model": "whisper-1", | |
}, | |
"Azure-OpenAI": { | |
"chat_model": "azure-gpt-35-turbo", | |
"embedding_model": "azure-text-embedding-ada-002", | |
"image2text_model": "azure-gpt-4-vision-preview", | |
"asr_model": "azure-whisper-1", | |
}, | |
"ZHIPU-AI": { | |
"chat_model": "glm-3-turbo", | |
"embedding_model": "embedding-2", | |
"image2text_model": "glm-4v", | |
"asr_model": "", | |
}, | |
"Ollama": { | |
"chat_model": "qwen-14B-chat", | |
"embedding_model": "flag-embedding", | |
"image2text_model": "", | |
"asr_model": "", | |
}, | |
"Moonshot": { | |
"chat_model": "moonshot-v1-8k", | |
"embedding_model": "", | |
"image2text_model": "", | |
"asr_model": "", | |
}, | |
"DeepSeek": { | |
"chat_model": "deepseek-chat", | |
"embedding_model": "", | |
"image2text_model": "", | |
"asr_model": "", | |
}, | |
"VolcEngine": { | |
"chat_model": "", | |
"embedding_model": "", | |
"image2text_model": "", | |
"asr_model": "", | |
}, | |
"BAAI": { | |
"chat_model": "", | |
"embedding_model": "BAAI/bge-large-zh-v1.5", | |
"image2text_model": "", | |
"asr_model": "", | |
"rerank_model": "BAAI/bge-reranker-v2-m3", | |
} | |
} | |
LLM = get_base_config("user_default_llm", {}) | |
LLM_FACTORY = LLM.get("factory", "Tongyi-Qianwen") | |
LLM_BASE_URL = LLM.get("base_url") | |
if LLM_FACTORY not in default_llm: | |
print( | |
"\33[91m【ERROR】\33[0m:", | |
f"LLM factory {LLM_FACTORY} has not supported yet, switch to 'Tongyi-Qianwen/QWen' automatically, and please check the API_KEY in service_conf.yaml.") | |
LLM_FACTORY = "Tongyi-Qianwen" | |
CHAT_MDL = default_llm[LLM_FACTORY]["chat_model"] | |
EMBEDDING_MDL = default_llm["BAAI"]["embedding_model"] | |
RERANK_MDL = default_llm["BAAI"]["rerank_model"] | |
ASR_MDL = default_llm[LLM_FACTORY]["asr_model"] | |
IMAGE2TEXT_MDL = default_llm[LLM_FACTORY]["image2text_model"] | |
API_KEY = LLM.get("api_key", "") | |
PARSERS = LLM.get( | |
"parsers", | |
"naive:General,qa:Q&A,resume:Resume,manual:Manual,table:Table,paper:Paper,book:Book,laws:Laws,presentation:Presentation,picture:Picture,one:One,audio:Audio,knowledge_graph:Knowledge Graph,email:Email") | |
# distribution | |
DEPENDENT_DISTRIBUTION = get_base_config("dependent_distribution", False) | |
RAG_FLOW_UPDATE_CHECK = False | |
HOST = get_base_config(RAG_FLOW_SERVICE_NAME, {}).get("host", "127.0.0.1") | |
HTTP_PORT = get_base_config(RAG_FLOW_SERVICE_NAME, {}).get("http_port") | |
SECRET_KEY = get_base_config( | |
RAG_FLOW_SERVICE_NAME, | |
{}).get( | |
"secret_key", | |
"infiniflow") | |
TOKEN_EXPIRE_IN = get_base_config( | |
RAG_FLOW_SERVICE_NAME, {}).get( | |
"token_expires_in", 3600) | |
NGINX_HOST = get_base_config( | |
RAG_FLOW_SERVICE_NAME, {}).get( | |
"nginx", {}).get("host") or HOST | |
NGINX_HTTP_PORT = get_base_config( | |
RAG_FLOW_SERVICE_NAME, {}).get( | |
"nginx", {}).get("http_port") or HTTP_PORT | |
RANDOM_INSTANCE_ID = get_base_config( | |
RAG_FLOW_SERVICE_NAME, {}).get( | |
"random_instance_id", False) | |
PROXY = get_base_config(RAG_FLOW_SERVICE_NAME, {}).get("proxy") | |
PROXY_PROTOCOL = get_base_config(RAG_FLOW_SERVICE_NAME, {}).get("protocol") | |
DATABASE = decrypt_database_config(name="mysql") | |
# Switch | |
# upload | |
UPLOAD_DATA_FROM_CLIENT = True | |
# authentication | |
AUTHENTICATION_CONF = get_base_config("authentication", {}) | |
# client | |
CLIENT_AUTHENTICATION = AUTHENTICATION_CONF.get( | |
"client", {}).get( | |
"switch", False) | |
HTTP_APP_KEY = AUTHENTICATION_CONF.get("client", {}).get("http_app_key") | |
GITHUB_OAUTH = get_base_config("oauth", {}).get("github") | |
FEISHU_OAUTH = get_base_config("oauth", {}).get("feishu") | |
WECHAT_OAUTH = get_base_config("oauth", {}).get("wechat") | |
# site | |
SITE_AUTHENTICATION = AUTHENTICATION_CONF.get("site", {}).get("switch", False) | |
# permission | |
PERMISSION_CONF = get_base_config("permission", {}) | |
PERMISSION_SWITCH = PERMISSION_CONF.get("switch") | |
COMPONENT_PERMISSION = PERMISSION_CONF.get("component") | |
DATASET_PERMISSION = PERMISSION_CONF.get("dataset") | |
HOOK_MODULE = get_base_config("hook_module") | |
HOOK_SERVER_NAME = get_base_config("hook_server_name") | |
ENABLE_MODEL_STORE = get_base_config('enable_model_store', False) | |
# authentication | |
USE_AUTHENTICATION = False | |
USE_DATA_AUTHENTICATION = False | |
AUTOMATIC_AUTHORIZATION_OUTPUT_DATA = True | |
USE_DEFAULT_TIMEOUT = False | |
AUTHENTICATION_DEFAULT_TIMEOUT = 7 * 24 * 60 * 60 # s | |
PRIVILEGE_COMMAND_WHITELIST = [] | |
CHECK_NODES_IDENTITY = False | |
retrievaler = search.Dealer(ELASTICSEARCH) | |
kg_retrievaler = kg_search.KGSearch(ELASTICSEARCH) | |
class CustomEnum(Enum): | |
def valid(cls, value): | |
try: | |
cls(value) | |
return True | |
except BaseException: | |
return False | |
def values(cls): | |
return [member.value for member in cls.__members__.values()] | |
def names(cls): | |
return [member.name for member in cls.__members__.values()] | |
class PythonDependenceName(CustomEnum): | |
Rag_Source_Code = "python" | |
Python_Env = "miniconda" | |
class ModelStorage(CustomEnum): | |
REDIS = "redis" | |
MYSQL = "mysql" | |
class RetCode(IntEnum, CustomEnum): | |
SUCCESS = 0 | |
NOT_EFFECTIVE = 10 | |
EXCEPTION_ERROR = 100 | |
ARGUMENT_ERROR = 101 | |
DATA_ERROR = 102 | |
OPERATING_ERROR = 103 | |
CONNECTION_ERROR = 105 | |
RUNNING = 106 | |
PERMISSION_ERROR = 108 | |
AUTHENTICATION_ERROR = 109 | |
UNAUTHORIZED = 401 | |
SERVER_ERROR = 500 | |