| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| import pathlib |
| import datetime |
|
|
| from api.db.services.dialog_service import keyword_extraction, label_question |
| from rag.app.qa import rmPrefix, beAdoc |
| from rag.nlp import rag_tokenizer |
| from api.db import LLMType, ParserType |
| from api.db.services.llm_service import TenantLLMService, LLMBundle |
| from api import settings |
| import xxhash |
| import re |
| from api.utils.api_utils import token_required |
| from api.db.db_models import Task |
| from api.db.services.task_service import TaskService, queue_tasks |
| from api.utils.api_utils import server_error_response |
| from api.utils.api_utils import get_result, get_error_data_result |
| from io import BytesIO |
| from flask import request, send_file |
| from api.db import FileSource, TaskStatus, FileType |
| from api.db.db_models import File |
| from api.db.services.document_service import DocumentService |
| from api.db.services.file2document_service import File2DocumentService |
| from api.db.services.file_service import FileService |
| from api.db.services.knowledgebase_service import KnowledgebaseService |
| from api.utils.api_utils import construct_json_result, get_parser_config |
| from rag.nlp import search |
| from rag.utils import rmSpace |
| from rag.utils.storage_factory import STORAGE_IMPL |
|
|
| from pydantic import BaseModel, Field, validator |
|
|
| MAXIMUM_OF_UPLOADING_FILES = 256 |
|
|
|
|
| class Chunk(BaseModel): |
| id: str = "" |
| content: str = "" |
| document_id: str = "" |
| docnm_kwd: str = "" |
| important_keywords: list = Field(default_factory=list) |
| questions: list = Field(default_factory=list) |
| question_tks: str = "" |
| image_id: str = "" |
| available: bool = True |
| positions: list[list[int]] = Field(default_factory=list) |
|
|
| @validator('positions') |
| def validate_positions(cls, value): |
| for sublist in value: |
| if len(sublist) != 5: |
| raise ValueError("Each sublist in positions must have a length of 5") |
| return value |
|
|
| @manager.route("/datasets/<dataset_id>/documents", methods=["POST"]) |
| @token_required |
| def upload(dataset_id, tenant_id): |
| """ |
| Upload documents to a dataset. |
| --- |
| tags: |
| - Documents |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| - in: formData |
| name: file |
| type: file |
| required: true |
| description: Document files to upload. |
| responses: |
| 200: |
| description: Successfully uploaded documents. |
| schema: |
| type: object |
| properties: |
| data: |
| type: array |
| items: |
| type: object |
| properties: |
| id: |
| type: string |
| description: Document ID. |
| name: |
| type: string |
| description: Document name. |
| chunk_count: |
| type: integer |
| description: Number of chunks. |
| token_count: |
| type: integer |
| description: Number of tokens. |
| dataset_id: |
| type: string |
| description: ID of the dataset. |
| chunk_method: |
| type: string |
| description: Chunking method used. |
| run: |
| type: string |
| description: Processing status. |
| """ |
| if "file" not in request.files: |
| return get_error_data_result( |
| message="No file part!", code=settings.RetCode.ARGUMENT_ERROR |
| ) |
| file_objs = request.files.getlist("file") |
| for file_obj in file_objs: |
| if file_obj.filename == "": |
| return get_result( |
| message="No file selected!", code=settings.RetCode.ARGUMENT_ERROR |
| ) |
| ''' |
| # total size |
| total_size = 0 |
| for file_obj in file_objs: |
| file_obj.seek(0, os.SEEK_END) |
| total_size += file_obj.tell() |
| file_obj.seek(0) |
| MAX_TOTAL_FILE_SIZE = 10 * 1024 * 1024 |
| if total_size > MAX_TOTAL_FILE_SIZE: |
| return get_result( |
| message=f"Total file size exceeds 10MB limit! ({total_size / (1024 * 1024):.2f} MB)", |
| code=settings.RetCode.ARGUMENT_ERROR, |
| ) |
| ''' |
| e, kb = KnowledgebaseService.get_by_id(dataset_id) |
| if not e: |
| raise LookupError(f"Can't find the dataset with ID {dataset_id}!") |
| err, files = FileService.upload_document(kb, file_objs, tenant_id) |
| if err: |
| return get_result(message="\n".join(err), code=settings.RetCode.SERVER_ERROR) |
| |
| renamed_doc_list = [] |
| for file in files: |
| doc = file[0] |
| key_mapping = { |
| "chunk_num": "chunk_count", |
| "kb_id": "dataset_id", |
| "token_num": "token_count", |
| "parser_id": "chunk_method", |
| } |
| renamed_doc = {} |
| for key, value in doc.items(): |
| new_key = key_mapping.get(key, key) |
| renamed_doc[new_key] = value |
| renamed_doc["run"] = "UNSTART" |
| renamed_doc_list.append(renamed_doc) |
| return get_result(data=renamed_doc_list) |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/documents/<document_id>", methods=["PUT"]) |
| @token_required |
| def update_doc(tenant_id, dataset_id, document_id): |
| """ |
| Update a document within a dataset. |
| --- |
| tags: |
| - Documents |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document to update. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| - in: body |
| name: body |
| description: Document update parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| name: |
| type: string |
| description: New name of the document. |
| parser_config: |
| type: object |
| description: Parser configuration. |
| chunk_method: |
| type: string |
| description: Chunking method. |
| responses: |
| 200: |
| description: Document updated successfully. |
| schema: |
| type: object |
| """ |
| req = request.json |
| if not KnowledgebaseService.query(id=dataset_id, tenant_id=tenant_id): |
| return get_error_data_result(message="You don't own the dataset.") |
| doc = DocumentService.query(kb_id=dataset_id, id=document_id) |
| if not doc: |
| return get_error_data_result(message="The dataset doesn't own the document.") |
| doc = doc[0] |
| if "chunk_count" in req: |
| if req["chunk_count"] != doc.chunk_num: |
| return get_error_data_result(message="Can't change `chunk_count`.") |
| if "token_count" in req: |
| if req["token_count"] != doc.token_num: |
| return get_error_data_result(message="Can't change `token_count`.") |
| if "progress" in req: |
| if req["progress"] != doc.progress: |
| return get_error_data_result(message="Can't change `progress`.") |
|
|
| if "name" in req and req["name"] != doc.name: |
| if ( |
| pathlib.Path(req["name"].lower()).suffix |
| != pathlib.Path(doc.name.lower()).suffix |
| ): |
| return get_result( |
| message="The extension of file can't be changed", |
| code=settings.RetCode.ARGUMENT_ERROR, |
| ) |
| for d in DocumentService.query(name=req["name"], kb_id=doc.kb_id): |
| if d.name == req["name"]: |
| return get_error_data_result( |
| message="Duplicated document name in the same dataset." |
| ) |
| if not DocumentService.update_by_id(document_id, {"name": req["name"]}): |
| return get_error_data_result(message="Database error (Document rename)!") |
|
|
| informs = File2DocumentService.get_by_document_id(document_id) |
| if informs: |
| e, file = FileService.get_by_id(informs[0].file_id) |
| FileService.update_by_id(file.id, {"name": req["name"]}) |
| if "parser_config" in req: |
| DocumentService.update_parser_config(doc.id, req["parser_config"]) |
| if "chunk_method" in req: |
| valid_chunk_method = { |
| "naive", |
| "manual", |
| "qa", |
| "table", |
| "paper", |
| "book", |
| "laws", |
| "presentation", |
| "picture", |
| "one", |
| "knowledge_graph", |
| "email", |
| "tag" |
| } |
| if req.get("chunk_method") not in valid_chunk_method: |
| return get_error_data_result( |
| f"`chunk_method` {req['chunk_method']} doesn't exist" |
| ) |
| if doc.parser_id.lower() == req["chunk_method"].lower(): |
| return get_result() |
|
|
| if doc.type == FileType.VISUAL or re.search(r"\.(ppt|pptx|pages)$", doc.name): |
| return get_error_data_result(message="Not supported yet!") |
|
|
| e = DocumentService.update_by_id( |
| doc.id, |
| { |
| "parser_id": req["chunk_method"], |
| "progress": 0, |
| "progress_msg": "", |
| "run": TaskStatus.UNSTART.value, |
| }, |
| ) |
| if not e: |
| return get_error_data_result(message="Document not found!") |
| req["parser_config"] = get_parser_config( |
| req["chunk_method"], req.get("parser_config") |
| ) |
| DocumentService.update_parser_config(doc.id, req["parser_config"]) |
| if doc.token_num > 0: |
| e = DocumentService.increment_chunk_num( |
| doc.id, |
| doc.kb_id, |
| doc.token_num * -1, |
| doc.chunk_num * -1, |
| doc.process_duation * -1, |
| ) |
| if not e: |
| return get_error_data_result(message="Document not found!") |
| settings.docStoreConn.delete({"doc_id": doc.id}, search.index_name(tenant_id), dataset_id) |
|
|
| return get_result() |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/documents/<document_id>", methods=["GET"]) |
| @token_required |
| def download(tenant_id, dataset_id, document_id): |
| """ |
| Download a document from a dataset. |
| --- |
| tags: |
| - Documents |
| security: |
| - ApiKeyAuth: [] |
| produces: |
| - application/octet-stream |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document to download. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Document file stream. |
| schema: |
| type: file |
| 400: |
| description: Error message. |
| schema: |
| type: object |
| """ |
| if not KnowledgebaseService.query(id=dataset_id, tenant_id=tenant_id): |
| return get_error_data_result(message=f"You do not own the dataset {dataset_id}.") |
| doc = DocumentService.query(kb_id=dataset_id, id=document_id) |
| if not doc: |
| return get_error_data_result( |
| message=f"The dataset not own the document {document_id}." |
| ) |
| |
| doc_id, doc_location = File2DocumentService.get_storage_address( |
| doc_id=document_id |
| ) |
| file_stream = STORAGE_IMPL.get(doc_id, doc_location) |
| if not file_stream: |
| return construct_json_result( |
| message="This file is empty.", code=settings.RetCode.DATA_ERROR |
| ) |
| file = BytesIO(file_stream) |
| |
| return send_file( |
| file, |
| as_attachment=True, |
| download_name=doc[0].name, |
| mimetype="application/octet-stream", |
| ) |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/documents", methods=["GET"]) |
| @token_required |
| def list_docs(dataset_id, tenant_id): |
| """ |
| List documents in a dataset. |
| --- |
| tags: |
| - Documents |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: query |
| name: id |
| type: string |
| required: false |
| description: Filter by document ID. |
| - in: query |
| name: page |
| type: integer |
| required: false |
| default: 1 |
| description: Page number. |
| - in: query |
| name: page_size |
| type: integer |
| required: false |
| default: 30 |
| description: Number of items per page. |
| - in: query |
| name: orderby |
| type: string |
| required: false |
| default: "create_time" |
| description: Field to order by. |
| - in: query |
| name: desc |
| type: boolean |
| required: false |
| default: true |
| description: Order in descending. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: List of documents. |
| schema: |
| type: object |
| properties: |
| total: |
| type: integer |
| description: Total number of documents. |
| docs: |
| type: array |
| items: |
| type: object |
| properties: |
| id: |
| type: string |
| description: Document ID. |
| name: |
| type: string |
| description: Document name. |
| chunk_count: |
| type: integer |
| description: Number of chunks. |
| token_count: |
| type: integer |
| description: Number of tokens. |
| dataset_id: |
| type: string |
| description: ID of the dataset. |
| chunk_method: |
| type: string |
| description: Chunking method used. |
| run: |
| type: string |
| description: Processing status. |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}. ") |
| id = request.args.get("id") |
| name = request.args.get("name") |
| if not DocumentService.query(id=id, kb_id=dataset_id): |
| return get_error_data_result(message=f"You don't own the document {id}.") |
| if not DocumentService.query(name=name, kb_id=dataset_id): |
| return get_error_data_result(message=f"You don't own the document {name}.") |
| page = int(request.args.get("page", 1)) |
| keywords = request.args.get("keywords", "") |
| page_size = int(request.args.get("page_size", 30)) |
| orderby = request.args.get("orderby", "create_time") |
| if request.args.get("desc") == "False": |
| desc = False |
| else: |
| desc = True |
| docs, tol = DocumentService.get_list( |
| dataset_id, page, page_size, orderby, desc, keywords, id, name |
| ) |
|
|
| |
| renamed_doc_list = [] |
| for doc in docs: |
| key_mapping = { |
| "chunk_num": "chunk_count", |
| "kb_id": "dataset_id", |
| "token_num": "token_count", |
| "parser_id": "chunk_method", |
| } |
| run_mapping = { |
| "0": "UNSTART", |
| "1": "RUNNING", |
| "2": "CANCEL", |
| "3": "DONE", |
| "4": "FAIL", |
| } |
| renamed_doc = {} |
| for key, value in doc.items(): |
| if key == "run": |
| renamed_doc["run"] = run_mapping.get(str(value)) |
| new_key = key_mapping.get(key, key) |
| renamed_doc[new_key] = value |
| if key == "run": |
| renamed_doc["run"] = run_mapping.get(value) |
| renamed_doc_list.append(renamed_doc) |
| return get_result(data={"total": tol, "docs": renamed_doc_list}) |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/documents", methods=["DELETE"]) |
| @token_required |
| def delete(tenant_id, dataset_id): |
| """ |
| Delete documents from a dataset. |
| --- |
| tags: |
| - Documents |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: body |
| name: body |
| description: Document deletion parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| ids: |
| type: array |
| items: |
| type: string |
| description: List of document IDs to delete. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Documents deleted successfully. |
| schema: |
| type: object |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}. ") |
| req = request.json |
| if not req: |
| doc_ids = None |
| else: |
| doc_ids = req.get("ids") |
| if not doc_ids: |
| doc_list = [] |
| docs = DocumentService.query(kb_id=dataset_id) |
| for doc in docs: |
| doc_list.append(doc.id) |
| else: |
| doc_list = doc_ids |
| root_folder = FileService.get_root_folder(tenant_id) |
| pf_id = root_folder["id"] |
| FileService.init_knowledgebase_docs(pf_id, tenant_id) |
| errors = "" |
| for doc_id in doc_list: |
| try: |
| e, doc = DocumentService.get_by_id(doc_id) |
| if not e: |
| return get_error_data_result(message="Document not found!") |
| tenant_id = DocumentService.get_tenant_id(doc_id) |
| if not tenant_id: |
| return get_error_data_result(message="Tenant not found!") |
|
|
| b, n = File2DocumentService.get_storage_address(doc_id=doc_id) |
|
|
| if not DocumentService.remove_document(doc, tenant_id): |
| return get_error_data_result( |
| message="Database error (Document removal)!" |
| ) |
|
|
| f2d = File2DocumentService.get_by_document_id(doc_id) |
| FileService.filter_delete( |
| [ |
| File.source_type == FileSource.KNOWLEDGEBASE, |
| File.id == f2d[0].file_id, |
| ] |
| ) |
| File2DocumentService.delete_by_document_id(doc_id) |
|
|
| STORAGE_IMPL.rm(b, n) |
| except Exception as e: |
| errors += str(e) |
|
|
| if errors: |
| return get_result(message=errors, code=settings.RetCode.SERVER_ERROR) |
|
|
| return get_result() |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/chunks", methods=["POST"]) |
| @token_required |
| def parse(tenant_id, dataset_id): |
| """ |
| Start parsing documents into chunks. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: body |
| name: body |
| description: Parsing parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| document_ids: |
| type: array |
| items: |
| type: string |
| description: List of document IDs to parse. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Parsing started successfully. |
| schema: |
| type: object |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| req = request.json |
| if not req.get("document_ids"): |
| return get_error_data_result("`document_ids` is required") |
| for id in req["document_ids"]: |
| doc = DocumentService.query(id=id, kb_id=dataset_id) |
| if not doc: |
| return get_error_data_result(message=f"You don't own the document {id}.") |
| if doc[0].progress != 0.0: |
| return get_error_data_result( |
| "Can't stop parsing document with progress at 0 or 100" |
| ) |
| info = {"run": "1", "progress": 0} |
| info["progress_msg"] = "" |
| info["chunk_num"] = 0 |
| info["token_num"] = 0 |
| DocumentService.update_by_id(id, info) |
| settings.docStoreConn.delete({"doc_id": id}, search.index_name(tenant_id), dataset_id) |
| TaskService.filter_delete([Task.doc_id == id]) |
| e, doc = DocumentService.get_by_id(id) |
| doc = doc.to_dict() |
| doc["tenant_id"] = tenant_id |
| bucket, name = File2DocumentService.get_storage_address(doc_id=doc["id"]) |
| queue_tasks(doc, bucket, name) |
| return get_result() |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/chunks", methods=["DELETE"]) |
| @token_required |
| def stop_parsing(tenant_id, dataset_id): |
| """ |
| Stop parsing documents into chunks. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: body |
| name: body |
| description: Stop parsing parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| document_ids: |
| type: array |
| items: |
| type: string |
| description: List of document IDs to stop parsing. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Parsing stopped successfully. |
| schema: |
| type: object |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| req = request.json |
| if not req.get("document_ids"): |
| return get_error_data_result("`document_ids` is required") |
| for id in req["document_ids"]: |
| doc = DocumentService.query(id=id, kb_id=dataset_id) |
| if not doc: |
| return get_error_data_result(message=f"You don't own the document {id}.") |
| if int(doc[0].progress) == 1 or doc[0].progress == 0: |
| return get_error_data_result( |
| "Can't stop parsing document with progress at 0 or 1" |
| ) |
| info = {"run": "2", "progress": 0, "chunk_num": 0} |
| DocumentService.update_by_id(id, info) |
| settings.docStoreConn.delete({"doc_id": doc.id}, search.index_name(tenant_id), dataset_id) |
| return get_result() |
|
|
|
|
| @manager.route("/datasets/<dataset_id>/documents/<document_id>/chunks", methods=["GET"]) |
| @token_required |
| def list_chunks(tenant_id, dataset_id, document_id): |
| """ |
| List chunks of a document. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document. |
| - in: query |
| name: page |
| type: integer |
| required: false |
| default: 1 |
| description: Page number. |
| - in: query |
| name: page_size |
| type: integer |
| required: false |
| default: 30 |
| description: Number of items per page. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: List of chunks. |
| schema: |
| type: object |
| properties: |
| total: |
| type: integer |
| description: Total number of chunks. |
| chunks: |
| type: array |
| items: |
| type: object |
| properties: |
| id: |
| type: string |
| description: Chunk ID. |
| content: |
| type: string |
| description: Chunk content. |
| document_id: |
| type: string |
| description: ID of the document. |
| important_keywords: |
| type: array |
| items: |
| type: string |
| description: Important keywords. |
| image_id: |
| type: string |
| description: Image ID associated with the chunk. |
| doc: |
| type: object |
| description: Document details. |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| doc = DocumentService.query(id=document_id, kb_id=dataset_id) |
| if not doc: |
| return get_error_data_result( |
| message=f"You don't own the document {document_id}." |
| ) |
| doc = doc[0] |
| req = request.args |
| doc_id = document_id |
| page = int(req.get("page", 1)) |
| size = int(req.get("page_size", 30)) |
| question = req.get("keywords", "") |
| query = { |
| "doc_ids": [doc_id], |
| "page": page, |
| "size": size, |
| "question": question, |
| "sort": True, |
| } |
| key_mapping = { |
| "chunk_num": "chunk_count", |
| "kb_id": "dataset_id", |
| "token_num": "token_count", |
| "parser_id": "chunk_method", |
| } |
| run_mapping = { |
| "0": "UNSTART", |
| "1": "RUNNING", |
| "2": "CANCEL", |
| "3": "DONE", |
| "4": "FAIL", |
| } |
| doc = doc.to_dict() |
| renamed_doc = {} |
| for key, value in doc.items(): |
| new_key = key_mapping.get(key, key) |
| renamed_doc[new_key] = value |
| if key == "run": |
| renamed_doc["run"] = run_mapping.get(str(value)) |
|
|
| res = {"total": 0, "chunks": [], "doc": renamed_doc} |
| if req.get("id"): |
| chunk = settings.docStoreConn.get(req.get("id"), search.index_name(tenant_id), [dataset_id]) |
| k = [] |
| for n in chunk.keys(): |
| if re.search(r"(_vec$|_sm_|_tks|_ltks)", n): |
| k.append(n) |
| for n in k: |
| del chunk[n] |
| if not chunk: |
| return get_error_data_result(f"Chunk `{req.get('id')}` not found.") |
| res['total'] = 1 |
| final_chunk = { |
| "id":chunk.get("id",chunk.get("chunk_id")), |
| "content":chunk["content_with_weight"], |
| "document_id":chunk.get("doc_id",chunk.get("document_id")), |
| "docnm_kwd":chunk["docnm_kwd"], |
| "important_keywords":chunk.get("important_kwd",[]), |
| "questions":chunk.get("question_kwd",[]), |
| "dataset_id":chunk.get("kb_id",chunk.get("dataset_id")), |
| "image_id":chunk["img_id"], |
| "available":bool(chunk.get("available_int",1)), |
| "positions":chunk.get("position_int",[]), |
| } |
| res["chunks"].append(final_chunk) |
| _ = Chunk(**final_chunk) |
|
|
| elif settings.docStoreConn.indexExist(search.index_name(tenant_id), dataset_id): |
| sres = settings.retrievaler.search(query, search.index_name(tenant_id), [dataset_id], emb_mdl=None, |
| highlight=True) |
| res["total"] = sres.total |
| for id in sres.ids: |
| d = { |
| "id": id, |
| "content": ( |
| rmSpace(sres.highlight[id]) |
| if question and id in sres.highlight |
| else sres.field[id].get("content_with_weight", "") |
| ), |
| "document_id": sres.field[id]["doc_id"], |
| "docnm_kwd": sres.field[id]["docnm_kwd"], |
| "important_keywords": sres.field[id].get("important_kwd", []), |
| "questions": sres.field[id].get("question_kwd", []), |
| "dataset_id": sres.field[id].get("kb_id", sres.field[id].get("dataset_id")), |
| "image_id": sres.field[id].get("img_id", ""), |
| "available": bool(sres.field[id].get("available_int", 1)), |
| "positions": sres.field[id].get("position_int",[]), |
| } |
| res["chunks"].append(d) |
| _ = Chunk(**d) |
| return get_result(data=res) |
|
|
|
|
| @manager.route( |
| "/datasets/<dataset_id>/documents/<document_id>/chunks", methods=["POST"] |
| ) |
| @token_required |
| def add_chunk(tenant_id, dataset_id, document_id): |
| """ |
| Add a chunk to a document. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document. |
| - in: body |
| name: body |
| description: Chunk data. |
| required: true |
| schema: |
| type: object |
| properties: |
| content: |
| type: string |
| required: true |
| description: Content of the chunk. |
| important_keywords: |
| type: array |
| items: |
| type: string |
| description: Important keywords. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Chunk added successfully. |
| schema: |
| type: object |
| properties: |
| chunk: |
| type: object |
| properties: |
| id: |
| type: string |
| description: Chunk ID. |
| content: |
| type: string |
| description: Chunk content. |
| document_id: |
| type: string |
| description: ID of the document. |
| important_keywords: |
| type: array |
| items: |
| type: string |
| description: Important keywords. |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| doc = DocumentService.query(id=document_id, kb_id=dataset_id) |
| if not doc: |
| return get_error_data_result( |
| message=f"You don't own the document {document_id}." |
| ) |
| doc = doc[0] |
| req = request.json |
| if not req.get("content"): |
| return get_error_data_result(message="`content` is required") |
| if "important_keywords" in req: |
| if not isinstance(req["important_keywords"], list): |
| return get_error_data_result( |
| "`important_keywords` is required to be a list" |
| ) |
| if "questions" in req: |
| if not isinstance(req["questions"], list): |
| return get_error_data_result( |
| "`questions` is required to be a list" |
| ) |
| chunk_id = xxhash.xxh64((req["content"] + document_id).encode("utf-8")).hexdigest() |
| d = { |
| "id": chunk_id, |
| "content_ltks": rag_tokenizer.tokenize(req["content"]), |
| "content_with_weight": req["content"], |
| } |
| d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) |
| d["important_kwd"] = req.get("important_keywords", []) |
| d["important_tks"] = rag_tokenizer.tokenize( |
| " ".join(req.get("important_keywords", [])) |
| ) |
| d["question_kwd"] = req.get("questions", []) |
| d["question_tks"] = rag_tokenizer.tokenize( |
| "\n".join(req.get("questions", [])) |
| ) |
| d["create_time"] = str(datetime.datetime.now()).replace("T", " ")[:19] |
| d["create_timestamp_flt"] = datetime.datetime.now().timestamp() |
| d["kb_id"] = dataset_id |
| d["docnm_kwd"] = doc.name |
| d["doc_id"] = document_id |
| embd_id = DocumentService.get_embd_id(document_id) |
| embd_mdl = TenantLLMService.model_instance( |
| tenant_id, LLMType.EMBEDDING.value, embd_id |
| ) |
| v, c = embd_mdl.encode([doc.name, req["content"] if not d["question_kwd"] else "\n".join(d["question_kwd"])]) |
| v = 0.1 * v[0] + 0.9 * v[1] |
| d["q_%d_vec" % len(v)] = v.tolist() |
| settings.docStoreConn.insert([d], search.index_name(tenant_id), dataset_id) |
|
|
| DocumentService.increment_chunk_num(doc.id, doc.kb_id, c, 1, 0) |
| |
| key_mapping = { |
| "id": "id", |
| "content_with_weight": "content", |
| "doc_id": "document_id", |
| "important_kwd": "important_keywords", |
| "question_kwd": "questions", |
| "kb_id": "dataset_id", |
| "create_timestamp_flt": "create_timestamp", |
| "create_time": "create_time", |
| "document_keyword": "document", |
| } |
| renamed_chunk = {} |
| for key, value in d.items(): |
| if key in key_mapping: |
| new_key = key_mapping.get(key, key) |
| renamed_chunk[new_key] = value |
| _ = Chunk(**renamed_chunk) |
| return get_result(data={"chunk": renamed_chunk}) |
| |
|
|
|
|
| @manager.route( |
| "datasets/<dataset_id>/documents/<document_id>/chunks", methods=["DELETE"] |
| ) |
| @token_required |
| def rm_chunk(tenant_id, dataset_id, document_id): |
| """ |
| Remove chunks from a document. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document. |
| - in: body |
| name: body |
| description: Chunk removal parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| chunk_ids: |
| type: array |
| items: |
| type: string |
| description: List of chunk IDs to remove. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Chunks removed successfully. |
| schema: |
| type: object |
| """ |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| req = request.json |
| condition = {"doc_id": document_id} |
| if "chunk_ids" in req: |
| condition["id"] = req["chunk_ids"] |
| chunk_number = settings.docStoreConn.delete(condition, search.index_name(tenant_id), dataset_id) |
| if chunk_number != 0: |
| DocumentService.decrement_chunk_num(document_id, dataset_id, 1, chunk_number, 0) |
| if "chunk_ids" in req and chunk_number != len(req["chunk_ids"]): |
| return get_error_data_result(message=f"rm_chunk deleted chunks {chunk_number}, expect {len(req['chunk_ids'])}") |
| return get_result(message=f"deleted {chunk_number} chunks") |
|
|
|
|
| @manager.route( |
| "/datasets/<dataset_id>/documents/<document_id>/chunks/<chunk_id>", methods=["PUT"] |
| ) |
| @token_required |
| def update_chunk(tenant_id, dataset_id, document_id, chunk_id): |
| """ |
| Update a chunk within a document. |
| --- |
| tags: |
| - Chunks |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: path |
| name: dataset_id |
| type: string |
| required: true |
| description: ID of the dataset. |
| - in: path |
| name: document_id |
| type: string |
| required: true |
| description: ID of the document. |
| - in: path |
| name: chunk_id |
| type: string |
| required: true |
| description: ID of the chunk to update. |
| - in: body |
| name: body |
| description: Chunk update parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| content: |
| type: string |
| description: Updated content of the chunk. |
| important_keywords: |
| type: array |
| items: |
| type: string |
| description: Updated important keywords. |
| available: |
| type: boolean |
| description: Availability status of the chunk. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Chunk updated successfully. |
| schema: |
| type: object |
| """ |
| chunk = settings.docStoreConn.get(chunk_id, search.index_name(tenant_id), [dataset_id]) |
| if chunk is None: |
| return get_error_data_result(f"Can't find this chunk {chunk_id}") |
| if not KnowledgebaseService.accessible(kb_id=dataset_id, user_id=tenant_id): |
| return get_error_data_result(message=f"You don't own the dataset {dataset_id}.") |
| doc = DocumentService.query(id=document_id, kb_id=dataset_id) |
| if not doc: |
| return get_error_data_result( |
| message=f"You don't own the document {document_id}." |
| ) |
| doc = doc[0] |
| req = request.json |
| if "content" in req: |
| content = req["content"] |
| else: |
| content = chunk.get("content_with_weight", "") |
| d = {"id": chunk_id, "content_with_weight": content} |
| d["content_ltks"] = rag_tokenizer.tokenize(d["content_with_weight"]) |
| d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) |
| if "important_keywords" in req: |
| if not isinstance(req["important_keywords"], list): |
| return get_error_data_result("`important_keywords` should be a list") |
| d["important_kwd"] = req.get("important_keywords", []) |
| d["important_tks"] = rag_tokenizer.tokenize(" ".join(req["important_keywords"])) |
| if "questions" in req: |
| if not isinstance(req["questions"], list): |
| return get_error_data_result("`questions` should be a list") |
| d["question_kwd"] = req.get("questions") |
| d["question_tks"] = rag_tokenizer.tokenize("\n".join(req["questions"])) |
| if "available" in req: |
| d["available_int"] = int(req["available"]) |
| embd_id = DocumentService.get_embd_id(document_id) |
| embd_mdl = TenantLLMService.model_instance( |
| tenant_id, LLMType.EMBEDDING.value, embd_id |
| ) |
| if doc.parser_id == ParserType.QA: |
| arr = [t for t in re.split(r"[\n\t]", d["content_with_weight"]) if len(t) > 1] |
| if len(arr) != 2: |
| return get_error_data_result( |
| message="Q&A must be separated by TAB/ENTER key." |
| ) |
| q, a = rmPrefix(arr[0]), rmPrefix(arr[1]) |
| d = beAdoc( |
| d, arr[0], arr[1], not any([rag_tokenizer.is_chinese(t) for t in q + a]) |
| ) |
|
|
| v, c = embd_mdl.encode([doc.name, d["content_with_weight"] if not d.get("question_kwd") else "\n".join(d["question_kwd"])]) |
| v = 0.1 * v[0] + 0.9 * v[1] if doc.parser_id != ParserType.QA else v[1] |
| d["q_%d_vec" % len(v)] = v.tolist() |
| settings.docStoreConn.update({"id": chunk_id}, d, search.index_name(tenant_id), dataset_id) |
| return get_result() |
|
|
|
|
| @manager.route("/retrieval", methods=["POST"]) |
| @token_required |
| def retrieval_test(tenant_id): |
| """ |
| Retrieve chunks based on a query. |
| --- |
| tags: |
| - Retrieval |
| security: |
| - ApiKeyAuth: [] |
| parameters: |
| - in: body |
| name: body |
| description: Retrieval parameters. |
| required: true |
| schema: |
| type: object |
| properties: |
| dataset_ids: |
| type: array |
| items: |
| type: string |
| required: true |
| description: List of dataset IDs to search in. |
| question: |
| type: string |
| required: true |
| description: Query string. |
| document_ids: |
| type: array |
| items: |
| type: string |
| description: List of document IDs to filter. |
| similarity_threshold: |
| type: number |
| format: float |
| description: Similarity threshold. |
| vector_similarity_weight: |
| type: number |
| format: float |
| description: Vector similarity weight. |
| top_k: |
| type: integer |
| description: Maximum number of chunks to return. |
| highlight: |
| type: boolean |
| description: Whether to highlight matched content. |
| - in: header |
| name: Authorization |
| type: string |
| required: true |
| description: Bearer token for authentication. |
| responses: |
| 200: |
| description: Retrieval results. |
| schema: |
| type: object |
| properties: |
| chunks: |
| type: array |
| items: |
| type: object |
| properties: |
| id: |
| type: string |
| description: Chunk ID. |
| content: |
| type: string |
| description: Chunk content. |
| document_id: |
| type: string |
| description: ID of the document. |
| dataset_id: |
| type: string |
| description: ID of the dataset. |
| similarity: |
| type: number |
| format: float |
| description: Similarity score. |
| """ |
| req = request.json |
| if not req.get("dataset_ids"): |
| return get_error_data_result("`dataset_ids` is required.") |
| kb_ids = req["dataset_ids"] |
| if not isinstance(kb_ids, list): |
| return get_error_data_result("`dataset_ids` should be a list") |
| for id in kb_ids: |
| if not KnowledgebaseService.accessible(kb_id=id, user_id=tenant_id): |
| return get_error_data_result(f"You don't own the dataset {id}.") |
| kbs = KnowledgebaseService.get_by_ids(kb_ids) |
| embd_nms = list(set([kb.embd_id for kb in kbs])) |
| if len(embd_nms) != 1: |
| return get_result( |
| message='Datasets use different embedding models."', |
| code=settings.RetCode.DATA_ERROR, |
| ) |
| if "question" not in req: |
| return get_error_data_result("`question` is required.") |
| page = int(req.get("page", 1)) |
| size = int(req.get("page_size", 30)) |
| question = req["question"] |
| doc_ids = req.get("document_ids", []) |
| use_kg = req.get("use_kg", False) |
| if not isinstance(doc_ids, list): |
| return get_error_data_result("`documents` should be a list") |
| doc_ids_list = KnowledgebaseService.list_documents_by_ids(kb_ids) |
| for doc_id in doc_ids: |
| if doc_id not in doc_ids_list: |
| return get_error_data_result( |
| f"The datasets don't own the document {doc_id}" |
| ) |
| similarity_threshold = float(req.get("similarity_threshold", 0.2)) |
| vector_similarity_weight = float(req.get("vector_similarity_weight", 0.3)) |
| top = int(req.get("top_k", 1024)) |
| if req.get("highlight") == "False" or req.get("highlight") == "false": |
| highlight = False |
| else: |
| highlight = True |
| try: |
| e, kb = KnowledgebaseService.get_by_id(kb_ids[0]) |
| if not e: |
| return get_error_data_result(message="Dataset not found!") |
| embd_mdl = LLMBundle(kb.tenant_id, LLMType.EMBEDDING, llm_name=kb.embd_id) |
|
|
| rerank_mdl = None |
| if req.get("rerank_id"): |
| rerank_mdl = LLMBundle(kb.tenant_id, LLMType.RERANK, llm_name=req["rerank_id"]) |
|
|
| if req.get("keyword", False): |
| chat_mdl = LLMBundle(kb.tenant_id, LLMType.CHAT) |
| question += keyword_extraction(chat_mdl, question) |
|
|
| ranks = settings.retrievaler.retrieval( |
| question, |
| embd_mdl, |
| kb.tenant_id, |
| kb_ids, |
| page, |
| size, |
| similarity_threshold, |
| vector_similarity_weight, |
| top, |
| doc_ids, |
| rerank_mdl=rerank_mdl, |
| highlight=highlight, |
| rank_feature=label_question(question, kbs) |
| ) |
| if use_kg: |
| ck = settings.kg_retrievaler.retrieval(question, |
| [k.tenant_id for k in kbs], |
| kb_ids, |
| embd_mdl, |
| LLMBundle(kb.tenant_id, LLMType.CHAT)) |
| if ck["content_with_weight"]: |
| ranks["chunks"].insert(0, ck) |
|
|
| for c in ranks["chunks"]: |
| c.pop("vector", None) |
|
|
| |
| renamed_chunks = [] |
| for chunk in ranks["chunks"]: |
| key_mapping = { |
| "chunk_id": "id", |
| "content_with_weight": "content", |
| "doc_id": "document_id", |
| "important_kwd": "important_keywords", |
| "question_kwd": "questions", |
| "docnm_kwd": "document_keyword", |
| "kb_id":"dataset_id" |
| } |
| rename_chunk = {} |
| for key, value in chunk.items(): |
| new_key = key_mapping.get(key, key) |
| rename_chunk[new_key] = value |
| renamed_chunks.append(rename_chunk) |
| ranks["chunks"] = renamed_chunks |
| return get_result(data=ranks) |
| except Exception as e: |
| if str(e).find("not_found") > 0: |
| return get_result( |
| message="No chunk found! Check the chunk status please!", |
| code=settings.RetCode.DATA_ERROR, |
| ) |
| return server_error_response(e) |
|
|