| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import copy |
| import re |
| from io import BytesIO |
| from xpinyin import Pinyin |
| import numpy as np |
| import pandas as pd |
| from openpyxl import load_workbook |
| from dateutil.parser import parse as datetime_parse |
|
|
| from api.db.services.knowledgebase_service import KnowledgebaseService |
| from deepdoc.parser.utils import get_text |
| from rag.nlp import rag_tokenizer, tokenize |
| from deepdoc.parser import ExcelParser |
|
|
|
|
| class Excel(ExcelParser): |
| def __call__(self, fnm, binary=None, from_page=0, |
| to_page=10000000000, callback=None): |
| if not binary: |
| wb = load_workbook(fnm) |
| else: |
| wb = load_workbook(BytesIO(binary)) |
| total = 0 |
| for sheetname in wb.sheetnames: |
| total += len(list(wb[sheetname].rows)) |
|
|
| res, fails, done = [], [], 0 |
| rn = 0 |
| for sheetname in wb.sheetnames: |
| ws = wb[sheetname] |
| rows = list(ws.rows) |
| if not rows: |
| continue |
| headers = [cell.value for cell in rows[0]] |
| missed = set([i for i, h in enumerate(headers) if h is None]) |
| headers = [ |
| cell.value for i, |
| cell in enumerate( |
| rows[0]) if i not in missed] |
| if not headers: |
| continue |
| data = [] |
| for i, r in enumerate(rows[1:]): |
| rn += 1 |
| if rn - 1 < from_page: |
| continue |
| if rn - 1 >= to_page: |
| break |
| row = [ |
| cell.value for ii, |
| cell in enumerate(r) if ii not in missed] |
| if len(row) != len(headers): |
| fails.append(str(i)) |
| continue |
| data.append(row) |
| done += 1 |
| if np.array(data).size == 0: |
| continue |
| res.append(pd.DataFrame(np.array(data), columns=headers)) |
|
|
| callback(0.3, ("Extract records: {}~{}".format(from_page + 1, min(to_page, from_page + rn)) + ( |
| f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) |
| return res |
|
|
|
|
| def trans_datatime(s): |
| try: |
| return datetime_parse(s.strip()).strftime("%Y-%m-%d %H:%M:%S") |
| except Exception: |
| pass |
|
|
|
|
| def trans_bool(s): |
| if re.match(r"(true|yes|是|\*|✓|✔|☑|✅|√)$", |
| str(s).strip(), flags=re.IGNORECASE): |
| return "yes" |
| if re.match(r"(false|no|否|⍻|×)$", str(s).strip(), flags=re.IGNORECASE): |
| return "no" |
|
|
|
|
| def column_data_type(arr): |
| arr = list(arr) |
| counts = {"int": 0, "float": 0, "text": 0, "datetime": 0, "bool": 0} |
| trans = {t: f for f, t in |
| [(int, "int"), (float, "float"), (trans_datatime, "datetime"), (trans_bool, "bool"), (str, "text")]} |
| for a in arr: |
| if a is None: |
| continue |
| if re.match(r"[+-]?[0-9]{,19}(\.0+)?$", str(a).replace("%%", "")): |
| counts["int"] += 1 |
| elif re.match(r"[+-]?[0-9.]{,19}$", str(a).replace("%%", "")): |
| counts["float"] += 1 |
| elif re.match(r"(true|yes|是|\*|✓|✔|☑|✅|√|false|no|否|⍻|×)$", str(a), flags=re.IGNORECASE): |
| counts["bool"] += 1 |
| elif trans_datatime(str(a)): |
| counts["datetime"] += 1 |
| else: |
| counts["text"] += 1 |
| counts = sorted(counts.items(), key=lambda x: x[1] * -1) |
| ty = counts[0][0] |
| for i in range(len(arr)): |
| if arr[i] is None: |
| continue |
| try: |
| arr[i] = trans[ty](str(arr[i])) |
| except Exception: |
| arr[i] = None |
| |
| |
| |
| return arr, ty |
|
|
|
|
| def chunk(filename, binary=None, from_page=0, to_page=10000000000, |
| lang="Chinese", callback=None, **kwargs): |
| """ |
| Excel and csv(txt) format files are supported. |
| For csv or txt file, the delimiter between columns is TAB. |
| The first line must be column headers. |
| Column headers must be meaningful terms inorder to make our NLP model understanding. |
| It's good to enumerate some synonyms using slash '/' to separate, and even better to |
| enumerate values using brackets like 'gender/sex(male, female)'. |
| Here are some examples for headers: |
| 1. supplier/vendor\tcolor(yellow, red, brown)\tgender/sex(male, female)\tsize(M,L,XL,XXL) |
| 2. 姓名/名字\t电话/手机/微信\t最高学历(高中,职高,硕士,本科,博士,初中,中技,中专,专科,专升本,MPA,MBA,EMBA) |
| |
| Every row in table will be treated as a chunk. |
| """ |
|
|
| if re.search(r"\.xlsx?$", filename, re.IGNORECASE): |
| callback(0.1, "Start to parse.") |
| excel_parser = Excel() |
| dfs = excel_parser( |
| filename, |
| binary, |
| from_page=from_page, |
| to_page=to_page, |
| callback=callback) |
| elif re.search(r"\.(txt|csv)$", filename, re.IGNORECASE): |
| callback(0.1, "Start to parse.") |
| txt = get_text(filename, binary) |
| lines = txt.split("\n") |
| fails = [] |
| headers = lines[0].split(kwargs.get("delimiter", "\t")) |
| rows = [] |
| for i, line in enumerate(lines[1:]): |
| if i < from_page: |
| continue |
| if i >= to_page: |
| break |
| row = [field for field in line.split(kwargs.get("delimiter", "\t"))] |
| if len(row) != len(headers): |
| fails.append(str(i)) |
| continue |
| rows.append(row) |
|
|
| callback(0.3, ("Extract records: {}~{}".format(from_page, min(len(lines), to_page)) + ( |
| f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) |
|
|
| dfs = [pd.DataFrame(np.array(rows), columns=headers)] |
|
|
| else: |
| raise NotImplementedError( |
| "file type not supported yet(excel, text, csv supported)") |
|
|
| res = [] |
| PY = Pinyin() |
| fieds_map = { |
| "text": "_tks", |
| "int": "_long", |
| "keyword": "_kwd", |
| "float": "_flt", |
| "datetime": "_dt", |
| "bool": "_kwd"} |
| for df in dfs: |
| for n in ["id", "_id", "index", "idx"]: |
| if n in df.columns: |
| del df[n] |
| clmns = df.columns.values |
| txts = list(copy.deepcopy(clmns)) |
| py_clmns = [ |
| PY.get_pinyins( |
| re.sub( |
| r"(/.*|([^()]+?)|\([^()]+?\))", |
| "", |
| str(n)), |
| '_')[0] for n in clmns] |
| clmn_tys = [] |
| for j in range(len(clmns)): |
| cln, ty = column_data_type(df[clmns[j]]) |
| clmn_tys.append(ty) |
| df[clmns[j]] = cln |
| if ty == "text": |
| txts.extend([str(c) for c in cln if c]) |
| clmns_map = [(py_clmns[i].lower() + fieds_map[clmn_tys[i]], str(clmns[i]).replace("_", " ")) |
| for i in range(len(clmns))] |
|
|
| eng = lang.lower() == "english" |
| for ii, row in df.iterrows(): |
| d = { |
| "docnm_kwd": filename, |
| "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) |
| } |
| row_txt = [] |
| for j in range(len(clmns)): |
| if row[clmns[j]] is None: |
| continue |
| if not str(row[clmns[j]]): |
| continue |
| if pd.isna(row[clmns[j]]): |
| continue |
| fld = clmns_map[j][0] |
| d[fld] = row[clmns[j]] if clmn_tys[j] != "text" else rag_tokenizer.tokenize( |
| row[clmns[j]]) |
| row_txt.append("{}:{}".format(clmns[j], row[clmns[j]])) |
| if not row_txt: |
| continue |
| tokenize(d, "; ".join(row_txt), eng) |
| res.append(d) |
|
|
| KnowledgebaseService.update_parser_config( |
| kwargs["kb_id"], {"field_map": {k: v for k, v in clmns_map}}) |
| callback(0.35, "") |
|
|
| return res |
|
|
|
|
| if __name__ == "__main__": |
| import sys |
|
|
| def dummy(prog=None, msg=""): |
| pass |
|
|
| chunk(sys.argv[1], callback=dummy) |
|
|