eson's picture
update
428b731
raw
history blame
No virus
10 kB
"""
## Dependency
pip install emoji --upgrade
##
https://github.com/CLUEbenchmark/CLUEPretrainedModels/blob/master/bert_dict.py
1. 更新langconv,新增: 余 吒 著 覆
2.
3. 删除30个阿拉伯字母 (阿拉伯语从右向左书写)
4. ok等字母
## TODO:
1. ##~ 这样的词典可以删除,对应要修改tokenizer。
2. 是否要加入空格 [SPACE] 这样的特殊符号。
a) 还原问题: 比如 new balance这样的词汇,会被合并。 会吗? 分词后是 new bal ##ance --> new balance 也能完全还原啊。
b) 语义问题: 同时,在一定意义上也能起到语义隔离的作用,比如 "剑南春 水晶剑 52度 单瓶装高度白酒 750ml 口感浓香型" https://item.jd.com/100006659994.html
[SEP] 也能work
"""
import codecs
import sys
import re
from langconv import *
import emoji
# 1. ℃ 这些符号在clue词典,但是"25℃" 不可分。策略一,加入词典 ##℃,策略二,更换分词器
# oov_clue = ['##°', '##~', '##℃', '##㎡', '##²', '##₂', '##×', '##x', '##+', '余', '覆', '著']
emoji_regex = emoji.get_emoji_regexp()
human_list = ['▲top', '▲topoct', '▲topmay', '▲topapr', '▲topmar', '▲topjun', '▲topdec', '▲topnov', '▲topaug', '▲topjul',
'▲topjan', '▲topsep', '▲topfeb', '¥799', '¥2899', '~~', '~~~', '##~6', '##~10', '~10', '##~5', '~5',
'##~20', '##~8', '##~17', '##~1', '~4', '##~3', '##~7', '~1', 'wedding', '×email', 'cp', '××', 'ok', 'a',
'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v',
'w', 'x', 'y', 'z', '##★', '##☆', '↓↓↓', '##●', '##♪', '▌♥', '##|',
'##d', '##▲', '##o', '★★', '##→', '#a', '⋯⋯', '##▼', '##○', '★★★★★', '##∥', '##◆', '##ω', '★★★', '##c',
'##s', '##e', '##p', '##■', '##↑', '##k', '##и', '◆◆', '##g', '##а', '±0', '##◎', '##─', '##r',
'##>', '##t', '★★★★', '##│', '##n', '##l', '##=', '##y', '☆☆☆', '##i', '##↓', 'ˋ▽ˊ', '##v', '↓↓',
'##f2016', '##q', '∟∣', '##я', '##←', '##◆◆', '##cm~', '##f', '##h', '##j', '##u', '##w',
'##z']
zhuyin_char = ['ㄅ', 'ㄆ', 'ㆠ', 'ㄇ', 'ㄈ', 'ㄪ', 'ㄉ', 'ㄊ', 'ㄋ', 'ㆹ', 'ㄌ', 'ㄍ', 'ㄎ', 'ㆣ', 'ㄫ', 'ㄏ', 'ㆸ', 'ㄐ', 'ㄑ', 'ㆢ', 'ㄬ',
'ㄒ', 'ㆺ', 'ㄓ', 'ㄔ', 'ㄕ', 'ㄖ', 'ㄗ', 'ㄘ', 'ㆡ', 'ㄙ', 'ㆡ', 'ㆪ', 'ㄨ', 'ㆫ', 'ㆨ', 'ㄩ', 'ㄚ', 'ㆩ', 'ㆦ', 'ㆧ', 'ㄛ',
'ㄜ', 'ㄝ', 'ㆤ', 'ㆥ', 'ㄞ', 'ㆮ', 'ㄟ', 'ㄠ', 'ㆯ', 'ㄡ', 'ㆰ', 'ㆱ', 'ㆬ', 'ㄢ', 'ㄣ', 'ㄯ', 'ㄤ', 'ㆲ', 'ㄥ', 'ㆭ', 'ㄦ',
'ㄭ']
special_token = ['[PAD]', '[UNK]', '[CLS]', '[SEP]', '[MASK]', '<S>', '<T>']
japan_chars = ['イ', 'ク', 'シ', 'ス', 'ト', 'ノ', 'フ', 'ラ', 'ル', 'ン']
korean_chars = ['ᄀ', 'ᄁ', 'ᄂ', 'ᄃ', 'ᄅ', 'ᄆ', 'ᄇ', 'ᄈ', 'ᄉ', 'ᄋ', 'ᄌ', 'ᄎ', 'ᄏ', 'ᄐ', 'ᄑ', 'ᄒ', 'ᅡ', 'ᅢ', 'ᅣ', 'ᅥ', 'ᅦ',
'ᅧ', 'ᅨ', 'ᅩ', 'ᅪ', 'ᅬ', 'ᅭ', 'ᅮ', 'ᅯ', 'ᅲ', 'ᅳ', 'ᅴ', 'ᅵ', 'ᆨ', 'ᆫ', 'ᆯ', 'ᆷ', 'ᆸ', 'ᆺ', 'ᆻ', 'ᆼ', 'ᗜ']
add_puns = ['”', '“', '—', '–', '…', '’', '‘']
# 单个“乾”被转化成了 “干”
add_cn_chars = [char for char in '呡乾绗楦硌袢钕蕞癀皲貉唛笕椴―胗旯鳙鲇鳐鳜鲅鳊鲳鲽鲣枞炝醅馊捯抻绉馐饧莜嘬腘肫鳟镊犽洌蝰铱' \
'髌锃镲锗甑戗裥弎粝霂猄轱苎偲兿铷栢帏黢洇沄誊忸怩蚬籺氚犇锒鸩噘偾髫']
#阿拉伯文 بسن
add_nums = ['10', '11', '12', '13', '14', '15', '16', '17', '18', '19', '20', '21', '22', '23', '24', '25', '26', '27',
'28', '29', '30', '31', '32', '33', '34', '35', '36', '37', '38', '39', '40', '41', '42', '43', '44', '45',
'46', '47', '48', '49', '50', '51', '52', '53', '54', '55', '56', '57', '58', '59', '60', '61', '62', '63',
'64', '65', '66', '67', '68', '69', '70', '71', '72', '73', '74', '75', '76', '77', '78', '79', '80', '81',
'82', '83', '84', '85', '86', '87', '88', '89', '90', '91', '92', '93', '94', '95', '96', '97', '98', '99',
'100', '120', '128', '180', '200', '256', '304', '360', '500', '512', '1000', '1080', '2000', '2014',
'2015', '2016', '2017', '2018', '2019', '2020', '2021', '2022']
cn_punc = ',。;:?!()~|' #
def q2b(uchar, skip_cn_punc=False):
# 有时,希望保留全角中文标点,例如cn_punc。
if skip_cn_punc and uchar in cn_punc:
return uchar
inside_code = ord(uchar)
if inside_code == 12288: # 全角空格直接转换
inside_code = 32
elif 65281 <= inside_code <= 65374: # 全角字符(除空格)根据关系转化
inside_code -= 65248
return chr(inside_code)
def str_q2b(ustring, skip_cn_punc=False):
""" 全角转半角 """
return ''.join([q2b(uchar, skip_cn_punc) for uchar in ustring])
with open('vocab.google.txt', 'r', encoding='utf-8') as fin, \
open('vocab.jd.txt.v2', 'w', encoding='utf-8') as fout:
cout_zh = 0
cout_en = 0
cout_jp = 0
cout_em = 0
cout_zh_res = 0
cout_zh_tra = 0
cout_zh_wp = 0
cout_en_del = 0
cout_en_res = 0
cout_num = 0
cout_num_del = 0
cout_num_res = 0
cout_hand_del = 0
cout_total = 0
cout_zhuyin = 0
cout_unused = 0
cout_special = 0
cout_jp = 0
cout_ko = 0
cout_ar = 0
for line in fin:
cout_total += 1
token = line.strip()
if not token:
continue
if token in ['|']:
print(token)
if token in human_list:
cout_hand_del += 1 # 13
continue
# chinese character
elif re.match(u'[\u4e00-\u9fa5]+', token.replace('##', '')):
cout_zh += 1 # 14642
token_simp = Converter('zh-hans').convert(token)
if token_simp != token:
cout_zh_tra += 1
continue
else:
if re.match(u'##', token):
# print(token)
cout_zh_wp += 1
continue
else:
cout_zh_res += 1
print(token, file=fout)
# korean character
elif re.match(u'[\uac00-\ud7ff]+', token.replace('##', '')):
# print(token)
cout_ko += 1
continue
# japanese character
elif re.match(u'[\u30a0-\u30ff\u3040-\u309f]+', token.replace('##', '')):
# print(token)
cout_jp += 1
continue
# arabic character
elif re.match(u'[\u0600-\u06ff\u0750-\u077f\ufb50-\ufbc1\ufbd3-\ufd3f'
u'\ufd50-\ufd8f\ufd50-\ufd8f\ufe70-\ufefc\uFDF0-\uFDFD]+', token.replace('##', '')):
cout_ar += 1
continue
# english character
elif re.match(u'[a-z]+', token.replace('##', '')):
# print(token)
cout_en += 1
print(token, file=fout)
continue
elif str_q2b(token, skip_cn_punc=True) != token:
print(token, '--', str_q2b(token, skip_cn_punc=True))
continue
# emoji character
elif re.match(emoji_regex, token.replace('##', '')):
# print(token)
cout_em += 1
continue
# multi-number characters
elif re.match(u'(##)?\d', token):
cout_num += 1
if len(token.replace('##', '')) == 1:
# print(token)
cout_num_res += 1
print(token, file=fout)
else:
cout_num_del += 1 # 这个操作应该还好
# print(token)
continue
elif token.replace('##', '') in zhuyin_char:
# print(token, file=fout)
cout_zhuyin += 1
continue
elif token.startswith('[unused'):
print(token, file=fout)
cout_unused += 1
elif token in special_token:
print(token, file=fout)
cout_special += 1
elif token.replace('##', '') in japan_chars:
cout_jp += 1
continue
elif token.replace('##', '') in korean_chars:
cout_ko += 1
continue
else:
# print(token)
print(token, file=fout)
# add tokens
if token == '"':
for token in add_puns:
print(token, file=fout)
if token == '9':
for token in add_nums:
cout_num_res += 1
print(token, file=fout)
if token == '龟':
for token in add_cn_chars:
print(token, file=fout)
print("cout_zh:{}".format(cout_zh)) # 14642
print("cout_zh_tra:{}".format(cout_zh_tra)) # 3264
print("cout_zh_wp:{}".format(cout_zh_wp)) # 5689
print("cout_zh_res:{}".format(cout_zh_res)) # 5689
print("cout_en:{}".format(cout_en)) # 3555
print("cout_en_del:{}".format(cout_en_del)) # 2235
print("cout_en_res:{}".format(cout_en_res)) # 1320
print("cout_num:{}".format(cout_num)) # 1179
print("cout_num_del:{}".format(cout_num_del)) # 1137
print("cout_num_res:{}".format(cout_num_res)) # 140
print("cout_hand_del:{}".format(cout_hand_del)) # 132
print("cout_zhuyin:{}".format(cout_zhuyin)) # 36
print("cout_unused:{}".format(cout_unused)) # 99
print("cout_special:{}".format(cout_special)) # 7
print("cout_jp:{}".format(cout_jp)) # 573
print("cout_ko:{}".format(cout_ko)) # 84
print("cout_ar:{}".format(cout_ar)) #
print("cout_em:{}".format(cout_em)) # 56