|
""" from https://github.com/keithito/tacotron """ |
|
|
|
''' |
|
Cleaners are transformations that run over the input text at both training and eval time. |
|
|
|
Cleaners can be selected by passing a comma-delimited list of cleaner names as the "cleaners" |
|
hyperparameter. Some cleaners are English-specific. You'll typically want to use: |
|
1. "english_cleaners" for English text |
|
2. "transliteration_cleaners" for non-English text that can be transliterated to ASCII using |
|
the Unidecode library (https://pypi.python.org/pypi/Unidecode) |
|
3. "basic_cleaners" if you do not want to transliterate (in this case, you should also update |
|
the symbols in symbols.py to match your data). |
|
''' |
|
|
|
import re |
|
from unidecode import unidecode |
|
import pyopenjtalk |
|
from jamo import h2j, j2hcj |
|
from pypinyin import lazy_pinyin, BOPOMOFO |
|
import jieba, cn2an |
|
|
|
|
|
|
|
_korean_classifiers = 'κ΅°λ° κΆ κ° κ·Έλ£¨ λ’ λ λ λ§λ¦¬ λͺ¨ λͺ¨κΈ λ λ° λ°μ§ λ°© λ² λ² λ³΄λ£¨ μ΄ μ μ μ μ μνΌ μ μ§ μ± μ² μ²© μΆ μΌ€λ ν¨ ν΅' |
|
|
|
|
|
_whitespace_re = re.compile(r'\s+') |
|
|
|
|
|
_japanese_characters = re.compile(r'[A-Za-z\d\u3005\u3040-\u30ff\u4e00-\u9fff\uff11-\uff19\uff21-\uff3a\uff41-\uff5a\uff66-\uff9d]') |
|
|
|
|
|
_japanese_marks = re.compile(r'[^A-Za-z\d\u3005\u3040-\u30ff\u4e00-\u9fff\uff11-\uff19\uff21-\uff3a\uff41-\uff5a\uff66-\uff9d]') |
|
|
|
|
|
_abbreviations = [(re.compile('\\b%s\\.' % x[0], re.IGNORECASE), x[1]) for x in [ |
|
('mrs', 'misess'), |
|
('mr', 'mister'), |
|
('dr', 'doctor'), |
|
('st', 'saint'), |
|
('co', 'company'), |
|
('jr', 'junior'), |
|
('maj', 'major'), |
|
('gen', 'general'), |
|
('drs', 'doctors'), |
|
('rev', 'reverend'), |
|
('lt', 'lieutenant'), |
|
('hon', 'honorable'), |
|
('sgt', 'sergeant'), |
|
('capt', 'captain'), |
|
('esq', 'esquire'), |
|
('ltd', 'limited'), |
|
('col', 'colonel'), |
|
('ft', 'fort'), |
|
]] |
|
|
|
|
|
_hangul_divided = [(re.compile('%s' % x[0]), x[1]) for x in [ |
|
('γ³', 'γ±γ
'), |
|
('γ΅', 'γ΄γ
'), |
|
('γΆ', 'γ΄γ
'), |
|
('γΊ', 'γΉγ±'), |
|
('γ»', 'γΉγ
'), |
|
('γΌ', 'γΉγ
'), |
|
('γ½', 'γΉγ
'), |
|
('γΎ', 'γΉγ
'), |
|
('γΏ', 'γΉγ
'), |
|
('γ
', 'γΉγ
'), |
|
('γ
', 'γ
γ
'), |
|
('γ
', 'γ
γ
'), |
|
('γ
', 'γ
γ
'), |
|
('γ
', 'γ
γ
£'), |
|
('γ
', 'γ
γ
'), |
|
('γ
', 'γ
γ
'), |
|
('γ
', 'γ
γ
£'), |
|
('γ
’', 'γ
‘γ
£'), |
|
('γ
', 'γ
£γ
'), |
|
('γ
', 'γ
£γ
'), |
|
('γ
', 'γ
£γ
'), |
|
('γ
', 'γ
£γ
'), |
|
('γ
', 'γ
£γ
'), |
|
('γ
', 'γ
£γ
') |
|
]] |
|
|
|
|
|
_latin_to_hangul = [(re.compile('%s' % x[0], re.IGNORECASE), x[1]) for x in [ |
|
('a', 'μμ΄'), |
|
('b', 'λΉ'), |
|
('c', 'μ'), |
|
('d', 'λ'), |
|
('e', 'μ΄'), |
|
('f', 'μν'), |
|
('g', 'μ§'), |
|
('h', 'μμ΄μΉ'), |
|
('i', 'μμ΄'), |
|
('j', 'μ μ΄'), |
|
('k', 'μΌμ΄'), |
|
('l', 'μ'), |
|
('m', 'μ '), |
|
('n', 'μ'), |
|
('o', 'μ€'), |
|
('p', 'νΌ'), |
|
('q', 'ν'), |
|
('r', 'μλ₯΄'), |
|
('s', 'μμ€'), |
|
('t', 'ν°'), |
|
('u', 'μ '), |
|
('v', 'λΈμ΄'), |
|
('w', 'λλΈμ '), |
|
('x', 'μμ€'), |
|
('y', 'μμ΄'), |
|
('z', 'μ νΈ') |
|
]] |
|
|
|
|
|
_latin_to_bopomofo = [(re.compile('%s' % x[0], re.IGNORECASE), x[1]) for x in [ |
|
('a', 'γΛ'), |
|
('b', 'γ
γ§Λ'), |
|
('c', 'γγ§Λ'), |
|
('d', 'γγ§Λ'), |
|
('e', 'γ§Λ'), |
|
('f', 'γΛγγ¨Λ'), |
|
('g', 'γγ§Λ'), |
|
('h', 'γΛγγ©Λ'), |
|
('i', 'γΛ'), |
|
('j', 'γγΛ'), |
|
('k', 'γγΛ'), |
|
('l', 'γΛγΛ'), |
|
('m', 'γΛγγ¨Λ'), |
|
('n', 'γ£Λ'), |
|
('o', 'γ‘Λ'), |
|
('p', 'γγ§Λ'), |
|
('q', 'γγ§γ‘Λ'), |
|
('r', 'γΛ'), |
|
('s', 'γΛγΛ'), |
|
('t', 'γγ§Λ'), |
|
('u', 'γ§γ‘Λ'), |
|
('v', 'γ¨γ§Λ'), |
|
('w', 'γγΛγ
γ¨Λγγ§γ‘Λ'), |
|
('x', 'γΛγγ¨ΛγΛ'), |
|
('y', 'γ¨γΛ'), |
|
('z', 'γγΛ') |
|
]] |
|
|
|
|
|
|
|
_bopomofo_to_romaji = [(re.compile('%s' % x[0], re.IGNORECASE), x[1]) for x in [ |
|
('γ
γ', 'pβΌwo'), |
|
('γγ', 'pΚ°wo'), |
|
('γγ', 'mwo'), |
|
('γγ', 'fwo'), |
|
('γ
', 'pβΌ'), |
|
('γ', 'pΚ°'), |
|
('γ', 'm'), |
|
('γ', 'f'), |
|
('γ', 'tβΌ'), |
|
('γ', 'tΚ°'), |
|
('γ', 'n'), |
|
('γ', 'l'), |
|
('γ', 'kβΌ'), |
|
('γ', 'kΚ°'), |
|
('γ', 'h'), |
|
('γ', 'Κ§βΌ'), |
|
('γ', 'Κ§Κ°'), |
|
('γ', 'Κ'), |
|
('γ', 'Κ¦`βΌ'), |
|
('γ', 'Κ¦`Κ°'), |
|
('γ', 's`'), |
|
('γ', 'ΙΉ`'), |
|
('γ', 'Κ¦βΌ'), |
|
('γ', 'Κ¦Κ°'), |
|
('γ', 's'), |
|
('γ', 'a'), |
|
('γ', 'o'), |
|
('γ', 'Ι'), |
|
('γ', 'e'), |
|
('γ', 'ai'), |
|
('γ', 'ei'), |
|
('γ ', 'au'), |
|
('γ‘', 'ou'), |
|
('γ§γ’', 'yeNN'), |
|
('γ’', 'aNN'), |
|
('γ§γ£', 'iNN'), |
|
('γ£', 'ΙNN'), |
|
('γ€', 'aNg'), |
|
('γ§γ₯', 'iNg'), |
|
('γ¨γ₯', 'uNg'), |
|
('γ©γ₯', 'yuNg'), |
|
('γ₯', 'ΙNg'), |
|
('γ¦', 'ΙΙ»'), |
|
('γ§', 'i'), |
|
('γ¨', 'u'), |
|
('γ©', 'Ι₯'), |
|
('Λ', 'β'), |
|
('Λ', 'β'), |
|
('Λ', 'ββ'), |
|
('Λ', 'β'), |
|
('Λ', ''), |
|
('οΌ', ','), |
|
('γ', '.'), |
|
('οΌ', '!'), |
|
('οΌ', '?'), |
|
('β', '-') |
|
]] |
|
|
|
|
|
def expand_abbreviations(text): |
|
for regex, replacement in _abbreviations: |
|
text = re.sub(regex, replacement, text) |
|
return text |
|
|
|
|
|
def lowercase(text): |
|
return text.lower() |
|
|
|
|
|
def collapse_whitespace(text): |
|
return re.sub(_whitespace_re, ' ', text) |
|
|
|
|
|
def convert_to_ascii(text): |
|
return unidecode(text) |
|
|
|
|
|
def japanese_to_romaji_with_accent(text): |
|
'''Reference https://r9y9.github.io/ttslearn/latest/notebooks/ch10_Recipe-Tacotron.html''' |
|
sentences = re.split(_japanese_marks, text) |
|
marks = re.findall(_japanese_marks, text) |
|
text = '' |
|
for i, sentence in enumerate(sentences): |
|
if re.match(_japanese_characters, sentence): |
|
if text!='': |
|
text+=' ' |
|
labels = pyopenjtalk.extract_fullcontext(sentence) |
|
for n, label in enumerate(labels): |
|
phoneme = re.search(r'\-([^\+]*)\+', label).group(1) |
|
if phoneme not in ['sil','pau']: |
|
text += phoneme.replace('ch','Κ§').replace('sh','Κ').replace('cl','Q') |
|
else: |
|
continue |
|
n_moras = int(re.search(r'/F:(\d+)_', label).group(1)) |
|
a1 = int(re.search(r"/A:(\-?[0-9]+)\+", label).group(1)) |
|
a2 = int(re.search(r"\+(\d+)\+", label).group(1)) |
|
a3 = int(re.search(r"\+(\d+)/", label).group(1)) |
|
if re.search(r'\-([^\+]*)\+', labels[n + 1]).group(1) in ['sil','pau']: |
|
a2_next=-1 |
|
else: |
|
a2_next = int(re.search(r"\+(\d+)\+", labels[n + 1]).group(1)) |
|
|
|
if a3 == 1 and a2_next == 1: |
|
text += ' ' |
|
|
|
elif a1 == 0 and a2_next == a2 + 1 and a2 != n_moras: |
|
text += 'β' |
|
|
|
elif a2 == 1 and a2_next == 2: |
|
text += 'β' |
|
if i<len(marks): |
|
text += unidecode(marks[i]).replace(' ','') |
|
return text |
|
|
|
|
|
def latin_to_hangul(text): |
|
for regex, replacement in _latin_to_hangul: |
|
text = re.sub(regex, replacement, text) |
|
return text |
|
|
|
|
|
def divide_hangul(text): |
|
for regex, replacement in _hangul_divided: |
|
text = re.sub(regex, replacement, text) |
|
return text |
|
|
|
|
|
def hangul_number(num, sino=True): |
|
'''Reference https://github.com/Kyubyong/g2pK''' |
|
num = re.sub(',', '', num) |
|
|
|
if num == '0': |
|
return 'μ' |
|
if not sino and num == '20': |
|
return 'μ€λ¬΄' |
|
|
|
digits = '123456789' |
|
names = 'μΌμ΄μΌμ¬μ€μ‘μΉ νꡬ' |
|
digit2name = {d: n for d, n in zip(digits, names)} |
|
|
|
modifiers = 'ν λ μΈ λ€ λ€μ― μ¬μ― μΌκ³± μ¬λ μν' |
|
decimals = 'μ΄ μ€λ¬Ό μλ₯Έ λ§ν μ° μμ μΌν μ¬λ μν' |
|
digit2mod = {d: mod for d, mod in zip(digits, modifiers.split())} |
|
digit2dec = {d: dec for d, dec in zip(digits, decimals.split())} |
|
|
|
spelledout = [] |
|
for i, digit in enumerate(num): |
|
i = len(num) - i - 1 |
|
if sino: |
|
if i == 0: |
|
name = digit2name.get(digit, '') |
|
elif i == 1: |
|
name = digit2name.get(digit, '') + 'μ' |
|
name = name.replace('μΌμ', 'μ') |
|
else: |
|
if i == 0: |
|
name = digit2mod.get(digit, '') |
|
elif i == 1: |
|
name = digit2dec.get(digit, '') |
|
if digit == '0': |
|
if i % 4 == 0: |
|
last_three = spelledout[-min(3, len(spelledout)):] |
|
if ''.join(last_three) == '': |
|
spelledout.append('') |
|
continue |
|
else: |
|
spelledout.append('') |
|
continue |
|
if i == 2: |
|
name = digit2name.get(digit, '') + 'λ°±' |
|
name = name.replace('μΌλ°±', 'λ°±') |
|
elif i == 3: |
|
name = digit2name.get(digit, '') + 'μ²' |
|
name = name.replace('μΌμ²', 'μ²') |
|
elif i == 4: |
|
name = digit2name.get(digit, '') + 'λ§' |
|
name = name.replace('μΌλ§', 'λ§') |
|
elif i == 5: |
|
name = digit2name.get(digit, '') + 'μ' |
|
name = name.replace('μΌμ', 'μ') |
|
elif i == 6: |
|
name = digit2name.get(digit, '') + 'λ°±' |
|
name = name.replace('μΌλ°±', 'λ°±') |
|
elif i == 7: |
|
name = digit2name.get(digit, '') + 'μ²' |
|
name = name.replace('μΌμ²', 'μ²') |
|
elif i == 8: |
|
name = digit2name.get(digit, '') + 'μ΅' |
|
elif i == 9: |
|
name = digit2name.get(digit, '') + 'μ' |
|
elif i == 10: |
|
name = digit2name.get(digit, '') + 'λ°±' |
|
elif i == 11: |
|
name = digit2name.get(digit, '') + 'μ²' |
|
elif i == 12: |
|
name = digit2name.get(digit, '') + 'μ‘°' |
|
elif i == 13: |
|
name = digit2name.get(digit, '') + 'μ' |
|
elif i == 14: |
|
name = digit2name.get(digit, '') + 'λ°±' |
|
elif i == 15: |
|
name = digit2name.get(digit, '') + 'μ²' |
|
spelledout.append(name) |
|
return ''.join(elem for elem in spelledout) |
|
|
|
|
|
def number_to_hangul(text): |
|
'''Reference https://github.com/Kyubyong/g2pK''' |
|
tokens = set(re.findall(r'(\d[\d,]*)([\uac00-\ud71f]+)', text)) |
|
for token in tokens: |
|
num, classifier = token |
|
if classifier[:2] in _korean_classifiers or classifier[0] in _korean_classifiers: |
|
spelledout = hangul_number(num, sino=False) |
|
else: |
|
spelledout = hangul_number(num, sino=True) |
|
text = text.replace(f'{num}{classifier}', f'{spelledout}{classifier}') |
|
|
|
digits = '0123456789' |
|
names = 'μμΌμ΄μΌμ¬μ€μ‘μΉ νꡬ' |
|
for d, n in zip(digits, names): |
|
text = text.replace(d, n) |
|
return text |
|
|
|
|
|
def number_to_chinese(text): |
|
numbers = re.findall(r'\d+(?:\.?\d+)?', text) |
|
for number in numbers: |
|
text = text.replace(number, cn2an.an2cn(number),1) |
|
return text |
|
|
|
|
|
def chinese_to_bopomofo(text): |
|
text=text.replace('γ','οΌ').replace('οΌ','οΌ').replace('οΌ','οΌ') |
|
words=jieba.lcut(text,cut_all=False) |
|
text='' |
|
for word in words: |
|
bopomofos=lazy_pinyin(word,BOPOMOFO) |
|
if not re.search('[\u4e00-\u9fff]',word): |
|
text+=word |
|
continue |
|
for i in range(len(bopomofos)): |
|
if re.match('[\u3105-\u3129]',bopomofos[i][-1]): |
|
bopomofos[i]+='Λ' |
|
if text!='': |
|
text+=' ' |
|
text+=''.join(bopomofos) |
|
return text |
|
|
|
|
|
def latin_to_bopomofo(text): |
|
for regex, replacement in _latin_to_bopomofo: |
|
text = re.sub(regex, replacement, text) |
|
return text |
|
|
|
|
|
def bopomofo_to_romaji(text): |
|
for regex, replacement in _bopomofo_to_romaji: |
|
text = re.sub(regex, replacement, text) |
|
return text |
|
|
|
|
|
def basic_cleaners(text): |
|
'''Basic pipeline that lowercases and collapses whitespace without transliteration.''' |
|
text = lowercase(text) |
|
text = collapse_whitespace(text) |
|
return text |
|
|
|
|
|
def transliteration_cleaners(text): |
|
'''Pipeline for non-English text that transliterates to ASCII.''' |
|
text = convert_to_ascii(text) |
|
text = lowercase(text) |
|
text = collapse_whitespace(text) |
|
return text |
|
|
|
|
|
def japanese_cleaners(text): |
|
text=japanese_to_romaji_with_accent(text) |
|
if re.match('[A-Za-z]',text[-1]): |
|
text += '.' |
|
return text |
|
|
|
|
|
def japanese_cleaners2(text): |
|
return japanese_cleaners(text).replace('ts','Κ¦').replace('...','β¦') |
|
|
|
|
|
def korean_cleaners(text): |
|
'''Pipeline for Korean text''' |
|
text = latin_to_hangul(text) |
|
text = number_to_hangul(text) |
|
text = j2hcj(h2j(text)) |
|
text = divide_hangul(text) |
|
if re.match('[\u3131-\u3163]',text[-1]): |
|
text += '.' |
|
return text |
|
|
|
|
|
def chinese_cleaners(text): |
|
'''Pipeline for Chinese text''' |
|
text=number_to_chinese(text) |
|
text=chinese_to_bopomofo(text) |
|
text=latin_to_bopomofo(text) |
|
if re.match('[ΛΛΛΛΛ]',text[-1]): |
|
text += 'γ' |
|
return text |
|
|
|
|
|
def zh_ja_mixture_cleaners(text): |
|
chinese_texts=re.findall(r'\[ZH\].*?\[ZH\]',text) |
|
japanese_texts=re.findall(r'\[JA\].*?\[JA\]',text) |
|
for chinese_text in chinese_texts: |
|
cleaned_text=number_to_chinese(chinese_text[4:-4]) |
|
cleaned_text=chinese_to_bopomofo(cleaned_text) |
|
cleaned_text=latin_to_bopomofo(cleaned_text) |
|
cleaned_text=bopomofo_to_romaji(cleaned_text) |
|
cleaned_text=re.sub('i[aoe]',lambda x:'y'+x.group(0)[1:],cleaned_text) |
|
cleaned_text=re.sub('u[aoΙe]',lambda x:'w'+x.group(0)[1:],cleaned_text) |
|
cleaned_text=re.sub('([Κ¦sΙΉ]`[βΌΚ°]?)([βββ]+)',lambda x:x.group(1)+'ΙΉ`'+x.group(2),cleaned_text).replace('Ι»','ΙΉ`') |
|
cleaned_text=re.sub('([Κ¦s][βΌΚ°]?)([βββ]+)',lambda x:x.group(1)+'ΙΉ'+x.group(2),cleaned_text) |
|
text = text.replace(chinese_text,cleaned_text+' ',1) |
|
for japanese_text in japanese_texts: |
|
cleaned_text=japanese_to_romaji_with_accent(japanese_text[4:-4]).replace('ts','Κ¦').replace('u','Ι―').replace('...','β¦') |
|
text = text.replace(japanese_text,cleaned_text+' ',1) |
|
text=text[:-1] |
|
if len(text) and re.match('[A-Za-zΙ―ΙΉΙΙ₯βββ]',text[-1]): |
|
text += '.' |
|
return text |