| import base64 |
| import os |
| from functools import lru_cache |
| from typing import Optional |
| import torch |
| from transformers import AutoTokenizer |
| import tiktoken |
|
|
| LANGUAGES = { |
| "en": "english", "zh": "chinese", "de": "german", "es": "spanish", "ru": "russian", |
| "ko": "korean", "fr": "french", "ja": "japanese", "pt": "portuguese", "tr": "turkish", |
| "pl": "polish", "ca": "catalan", "nl": "dutch", "ar": "arabic", "sv": "swedish", "it": "italian", |
| "id": "indonesian", "hi": "hindi", "fi": "finnish", "vi": "vietnamese", "he": "hebrew", |
| "uk": "ukrainian", "el": "greek", "ms": "malay", "cs": "czech", "ro": "romanian", "da": "danish", |
| "hu": "hungarian", "ta": "tamil", "no": "norwegian", "th": "thai", "ur": "urdu", "hr": "croatian", |
| "bg": "bulgarian", "lt": "lithuanian", "la": "latin", "mi": "maori", "ml": "malayalam", "cy": "welsh", |
| "sk": "slovak", "te": "telugu", "fa": "persian", "lv": "latvian", "bn": "bengali", "sr": "serbian", |
| "az": "azerbaijani", "sl": "slovenian", "kn": "kannada", "et": "estonian", "mk": "macedonian", |
| "br": "breton", "eu": "basque", "is": "icelandic", "hy": "armenian", "ne": "nepali", "mn": "mongolian", |
| "bs": "bosnian", "kk": "kazakh", "sq": "albanian", "sw": "swahili", "gl": "galician", "mr": "marathi", |
| "pa": "punjabi", "si": "sinhala", "km": "khmer", "sn": "shona", "yo": "yoruba", "so": "somali", |
| "af": "afrikaans", "oc": "occitan", "ka": "georgian", "be": "belarusian", "tg": "tajik", |
| "sd": "sindhi", "gu": "gujarati", "am": "amharic", "yi": "yiddish", "lo": "lao", "uz": "uzbek", |
| "fo": "faroese", "ht": "haitian creole", "ps": "pashto", "tk": "turkmen", "nn": "nynorsk", |
| "mt": "maltese", "sa": "sanskrit", "lb": "luxembourgish", "my": "myanmar", "bo": "tibetan", |
| "tl": "tagalog", "mg": "malagasy", "as": "assamese", "tt": "tatar", "haw": "hawaiian", |
| "ln": "lingala", "ha": "hausa", "ba": "bashkir", "jw": "javanese", "su": "sundanese", |
| "yue": "cantonese", "minnan": "minnan", "wuyu": "wuyu", "dialect": "dialect", "zh/en": "zh/en", "en/zh": "en/zh" |
| } |
|
|
| TO_LANGUAGE_CODE = { |
| **{language: code for code, language in LANGUAGES.items()}, |
| "burmese": "my", "valencian": "ca", "flemish": "nl", "haitian": "ht", "letzeburgesch": "lb", |
| "pushto": "ps", "panjabi": "pa", "moldavian": "ro", "moldovan": "ro", "sinhalese": "si", |
| "castilian": "es", "mandarin": "zh", |
| } |
|
|
| AUDIO_EVENT = { |
| "ASR": "ASR", "AED": "AED", "SER": "SER", "Speech": "Speech", "/Speech": "/Speech", |
| "BGM": "BGM", "/BGM": "/BGM", "Laughter": "Laughter", "/Laughter": "/Laughter", |
| "Applause": "Applause", "/Applause": "/Applause", |
| } |
|
|
| EMOTION = { |
| "HAPPY": "HAPPY", "SAD": "SAD", "ANGRY": "ANGRY", "NEUTRAL": "NEUTRAL", |
| } |
|
|
| TTS_Vocal_Token = { |
| "TTS/B": "TTS/B", "TTS/O": "TTS/O", "TTS/Q": "TTS/Q", "TTS/A": "TTS/A", "TTS/CO": "TTS/CO", |
| "TTS/CL": "TTS/CL", "TTS/H": "TTS/H", **{f"TTS/SP{i:02d}": f"TTS/SP{i:02d}" for i in range(1, 14)} |
| } |
|
|
| |
| @lru_cache(maxsize=None) |
| def get_encoding(name: str = "gpt2", num_languages: int = 99): |
| vocab_path = os.path.join(os.path.dirname(__file__), "assets", f"{name}.tiktoken") |
| ranks = { |
| base64.b64decode(token): int(rank) |
| for token, rank in (line.split() for line in open(vocab_path) if line) |
| } |
| n_vocab = len(ranks) |
| special_tokens = {} |
| specials = [ |
| "<|endoftext|>", "<|startoftranscript|>", |
| *[f"<|{lang}|>" for lang in list(LANGUAGES.keys())[:num_languages]], |
| *[f"<|{audio_event}|>" for audio_event in list(AUDIO_EVENT.keys())], |
| *[f"<|{emotion}|>" for emotion in list(EMOTION.keys())], |
| "<|translate|>", "<|transcribe|>", "<|startoflm|>", "<|startofprev|>", |
| "<|nospeech|>", "<|notimestamps|>", |
| *[f"<|SPECIAL_TOKEN_{i}|>" for i in range(1, 31)], |
| *[f"<|{tts}|>" for tts in list(TTS_Vocal_Token.keys())], |
| *[f"<|{i * 0.02:.2f}|>" for i in range(1501)], |
| ] |
| for token in specials: |
| special_tokens[token] = n_vocab |
| n_vocab += 1 |
| return tiktoken.Encoding( |
| name=os.path.basename(vocab_path), |
| explicit_n_vocab=n_vocab, |
| pat_str=r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""", |
| mergeable_ranks=ranks, |
| special_tokens=special_tokens, |
| ) |
|
|
| class SimpleTokenizer: |
| def __init__(self, encoding, num_languages: int = 99, language: Optional[str] = None, task: Optional[str] = None): |
| self.encoding = encoding |
| self.num_languages = num_languages |
| self.language = language |
| self.task = task |
| def encode(self, text: str): |
| return self.encoding.encode(text) |
| def decode(self, tokens: list): |
| return self.encoding.decode(tokens) |
|
|
| @lru_cache(maxsize=None) |
| def get_tokenizer( |
| multilingual: bool, |
| *, |
| num_languages: int = 99, |
| language: Optional[str] = None, |
| task: Optional[str] = None, |
| ) -> SimpleTokenizer: |
| if language is not None: |
| language = language.lower() |
| if language not in LANGUAGES: |
| if language in TO_LANGUAGE_CODE: |
| language = TO_LANGUAGE_CODE[language] |
| else: |
| raise ValueError(f"Unsupported language: {language}") |
| if multilingual: |
| encoding_name = "multilingual_zh_ja_yue_char_del" |
| language = language or "en" |
| task = task or "transcribe" |
| else: |
| encoding_name = "gpt2" |
| language = None |
| task = None |
| encoding = get_encoding(name=encoding_name, num_languages=num_languages) |
| return SimpleTokenizer(encoding=encoding, num_languages=num_languages, language=language, task=task) |
|
|
| class QwenTokenizer(): |
| def __init__(self, token_path, skip_special_tokens=True): |
| super().__init__() |
| special_tokens = { |
| 'eos_token': '<|endoftext|>', |
| 'pad_token': '<|endoftext|>', |
| 'additional_special_tokens': [ |
| '<|im_start|>', '<|im_end|>', '<|endofprompt|>', |
| '[breath]', '<strong>', '</strong>', '[noise]', |
| '[laughter]', '[cough]', '[clucking]', '[accent]', |
| '[quick_breath]', |
| "<laughter>", "</laughter>", |
| "[hissing]", "[sigh]", "[vocalized-noise]", |
| "[lipsmack]", "[mn]" |
| ] |
| } |
| self.special_tokens = special_tokens |
| self.tokenizer = AutoTokenizer.from_pretrained(token_path) |
| self.tokenizer.add_special_tokens(special_tokens) |
| self.skip_special_tokens = skip_special_tokens |
| def encode(self, text, **kwargs): |
| tokens = self.tokenizer([text], return_tensors="pt") |
| return tokens["input_ids"][0].cpu().tolist() |
| def decode(self, tokens): |
| tokens = torch.tensor(tokens, dtype=torch.int64) |
| return self.tokenizer.batch_decode([tokens], skip_special_tokens=self.skip_special_tokens)[0] |
|
|
| @lru_cache(maxsize=None) |
| def get_qwen_tokenizer(token_path: str, skip_special_tokens: bool) -> QwenTokenizer: |
| return QwenTokenizer(token_path=token_path, skip_special_tokens=skip_special_tokens) |