| |
| |
| |
| |
| """ |
| Utility functions for tokenize module. |
| """ |
|
|
| import re |
| from typing import Callable, List |
|
|
| _DIGITS_WITH_SEPARATOR = re.compile(r"(\d+[\.\,:])+\d+") |
|
|
|
|
| def apply_postprocessors( |
| segments: List[str], postprocessors: Callable[[List[str]], List[str]] |
| ) -> List[str]: |
| """ |
| A list of callables to apply to a raw segmentation result. |
| """ |
| for func in postprocessors: |
| segments = func(segments) |
|
|
| return segments |
|
|
|
|
| def rejoin_formatted_num(segments: List[str]) -> List[str]: |
| """ |
| Rejoin well-known formatted numeric that are over-tokenized. |
| The formatted numeric are numbers separated by ":", ",", or ".", |
| such as time, decimal numbers, comma-added numbers, and IP addresses. |
| |
| :param List[str] segments: result from word tokenizer |
| :return: a list of fixed tokens |
| :rtype: List[str] |
| |
| :Example: |
| tokens = ['ขณะ', 'นี้', 'เวลา', ' ', '12', ':', '00น', ' ', 'อัตรา', |
| 'แลกเปลี่ยน', ' ', '1', ',', '234', '.', '5', ' ', 'baht/zeny'] |
| rejoin_formatted_num(tokens) |
| # output: |
| # ['ขณะ', 'นี้', 'เวลา', ' ', '12:00น', ' ', 'อัตรา', 'แลกเปลี่ยน', ' ', '1,234.5', ' ', 'baht/zeny'] |
| |
| tokens = ['IP', ' ', 'address', ' ', 'ของ', 'คุณ', 'คือ', ' ', '127', '.', '0', '.', '0', '.', '1', ' ', 'ครับ'] |
| rejoin_formatted_num(tokens) |
| # output: |
| # ['IP', ' ', 'address', ' ', 'ของ', 'คุณ', 'คือ', ' ', '127.0.0.1', ' ', 'ครับ'] |
| """ |
| original = "".join(segments) |
| matching_results = _DIGITS_WITH_SEPARATOR.finditer(original) |
| tokens_joined = [] |
| pos = 0 |
| segment_idx = 0 |
|
|
| match = next(matching_results, None) |
| while segment_idx < len(segments) and match: |
| is_span_beginning = pos >= match.start() |
| token = segments[segment_idx] |
| if is_span_beginning: |
| connected_token = "" |
| while pos < match.end() and segment_idx < len(segments): |
| connected_token += segments[segment_idx] |
| pos += len(segments[segment_idx]) |
| segment_idx += 1 |
| if connected_token: |
| tokens_joined.append(connected_token) |
| match = next(matching_results, None) |
| else: |
| tokens_joined.append(token) |
| segment_idx += 1 |
| pos += len(token) |
| tokens_joined += segments[segment_idx:] |
| return tokens_joined |
|
|
|
|
| def strip_whitespace(segments: List[str]) -> List[str]: |
| """ |
| Strip whitespace(s) off each token and remove whitespace tokens. |
| :param List[str] segments: result from word tokenizer |
| :return: a list of tokens |
| :rtype: List[str] |
| |
| :Example: |
| tokens = [" ", "วันนี้ ", "เวลา ", "19.00น"] |
| strip_whitespace(tokens) |
| # ["วันนี้", "เวลา", "19.00น"] |
| |
| """ |
| segments = [token.strip(" ") for token in segments if token.strip(" ")] |
| return segments |
|
|