peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/tokenizers
/trainers
/__init__.pyi
| # Generated content DO NOT EDIT | |
| class Trainer: | |
| """ | |
| Base class for all trainers | |
| This class is not supposed to be instantiated directly. Instead, any implementation of a | |
| Trainer will return an instance of this class when instantiated. | |
| """ | |
| class BpeTrainer(Trainer): | |
| """ | |
| Trainer capable of training a BPE model | |
| Args: | |
| vocab_size (:obj:`int`, `optional`): | |
| The size of the final vocabulary, including all tokens and alphabet. | |
| min_frequency (:obj:`int`, `optional`): | |
| The minimum frequency a pair should have in order to be merged. | |
| show_progress (:obj:`bool`, `optional`): | |
| Whether to show progress bars while training. | |
| special_tokens (:obj:`List[Union[str, AddedToken]]`, `optional`): | |
| A list of special tokens the model should know of. | |
| limit_alphabet (:obj:`int`, `optional`): | |
| The maximum different characters to keep in the alphabet. | |
| initial_alphabet (:obj:`List[str]`, `optional`): | |
| A list of characters to include in the initial alphabet, even | |
| if not seen in the training dataset. | |
| If the strings contain more than one character, only the first one | |
| is kept. | |
| continuing_subword_prefix (:obj:`str`, `optional`): | |
| A prefix to be used for every subword that is not a beginning-of-word. | |
| end_of_word_suffix (:obj:`str`, `optional`): | |
| A suffix to be used for every subword that is a end-of-word. | |
| max_token_length (:obj:`int`, `optional`): | |
| Prevents creating tokens longer than the specified size. | |
| This can help with reducing polluting your vocabulary with | |
| highly repetitive tokens like `======` for wikipedia | |
| """ | |
| class UnigramTrainer(Trainer): | |
| """ | |
| Trainer capable of training a Unigram model | |
| Args: | |
| vocab_size (:obj:`int`): | |
| The size of the final vocabulary, including all tokens and alphabet. | |
| show_progress (:obj:`bool`): | |
| Whether to show progress bars while training. | |
| special_tokens (:obj:`List[Union[str, AddedToken]]`): | |
| A list of special tokens the model should know of. | |
| initial_alphabet (:obj:`List[str]`): | |
| A list of characters to include in the initial alphabet, even | |
| if not seen in the training dataset. | |
| If the strings contain more than one character, only the first one | |
| is kept. | |
| shrinking_factor (:obj:`float`): | |
| The shrinking factor used at each step of the training to prune the | |
| vocabulary. | |
| unk_token (:obj:`str`): | |
| The token used for out-of-vocabulary tokens. | |
| max_piece_length (:obj:`int`): | |
| The maximum length of a given token. | |
| n_sub_iterations (:obj:`int`): | |
| The number of iterations of the EM algorithm to perform before | |
| pruning the vocabulary. | |
| """ | |
| def __init__( | |
| self, | |
| vocab_size=8000, | |
| show_progress=True, | |
| special_tokens=[], | |
| shrinking_factor=0.75, | |
| unk_token=None, | |
| max_piece_length=16, | |
| n_sub_iterations=2, | |
| ): | |
| pass | |
| class WordLevelTrainer(Trainer): | |
| """ | |
| Trainer capable of training a WorldLevel model | |
| Args: | |
| vocab_size (:obj:`int`, `optional`): | |
| The size of the final vocabulary, including all tokens and alphabet. | |
| min_frequency (:obj:`int`, `optional`): | |
| The minimum frequency a pair should have in order to be merged. | |
| show_progress (:obj:`bool`, `optional`): | |
| Whether to show progress bars while training. | |
| special_tokens (:obj:`List[Union[str, AddedToken]]`): | |
| A list of special tokens the model should know of. | |
| """ | |
| class WordPieceTrainer(Trainer): | |
| """ | |
| Trainer capable of training a WordPiece model | |
| Args: | |
| vocab_size (:obj:`int`, `optional`): | |
| The size of the final vocabulary, including all tokens and alphabet. | |
| min_frequency (:obj:`int`, `optional`): | |
| The minimum frequency a pair should have in order to be merged. | |
| show_progress (:obj:`bool`, `optional`): | |
| Whether to show progress bars while training. | |
| special_tokens (:obj:`List[Union[str, AddedToken]]`, `optional`): | |
| A list of special tokens the model should know of. | |
| limit_alphabet (:obj:`int`, `optional`): | |
| The maximum different characters to keep in the alphabet. | |
| initial_alphabet (:obj:`List[str]`, `optional`): | |
| A list of characters to include in the initial alphabet, even | |
| if not seen in the training dataset. | |
| If the strings contain more than one character, only the first one | |
| is kept. | |
| continuing_subword_prefix (:obj:`str`, `optional`): | |
| A prefix to be used for every subword that is not a beginning-of-word. | |
| end_of_word_suffix (:obj:`str`, `optional`): | |
| A suffix to be used for every subword that is a end-of-word. | |
| """ | |
| def __init__( | |
| self, | |
| vocab_size=30000, | |
| min_frequency=0, | |
| show_progress=True, | |
| special_tokens=[], | |
| limit_alphabet=None, | |
| initial_alphabet=[], | |
| continuing_subword_prefix="##", | |
| end_of_word_suffix=None, | |
| ): | |
| pass | |