# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2021, PaddleNLP
# This file is distributed under the same license as the PaddleNLP package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2022.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PaddleNLP \n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2022-03-18 21:31+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.9.0\n"

#: ../source/paddlenlp.transformers.ernie_doc.tokenizer.rst:2
msgid "tokenizer"
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:1
msgid "基类：:class:`paddlenlp.transformers.ernie.tokenizer.ErnieTokenizer`"
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:1
msgid ""
"Constructs an ERNIE-Doc tokenizer. It uses a basic tokenizer to do "
"punctuation splitting, lower casing and so on, and follows a WordPiece "
"tokenizer to tokenize as subwords."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:5
msgid ""
"This tokenizer inherits from "
":class:`~paddlenlp.transformers.ernie.tokenizer.ErnieTokenizer`. For more"
" information regarding those methods, please refer to this superclass."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer
msgid "参数"
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:8
msgid ""
"The vocabulary file path (ends with '.txt') required to instantiate a "
"`WordpieceTokenizer`."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:11
msgid "Whether or not to lowercase the input when tokenizing. Defaults to`True`."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:13
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:14
msgid ""
"A special token representing the *unknown (out-of-vocabulary)* token. An "
"unknown token is set to be `unk_token` inorder to be converted to an ID. "
"Defaults to \"[UNK]\"."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:17
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:18
msgid ""
"A special token separating two different sentences in the same input. "
"Defaults to \"[SEP]\"."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:20
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:21
msgid ""
"A special token used to make arrays of tokens the same size for batching "
"purposes. Defaults to \"[PAD]\"."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:23
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:24
msgid ""
"A special token used for sequence classification. It is the last token of"
" the sequence when built with special tokens. Defaults to \"[CLS]\"."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:26
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:27
msgid ""
"A special token representing a masked token. This is the token used in "
"the masked language modeling task which the model tries to predict the "
"original unmasked ones. Defaults to \"[MASK]\"."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:32
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocTokenizer:33
msgid "实际案例"
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:1
msgid "基类：:class:`paddlenlp.transformers.tokenizer_utils.BPETokenizer`"
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:1
msgid ""
"Constructs an ERNIE-Doc BPE tokenizer. It uses a bpe tokenizer to do "
"punctuation splitting, lower casing and so on, then tokenize words as "
"subwords."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:4
msgid ""
"This tokenizer inherits from "
":class:`~paddlenlp.transformers.tokenizer_utils.BPETokenizer`. For more "
"information regarding those methods, please refer to this superclass."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:7
msgid "File path of the vocabulary."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:9
msgid "File path of the id to vocab."
msgstr ""

#: of paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer:11
msgid "File path of word merge text."
msgstr ""

#: of
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer.vocab_size:1
msgid "Return the size of vocabulary."
msgstr ""

#: of
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer.vocab_size
msgid "返回"
msgstr ""

#: of
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer.vocab_size:3
msgid "The size of vocabulary."
msgstr ""

#: of
#: paddlenlp.transformers.ernie_doc.tokenizer.ErnieDocBPETokenizer.vocab_size
msgid "返回类型"
msgstr ""

