# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2021, PaddleNLP
# This file is distributed under the same license as the PaddleNLP package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2022.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PaddleNLP \n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2022-03-18 21:31+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.9.0\n"

#: ../source/paddlenlp.transformers.blenderbot.modeling.rst:2
msgid "modeling"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotDecoder:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotEncoder:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:1
msgid "基类：:class:`paddlenlp.transformers.blenderbot.modeling.BlenderbotPretrainedModel`"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:1
msgid "Construct a bare Blenderbot Model."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:3
msgid ""
"This model inherits from "
":class:`~paddlenlp.transformers.model_utils.PretrainedModel`. Check the "
"superclass documentation for the generic methods and the library "
"implements for all its model."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:6
msgid ""
"This model is also a Paddle `paddle.nn.Layer "
"<https://www.paddlepaddle.org.cn/documentation "
"/docs/en/api/paddle/fluid/dygraph/layers/Layer_en.html>`__ subclass. Use "
"it as a regular Paddle Layer and refer to the Paddle documentation for "
"all matter related to general usage and behavior."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward
msgid "参数"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:10
msgid "Vocabulary size of the Blenderbot model."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:12
msgid "The id for begging of sentences token. Defaults to ``1``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:14
msgid "The id for padding token. Defaults to ``0``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:16
msgid "The id for end of sentence token. Defaults to ``2``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:18
msgid "The id indicating the start of decoding sentence. Defaults to ``1``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:20
msgid "Dimensionality of the layers and the pooler layer. Defaults to ``1280``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:22
msgid ""
"Number of Transformer encoder layers for BlenderbotEncoder. Defaults to "
"``2``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:24
msgid ""
"Number of Transformer decoder layers for BlenderbotDecoder. Defaults to "
"``12``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:26
msgid ""
"Number of attention heads for each Transformer encoder layer in "
"BlenderbotEncoder. Defaults to ``32``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:29
msgid ""
"Number of attention heads for each Transformer decoder layer in "
"BlenderbotDecoder. Defaults to ``32``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:32
msgid ""
"Dimensionality of the feed-forward layer for each Transformer encoder "
"layer in BlenderbotEncoder. Defaults to ``5120``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:35
msgid ""
"Dimensionality of the feed-forward layer for each Transformer dncoder "
"layer in BlenderbotDncoder. Defaults to ``5120``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:38
msgid ""
"The dropout probability for all fully connected layers in the embeddings,"
" encoder, and pooler. Defaults to ``0.1``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:41
msgid ""
"The non-linear activation function (function or string) in the encoder "
"and pooler. ``\"gelu\"``, ``\"relu\"`` and any other paddle supported "
"activation functions are supported. Defaults to ``\"gelu\"``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:45
msgid "The dropout ratio for the attention probabilities. Defaults to ``0.0``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:48
msgid "The dropout ratio for activations inside the fully connected layer."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:50
msgid ", The max position index of an input sequence. Defaults to ``128``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:53
msgid ""
"The standard deviation of the truncated_normal_initializer for "
"initializing all weight matrices. Defaults to ``0.02``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:56
msgid ""
"Indicate whether to scale embeddings by diving by sqrt(d_model). Defaults"
" to ``True``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel:58
msgid ""
"Indicate whether to put layer normalization into preprocessing of MHA and"
" FFN sub-layers. If True, pre-process is layer normalization and post-"
"precess includes dropout, residual connection. Otherwise, no pre-process "
"and post-precess includes dropout, residual connection, layer "
"normalization. Defaults to ``True``."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:1
msgid ""
"Indices of input sequence tokens in the vocabulary. They are numerical "
"representations of tokens that build the input sequence. It's data type "
"should be `int64` and has a shape of [batch_size, sequence_length]."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:5
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:5
msgid ""
"Mask to indicate whether to perform attention on each input token or not."
" The values should be either 0 or 1. The attention scores will be set to "
"**-infinity** for any positions in the mask that are **0**, and will be "
"**unchanged** for positions that are **1**.  - **1** for tokens that are "
"**not masked**, - **0** for tokens that are **masked**.  It's data type "
"should be `float32` and has a shape of [batch_size, sequence_length]. "
"Defaults to `None`."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:5
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:5
msgid ""
"Mask to indicate whether to perform attention on each input token or not."
" The values should be either 0 or 1. The attention scores will be set to "
"**-infinity** for any positions in the mask that are **0**, and will be "
"**unchanged** for positions that are **1**."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:10
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:10
msgid "**1** for tokens that are **not masked**,"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:11
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:11
msgid "**0** for tokens that are **masked**."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:13
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:13
msgid ""
"It's data type should be `float32` and has a shape of [batch_size, "
"sequence_length]. Defaults to `None`."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:16
msgid ""
"If not provided, ``decoder_input_ids`` will be automatically generated "
"based on ``decoder_start_token_id`` and ``input_ids``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:19
msgid ""
"If not provided, the default ``decoder_attention_mask`` will be a tensor "
"with upper triangular part being ``-np.inf``. the shape will be "
"``(decoder_length, decoder_length)``"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:22
msgid ""
"The output of encoder. If not provided, a ``encoder_output`` will be "
"generated from BlenderbotEncoder. Defaults to ``None``."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:16
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:25
msgid "Indicates whether to use cache to speed up decoding. Defaults to ``False``"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:18
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:27
msgid ""
"It is a list, and each element in the list is a tuple( "
":code:`(incremental_cache, static_cache)` ). See "
"`paddle.nn.TransformerDecoder.gen_cache` for more details. It is only "
"used for inference and should be None for training. Default None."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotEncoder.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward
msgid "返回"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:33
msgid ""
"If ``use_cache=False``, the return will be the last hidden state of "
"decoder with shape of [batch_size, seq_lens, hidden_size]. ``seq_lens`` "
"corresponds to the length of input sequence. Otherwise, the return will "
"be a tuple of ``(decoder_output, cache)``. Please refer to class "
":class:`paddle.nn.TransformerDecoder` for more information regarding "
"``cache``."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotEncoder.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward
msgid "返回类型"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:32
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:11
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.forward:40
msgid "示例"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.get_encoder:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotModel.get_encoder:1
msgid "This method is required for model with encoder-decoder architecture."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotPretrainedModel:1
msgid "基类：:class:`paddlenlp.transformers.model_utils.PretrainedModel`"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotPretrainedModel:1
msgid ""
"An abstract class for pretrained Blenderbot models. It provides "
"Blenderbot related `model_config_file`, `resource_files_names`, "
"`pretrained_resource_files_map`, `pretrained_init_configuration`, "
"`base_model_prefix` for downloading and loading pretrained models. Refer "
"to :class:`~paddlenlp.transformers.model_utils.PretrainedModel` for more "
"details."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotPretrainedModel.init_weights:1
msgid "Initialization hook"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotEncoder:1
msgid ""
"The encoder of Blenderbot Model. Please refer to "
":class:`~paddlenlp.transformers.model_utils.PretrainedModel` or "
":class:`~paddlenlp.transformers.Blenderbot.BlenderbotModel` for more "
"information regarding methods and arguments."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotEncoder.forward:1
msgid ""
"The last hidden states at the last layer of the encoder. It's data type "
"should be `float` and has a shape of `(batch_size, seq_lens, "
"hidden_size)`. ``seq_lens`` corresponds to the length of input sequence."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotDecoder:1
msgid ""
"The decoder of Blenderbot Model. Please refer to "
":class:`~paddlenlp.transformers.model_utils.PretrainedModel` and "
":class:`~paddlenlp.transformers.Blenderbot.BlenderbotModel` for more "
"information regarding methods and arguments."
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotDecoder.forward:1
msgid ""
"Please refer to "
":class:`~paddlenlp.transformers.Blenderbot.BlenderbotModel` for more "
"information regarding the arguments."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:1
msgid ""
"Please refer to "
":class:`~paddlenlp.transformers.Blenderbot.BlenderbotModel` for more "
"information regarding arguments. :returns:"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:27
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:6
msgid "If ``use_cache=False``, the return will be a tensor with shape of"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:6
msgid ""
"[batch_size, seq_lens, hidden_size]. Otherwise, the return will be a "
"tuple of ``(decoder_output, cache)``."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:14
msgid ""
"import paddle from paddlenlp.transformers import BlenderbotTokenizer, "
"BlenderbotForConditionalGeneration"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:17
msgid ""
"pretrained_model_name = \"blenderbot-400M-distill\" tokenizer = "
"BlenderbotTokenizer.from_pretrained(pretrained_model_name) model = "
"BlenderbotForConditionalGeneration.from_pretrained(pretrained_model_name)"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:21
msgid ""
"sample_text = \"My friends are cool but they eat too many carbs.\" inputs"
" = tokenizer(sample_text, return_attention_mask=True, "
"return_token_type_ids=False) inputs = {k: paddle.to_tensor([v]) for (k, "
"v) in inputs.items()}"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:25
msgid ""
"# Generate response using beam search result_ids, scores = "
"model.generate(input_ids=inputs['input_ids'],"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:27
msgid ""
"max_length=60, min_length=20, decode_strategy='beam_search', "
"num_beams=10, length_penalty=0.65)"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:34
msgid "for sequence_ids in result_ids.numpy().tolist():"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.forward:33
msgid ""
"print(\"User:    \", sample_text) print(\"bot:     \", "
"tokenizer.convert_ids_to_string(sequence_ids)) # \"bot:   That's "
"unfortunate. Are they trying to lose weight?\""
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.prepare_inputs_for_generation:1
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForConditionalGeneration.prepare_inputs_for_generation:1
msgid ""
"Prepare inputs for decoder to generate sentences. :returns: A dictionary "
"containing necessary inputs for generating next token. :rtype: dict"
msgstr ""

#: of paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM:1
msgid ""
"Constructs BLenderbot For Causal Language Model. This model is equivalent"
" to the blenderbot decoder without cross-attention."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:24
msgid ""
"If ``use_cache=False``, the return will be a tensor with shape of     "
"[batch_size, seq_lens, hidden_size]. Otherwise, the return will be a "
"tuple     of ``(lm_logits, cache)``."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:27
msgid ""
"[batch_size, seq_lens, hidden_size]. Otherwise, the return will be a "
"tuple of ``(lm_logits, cache)``."
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:35
msgid ""
"import paddle from paddlenlp.transformers import BlenderbotTokenizer, "
"BlenderbotForCausalLM use_cache = False text = \"My friends are cool but "
"they eat too many carbs.\" model_name = \"blenderbot-400M-distill\" "
"tokenizer = BlenderbotTokenizer.from_pretrained(model_name) model = "
"BlenderbotForCausalLM.from_pretrained(model_name) model.eval() inputs = "
"tokenizer(text) inputs = {k: paddle.to_tensor([v]) for (k, v) in "
"inputs.items()}"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:47
msgid "with paddle.no_grad():"
msgstr ""

#: of
#: paddlenlp.transformers.blenderbot.modeling.BlenderbotForCausalLM.forward:47
msgid ""
"outputs = model(**inputs, use_cache=use_cache) # outputs is a tuple of "
"(lm_logits, cache) if ``use_cache=True``."
msgstr ""

