# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2021, PaddleNLP
# This file is distributed under the same license as the PaddleNLP package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2022.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PaddleNLP \n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2022-03-18 21:31+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.9.0\n"

#: ../source/paddlenlp.ops.fast_transformer.transformer.decoder.rst:2
msgid "decoder"
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:1
msgid "FasterTransformer decoder block."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder
#: paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder.forward
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder.forward
msgid "参数"
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:3
msgid "Transformer decoder block."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:13
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:5
msgid "The number of head used in multi-head attention."
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:7
msgid "The size of per head used in multi-head attention."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:31
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:9
msgid "The path to decoder_lib. Default to None."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:33
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder:11
msgid "Whether to use fp16 for decoder. Default to False."
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder.forward:1
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder.forward:1
msgid ""
"Defines the computation performed at every call. Should be overridden by "
"all subclasses."
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder.forward:4
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder.forward:4
msgid "unpacked tuple arguments"
msgstr ""

#: of
#: paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder.forward:6
#: paddlenlp.ops.fast_transformer.transformer.decoder.InferTransformerDecoder.forward:6
msgid "unpacked dict arguments"
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:1
msgid "FasterTransformer decoder for auto-regressive generation."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:3
msgid "The size of source vocabulary."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:5
msgid "The size of target vocabulary."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:7
msgid "The maximum length of input sequences."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:9
msgid "The number of sub-layers to be stacked in the encoder."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:11
msgid "The number of sub-layers to be stacked in the decoder."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:15
msgid ""
"The dimension for word embeddings, which is also the last dimension of "
"the input and output of multi-head attention, position-wise feed-forward "
"networks, encoder and decoder."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:19
msgid "Size of the hidden layer in position-wise feed-forward networks."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:21
msgid "Dropout rates. Used for pre-process, activation and inside attention."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:23
msgid "Whether to use weight sharing."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:25
msgid "The start token id and also is used as padding id. Defaults to 0."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:27
msgid "The end token id. Defaults to 1."
msgstr ""

#: of paddlenlp.ops.fast_transformer.transformer.decoder.FasterDecoder:29
msgid "The maximum output length. Defaults to 256."
msgstr ""

