-
Notifications
You must be signed in to change notification settings - Fork 773
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Docathon][Add CN Doc No.56-57] #6358
Changes from 1 commit
fb40183
31d53df
358b214
1f88fa8
3f8be7e
6c6f0cf
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,37 @@ | ||
.. _cn_api_paddle_incubate_nn_FusedBiasDropoutResidualLayerNorm: | ||
|
||
FusedBiasDropoutResidualLayerNorm | ||
------------------------------- | ||
|
||
.. py:class:: paddle.incubate.nn.FusedBiasDropoutResidualLayerNorm(embed_dim, dropout_rate=0.5, weight_attr=None, bias_attr=None, epsilon=1e-05, name=None) | ||
|
||
应用融合偏置、Dropout 和残差层归一化操作。 | ||
|
||
参数 | ||
:::::::::::: | ||
- **embed_dim** (int) - 输入和输出中预期的特征大小。 | ||
- **dropout_rate** (float,可选) - 在注意力权重上使用的 Dropout 概率,用于在注意力后的 Dropout 过程中丢弃一些注意力目标。0 表示无 Dropout。默认为 0.5。 | ||
- **bias_attr** (ParamAttr|bool,可选) - 指定偏置参数的属性。默认为 None,意味着使用默认的偏置参数属性。如果设置为 False,则该层不会有可训练的偏置参数。具体用法请参见 :ref:`cn_api_paddle_ParamAttr` 。 | ||
- **epsilon** (float,可选) - 添加到方差中的小值,以防止除零。默认为 1e-05。 | ||
|
||
代码示例 | ||
:::::::::::: | ||
|
||
COPY-FROM: paddle.incubate.nn.FusedBiasDropoutResidualLayerNorm | ||
|
||
forward(src, src_mask=None, cache=None) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 此处和英文文档无法对应,请再次确定来源是否有误 应修改为forward(x, residual),代码中也是此格式,请本人也重新确认一下。 There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
||
:::::::::::: | ||
应用融合偏置、Dropout 和残差层归一化操作。 | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 使用原算子名,无需翻译,如果要解释操作,请在后续补充。 There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
||
|
||
参数 | ||
:::::::::::: | ||
- **x** (Tensor) - 输入张量。它是一个形状为 `[batch_size, seq_len, embed_dim]` 的张量。数据类型应为 float32 或 float64。 | ||
- **residual** (Tensor,可选) - 残差张量。它是一个形状为 `[batch_size, value_length, vdim]` 的张量。数据类型应为 float32 或 float64。 | ||
|
||
返回 | ||
:::::::::::: | ||
Tensor|tuple:数据类型与 ``x`` 一样。 | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 根据英文文档 应翻译为 “与x具有相同数据类型和形状的张量” There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
||
|
||
extra_repr() | ||
:::::::::::: | ||
返回当前层的额外信息。 | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 根据英文文档 应翻译为“当前层的额外表示,可以有使用者建立的层的自定义实现”(可以想一下更好的翻译) There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,43 @@ | ||
.. _cn_api_paddle_incubate_nn_functional_fused_bias_dropout_residual_layer_norm: | ||
|
||
fused_bias_dropout_residual_layer_norm | ||
------------------------------- | ||
|
||
.. py:function:: paddle.incubate.nn.functional.fused_bias_dropout_residual_layer_norm(x, residual, bias=None, ln_scale=None, ln_bias=None, dropout_rate=0.5, ln_epsilon=1e-05, training=True, mode='upscale_in_train', name=None) | ||
|
||
融合偏置、Dropout 和残差层归一化操作符。其伪代码如下: | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. ”融合偏置、Dropout 和残差层归一化“不需要翻译,直接饮用原操作符名。如改为“fused_bias_dropout_residual_layer_norm操作符。” 若想增加功能介绍,可修为形式 ”fused_bias_dropout_residual_layer_norm操作符, 包含融合偏置、Dropout 和残差层归一化。“ There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
||
|
||
.. code-block:: text | ||
|
||
>>> y = layer_norm(residual + dropout(bias + x)) | ||
|
||
参数 | ||
:::::::::::: | ||
- **x** (Tensor) - 输入张量。其形状为 `[*, embed_dim]`。 | ||
- **residual** (Tensor) - 残差张量。其形状与 x 相同。 | ||
- **bias** (Tensor,可选) - 线性的偏置。其形状为 `[embed_dim]`。默认为 None。 | ||
- **ln_scale** (Tensor,可选) - 层归一化的权重张量。其形状为 `[embed_dim]`。默认为 None。 | ||
- **ln_bias** (Tensor,可选) - 层归一化的偏置张量。其形状为 `[embed_dim]`。默认为 None。 | ||
- **dropout_rate** (float,可选) - 在注意力权重上使用的 Dropout 概率,用于在注意力后的 Dropout 过程中丢弃一些注意力目标。0 表示无 Dropout。默认为 0.5。 | ||
- **ln_epsilon** (float,可选) - 在层归一化的分母中添加的小浮点数,用于避免除以零。默认为 1e-5。 | ||
- **training** (bool,可选) - 表示是否处于训练阶段的标志。默认为 True。 | ||
- **mode** (str,可选) - ['upscale_in_train'(默认) | 'downscale_in_infer'],有两种模式: | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. “有两种模式” -> "两种模式分别为:" There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done |
||
|
||
1. upscale_in_train(默认),在训练时上调输出 | ||
- 训练:out = input * mask / (1.0 - p) | ||
- 推理:out = input | ||
zade23 marked this conversation as resolved.
Show resolved
Hide resolved
|
||
|
||
2. downscale_in_infer,在推理时下调输出 | ||
- 训练:out = input * mask | ||
- 推理:out = input * (1.0 - p) | ||
zade23 marked this conversation as resolved.
Show resolved
Hide resolved
|
||
- **name** (str,可选) - 操作的名称(可选,默认为 None)。更多信息,请参考:ref:`api_guide_Name`。 | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 此处的“:ref: There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. @Vvsmile 开发者您好 若是我理解错您的意思了,或有其他方面的修改需求,烦请更新Review There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 细节问题,统一一下ref前是中文冒号还是英文冒号,表述不清引起的额外工作量,烦请见谅 There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
了解了,目前修改中已经统一使用英文冒号了 There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 辛苦 |
||
|
||
返回 | ||
:::::::::::: | ||
- Tensor,输出张量,数据类型和形状与 `x` 相同。 | ||
|
||
|
||
代码示例 | ||
:::::::::::: | ||
|
||
COPY-FROM: paddle.incubate.nn.functional.fused_bias_dropout_residual_layer_norm |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
使用原算子名,无需翻译,如果要解释操作,请在后续补充。
推荐修改为
”应用fused_bias_dropout_residual_layer_norm操作符,包含融合偏置、Dropout 和残差层归一化操作。“
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done
修改前:
应用融合偏置、Dropout 和残差层归一化操作。
修改后:
应用fused_bias_dropout_residual_layer_norm操作符,包含融合偏置、Dropout 和残差层归一化操作。