From 87b1bff4b04d1a68f264007674d5327ce0259a56 Mon Sep 17 00:00:00 2001 From: limin2021 <1121099234@qq.com> Date: Tue, 9 Nov 2021 08:07:09 +0000 Subject: [PATCH] Polish formats. --- paddle/fluid/operators/fused/fused_attention_op.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/paddle/fluid/operators/fused/fused_attention_op.cc b/paddle/fluid/operators/fused/fused_attention_op.cc index f2d6e180596f6..11601a5ce40d5 100644 --- a/paddle/fluid/operators/fused/fused_attention_op.cc +++ b/paddle/fluid/operators/fused/fused_attention_op.cc @@ -325,7 +325,7 @@ class FusedAttentionOpMaker : public framework::OpProtoAndCheckerMaker { out = layer_norm(input); out = compute_qkv(out) + bias; // fmha module - { + { out = transpose(out, perm=[2, 0, 3, 1, 4]); out = q * k^t; out = attn_mask + out;