Skip to content

Commit 9ddb9f6

Browse files
Merge pull request #234 from ooooo-create/fix_rrelu
[Accuracy diff No.48-49] Fix accuracy diff for rrelu API
2 parents 42fc091 + 5c12d23 commit 9ddb9f6

File tree

5 files changed

+7
-6
lines changed

5 files changed

+7
-6
lines changed

tester/api_config/2_paddle_only_random/random_calculation.txt

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25604,4 +25604,9 @@ paddle.put_along_axis(Tensor([6, 8000],"float32"), Tensor([6, 799],"int64"), Ten
2560425604
paddle.put_along_axis(Tensor([7, 8000],"float32"), Tensor([7, 799],"int64"), Tensor([7, 799],"float32"), 1, )
2560525605
paddle.put_along_axis(Tensor([8, 8000],"float32"), Tensor([8, 799],"int64"), Tensor([8, 799],"float32"), 1, )
2560625606
paddle.put_along_axis(Tensor([9, 8000],"float32"), Tensor([9, 799],"int64"), Tensor([9, 799],"float32"), 1, )
25607-
paddle.topk(Tensor([128, 1000],"float16"), k=5, )
25607+
paddle.topk(Tensor([128, 1000],"float16"), k=5, )
25608+
paddle.nn.functional.rrelu(Tensor([1, 2, 3, 4],"float64"), lower=0.05, upper=0.25, training=True, )
25609+
paddle.nn.functional.rrelu(Tensor([],"float32"), )
25610+
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float32"), 0.3, 0.300000009, training=True, )
25611+
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float64"), 0.3, 0.300000009, training=True, )
25612+
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float16"), 0.3, 0.300000009, training=True, )

tester/api_config/5_accuracy/accuracy_7.txt

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -153424,9 +153424,7 @@ paddle.nn.functional.relu6(x=Tensor([4],"float64"), )
153424153424
paddle.nn.functional.rrelu(Tensor([1, 2, 3, 4],"float64"), 0.05, 0.25, training=False, )
153425153425
paddle.nn.functional.rrelu(Tensor([1, 2, 3, 4],"float64"), 0.1, 0.33, training=False, )
153426153426
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float32"), 0.1, 0.3, training=False, )
153427-
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float32"), 0.3, 0.300000009, training=True, )
153428153427
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float64"), 0.1, 0.3, training=False, )
153429-
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float64"), 0.3, 0.300000009, training=True, )
153430153428
paddle.nn.functional.selu(Tensor([2, 2],"float32"), 1.0507009873554805, 1.6732632423543772, None, )
153431153429
paddle.nn.functional.selu(Tensor([2, 2],"float64"), 1.0507009873554805, 1.6732632423543772, None, )
153432153430
paddle.nn.functional.selu(Tensor([3, 3, 3],"float64"), 1.0507009873554805, 0, None, )

tester/api_config/5_accuracy/accuracy_cpu_kernel.txt

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38158,7 +38158,6 @@ paddle.nn.functional.relu6(Tensor([64, 906, 7, 7],"float16"), None, )
3815838158
paddle.nn.functional.relu6(Tensor([64, 96, 56, 56],"float16"), None, )
3815938159
paddle.nn.functional.relu6(Tensor([64, 972, 7, 7],"float16"), None, )
3816038160
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float16"), 0.1, 0.3, training=False, )
38161-
paddle.nn.functional.rrelu(Tensor([2, 3, 4, 5],"float16"), 0.3, 0.300000009, training=True, )
3816238161
paddle.nn.functional.sigmoid(Tensor([1, 3, 5, 2, 4],"float16"), )
3816338162
paddle.nn.functional.sigmoid(Tensor([10, 499, 4],"float16"), None, )
3816438163
paddle.nn.functional.sigmoid(Tensor([100, 200, 304],"float16"), )

tester/api_config/5_accuracy/accuracy_gpu_error.txt

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19940,8 +19940,6 @@ paddle.logit(Tensor([],"float32"), )
1994019940
paddle.nextafter(Tensor([],"float32"), Tensor([2, 3, 4],"float32"), )
1994119941
paddle.nn.functional.cosine_similarity(Tensor([23, 12, 1],"float32"), Tensor([23, 1, 10],"float32"), axis=2, eps=1e-06, )
1994219942
paddle.nn.functional.cosine_similarity(Tensor([5, 1, 3],"float64"), Tensor([1, 3],"float64"), axis=0, eps=1e-08, )
19943-
paddle.nn.functional.rrelu(Tensor([1, 2, 3, 4],"float64"), lower=0.05, upper=0.25, training=True, )
19944-
paddle.nn.functional.rrelu(Tensor([],"float32"), )
1994519943
paddle.reciprocal(Tensor([16, 1, 640, 640],"float16"), )
1994619944
paddle.reciprocal(Tensor([4, 1, 640, 640],"float16"), )
1994719945
paddle.reciprocal(Tensor([8, 1, 640, 640],"float16"), )

tester/base.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -67,6 +67,7 @@
6767
# "paddle.nn.functional.feature_alpha_dropout",
6868
# "paddle.incubate.nn.functional.fused_multi_head_attention", # If parameter "dropout_rate=0.5, attn_dropout_rate=0.5 (default value)" is not equal to 0.0 or 1.0, the result involves random calculation.
6969
# "paddle.nn.functional.scaled_dot_product_attention", # If parameter "dropout_p=0.0" is not equal to 0.0 or 1.0, the result involves random calculation.
70+
# "paddle.nn.functional.rrelu", # If parameter "training=True" is set, the result involves random calculation.
7071
]
7172
)
7273

0 commit comments

Comments
 (0)