diff --git a/python/paddle/nn/functional/activation.py b/python/paddle/nn/functional/activation.py index fa0ee722a072dc..010f6ff93ec8a1 100644 --- a/python/paddle/nn/functional/activation.py +++ b/python/paddle/nn/functional/activation.py @@ -1448,7 +1448,7 @@ def swish(x, name=None): Tensor(shape=[3], dtype=float32, place=Place(cpu), stop_gradient=True, [-0.23840584, 0. , 0.73105860]) """ - if in_dynamic_mode(): + if in_dynamic_or_pir_mode(): return _C_ops.swish(x) else: check_variable_and_dtype( diff --git a/test/legacy_test/test_activation_op.py b/test/legacy_test/test_activation_op.py index a96f892c3f1bd9..3ade491ee13e83 100644 --- a/test/legacy_test/test_activation_op.py +++ b/test/legacy_test/test_activation_op.py @@ -4428,6 +4428,7 @@ def test_check_grad(self): self.check_grad( ['X'], 'Out', + check_pir=True, ) @@ -4447,6 +4448,7 @@ def setUp(self): else paddle.CPUPlace() ) + @test_with_pir_api def test_static_api(self): with static_guard(): with paddle.static.program_guard(paddle.static.Program()): @@ -4470,6 +4472,7 @@ def test_dygraph_api(self): for r in [out1, out2]: np.testing.assert_allclose(out_ref, r.numpy(), rtol=1e-05) + @test_with_pir_api def test_base_api(self): with static_guard(): with base.program_guard(base.Program()):