diff --git a/test/legacy_test/test_imperative_hook_for_layer.py b/test/legacy_test/test_imperative_hook_for_layer.py index d5ce15c58233a8..ea59b930318745 100644 --- a/test/legacy_test/test_imperative_hook_for_layer.py +++ b/test/legacy_test/test_imperative_hook_for_layer.py @@ -17,6 +17,7 @@ import numpy as np from test_imperative_lod_tensor_to_selected_rows import SimpleNet +import paddle from paddle import base from paddle.base import core from paddle.base.dygraph import base as imperative_base @@ -55,8 +56,7 @@ def test_forward_hook_return_value(self): for place in places: with base.dygraph.guard(place): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) base.set_flags({'FLAGS_sort_sum_gradient': True}) input_word = ( @@ -137,8 +137,7 @@ def test_forward_hook(self): for place in places: with base.dygraph.guard(place): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) base.set_flags({'FLAGS_sort_sum_gradient': True}) global call_forward_post_hook diff --git a/test/legacy_test/test_imperative_mnist.py b/test/legacy_test/test_imperative_mnist.py index 614081effe90d8..c6717001d2a662 100644 --- a/test/legacy_test/test_imperative_mnist.py +++ b/test/legacy_test/test_imperative_mnist.py @@ -120,8 +120,7 @@ def test_mnist_float32(self): traced_layer = None with base.dygraph.guard(): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) mnist = MNIST() sgd = paddle.optimizer.SGD( @@ -177,8 +176,7 @@ def test_mnist_float32(self): dy_param_value[param.name] = param.numpy() with new_program_scope(): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) exe = base.Executor( base.CPUPlace() diff --git a/test/legacy_test/test_imperative_mnist_sorted_gradient.py b/test/legacy_test/test_imperative_mnist_sorted_gradient.py index 9629858b2d8bf6..4322f390e25504 100644 --- a/test/legacy_test/test_imperative_mnist_sorted_gradient.py +++ b/test/legacy_test/test_imperative_mnist_sorted_gradient.py @@ -30,8 +30,7 @@ def test_mnist_sort_gradient_float32(self): epoch_num = 1 with base.dygraph.guard(): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) base.set_flags({'FLAGS_sort_sum_gradient': True}) mnist2 = MNIST() @@ -82,8 +81,7 @@ def test_mnist_sort_gradient_float32(self): break with new_program_scope(): - base.default_startup_program().random_seed = seed - base.default_main_program().random_seed = seed + paddle.seed(seed) exe = base.Executor( base.CPUPlace() diff --git a/test/quantization/test_imperative_out_scale.py b/test/quantization/test_imperative_out_scale.py index a8317addd6edd0..2f7917ab5e1dce 100644 --- a/test/quantization/test_imperative_out_scale.py +++ b/test/quantization/test_imperative_out_scale.py @@ -145,8 +145,7 @@ def test_out_scale_acc(self): with base.dygraph.guard(): np.random.seed(seed) - paddle.static.default_main_program().random_seed = seed - paddle.static.default_startup_program().random_seed = seed + paddle.seed(seed) lenet = ImperativeLenet() lenet = fix_model_dict(lenet) diff --git a/test/quantization/test_imperative_ptq.py b/test/quantization/test_imperative_ptq.py index 189be58754c9f9..e01482c9576e69 100644 --- a/test/quantization/test_imperative_ptq.py +++ b/test/quantization/test_imperative_ptq.py @@ -86,8 +86,7 @@ def setUpClass(cls): seed = 1 np.random.seed(seed) - paddle.static.default_main_program().random_seed = seed - paddle.static.default_startup_program().random_seed = seed + paddle.seed(seed) def cache_unzipping(self, target_folder, zip_path): if not os.path.exists(target_folder):