Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 3 additions & 4 deletions test/legacy_test/test_imperative_hook_for_layer.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,7 @@
import numpy as np
from test_imperative_lod_tensor_to_selected_rows import SimpleNet

import paddle
from paddle import base
from paddle.base import core
from paddle.base.dygraph import base as imperative_base
Expand Down Expand Up @@ -55,8 +56,7 @@ def test_forward_hook_return_value(self):

for place in places:
with base.dygraph.guard(place):
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)
base.set_flags({'FLAGS_sort_sum_gradient': True})

input_word = (
Expand Down Expand Up @@ -137,8 +137,7 @@ def test_forward_hook(self):

for place in places:
with base.dygraph.guard(place):
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)
base.set_flags({'FLAGS_sort_sum_gradient': True})

global call_forward_post_hook
Expand Down
6 changes: 2 additions & 4 deletions test/legacy_test/test_imperative_mnist.py
Original file line number Diff line number Diff line change
Expand Up @@ -120,8 +120,7 @@ def test_mnist_float32(self):
traced_layer = None

with base.dygraph.guard():
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)

mnist = MNIST()
sgd = paddle.optimizer.SGD(
Expand Down Expand Up @@ -177,8 +176,7 @@ def test_mnist_float32(self):
dy_param_value[param.name] = param.numpy()

with new_program_scope():
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)

exe = base.Executor(
base.CPUPlace()
Expand Down
6 changes: 2 additions & 4 deletions test/legacy_test/test_imperative_mnist_sorted_gradient.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,7 @@ def test_mnist_sort_gradient_float32(self):
epoch_num = 1

with base.dygraph.guard():
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)
base.set_flags({'FLAGS_sort_sum_gradient': True})

mnist2 = MNIST()
Expand Down Expand Up @@ -82,8 +81,7 @@ def test_mnist_sort_gradient_float32(self):
break

with new_program_scope():
base.default_startup_program().random_seed = seed
base.default_main_program().random_seed = seed
paddle.seed(seed)

exe = base.Executor(
base.CPUPlace()
Expand Down
3 changes: 1 addition & 2 deletions test/quantization/test_imperative_out_scale.py
Original file line number Diff line number Diff line change
Expand Up @@ -145,8 +145,7 @@ def test_out_scale_acc(self):

with base.dygraph.guard():
np.random.seed(seed)
paddle.static.default_main_program().random_seed = seed
paddle.static.default_startup_program().random_seed = seed
paddle.seed(seed)

lenet = ImperativeLenet()
lenet = fix_model_dict(lenet)
Expand Down
3 changes: 1 addition & 2 deletions test/quantization/test_imperative_ptq.py
Original file line number Diff line number Diff line change
Expand Up @@ -86,8 +86,7 @@ def setUpClass(cls):

seed = 1
np.random.seed(seed)
paddle.static.default_main_program().random_seed = seed
paddle.static.default_startup_program().random_seed = seed
paddle.seed(seed)

def cache_unzipping(self, target_folder, zip_path):
if not os.path.exists(target_folder):
Expand Down