Skip to content

Commit e4bc755

Browse files
committed
fix ut
1 parent 311a5f0 commit e4bc755

File tree

2 files changed

+4
-1
lines changed

2 files changed

+4
-1
lines changed

python/paddle/fluid/optimizer.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@
4242
from functools import cmp_to_key
4343
from .wrapped_decorator import signature_safe_contextmanager
4444
from .. import compat as cpt
45+
import warnings
4546

4647
__all__ = [
4748
'SGD', 'Momentum', 'Adagrad', 'Adam', 'Adamax', 'Dpsgd', 'DecayedAdagrad',
@@ -993,7 +994,7 @@ def flatten_param_grads(self, params_grads):
993994
p, 'regularizer', None) is not None:
994995
warnings.warn(
995996
"flatten_param_grads=True will be discarded since paramter '{}''s need_clip is False or "
996-
"the regularizer is set for parameter".format(p.name))
997+
"the regularizer is set".format(p.name))
997998
self._flatten_param_grads = False
998999
return params_grads
9991000

python/paddle/fluid/tests/unittests/test_adam_op.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -785,6 +785,7 @@ def test_adam_api(self):
785785

786786
def test_adam_flatten_param_grads_with_regularizer(self):
787787
# flatten_param_grads + regularizer is not supported yet.
788+
paddle.enable_static()
788789
main = fluid.Program()
789790
weight_attr = paddle.ParamAttr(
790791
name="weight1",
@@ -805,6 +806,7 @@ def test_adam_flatten_param_grads_with_regularizer(self):
805806
adam = fluid.optimizer.AdamOptimizer(
806807
0.01, flatten_param_grads=True, align_size=256)
807808
adam.minimize(avg_cost)
809+
paddle.disable_static()
808810

809811
self.assertEqual(adam._flatten_param_grads, False)
810812

0 commit comments

Comments
 (0)