From 9608a9318a7beb6b5c75a18e0b65d7be28552c3f Mon Sep 17 00:00:00 2001 From: JiabinYang <360788950@qq.com> Date: Mon, 25 Oct 2021 09:11:35 +0000 Subject: [PATCH 1/5] remove additional warnning in layer.to --- python/paddle/fluid/dygraph/layers.py | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index 694f9dc25e80c5..dcac591d5df58b 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -1479,17 +1479,19 @@ def _apply(self, func, device, dtype, blocking): param_applied.is_distributed = param.is_distributed self._parameters[key] = param_applied - if param.grad is not None: - with no_grad(): - grad_applied = func(param._grad_ivar(), device, dtype, - blocking) - - grad_applied.stop_gradient = param._grad_ivar( - ).stop_gradient - if hasattr(param._grad_ivar(), 'is_distributed'): - grad_applied.is_distributed = param._grad_ivar( - ).is_distributed - self._parameters[key]._set_grad_ivar(grad_applied) + with warnings.catch_warnings(): + warnings.filterwarnings("ignore", category=UserWarning) + if param.grad is not None: + with no_grad(): + grad_applied = func(param._grad_ivar(), device, + dtype, blocking) + + grad_applied.stop_gradient = param._grad_ivar( + ).stop_gradient + if hasattr(param._grad_ivar(), 'is_distributed'): + grad_applied.is_distributed = param._grad_ivar( + ).is_distributed + self._parameters[key]._set_grad_ivar(grad_applied) self._parameters_transform_map[id(param)] = [param_applied, key] From 417b6ffdf430ab2dd67176acdc5b1a69ff91d287 Mon Sep 17 00:00:00 2001 From: JiabinYang <360788950@qq.com> Date: Mon, 25 Oct 2021 09:14:52 +0000 Subject: [PATCH 2/5] remove additional warnning in layer.to --- python/paddle/fluid/dygraph/layers.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index dcac591d5df58b..95451a4843f12e 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -1479,8 +1479,6 @@ def _apply(self, func, device, dtype, blocking): param_applied.is_distributed = param.is_distributed self._parameters[key] = param_applied - with warnings.catch_warnings(): - warnings.filterwarnings("ignore", category=UserWarning) if param.grad is not None: with no_grad(): grad_applied = func(param._grad_ivar(), device, @@ -1575,7 +1573,10 @@ def transform(t, device, dtype, blocking): if dtype is None: dtype = t.dtype - new_t = t._copy_to(device, blocking) + with warnings.catch_warnings(): + warnings.filterwarnings("ignore", category=UserWarning) + new_t = t._copy_to(device, blocking) + if isinstance(t, framework.ParamBase): if dtype is not None and dtype != t.dtype: framework._dygraph_tracer().trace_op( From 4f12d7ee68b4e94202235fe72a768b0f37a8bbe1 Mon Sep 17 00:00:00 2001 From: JiabinYang <360788950@qq.com> Date: Mon, 25 Oct 2021 09:19:33 +0000 Subject: [PATCH 3/5] remove additional warnning in layer.to --- python/paddle/fluid/dygraph/layers.py | 11 +++++------ 1 file changed, 5 insertions(+), 6 deletions(-) diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index 95451a4843f12e..cac1b840916a39 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -1572,11 +1572,7 @@ def transform(t, device, dtype, blocking): device = t.place if dtype is None: dtype = t.dtype - - with warnings.catch_warnings(): - warnings.filterwarnings("ignore", category=UserWarning) - new_t = t._copy_to(device, blocking) - + new_t = t._copy_to(device, blocking) if isinstance(t, framework.ParamBase): if dtype is not None and dtype != t.dtype: framework._dygraph_tracer().trace_op( @@ -1593,7 +1589,10 @@ def transform(t, device, dtype, blocking): return new_t - self._apply(transform, device, dtype, blocking) + with warnings.catch_warnings(): + warnings.filterwarnings("ignore", category=UserWarning) + self._apply(transform, device, dtype, blocking) + self._dtype = dtype # [aliases] Compatible with old method names From 3fd7e25ecedd929202530f5cee11ffb59cc4634c Mon Sep 17 00:00:00 2001 From: JiabinYang <360788950@qq.com> Date: Mon, 25 Oct 2021 09:21:05 +0000 Subject: [PATCH 4/5] remove additional warnning in layer.to --- python/paddle/fluid/dygraph/layers.py | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index cac1b840916a39..b56c34b8a3def6 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -1479,17 +1479,17 @@ def _apply(self, func, device, dtype, blocking): param_applied.is_distributed = param.is_distributed self._parameters[key] = param_applied - if param.grad is not None: - with no_grad(): - grad_applied = func(param._grad_ivar(), device, - dtype, blocking) - - grad_applied.stop_gradient = param._grad_ivar( - ).stop_gradient - if hasattr(param._grad_ivar(), 'is_distributed'): - grad_applied.is_distributed = param._grad_ivar( - ).is_distributed - self._parameters[key]._set_grad_ivar(grad_applied) + if param.grad is not None: + with no_grad(): + grad_applied = func(param._grad_ivar(), device, dtype, + blocking) + + grad_applied.stop_gradient = param._grad_ivar( + ).stop_gradient + if hasattr(param._grad_ivar(), 'is_distributed'): + grad_applied.is_distributed = param._grad_ivar( + ).is_distributed + self._parameters[key]._set_grad_ivar(grad_applied) self._parameters_transform_map[id(param)] = [param_applied, key] From 7d00f1ac271aef0dd21733bc6ea9884d24128e79 Mon Sep 17 00:00:00 2001 From: JiabinYang <360788950@qq.com> Date: Mon, 25 Oct 2021 09:33:04 +0000 Subject: [PATCH 5/5] remove additional warnning in layer.to --- python/paddle/fluid/dygraph/layers.py | 1 + 1 file changed, 1 insertion(+) diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index b56c34b8a3def6..e1855ee6db9af8 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -1572,6 +1572,7 @@ def transform(t, device, dtype, blocking): device = t.place if dtype is None: dtype = t.dtype + new_t = t._copy_to(device, blocking) if isinstance(t, framework.ParamBase): if dtype is not None and dtype != t.dtype: