From 9afe4585afe3c1d15a4dc22cef89b4c752edc066 Mon Sep 17 00:00:00 2001 From: kangguangli Date: Mon, 5 Dec 2022 16:26:35 +0800 Subject: [PATCH 1/3] replace layers.cross_entropy with paddle.entropy --- .../fluid/contrib/slim/tests/imperative_test_utils.py | 2 +- python/paddle/fluid/contrib/slim/tests/test_graph.py | 2 +- .../fluid/contrib/slim/tests/test_imperative_qat.py | 2 +- .../fluid/contrib/slim/tests/test_imperative_qat_amp.py | 4 ++-- .../fluid/contrib/slim/tests/test_imperative_qat_lsq.py | 2 +- .../contrib/slim/tests/test_quantization_mkldnn_pass.py | 2 +- .../fluid/contrib/slim/tests/test_quantization_pass.py | 8 ++++---- .../contrib/slim/tests/test_quantization_scale_pass.py | 2 +- .../contrib/slim/tests/test_quantize_transpiler_v2.py | 2 +- .../contrib/slim/tests/test_user_defined_quantization.py | 2 +- .../fluid/contrib/tests/test_quantize_transpiler.py | 6 +++--- .../fluid/contrib/tests/test_weight_decay_extend.py | 2 +- .../paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py | 2 +- .../fluid/tests/book/notest_understand_sentiment.py | 4 ++-- .../paddle/fluid/tests/book/test_image_classification.py | 2 +- python/paddle/fluid/tests/book/test_recognize_digits.py | 2 +- python/paddle/fluid/tests/book/test_word2vec_book.py | 4 +++- python/paddle/fluid/tests/test_error_clip.py | 2 +- python/paddle/fluid/tests/test_if_else_op.py | 6 ++++-- 19 files changed, 31 insertions(+), 27 deletions(-) diff --git a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py index c0637bc7de480e..de734f27fd9aca 100644 --- a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py +++ b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py @@ -70,7 +70,7 @@ def train_lenet(lenet, reader, optimizer): label = paddle.to_tensor(y_data) out = lenet(img) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_graph.py b/python/paddle/fluid/contrib/slim/tests/test_graph.py index 482c7237bfce8a..fa4dd5642c1844 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_graph.py +++ b/python/paddle/fluid/contrib/slim/tests/test_graph.py @@ -47,7 +47,7 @@ def conv_block(): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return [img, label], avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py index 0e0fbd752b2874..23d1e11ce14f2e 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py @@ -130,7 +130,7 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label) avg_loss = paddle.mean(loss) avg_loss.backward() adam.minimize(avg_loss) diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py index d01fc2e63cfd10..f51bbe7134158e 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py @@ -119,7 +119,7 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): with paddle.amp.auto_cast(): out = model(img) acc = paddle.static.accuracy(out, label) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label) avg_loss = paddle.mean(loss) scaled_loss = scaler.scale(avg_loss) scaled_loss.backward() @@ -129,7 +129,7 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): else: out = model(img) acc = paddle.static.accuracy(out, label) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py index 2b06ee5bf06d87..398f7738bfe669 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py @@ -171,7 +171,7 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py index 23b89512454a7e..4d87aee1a1a78f 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py @@ -47,7 +47,7 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py index 5aabeee1197993..6f3f60f7c0c0e7 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py @@ -39,7 +39,7 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = fluid.layers.cross_entropy(input=hidden, label=label) + loss = paddle.nn.functional.cross_entropy(input=hidden, label=label) loss = paddle.mean(loss) return loss @@ -87,7 +87,7 @@ def conv_bn_layer( input=hidden, pool_size=2, pool_type='avg', pool_stride=2 ) fc = fluid.layers.fc(input=pool, size=10) - loss = fluid.layers.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label) loss = paddle.mean(loss) return loss @@ -115,7 +115,7 @@ def conv_net(img, label, quant_skip_pattern): hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') with fluid.name_scope(quant_skip_pattern): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss @@ -756,7 +756,7 @@ def conv_bn_layer( ) pool_add = paddle.nn.functional.relu(paddle.add(x=pool1, y=pool2)) fc = fluid.layers.fc(input=pool_add, size=10) - loss = fluid.layers.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label) loss = paddle.mean(loss) return loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py index 46e37002460372..38cbee296d9899 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py @@ -55,7 +55,7 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py index 3b487ae1fae2b0..167be8853585ef 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py @@ -52,7 +52,7 @@ def conv_net(img, label): with fluid.name_scope("skip_quant"): hidden = fluid.layers.fc(input=conv_pool_1, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py index 25656278137a7c..2abfa3e07f1e81 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py +++ b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py @@ -57,7 +57,7 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py index cdbd65fad68a62..fb004b54084e62 100644 --- a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py +++ b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py @@ -30,7 +30,7 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = fluid.layers.cross_entropy(input=hidden, label=label) + loss = paddle.nn.functional.cross_entropy(input=hidden, label=label) loss = paddle.mean(loss) return loss @@ -58,7 +58,7 @@ def conv_bn_layer( short = conv_bn_layer(hidden, 16, 1, 1, 0, act=None) hidden = paddle.nn.functional.relu(paddle.add(x=conv, y=short)) fc = fluid.layers.fc(input=hidden, size=10) - loss = fluid.layers.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label) loss = paddle.mean(loss) return loss @@ -82,7 +82,7 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py index 7b40d513f2e5c3..e547ea7300be17 100644 --- a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py +++ b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py @@ -86,7 +86,7 @@ def bow_net( fc_1 = fluid.layers.fc(input=bow_tanh, size=hid_dim, act="tanh") fc_2 = fluid.layers.fc(input=fc_1, size=hid_dim2, act="tanh") prediction = fluid.layers.fc(input=[fc_2], size=class_dim, act="softmax") - cost = fluid.layers.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_cost = paddle.mean(x=cost) return avg_cost diff --git a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py index 3c034d60f37537..50bee3b7610b00 100644 --- a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py +++ b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py @@ -150,7 +150,7 @@ def model(): auc_var, batch_auc_var, auc_states = paddle.static.auc( input=predict, label=label ) - cost = fluid.layers.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label) avg_cost = paddle.mean(x=cost) return datas, avg_cost, predict, train_file_path diff --git a/python/paddle/fluid/tests/book/notest_understand_sentiment.py b/python/paddle/fluid/tests/book/notest_understand_sentiment.py index 3809e797f3c7ca..eedca5e56a84b9 100644 --- a/python/paddle/fluid/tests/book/notest_understand_sentiment.py +++ b/python/paddle/fluid/tests/book/notest_understand_sentiment.py @@ -47,7 +47,7 @@ def convolution_net( prediction = fluid.layers.fc( input=[conv_3, conv_4], size=class_dim, act="softmax" ) - cost = fluid.layers.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction @@ -82,7 +82,7 @@ def stacked_lstm_net( prediction = fluid.layers.fc( input=[fc_last, lstm_last], size=class_dim, act='softmax' ) - cost = fluid.layers.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction diff --git a/python/paddle/fluid/tests/book/test_image_classification.py b/python/paddle/fluid/tests/book/test_image_classification.py index 3a401df20370de..ef3db8ee84905d 100644 --- a/python/paddle/fluid/tests/book/test_image_classification.py +++ b/python/paddle/fluid/tests/book/test_image_classification.py @@ -119,7 +119,7 @@ def train(net_type, use_cuda, save_dirname, is_local): raise ValueError("%s network is not supported" % net_type) predict = fluid.layers.fc(input=net, size=classdim, act='softmax') - cost = fluid.layers.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label) avg_cost = paddle.mean(cost) acc = paddle.static.accuracy(input=predict, label=label) diff --git a/python/paddle/fluid/tests/book/test_recognize_digits.py b/python/paddle/fluid/tests/book/test_recognize_digits.py index a89cb1617a12a2..af2be2ab6c7073 100644 --- a/python/paddle/fluid/tests/book/test_recognize_digits.py +++ b/python/paddle/fluid/tests/book/test_recognize_digits.py @@ -30,7 +30,7 @@ def loss_net(hidden, label): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = fluid.layers.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) avg_loss = paddle.mean(loss) acc = paddle.static.accuracy(input=prediction, label=label) return prediction, avg_loss, acc diff --git a/python/paddle/fluid/tests/book/test_word2vec_book.py b/python/paddle/fluid/tests/book/test_word2vec_book.py index ab89e79f60a897..e503ba8ec7b422 100644 --- a/python/paddle/fluid/tests/book/test_word2vec_book.py +++ b/python/paddle/fluid/tests/book/test_word2vec_book.py @@ -96,7 +96,9 @@ def __network__(words): predict_word = fluid.layers.fc( input=hidden1, size=dict_size, act='softmax' ) - cost = fluid.layers.cross_entropy(input=predict_word, label=words[4]) + cost = paddle.nn.functional.cross_entropy( + input=predict_word, label=words[4] + ) avg_cost = paddle.mean(cost) return avg_cost, predict_word diff --git a/python/paddle/fluid/tests/test_error_clip.py b/python/paddle/fluid/tests/test_error_clip.py index f6e1122e04562b..241e48bc68dee2 100644 --- a/python/paddle/fluid/tests/test_error_clip.py +++ b/python/paddle/fluid/tests/test_error_clip.py @@ -31,7 +31,7 @@ label = fluid.layers.data(name='y', shape=[1], dtype='int64') - cost = fluid.layers.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label) avg_cost = paddle.mean(cost) prog_clip = prog.clone() diff --git a/python/paddle/fluid/tests/test_if_else_op.py b/python/paddle/fluid/tests/test_if_else_op.py index 24857164dc30b8..cd015ad0ab452b 100644 --- a/python/paddle/fluid/tests/test_if_else_op.py +++ b/python/paddle/fluid/tests/test_if_else_op.py @@ -65,7 +65,7 @@ def not_test_raw_api(self): prob = merge_lod_tensor( in_true=true_out, in_false=false_out, mask=cond, x=image ) - loss = layers.cross_entropy(input=prob, label=label) + loss = paddle.nn.functional.cross_entropy(input=prob, label=label) avg_loss = paddle.mean(loss) optimizer = MomentumOptimizer(learning_rate=0.001, momentum=0.9) @@ -121,7 +121,9 @@ def not_test_ifelse(self): ie.output(prob) prob = ie() - loss = layers.cross_entropy(input=prob[0], label=label) + loss = paddle.nn.functional.cross_entropy( + input=prob[0], label=label + ) avg_loss = paddle.mean(loss) optimizer = MomentumOptimizer(learning_rate=0.001, momentum=0.9) From e94b5483b02520bc8861b5d87cca463e893b1dff Mon Sep 17 00:00:00 2001 From: kangguangli Date: Wed, 7 Dec 2022 03:55:13 +0000 Subject: [PATCH 2/3] fix args --- .../fluid/contrib/slim/tests/imperative_test_utils.py | 2 +- python/paddle/fluid/contrib/slim/tests/test_graph.py | 2 +- .../fluid/contrib/slim/tests/test_imperative_qat.py | 2 +- .../fluid/contrib/slim/tests/test_imperative_qat_amp.py | 4 ++-- .../fluid/contrib/slim/tests/test_imperative_qat_lsq.py | 2 +- .../contrib/slim/tests/test_quantization_mkldnn_pass.py | 2 +- .../fluid/contrib/slim/tests/test_quantization_pass.py | 8 ++++---- .../contrib/slim/tests/test_quantization_scale_pass.py | 2 +- .../contrib/slim/tests/test_quantize_transpiler_v2.py | 2 +- .../contrib/slim/tests/test_user_defined_quantization.py | 2 +- .../fluid/contrib/tests/test_quantize_transpiler.py | 6 +++--- .../fluid/contrib/tests/test_weight_decay_extend.py | 2 +- .../paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py | 2 +- .../fluid/tests/book/notest_understand_sentiment.py | 4 ++-- .../paddle/fluid/tests/book/test_image_classification.py | 2 +- python/paddle/fluid/tests/book/test_recognize_digits.py | 2 +- python/paddle/fluid/tests/book/test_word2vec_book.py | 2 +- python/paddle/fluid/tests/test_error_clip.py | 2 +- python/paddle/fluid/tests/test_if_else_op.py | 4 ++-- 19 files changed, 27 insertions(+), 27 deletions(-) diff --git a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py index de734f27fd9aca..b61b69c9af4ed4 100644 --- a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py +++ b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py @@ -70,7 +70,7 @@ def train_lenet(lenet, reader, optimizer): label = paddle.to_tensor(y_data) out = lenet(img) - loss = paddle.nn.functional.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_graph.py b/python/paddle/fluid/contrib/slim/tests/test_graph.py index fa4dd5642c1844..cf1430662d0a28 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_graph.py +++ b/python/paddle/fluid/contrib/slim/tests/test_graph.py @@ -47,7 +47,7 @@ def conv_block(): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return [img, label], avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py index 23d1e11ce14f2e..6b9e3fbdb83363 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py @@ -130,7 +130,7 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) avg_loss.backward() adam.minimize(avg_loss) diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py index f51bbe7134158e..097d90dc4c8eb3 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py @@ -119,7 +119,7 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): with paddle.amp.auto_cast(): out = model(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) scaled_loss = scaler.scale(avg_loss) scaled_loss.backward() @@ -129,7 +129,7 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): else: out = model(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py index 398f7738bfe669..9a161c6f466801 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py @@ -171,7 +171,7 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py index 4d87aee1a1a78f..89d50a5cbfe235 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py @@ -47,7 +47,7 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py index fd9015dda5b804..f1c45c5e694da3 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py @@ -39,7 +39,7 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = paddle.nn.functional.cross_entropy(input=hidden, label=label) + loss = paddle.nn.functional.cross_entropy(input=hidden, label=label, reduction='none', use_softmax=False) loss = paddle.mean(loss) return loss @@ -87,7 +87,7 @@ def conv_bn_layer( input=hidden, pool_size=2, pool_type='avg', pool_stride=2 ) fc = fluid.layers.fc(input=pool, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) loss = paddle.mean(loss) return loss @@ -115,7 +115,7 @@ def conv_net(img, label, quant_skip_pattern): hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') with fluid.name_scope(quant_skip_pattern): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss @@ -756,7 +756,7 @@ def conv_bn_layer( ) pool_add = paddle.nn.functional.relu(paddle.add(x=pool1, y=pool2)) fc = fluid.layers.fc(input=pool_add, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) loss = paddle.mean(loss) return loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py index 38cbee296d9899..5f15936c31f9fc 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py @@ -55,7 +55,7 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py index 167be8853585ef..1e1536e2d318a5 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py @@ -52,7 +52,7 @@ def conv_net(img, label): with fluid.name_scope("skip_quant"): hidden = fluid.layers.fc(input=conv_pool_1, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py index 2abfa3e07f1e81..a064cfd48cf9d4 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py +++ b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py @@ -57,7 +57,7 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py index fb004b54084e62..c49a8d8f3476ff 100644 --- a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py +++ b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py @@ -30,7 +30,7 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = paddle.nn.functional.cross_entropy(input=hidden, label=label) + loss = paddle.nn.functional.cross_entropy(input=hidden, label=label, reduction='none', use_softmax=False) loss = paddle.mean(loss) return loss @@ -58,7 +58,7 @@ def conv_bn_layer( short = conv_bn_layer(hidden, 16, 1, 1, 0, act=None) hidden = paddle.nn.functional.relu(paddle.add(x=conv, y=short)) fc = fluid.layers.fc(input=hidden, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label) + loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) loss = paddle.mean(loss) return loss @@ -82,7 +82,7 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py index e547ea7300be17..821e7b4b164d6e 100644 --- a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py +++ b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py @@ -86,7 +86,7 @@ def bow_net( fc_1 = fluid.layers.fc(input=bow_tanh, size=hid_dim, act="tanh") fc_2 = fluid.layers.fc(input=fc_1, size=hid_dim2, act="tanh") prediction = fluid.layers.fc(input=[fc_2], size=class_dim, act="softmax") - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(x=cost) return avg_cost diff --git a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py index 50bee3b7610b00..cd5390bc59b628 100644 --- a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py +++ b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py @@ -150,7 +150,7 @@ def model(): auc_var, batch_auc_var, auc_states = paddle.static.auc( input=predict, label=label ) - cost = paddle.nn.functional.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(x=cost) return datas, avg_cost, predict, train_file_path diff --git a/python/paddle/fluid/tests/book/notest_understand_sentiment.py b/python/paddle/fluid/tests/book/notest_understand_sentiment.py index eedca5e56a84b9..d1b620d4f56866 100644 --- a/python/paddle/fluid/tests/book/notest_understand_sentiment.py +++ b/python/paddle/fluid/tests/book/notest_understand_sentiment.py @@ -47,7 +47,7 @@ def convolution_net( prediction = fluid.layers.fc( input=[conv_3, conv_4], size=class_dim, act="softmax" ) - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction @@ -82,7 +82,7 @@ def stacked_lstm_net( prediction = fluid.layers.fc( input=[fc_last, lstm_last], size=class_dim, act='softmax' ) - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label) + cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction diff --git a/python/paddle/fluid/tests/book/test_image_classification.py b/python/paddle/fluid/tests/book/test_image_classification.py index ef3db8ee84905d..af41a59f5230fa 100644 --- a/python/paddle/fluid/tests/book/test_image_classification.py +++ b/python/paddle/fluid/tests/book/test_image_classification.py @@ -119,7 +119,7 @@ def train(net_type, use_cuda, save_dirname, is_local): raise ValueError("%s network is not supported" % net_type) predict = fluid.layers.fc(input=net, size=classdim, act='softmax') - cost = paddle.nn.functional.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(cost) acc = paddle.static.accuracy(input=predict, label=label) diff --git a/python/paddle/fluid/tests/book/test_recognize_digits.py b/python/paddle/fluid/tests/book/test_recognize_digits.py index af2be2ab6c7073..fc128c2a1f2428 100644 --- a/python/paddle/fluid/tests/book/test_recognize_digits.py +++ b/python/paddle/fluid/tests/book/test_recognize_digits.py @@ -30,7 +30,7 @@ def loss_net(hidden, label): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label) + loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) acc = paddle.static.accuracy(input=prediction, label=label) return prediction, avg_loss, acc diff --git a/python/paddle/fluid/tests/book/test_word2vec_book.py b/python/paddle/fluid/tests/book/test_word2vec_book.py index e503ba8ec7b422..e12441bda4caf7 100644 --- a/python/paddle/fluid/tests/book/test_word2vec_book.py +++ b/python/paddle/fluid/tests/book/test_word2vec_book.py @@ -97,7 +97,7 @@ def __network__(words): input=hidden1, size=dict_size, act='softmax' ) cost = paddle.nn.functional.cross_entropy( - input=predict_word, label=words[4] + input=predict_word, label=words[4], reduction='none', use_softmax=False ) avg_cost = paddle.mean(cost) return avg_cost, predict_word diff --git a/python/paddle/fluid/tests/test_error_clip.py b/python/paddle/fluid/tests/test_error_clip.py index 241e48bc68dee2..fc4e157441bb76 100644 --- a/python/paddle/fluid/tests/test_error_clip.py +++ b/python/paddle/fluid/tests/test_error_clip.py @@ -31,7 +31,7 @@ label = fluid.layers.data(name='y', shape=[1], dtype='int64') - cost = paddle.nn.functional.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) avg_cost = paddle.mean(cost) prog_clip = prog.clone() diff --git a/python/paddle/fluid/tests/test_if_else_op.py b/python/paddle/fluid/tests/test_if_else_op.py index cd015ad0ab452b..67be3e20c81045 100644 --- a/python/paddle/fluid/tests/test_if_else_op.py +++ b/python/paddle/fluid/tests/test_if_else_op.py @@ -65,7 +65,7 @@ def not_test_raw_api(self): prob = merge_lod_tensor( in_true=true_out, in_false=false_out, mask=cond, x=image ) - loss = paddle.nn.functional.cross_entropy(input=prob, label=label) + loss = paddle.nn.functional.cross_entropy(input=prob, label=label, reduction='none', use_softmax=False) avg_loss = paddle.mean(loss) optimizer = MomentumOptimizer(learning_rate=0.001, momentum=0.9) @@ -122,7 +122,7 @@ def not_test_ifelse(self): prob = ie() loss = paddle.nn.functional.cross_entropy( - input=prob[0], label=label + input=prob[0], label=label, reduction='none', use_softmax=False ) avg_loss = paddle.mean(loss) From 664b07ac67a999e61f29c138ffb376e4a5ac1551 Mon Sep 17 00:00:00 2001 From: kangguangli Date: Wed, 7 Dec 2022 06:21:23 +0000 Subject: [PATCH 3/3] fix codestyle --- .../contrib/slim/tests/imperative_test_utils.py | 4 +++- .../fluid/contrib/slim/tests/test_graph.py | 4 +++- .../contrib/slim/tests/test_imperative_qat.py | 4 +++- .../slim/tests/test_imperative_qat_amp.py | 8 ++++++-- .../slim/tests/test_imperative_qat_lsq.py | 4 +++- .../slim/tests/test_quantization_mkldnn_pass.py | 4 +++- .../contrib/slim/tests/test_quantization_pass.py | 16 ++++++++++++---- .../slim/tests/test_quantization_scale_pass.py | 4 +++- .../slim/tests/test_quantize_transpiler_v2.py | 4 +++- .../slim/tests/test_user_defined_quantization.py | 4 +++- .../contrib/tests/test_quantize_transpiler.py | 12 +++++++++--- .../contrib/tests/test_weight_decay_extend.py | 4 +++- .../fluid/incubate/fleet/tests/fleet_deep_ctr.py | 4 +++- .../tests/book/notest_understand_sentiment.py | 8 ++++++-- .../tests/book/test_image_classification.py | 4 +++- .../fluid/tests/book/test_recognize_digits.py | 4 +++- .../fluid/tests/book/test_word2vec_book.py | 5 ++++- python/paddle/fluid/tests/test_error_clip.py | 4 +++- python/paddle/fluid/tests/test_if_else_op.py | 4 +++- 19 files changed, 79 insertions(+), 26 deletions(-) diff --git a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py index b61b69c9af4ed4..744bd3690309b2 100644 --- a/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py +++ b/python/paddle/fluid/contrib/slim/tests/imperative_test_utils.py @@ -70,7 +70,9 @@ def train_lenet(lenet, reader, optimizer): label = paddle.to_tensor(y_data) out = lenet(img) - loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_graph.py b/python/paddle/fluid/contrib/slim/tests/test_graph.py index cf1430662d0a28..ce014c949afd89 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_graph.py +++ b/python/paddle/fluid/contrib/slim/tests/test_graph.py @@ -47,7 +47,9 @@ def conv_block(): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return [img, label], avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py index 6b9e3fbdb83363..cbf5b30976d8c7 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat.py @@ -130,7 +130,9 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) avg_loss.backward() adam.minimize(avg_loss) diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py index 097d90dc4c8eb3..ea0a38ea5519a2 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_amp.py @@ -119,7 +119,9 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): with paddle.amp.auto_cast(): out = model(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) scaled_loss = scaler.scale(avg_loss) scaled_loss.backward() @@ -129,7 +131,9 @@ def model_train(self, model, batch_num=-1, batch_size=32, use_amp=False): else: out = model(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py index 9a161c6f466801..caf9648b3c3369 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_lsq.py @@ -171,7 +171,9 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = paddle.nn.functional.cross_entropy(out, label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py index 89d50a5cbfe235..734c96b6d34936 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_mkldnn_pass.py @@ -47,7 +47,9 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py index f1c45c5e694da3..7fea0da68561d5 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py @@ -39,7 +39,9 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = paddle.nn.functional.cross_entropy(input=hidden, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=hidden, label=label, reduction='none', use_softmax=False + ) loss = paddle.mean(loss) return loss @@ -87,7 +89,9 @@ def conv_bn_layer( input=hidden, pool_size=2, pool_type='avg', pool_stride=2 ) fc = fluid.layers.fc(input=pool, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=fc, label=label, reduction='none', use_softmax=False + ) loss = paddle.mean(loss) return loss @@ -115,7 +119,9 @@ def conv_net(img, label, quant_skip_pattern): hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') with fluid.name_scope(quant_skip_pattern): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss @@ -756,7 +762,9 @@ def conv_bn_layer( ) pool_add = paddle.nn.functional.relu(paddle.add(x=pool1, y=pool2)) fc = fluid.layers.fc(input=pool_add, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=fc, label=label, reduction='none', use_softmax=False + ) loss = paddle.mean(loss) return loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py index 5f15936c31f9fc..264d9a7550b524 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_scale_pass.py @@ -55,7 +55,9 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py index 1e1536e2d318a5..0c5c957f776441 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantize_transpiler_v2.py @@ -52,7 +52,9 @@ def conv_net(img, label): with fluid.name_scope("skip_quant"): hidden = fluid.layers.fc(input=conv_pool_1, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py index a064cfd48cf9d4..c355a98cd82cf6 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py +++ b/python/paddle/fluid/contrib/slim/tests/test_user_defined_quantization.py @@ -57,7 +57,9 @@ def conv_net(img, label): ) hidden = fluid.layers.fc(input=conv_pool_2, size=100, act='relu') prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py index c49a8d8f3476ff..72e9b172c843f4 100644 --- a/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py +++ b/python/paddle/fluid/contrib/tests/test_quantize_transpiler.py @@ -30,7 +30,9 @@ def linear_fc(num): hidden = data for _ in range(num): hidden = fluid.layers.fc(hidden, size=128, act='relu') - loss = paddle.nn.functional.cross_entropy(input=hidden, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=hidden, label=label, reduction='none', use_softmax=False + ) loss = paddle.mean(loss) return loss @@ -58,7 +60,9 @@ def conv_bn_layer( short = conv_bn_layer(hidden, 16, 1, 1, 0, act=None) hidden = paddle.nn.functional.relu(paddle.add(x=conv, y=short)) fc = fluid.layers.fc(input=hidden, size=10) - loss = paddle.nn.functional.cross_entropy(input=fc, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=fc, label=label, reduction='none', use_softmax=False + ) loss = paddle.mean(loss) return loss @@ -82,7 +86,9 @@ def conv_net(img, label): act="relu", ) prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) return avg_loss diff --git a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py index 821e7b4b164d6e..4f32eaffc3e52b 100644 --- a/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py +++ b/python/paddle/fluid/contrib/tests/test_weight_decay_extend.py @@ -86,7 +86,9 @@ def bow_net( fc_1 = fluid.layers.fc(input=bow_tanh, size=hid_dim, act="tanh") fc_2 = fluid.layers.fc(input=fc_1, size=hid_dim2, act="tanh") prediction = fluid.layers.fc(input=[fc_2], size=class_dim, act="softmax") - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(x=cost) return avg_cost diff --git a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py index cd5390bc59b628..e4f3cc959f314f 100644 --- a/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py +++ b/python/paddle/fluid/incubate/fleet/tests/fleet_deep_ctr.py @@ -150,7 +150,9 @@ def model(): auc_var, batch_auc_var, auc_states = paddle.static.auc( input=predict, label=label ) - cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=predict, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(x=cost) return datas, avg_cost, predict, train_file_path diff --git a/python/paddle/fluid/tests/book/notest_understand_sentiment.py b/python/paddle/fluid/tests/book/notest_understand_sentiment.py index d1b620d4f56866..1d952bffc4d33d 100644 --- a/python/paddle/fluid/tests/book/notest_understand_sentiment.py +++ b/python/paddle/fluid/tests/book/notest_understand_sentiment.py @@ -47,7 +47,9 @@ def convolution_net( prediction = fluid.layers.fc( input=[conv_3, conv_4], size=class_dim, act="softmax" ) - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction @@ -82,7 +84,9 @@ def stacked_lstm_net( prediction = fluid.layers.fc( input=[fc_last, lstm_last], size=class_dim, act='softmax' ) - cost = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) accuracy = paddle.static.accuracy(input=prediction, label=label) return avg_cost, accuracy, prediction diff --git a/python/paddle/fluid/tests/book/test_image_classification.py b/python/paddle/fluid/tests/book/test_image_classification.py index af41a59f5230fa..96afda40d33837 100644 --- a/python/paddle/fluid/tests/book/test_image_classification.py +++ b/python/paddle/fluid/tests/book/test_image_classification.py @@ -119,7 +119,9 @@ def train(net_type, use_cuda, save_dirname, is_local): raise ValueError("%s network is not supported" % net_type) predict = fluid.layers.fc(input=net, size=classdim, act='softmax') - cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=predict, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) acc = paddle.static.accuracy(input=predict, label=label) diff --git a/python/paddle/fluid/tests/book/test_recognize_digits.py b/python/paddle/fluid/tests/book/test_recognize_digits.py index fc128c2a1f2428..b31827cd824a61 100644 --- a/python/paddle/fluid/tests/book/test_recognize_digits.py +++ b/python/paddle/fluid/tests/book/test_recognize_digits.py @@ -30,7 +30,9 @@ def loss_net(hidden, label): prediction = fluid.layers.fc(input=hidden, size=10, act='softmax') - loss = paddle.nn.functional.cross_entropy(input=prediction, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prediction, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) acc = paddle.static.accuracy(input=prediction, label=label) return prediction, avg_loss, acc diff --git a/python/paddle/fluid/tests/book/test_word2vec_book.py b/python/paddle/fluid/tests/book/test_word2vec_book.py index e12441bda4caf7..640a33751a4740 100644 --- a/python/paddle/fluid/tests/book/test_word2vec_book.py +++ b/python/paddle/fluid/tests/book/test_word2vec_book.py @@ -97,7 +97,10 @@ def __network__(words): input=hidden1, size=dict_size, act='softmax' ) cost = paddle.nn.functional.cross_entropy( - input=predict_word, label=words[4], reduction='none', use_softmax=False + input=predict_word, + label=words[4], + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(cost) return avg_cost, predict_word diff --git a/python/paddle/fluid/tests/test_error_clip.py b/python/paddle/fluid/tests/test_error_clip.py index fc4e157441bb76..dcfe477a76b3e2 100644 --- a/python/paddle/fluid/tests/test_error_clip.py +++ b/python/paddle/fluid/tests/test_error_clip.py @@ -31,7 +31,9 @@ label = fluid.layers.data(name='y', shape=[1], dtype='int64') - cost = paddle.nn.functional.cross_entropy(input=predict, label=label, reduction='none', use_softmax=False) + cost = paddle.nn.functional.cross_entropy( + input=predict, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) prog_clip = prog.clone() diff --git a/python/paddle/fluid/tests/test_if_else_op.py b/python/paddle/fluid/tests/test_if_else_op.py index 67be3e20c81045..4da28e7cfd6c99 100644 --- a/python/paddle/fluid/tests/test_if_else_op.py +++ b/python/paddle/fluid/tests/test_if_else_op.py @@ -65,7 +65,9 @@ def not_test_raw_api(self): prob = merge_lod_tensor( in_true=true_out, in_false=false_out, mask=cond, x=image ) - loss = paddle.nn.functional.cross_entropy(input=prob, label=label, reduction='none', use_softmax=False) + loss = paddle.nn.functional.cross_entropy( + input=prob, label=label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) optimizer = MomentumOptimizer(learning_rate=0.001, momentum=0.9)