diff --git a/paddle/fluid/framework/ir/mkldnn/cpu_quantize_squash_pass_tester.cc b/paddle/fluid/framework/ir/mkldnn/cpu_quantize_squash_pass_tester.cc index 5b1cd5fe87aed0..90ed3009749ad5 100644 --- a/paddle/fluid/framework/ir/mkldnn/cpu_quantize_squash_pass_tester.cc +++ b/paddle/fluid/framework/ir/mkldnn/cpu_quantize_squash_pass_tester.cc @@ -1018,7 +1018,7 @@ TEST(CpuQuantizeSquashPass, fc_dequant_more_than_one_op_after_dequant) { // a->Concat1->b // b->Concat2 -// b->Quatize1(Scale)->c +// b->Quantize1(Scale)->c // c->Fc1 // c->Fc2 TEST(CpuQuantizeSquashPass, quatize_with_same_scale) { diff --git a/paddle/fluid/framework/ir/mkldnn/self_attention_fuse_pass.h b/paddle/fluid/framework/ir/mkldnn/self_attention_fuse_pass.h index ade48f398e3b6c..a264795bd78fbb 100644 --- a/paddle/fluid/framework/ir/mkldnn/self_attention_fuse_pass.h +++ b/paddle/fluid/framework/ir/mkldnn/self_attention_fuse_pass.h @@ -24,7 +24,7 @@ namespace paddle { namespace framework { namespace ir { -// Fusing of self-attetion structure +// Fusing of self-attention structure class Graph; diff --git a/paddle/fluid/framework/ir/mkldnn/shuffle_channel_mkldnn_detect_pass.cc b/paddle/fluid/framework/ir/mkldnn/shuffle_channel_mkldnn_detect_pass.cc index 5cf9a973061f04..764712a2fcd8a3 100644 --- a/paddle/fluid/framework/ir/mkldnn/shuffle_channel_mkldnn_detect_pass.cc +++ b/paddle/fluid/framework/ir/mkldnn/shuffle_channel_mkldnn_detect_pass.cc @@ -174,7 +174,7 @@ void ShuffleChannelMKLDNNDetectPass::ApplyImpl(ir::Graph* graph) const { } } - // shuffle_channel dosen't change shape + // shuffle_channel doesn't change shape if ((reshape2_shape[0] != -1) && (x_shape1[0] != reshape2_shape[0])) { return; } diff --git a/paddle/fluid/framework/ir/multi_devices_graph_pass/all_reduce_deps_pass.cc b/paddle/fluid/framework/ir/multi_devices_graph_pass/all_reduce_deps_pass.cc index 0dcf316c33c696..6327929663ab45 100644 --- a/paddle/fluid/framework/ir/multi_devices_graph_pass/all_reduce_deps_pass.cc +++ b/paddle/fluid/framework/ir/multi_devices_graph_pass/all_reduce_deps_pass.cc @@ -150,7 +150,7 @@ class AllReduceDepsPass : public ir::Pass { const std::vector& all_reduce_op_handles) const { // get vars order std::map> vars = - GetSoredGradientsFromStaleProgram(graph); + GetSortedGradientsFromStaleProgram(graph); std::stringstream out; size_t grads_of_stale_program = 0; out << "Get Order From details::kStaleProgramOpDescs: "; @@ -188,7 +188,7 @@ class AllReduceDepsPass : public ir::Pass { } } - std::map> GetSoredGradientsFromStaleProgram( + std::map> GetSortedGradientsFromStaleProgram( const ir::Graph& graph) const { std::map> vars; auto ops = diff --git a/paddle/fluid/framework/ir/multi_devices_graph_pass/backward_optimizer_op_deps_pass.cc b/paddle/fluid/framework/ir/multi_devices_graph_pass/backward_optimizer_op_deps_pass.cc index 82e8dada115560..a4feed4693a624 100644 --- a/paddle/fluid/framework/ir/multi_devices_graph_pass/backward_optimizer_op_deps_pass.cc +++ b/paddle/fluid/framework/ir/multi_devices_graph_pass/backward_optimizer_op_deps_pass.cc @@ -68,7 +68,7 @@ class BackWardOpDepsPass : public ir::Pass { AddDep(graph, opt_handles[i - 1], opt_handles[i]); } - VLOG(10) << "add deps between backward and optimze:"; + VLOG(10) << "add deps between backward and optimize:"; AddDep(graph, backward_op_handles[backward_op_handles.size() - 1], opt_handles[0]); diff --git a/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_check_pass.cc b/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_check_pass.cc index d08bbc1720de6c..95f64e896f77bd 100644 --- a/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_check_pass.cc +++ b/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_check_pass.cc @@ -19,14 +19,14 @@ namespace paddle { namespace framework { namespace ir { -class SSAGraghBuilderWithChecker : public ir::Pass { +class SSAGraphBuilderWithChecker : public ir::Pass { protected: void ApplyImpl(ir::Graph *graph) const override { PADDLE_ENFORCE_EQ( IsValidGraph(graph), true, platform::errors::InvalidArgument( - "In SSAGraghBuilderWithChecker, invalid Graph input.")); + "In SSAGraphBuilderWithChecker, invalid Graph input.")); } bool IsValidGraph(const ir::Graph *graph) const { @@ -99,6 +99,6 @@ class SSAGraghBuilderWithChecker : public ir::Pass { } // namespace paddle REGISTER_PASS(multi_devices_check_pass, - paddle::framework::ir::SSAGraghBuilderWithChecker) + paddle::framework::ir::SSAGraphBuilderWithChecker) .RequireGraphAttr(paddle::framework::details::kGraphVars) .RequireGraphAttr(paddle::framework::details::kGraphDepVars); diff --git a/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_print_pass.cc b/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_print_pass.cc index 97454b7583de22..6005c7de5c551d 100644 --- a/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_print_pass.cc +++ b/paddle/fluid/framework/ir/multi_devices_graph_pass/multi_devices_graph_print_pass.cc @@ -19,7 +19,7 @@ namespace paddle { namespace framework { namespace ir { -class SSAGraghBuilderWithPrinterPass : public ir::Pass { +class SSAGraphBuilderWithPrinterPass : public ir::Pass { protected: void ApplyImpl(ir::Graph *graph) const override { std::unique_ptr fout( @@ -102,5 +102,5 @@ void GraphvizSSAGraphPrinter::Print(const ir::Graph &graph, } // namespace paddle REGISTER_PASS(multi_devices_print_pass, - paddle::framework::ir::SSAGraghBuilderWithPrinterPass) + paddle::framework::ir::SSAGraphBuilderWithPrinterPass) .RequirePassAttr(paddle::framework::ir::kGraphvizPath); diff --git a/paddle/fluid/framework/ir/xpu/add_layernorm_xpu_fuse_pass.cc b/paddle/fluid/framework/ir/xpu/add_layernorm_xpu_fuse_pass.cc index 5db915c4e17f5f..55e8786f73c4d8 100644 --- a/paddle/fluid/framework/ir/xpu/add_layernorm_xpu_fuse_pass.cc +++ b/paddle/fluid/framework/ir/xpu/add_layernorm_xpu_fuse_pass.cc @@ -59,7 +59,7 @@ After the pass is applied: scale---- add_layernorm_fusion ---- bias / | \ \ / | \ \ - variance | meam z_add + variance | mean z_add Output */ struct AddLayernormXPUPattern : public PatternBase { diff --git a/paddle/fluid/framework/ir/xpu/fused_multi_transformer_int8_xpu_quant_pass.cc b/paddle/fluid/framework/ir/xpu/fused_multi_transformer_int8_xpu_quant_pass.cc index 4f9af98495c377..7c5b7c9e5e4e72 100755 --- a/paddle/fluid/framework/ir/xpu/fused_multi_transformer_int8_xpu_quant_pass.cc +++ b/paddle/fluid/framework/ir/xpu/fused_multi_transformer_int8_xpu_quant_pass.cc @@ -525,7 +525,7 @@ int FusedMultiTransformerInt8XPUQuantPass::FusedMultiTransformerInt8( id++; } }; - // genereate input node + // generate input node attr2weight( "qkv_in_scale", &(input_max_nodes_vec[0]), &(input_max_names_vec[0])); attr2weight("out_linear_in_scale", diff --git a/paddle/fluid/framework/ir/xpu/multi_encoder_xpu_adaptive_seqlen_fuse_pass.h b/paddle/fluid/framework/ir/xpu/multi_encoder_xpu_adaptive_seqlen_fuse_pass.h index a21d6498dea8ec..22910c21205300 100644 --- a/paddle/fluid/framework/ir/xpu/multi_encoder_xpu_adaptive_seqlen_fuse_pass.h +++ b/paddle/fluid/framework/ir/xpu/multi_encoder_xpu_adaptive_seqlen_fuse_pass.h @@ -40,7 +40,7 @@ class MultiEncoderXPUAdaptiveSeqlenFusePass : public FusePassBase { /* adaptive seqlen V1, before: - inpu_var* mask_var* + input_var* mask_var* | | | | embedding_xpu matmul @@ -59,7 +59,7 @@ class MultiEncoderXPUAdaptiveSeqlenFusePass : public FusePassBase { after: - inpu_var* mask_var* + input_var* mask_var* \ / \ / embedding_xpu @@ -81,7 +81,7 @@ class MultiEncoderXPUAdaptiveSeqlenFusePass : public FusePassBase { /* adaptive seqlen V2, before: - inpu_var* mask_var* + input_var* mask_var* | | | | embedding_xpu not_equal @@ -115,7 +115,7 @@ class MultiEncoderXPUAdaptiveSeqlenFusePass : public FusePassBase { after: - inpu_var* mask_var* + input_var* mask_var* \ / \ / embedding_xpu