diff --git a/paddle/fluid/framework/fleet/gloo_wrapper.cc b/paddle/fluid/framework/fleet/gloo_wrapper.cc index 140de78bf01dc5..277004b6dc1642 100644 --- a/paddle/fluid/framework/fleet/gloo_wrapper.cc +++ b/paddle/fluid/framework/fleet/gloo_wrapper.cc @@ -141,7 +141,7 @@ std::vector HdfsStore::get(const std::string& key) { PADDLE_ENFORCE_EQ(read_status, 0, paddle::platform::errors::Fatal( - "HdfsStore::get, path read faied: " + path)); + "HdfsStore::get, path read failed: " + path)); #endif return result; } diff --git a/paddle/fluid/framework/fleet/gloo_wrapper.h b/paddle/fluid/framework/fleet/gloo_wrapper.h index b1f236fb60548d..fa352fb8eb99f6 100644 --- a/paddle/fluid/framework/fleet/gloo_wrapper.h +++ b/paddle/fluid/framework/fleet/gloo_wrapper.h @@ -95,7 +95,7 @@ class ParallelConnectContext : public gloo::rendezvous::Context { : gloo::rendezvous::Context(rank, size, base) {} virtual ~ParallelConnectContext() {} // in gloo::rendezvous::Context wait&get one by one, - // slowly in case big size, especialy in HdfsStore + // slowly in case big size, especially in HdfsStore void connectFullMesh(Store& store, // NOLINT std::shared_ptr& dev); // NOLINT struct Impl { diff --git a/paddle/fluid/framework/fleet/heter_ps/feature_value.h b/paddle/fluid/framework/fleet/heter_ps/feature_value.h index 7cdb20bb978bcf..0ab35f4a2beca9 100644 --- a/paddle/fluid/framework/fleet/heter_ps/feature_value.h +++ b/paddle/fluid/framework/fleet/heter_ps/feature_value.h @@ -75,7 +75,7 @@ class CommonFeatureValueAccessor { __host__ __device__ int EmbedDim() const { return embed_sgd_dim; } __host__ __device__ int EmbedXDim() const { return embedx_sgd_dim; } __host__ __device__ int EmbedWDim() const { return embedx_dim; } - __host__ __device__ int CpuPtrIndex() const { return 0; } // cpuprt uint64 + __host__ __device__ int CpuPtrIndex() const { return 0; } // cpu_ptr uint64 __host__ __device__ int DeltaScoreIndex() const { return CpuPtrIndex() + 2; } @@ -686,7 +686,7 @@ class CommonFeatureValueAccessor { std::vector embedx_w; */ std::stringstream os; - os << "cpuptr: " << common_feature_value.CpuPtr(const_cast(v)) + os << "cpu_ptr: " << common_feature_value.CpuPtr(const_cast(v)) << " delta_score: " << v[2] << " show: " << v[3] << " click: " << v[4] << " embed_w:" << v[5] << " embed_g2sum:"; for (int i = common_feature_value.EmbedG2SumIndex(); @@ -732,7 +732,7 @@ struct FeatureValue { friend std::ostream& operator<<(std::ostream& out, FeatureValue& val) { out << "show: " << val.show << " clk: " << val.clk << " slot: " << val.slot << " lr: " << val.lr << " mf_dim: " << val.mf_dim - << "cpuptr: " << val.cpu_ptr << " mf_size: " << val.mf_size << " mf:"; + << "cpu_ptr: " << val.cpu_ptr << " mf_size: " << val.mf_size << " mf:"; for (int i = 0; i < val.mf_dim + 1; ++i) { out << " " << val.mf[i]; } diff --git a/paddle/fluid/framework/fleet/heter_ps/gpu_graph_utils.h b/paddle/fluid/framework/fleet/heter_ps/gpu_graph_utils.h index 7523b349f11908..ed0c23a0fa8dcf 100644 --- a/paddle/fluid/framework/fleet/heter_ps/gpu_graph_utils.h +++ b/paddle/fluid/framework/fleet/heter_ps/gpu_graph_utils.h @@ -96,7 +96,7 @@ inline void debug_gpu_memory_info(int gpu_id, const char* desc) { err, cudaSuccess, platform::errors::InvalidArgument("cudaMemGetInfo failed!")); - VLOG(0) << "updatex gpu memory on device " << gpu_id << ", " + VLOG(0) << "update gpu memory on device " << gpu_id << ", " << "avail=" << avail / 1024.0 / 1024.0 / 1024.0 << "g, " << "total=" << total / 1024.0 / 1024.0 / 1024.0 << "g, " << "use_rate=" << (total - avail) / static_cast(total) diff --git a/paddle/fluid/framework/fleet/heter_ps/heter_comm.h b/paddle/fluid/framework/fleet/heter_ps/heter_comm.h index 18e3966b220c0c..1ccca57cd2979d 100644 --- a/paddle/fluid/framework/fleet/heter_ps/heter_comm.h +++ b/paddle/fluid/framework/fleet/heter_ps/heter_comm.h @@ -278,7 +278,7 @@ class HeterComm { h_push_fea_sizes.resize(node_size * node_size); } }; - // pull parition shard key by devices + // pull partition shard key by devices struct PullResource { size_t h_recv_fea_num = 0; uint32_t* d_restore_keys_idx = nullptr; diff --git a/paddle/fluid/framework/io/crypto/aes_cipher.cc b/paddle/fluid/framework/io/crypto/aes_cipher.cc index 5a1258fa432f78..8802dc1b121584 100644 --- a/paddle/fluid/framework/io/crypto/aes_cipher.cc +++ b/paddle/fluid/framework/io/crypto/aes_cipher.cc @@ -12,19 +12,6 @@ // See the License for the specific language governing permissions and // limitations under the License. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - #include "paddle/fluid/framework/io/crypto/aes_cipher.h" #include diff --git a/paddle/fluid/framework/io/crypto/aes_cipher.h b/paddle/fluid/framework/io/crypto/aes_cipher.h index a60fe1de45d7c4..5dfdbd49f4f96b 100644 --- a/paddle/fluid/framework/io/crypto/aes_cipher.h +++ b/paddle/fluid/framework/io/crypto/aes_cipher.h @@ -12,19 +12,6 @@ // See the License for the specific language governing permissions and // limitations under the License. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - #pragma once #include diff --git a/paddle/fluid/framework/ir/fusion_group/code_generator.cc b/paddle/fluid/framework/ir/fusion_group/code_generator.cc index f18b5fd6686593..e59c495f2dd8de 100644 --- a/paddle/fluid/framework/ir/fusion_group/code_generator.cc +++ b/paddle/fluid/framework/ir/fusion_group/code_generator.cc @@ -366,7 +366,7 @@ std::unordered_map CodeGenerator::EncodeVarNodes( // Encoding output vars. for (auto* out : output_var_nodes) { - VLOG(3) << "Ecoding output names:" << out->Name() << "(" << out + VLOG(3) << "Encoding output names:" << out->Name() << "(" << out << "), id:" << id; if (var_ids.find(out) == var_ids.end()) { var_ids[out] = id++; diff --git a/paddle/fluid/framework/ir/fusion_group/fusion_group_pass.cc b/paddle/fluid/framework/ir/fusion_group/fusion_group_pass.cc index 1f8f3b63598ce2..4eba0a73ae787f 100644 --- a/paddle/fluid/framework/ir/fusion_group/fusion_group_pass.cc +++ b/paddle/fluid/framework/ir/fusion_group/fusion_group_pass.cc @@ -37,7 +37,7 @@ void FusionGroupPass::ApplyImpl(ir::Graph* graph) const { // if (!phi::GPUDeviceCode::IsAvailable()) { // LOG(WARNING) // << "Disable fusion_group because CUDA Driver or NVRTC is not - // avaiable."; + // available."; // return 0; // }