Skip to content

Commit 0d2235a

Browse files
dzhwinterQiJune
authored andcommitted
GPUPlace to CUDAPlace (#6960)
1 parent 87aae57 commit 0d2235a

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

56 files changed

+179
-180
lines changed

paddle/framework/init.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ bool InitDevices(const std::vector<std::string> &devices) {
5454
#ifdef PADDLE_WITH_CUDA
5555
auto pos = string::RFind(p, ':', string::Piece::npos);
5656
auto number = device.substr(pos + 1);
57-
places.emplace_back(platform::GPUPlace(std::stoi(number)));
57+
places.emplace_back(platform::CUDAPlace(std::stoi(number)));
5858
#else
5959
LOG(WARNING)
6060
<< "'GPU' is not supported, Please re-compile with WITH_GPU option";

paddle/framework/lod_tensor.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -224,7 +224,7 @@ void SerializeToStream(std::ostream &os, const LoDTensor &tensor,
224224
while (size != 0) {
225225
size_t size_to_write = std::min(kBufSize, static_cast<size_t>(size));
226226
memory::Copy(cpu, buf.get(),
227-
boost::get<platform::GPUPlace>(tensor.place()),
227+
boost::get<platform::CUDAPlace>(tensor.place()),
228228
reinterpret_cast<const void *>(data), size_to_write,
229229
gpu_dev_ctx.stream());
230230
gpu_dev_ctx.Wait();

paddle/framework/lod_tensor_test.cu

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ __global__ void test(size_t* a, int size) {
2727

2828
TEST(LoDTensor, LoDInGPU) {
2929
paddle::framework::LoDTensor lod_tensor;
30-
paddle::platform::GPUPlace place(0);
30+
paddle::platform::CUDAPlace place(0);
3131

3232
paddle::framework::LoD src_lod;
3333
src_lod.push_back(std::vector<size_t>{0, 2, 4, 6, 8, 10, 12, 14});

paddle/framework/op_kernel_type_test.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,13 +37,13 @@ TEST(OpKernelType, Hash) {
3737
using OpKernelType = paddle::framework::OpKernelType;
3838
using DataType = paddle::framework::proto::DataType;
3939
using CPUPlace = paddle::platform::CPUPlace;
40-
using GPUPlace = paddle::platform::GPUPlace;
40+
using CUDAPlace = paddle::platform::CUDAPlace;
4141
using DataLayout = paddle::framework::DataLayout;
4242
using LibraryType = paddle::framework::LibraryType;
4343

4444
OpKernelType op_kernel_type_1(DataType::FP32, CPUPlace(), DataLayout::kNCHW,
4545
LibraryType::kCUDNN);
46-
OpKernelType op_kernel_type_2(DataType::FP32, GPUPlace(0), DataLayout::kNCHW,
46+
OpKernelType op_kernel_type_2(DataType::FP32, CUDAPlace(0), DataLayout::kNCHW,
4747
LibraryType::kCUDNN);
4848

4949
OpKernelType::Hash hasher;

paddle/framework/op_registry.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -188,7 +188,7 @@ class OpKernelRegistrar : public Registrar {
188188
}
189189

190190
#define REGISTER_OP_CUDA_KERNEL(op_type, ...) \
191-
REGISTER_OP_KERNEL(op_type, CUDA, ::paddle::platform::GPUPlace, __VA_ARGS__)
191+
REGISTER_OP_KERNEL(op_type, CUDA, ::paddle::platform::CUDAPlace, __VA_ARGS__)
192192

193193
#define REGISTER_OP_CPU_KERNEL(op_type, ...) \
194194
REGISTER_OP_KERNEL(op_type, CPU, ::paddle::platform::CPUPlace, __VA_ARGS__)

paddle/framework/tensor.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ private:
7171
```
7272

7373
```c++
74-
typedef boost::variant<GpuPlace, CpuPlace> Place;
74+
typedef boost::variant<CUDAPlace, CpuPlace> Place;
7575
typedef boost::variant<Dim<1>, Dim<2>, Dim<3>, Dim<4>, Dim<5>,
7676
Dim<6>, Dim<7>, Dim<8>, Dim<9>> DDimVar;
7777
typedef boost::variant<

paddle/framework/tensor_impl.h

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -125,11 +125,11 @@ inline void* Tensor::mutable_data(platform::Place place, std::type_index type) {
125125
boost::get<platform::CPUPlace>(place), size, type));
126126
} else if (platform::is_gpu_place(place)) {
127127
#ifndef PADDLE_WITH_CUDA
128-
PADDLE_THROW("'GPUPlace' is not supported in CPU only device.");
128+
PADDLE_THROW("'CUDAPlace' is not supported in CPU only device.");
129129
}
130130
#else
131-
holder_.reset(new PlaceholderImpl<platform::GPUPlace>(
132-
boost::get<platform::GPUPlace>(place), size, type));
131+
holder_.reset(new PlaceholderImpl<platform::CUDAPlace>(
132+
boost::get<platform::CUDAPlace>(place), size, type));
133133
}
134134
#endif
135135
offset_ = 0;

paddle/framework/tensor_test.cc

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -80,20 +80,20 @@ TEST(Tensor, MutableData) {
8080
float* p1 = nullptr;
8181
float* p2 = nullptr;
8282
// initialization
83-
p1 = src_tensor.mutable_data<float>(make_ddim({1, 2, 3}), GPUPlace());
83+
p1 = src_tensor.mutable_data<float>(make_ddim({1, 2, 3}), CUDAPlace());
8484
EXPECT_NE(p1, nullptr);
8585
// set src_tensor a new dim with large size
8686
// momery is supposed to be re-allocated
87-
p2 = src_tensor.mutable_data<float>(make_ddim({3, 4}), GPUPlace());
87+
p2 = src_tensor.mutable_data<float>(make_ddim({3, 4}), CUDAPlace());
8888
EXPECT_NE(p2, nullptr);
8989
EXPECT_NE(p1, p2);
9090
// set src_tensor a new dim with same size
9191
// momery block is supposed to be unchanged
92-
p1 = src_tensor.mutable_data<float>(make_ddim({2, 2, 3}), GPUPlace());
92+
p1 = src_tensor.mutable_data<float>(make_ddim({2, 2, 3}), CUDAPlace());
9393
EXPECT_EQ(p1, p2);
9494
// set src_tensor a new dim with smaller size
9595
// momery block is supposed to be unchanged
96-
p2 = src_tensor.mutable_data<float>(make_ddim({2, 2}), GPUPlace());
96+
p2 = src_tensor.mutable_data<float>(make_ddim({2, 2}), CUDAPlace());
9797
EXPECT_EQ(p1, p2);
9898
}
9999
#endif
@@ -130,7 +130,7 @@ TEST(Tensor, ShareDataWith) {
130130
{
131131
Tensor src_tensor;
132132
Tensor dst_tensor;
133-
src_tensor.mutable_data<int>(make_ddim({2, 3, 4}), GPUPlace());
133+
src_tensor.mutable_data<int>(make_ddim({2, 3, 4}), CUDAPlace());
134134
dst_tensor.ShareDataWith(src_tensor);
135135
ASSERT_EQ(src_tensor.data<int>(), dst_tensor.data<int>());
136136
}
@@ -166,7 +166,7 @@ TEST(Tensor, Slice) {
166166
#ifdef PADDLE_WITH_CUDA
167167
{
168168
Tensor src_tensor;
169-
src_tensor.mutable_data<double>(make_ddim({6, 9}), GPUPlace());
169+
src_tensor.mutable_data<double>(make_ddim({6, 9}), CUDAPlace());
170170
Tensor slice_tensor = src_tensor.Slice(2, 6);
171171
DDim slice_dims = slice_tensor.dims();
172172
ASSERT_EQ(arity(slice_dims), 2);
@@ -176,11 +176,11 @@ TEST(Tensor, Slice) {
176176
uintptr_t src_data_address =
177177
reinterpret_cast<uintptr_t>(src_tensor.data<double>());
178178
uintptr_t src_mutable_data_address = reinterpret_cast<uintptr_t>(
179-
src_tensor.mutable_data<double>(src_tensor.dims(), GPUPlace()));
179+
src_tensor.mutable_data<double>(src_tensor.dims(), CUDAPlace()));
180180
uintptr_t slice_data_address =
181181
reinterpret_cast<uintptr_t>(slice_tensor.data<double>());
182182
uintptr_t slice_mutable_data_address = reinterpret_cast<uintptr_t>(
183-
slice_tensor.mutable_data<double>(slice_tensor.dims(), GPUPlace()));
183+
slice_tensor.mutable_data<double>(slice_tensor.dims(), CUDAPlace()));
184184
EXPECT_EQ(src_data_address, src_mutable_data_address);
185185
EXPECT_EQ(slice_data_address, slice_mutable_data_address);
186186
EXPECT_EQ(src_data_address + 9 * 2 * sizeof(double), slice_data_address);

paddle/framework/tensor_util.h

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -47,33 +47,33 @@ inline void CopyFrom(const Tensor& src, const platform::Place& dst_place,
4747
#ifdef PADDLE_WITH_CUDA
4848
else if (platform::is_gpu_place(src_place) && // NOLINT
4949
platform::is_cpu_place(dst_place)) {
50-
auto src_gpu_place = boost::get<platform::GPUPlace>(src_place);
50+
auto src_gpu_place = boost::get<platform::CUDAPlace>(src_place);
5151
auto dst_cpu_place = boost::get<platform::CPUPlace>(dst_place);
5252
auto ctx_place = ctx.GetPlace();
5353
PADDLE_ENFORCE(platform::is_gpu_place(ctx_place));
54-
auto ctx_gpu_place = boost::get<platform::GPUPlace>(ctx_place);
54+
auto ctx_gpu_place = boost::get<platform::CUDAPlace>(ctx_place);
5555
PADDLE_ENFORCE_EQ(src_gpu_place, ctx_gpu_place);
5656
memory::Copy(
5757
dst_cpu_place, dst_ptr, src_gpu_place, src_ptr, size,
5858
reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
5959
} else if (platform::is_cpu_place(src_place) &&
6060
platform::is_gpu_place(dst_place)) {
6161
auto src_cpu_place = boost::get<platform::CPUPlace>(src_place);
62-
auto dst_gpu_place = boost::get<platform::GPUPlace>(dst_place);
62+
auto dst_gpu_place = boost::get<platform::CUDAPlace>(dst_place);
6363
auto ctx_place = ctx.GetPlace();
6464
PADDLE_ENFORCE(platform::is_gpu_place(ctx_place));
65-
auto ctx_gpu_place = boost::get<platform::GPUPlace>(ctx_place);
65+
auto ctx_gpu_place = boost::get<platform::CUDAPlace>(ctx_place);
6666
PADDLE_ENFORCE_EQ(dst_gpu_place, ctx_gpu_place);
6767
memory::Copy(
6868
dst_gpu_place, dst_ptr, src_cpu_place, src_ptr, size,
6969
reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
7070
} else if (platform::is_gpu_place(src_place) &&
7171
platform::is_gpu_place(dst_place)) {
72-
auto src_gpu_place = boost::get<platform::GPUPlace>(src_place);
73-
auto dst_gpu_place = boost::get<platform::GPUPlace>(dst_place);
72+
auto src_gpu_place = boost::get<platform::CUDAPlace>(src_place);
73+
auto dst_gpu_place = boost::get<platform::CUDAPlace>(dst_place);
7474
auto ctx_place = ctx.GetPlace();
7575
PADDLE_ENFORCE(platform::is_gpu_place(ctx_place));
76-
auto ctx_gpu_place = boost::get<platform::GPUPlace>(ctx_place);
76+
auto ctx_gpu_place = boost::get<platform::CUDAPlace>(ctx_place);
7777
PADDLE_ENFORCE_EQ(src_gpu_place, ctx_gpu_place);
7878
memory::Copy(
7979
dst_gpu_place, dst_ptr, src_gpu_place, src_ptr, size,
@@ -108,7 +108,7 @@ inline void CopyFromVector(const std::vector<T>& src,
108108
#ifdef PADDLE_WITH_CUDA
109109
else if (platform::is_gpu_place(dst_place)) { // NOLINT
110110
memory::Copy(
111-
boost::get<platform::GPUPlace>(dst_place), dst_ptr, src_place, src_ptr,
111+
boost::get<platform::CUDAPlace>(dst_place), dst_ptr, src_place, src_ptr,
112112
size,
113113
reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
114114
}
@@ -141,7 +141,7 @@ inline void CopyToVector(const Tensor& src, const platform::DeviceContext& ctx,
141141
#ifdef PADDLE_WITH_CUDA
142142
else if (platform::is_gpu_place(src.place())) { // NOLINT
143143
memory::Copy(
144-
dst_place, dst_ptr, boost::get<platform::GPUPlace>(src.place()),
144+
dst_place, dst_ptr, boost::get<platform::CUDAPlace>(src.place()),
145145
src_ptr, size,
146146
reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
147147
}

paddle/framework/tensor_util_test.cc

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ TEST(CopyFrom, Tensor) {
5858
memcpy(src_ptr, arr, 9 * sizeof(int));
5959

6060
// CPU Tensor to GPU Tensor
61-
auto gpu_place = new platform::GPUPlace(0);
61+
auto gpu_place = new platform::CUDAPlace(0);
6262
platform::CUDADeviceContext gpu_ctx(*gpu_place);
6363
CopyFrom(src_tensor, *gpu_place, gpu_ctx, &gpu_tensor);
6464

@@ -143,7 +143,7 @@ TEST(CopyFromVector, Tensor) {
143143

144144
// Copy to GPUTensor
145145
gpu_tensor.Resize(make_ddim({3, 3}));
146-
auto gpu_place = new paddle::platform::GPUPlace();
146+
auto gpu_place = new paddle::platform::CUDAPlace();
147147
CUDADeviceContext gpu_ctx(*gpu_place);
148148
CopyFromVector<int>(src_vec, gpu_ctx, &gpu_tensor);
149149
// Copy from GPU to CPU tensor for comparison
@@ -210,7 +210,7 @@ TEST(CopyToVector, Tensor) {
210210
{
211211
std::vector<int> src_vec = {1, 2, 3, 4, 5, 6, 7, 8, 9};
212212
Tensor gpu_tensor;
213-
GPUPlace place;
213+
CUDAPlace place;
214214
CUDADeviceContext gpu_ctx(place);
215215
CopyFromVector<int>(src_vec, gpu_ctx, &gpu_tensor);
216216

0 commit comments

Comments
 (0)