From 6a4e2d5cffb7906c100920d1a5d21dfdf33787a3 Mon Sep 17 00:00:00 2001 From: co63oc <4617245+co63oc@users.noreply.github.com> Date: Sun, 4 Jan 2026 10:07:27 +0800 Subject: [PATCH] add is_gpu_pinned_place --- paddle/fluid/pybind/place.cc | 2 ++ paddle/fluid/pybind/tensor.cc | 4 ++-- paddle/phi/common/place.cc | 5 ++++- paddle/phi/common/place.h | 4 +++- paddle/phi/core/platform/profiler.cc | 8 ++++---- 5 files changed, 15 insertions(+), 8 deletions(-) diff --git a/paddle/fluid/pybind/place.cc b/paddle/fluid/pybind/place.cc index bd9722a750e083..a77f51ac0137d0 100644 --- a/paddle/fluid/pybind/place.cc +++ b/paddle/fluid/pybind/place.cc @@ -223,6 +223,8 @@ void BindPlace(pybind11::module &m) { // NOLINT [](phi::Place &self) { return phi::is_ipu_place(self); }) .def("is_cuda_pinned_place", [](phi::Place &self) { return phi::is_cuda_pinned_place(self); }) + .def("is_gpu_pinned_place", + [](phi::Place &self) { return phi::is_gpu_pinned_place(self); }) .def("is_xpu_pinned_place", [](phi::Place &self) { return phi::is_xpu_pinned_place(self); }) .def("is_custom_place", diff --git a/paddle/fluid/pybind/tensor.cc b/paddle/fluid/pybind/tensor.cc index 9f838e7350975c..b99b45749efff1 100644 --- a/paddle/fluid/pybind/tensor.cc +++ b/paddle/fluid/pybind/tensor.cc @@ -1273,7 +1273,7 @@ void BindTensor(pybind11::module &m) { // NOLINT auto holder = self.Holder(); PADDLE_ENFORCE_EQ( phi::is_cpu_place(holder->place()) || - phi::is_cuda_pinned_place(holder->place()), + phi::is_gpu_pinned_place(holder->place()), true, common::errors::InvalidArgument( "Tensor is not on CPU. share_filename only " "support CPU Tensor.")); @@ -1309,7 +1309,7 @@ void BindTensor(pybind11::module &m) { // NOLINT handle, shared_fd, flags, data_size, find_id); // copy data & reset holder - if (phi::is_cuda_pinned_place(holder->place())) { + if (phi::is_gpu_pinned_place(holder->place())) { #ifdef PADDLE_WITH_CUDA memory::Copy(phi::CPUPlace(), shared_holder->ptr(), phi::GPUPinnedPlace(), data_ptr, data_size); diff --git a/paddle/phi/common/place.cc b/paddle/phi/common/place.cc index d290aecbe4abab..80ec6880a0c4cc 100644 --- a/paddle/phi/common/place.cc +++ b/paddle/phi/common/place.cc @@ -200,6 +200,9 @@ bool is_cuda_pinned_place(const Place &p) { return p.GetType() == phi::AllocationType::GPUPINNED; } +bool is_gpu_pinned_place(const Place &p) { + return p.GetType() == phi::AllocationType::GPUPINNED; +} bool is_xpu_pinned_place(const Place &p) { return p.GetType() == phi::AllocationType::XPUPINNED; } @@ -230,7 +233,7 @@ bool places_are_same_class(const Place &p1, const Place &p2) { bool is_same_place(const Place &p1, const Place &p2) { if (places_are_same_class(p1, p2)) { - if (is_cpu_place(p1) || is_cuda_pinned_place(p1) || + if (is_cpu_place(p1) || is_gpu_pinned_place(p1) || is_xpu_pinned_place(p1)) { return true; } else { diff --git a/paddle/phi/common/place.h b/paddle/phi/common/place.h index 3437d49e3701f0..8637b64fe8562c 100644 --- a/paddle/phi/common/place.h +++ b/paddle/phi/common/place.h @@ -226,7 +226,9 @@ PADDLE_API bool is_xpu_place(const Place&); PADDLE_API bool is_ipu_place(const Place&); PADDLE_API bool is_cpu_place(const Place&); PADDLE_API bool is_pinned_place(const Place&); -PADDLE_API bool is_cuda_pinned_place(const Place&); +PADDLE_API bool is_cuda_pinned_place( + const Place&); // Deprecated, use is_gpu_pinned_place instead. +PADDLE_API bool is_gpu_pinned_place(const Place&); PADDLE_API bool is_xpu_pinned_place(const Place&); PADDLE_API bool is_custom_place(const Place& p); PADDLE_API bool is_accelerat_place(const Place& p); diff --git a/paddle/phi/core/platform/profiler.cc b/paddle/phi/core/platform/profiler.cc index 95bb0cb623a87f..ff4dd62bb4d476 100644 --- a/paddle/phi/core/platform/profiler.cc +++ b/paddle/phi/core/platform/profiler.cc @@ -99,7 +99,7 @@ RecordMemEvent::RecordMemEvent(const void *ptr, uint64_t peak_allocated = 0; uint64_t current_reserved = 0; // 0 means keep the same as before uint64_t peak_reserved = 0; // 0 means keep the same as before - if (phi::is_cpu_place(place) || phi::is_cuda_pinned_place(place)) { + if (phi::is_cpu_place(place) || phi::is_gpu_pinned_place(place)) { if (RecordMemEvent::has_initialized["cpu"][place.GetDeviceId()] == false) { RecordMemEvent::size_cache["cpu"][place.GetDeviceId()].push_back( @@ -182,7 +182,7 @@ RecordMemEvent::RecordMemEvent(const void *ptr, uint64_t peak_reserved = 0; uint64_t current_allocated = 0; // 0 means keep the same as before uint64_t peak_allocated = 0; // 0 means keep the same as before - if (phi::is_cpu_place(place) || phi::is_cuda_pinned_place(place)) { + if (phi::is_cpu_place(place) || phi::is_gpu_pinned_place(place)) { if (RecordMemEvent::has_initialized["cpu"][place.GetDeviceId()] == false) { RecordMemEvent::size_cache["cpu"][place.GetDeviceId()].push_back( @@ -264,7 +264,7 @@ RecordMemEvent::RecordMemEvent(const void *ptr, uint64_t peak_allocated = 0; uint64_t current_reserved = 0; // 0 means keep the same as before uint64_t peak_reserved = 0; // 0 means keep the same as before - if (phi::is_cpu_place(place) || phi::is_cuda_pinned_place(place)) { + if (phi::is_cpu_place(place) || phi::is_gpu_pinned_place(place)) { if (RecordMemEvent::has_initialized["cpu"][place.GetDeviceId()] == false) { RecordMemEvent::size_cache["cpu"][place.GetDeviceId()].push_back( @@ -346,7 +346,7 @@ RecordMemEvent::RecordMemEvent(const void *ptr, uint64_t peak_reserved = 0; uint64_t current_allocated = 0; // 0 means keep the same as before uint64_t peak_allocated = 0; // 0 means keep the same as before - if (phi::is_cpu_place(place) || phi::is_cuda_pinned_place(place)) { + if (phi::is_cpu_place(place) || phi::is_gpu_pinned_place(place)) { if (RecordMemEvent::has_initialized["cpu"][place.GetDeviceId()] == false) { RecordMemEvent::size_cache["cpu"][place.GetDeviceId()].push_back(