Skip to content

Commit

Permalink
remove mlu(#53007)
Browse files Browse the repository at this point in the history
  • Loading branch information
Liyulingyue authored Apr 18, 2023
1 parent 240e13a commit 4d5a3ad
Show file tree
Hide file tree
Showing 37 changed files with 11 additions and 114 deletions.
10 changes: 3 additions & 7 deletions cmake/operators.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,6 @@ function(op_library TARGET)
set(hip_cc_srcs)
set(xpu_cc_srcs)
set(xpu_kp_cc_srcs)
set(mlu_cc_srcs)
set(cudnn_cu_cc_srcs)
set(miopen_cu_cc_srcs)
set(cudnn_cu_srcs)
Expand Down Expand Up @@ -307,9 +306,8 @@ function(op_library TARGET)
# Unity Build relies on global option `WITH_UNITY_BUILD` and local option `UNITY`.
if(WITH_UNITY_BUILD AND op_library_UNITY)
# Combine the cc source files.
compose_unity_target_sources(
${UNITY_TARGET} cc ${cc_srcs} ${mkldnn_cc_srcs} ${xpu_cc_srcs}
${mlu_cc_srcs})
compose_unity_target_sources(${UNITY_TARGET} cc ${cc_srcs}
${mkldnn_cc_srcs} ${xpu_cc_srcs})
if(TARGET ${UNITY_TARGET})
# If `UNITY_TARGET` exists, add source files to `UNITY_TARGET`.
target_sources(${UNITY_TARGET} PRIVATE ${unity_target_cc_sources})
Expand All @@ -325,7 +323,7 @@ function(op_library TARGET)
else()
cc_library(
${TARGET}
SRCS ${cc_srcs} ${mkldnn_cc_srcs} ${xpu_cc_srcs} ${mlu_cc_srcs}
SRCS ${cc_srcs} ${mkldnn_cc_srcs} ${xpu_cc_srcs}
DEPS ${op_library_DEPS} ${op_common_deps})
endif()
endif()
Expand All @@ -337,7 +335,6 @@ function(op_library TARGET)
list(LENGTH mkldnn_cc_srcs mkldnn_cc_srcs_len)
list(LENGTH xpu_cc_srcs xpu_cc_srcs_len)
list(LENGTH miopen_cu_cc_srcs miopen_cu_cc_srcs_len)
list(LENGTH mlu_cc_srcs mlu_cc_srcs_len)

# Define operators that don't need pybind here.
foreach(
Expand Down Expand Up @@ -562,7 +559,6 @@ function(register_operators)
"*_op.cc")
string(REPLACE "_mkldnn" "" OPS "${OPS}")
string(REPLACE "_xpu" "" OPS "${OPS}")
string(REPLACE "_mlu" "" OPS "${OPS}")
string(REPLACE ".cc" "" OPS "${OPS}")
list(REMOVE_DUPLICATES OPS)
list(LENGTH register_operators_DEPS register_operators_DEPS_len)
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/eager/amp_auto_cast.h
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,6 @@ static inline bool NeedCast(const paddle::Tensor& tensor,
if (paddle::platform::is_gpu_place(place) ||
paddle::platform::is_cuda_pinned_place(place) ||
paddle::platform::is_xpu_place(place) ||
paddle::platform::is_mlu_place(place) ||
paddle::platform::is_npu_place(place) ||
paddle::platform::is_npu_pinned_place(place) ||
paddle::platform::is_custom_place(place)) {
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/eager/amp_utils.h
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,6 @@ inline phi::DataType GetDtypeWithPlace(
is_right_place = (paddle::platform::is_gpu_place(place) ||
paddle::platform::is_cuda_pinned_place(place) ||
paddle::platform::is_xpu_place(place) ||
paddle::platform::is_mlu_place(place) ||
paddle::platform::is_npu_place(place) ||
paddle::platform::is_npu_pinned_place(place) ||
paddle::platform::is_custom_place(place));
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/eager/eager_amp_auto_cast.h
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,6 @@ static inline bool NeedCast(const paddle::Tensor& tensor,
if (paddle::platform::is_gpu_place(place) ||
paddle::platform::is_cuda_pinned_place(place) ||
paddle::platform::is_xpu_place(place) ||
paddle::platform::is_mlu_place(place) ||
paddle::platform::is_npu_place(place) ||
paddle::platform::is_npu_pinned_place(place) ||
paddle::platform::is_custom_place(place) ||
Expand Down
2 changes: 0 additions & 2 deletions paddle/fluid/framework/library_type.h
Original file line number Diff line number Diff line change
Expand Up @@ -72,8 +72,6 @@ inline LibraryType StringToLibraryType(const char* ctype) {
return LibraryType::kPlain;
} else if (s == std::string("CUDA")) {
return LibraryType::kPlain;
} else if (s == std::string("MLU")) {
return LibraryType::kPlain;
} else {
PADDLE_THROW(platform::errors::Unimplemented(
"Unknown LibraryType string (%s), only support library type string "
Expand Down
3 changes: 0 additions & 3 deletions paddle/fluid/framework/tensor_util.cc
Original file line number Diff line number Diff line change
Expand Up @@ -276,7 +276,6 @@ void TensorCopyImpl(const TENSOR& src,
platform::DeviceContextPool& pool = platform::DeviceContextPool::Instance();
const platform::DeviceContext* dev_ctx;
if (platform::is_gpu_place(dst_place) || platform::is_npu_place(dst_place) ||
platform::is_mlu_place(dst_place) ||
platform::is_custom_place(dst_place)) {
dev_ctx = pool.Get(dst_place);
} else {
Expand Down Expand Up @@ -615,7 +614,6 @@ void TensorFromStream(std::istream& is,
size_t size = tensor->numel() * framework::SizeOfType(desc.data_type());
if (platform::is_gpu_place(dev_ctx.GetPlace()) ||
platform::is_xpu_place(dev_ctx.GetPlace()) ||
platform::is_mlu_place(dev_ctx.GetPlace()) ||
platform::is_npu_place(dev_ctx.GetPlace()) ||
platform::is_custom_place(dev_ctx.GetPlace())) {
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) || \
Expand Down Expand Up @@ -691,7 +689,6 @@ void TensorFromStream(std::istream& is,
size_t size = tensor->numel() * framework::SizeOfType(desc.data_type());
if (platform::is_gpu_place(dev_ctx.GetPlace()) ||
platform::is_xpu_place(dev_ctx.GetPlace()) ||
platform::is_mlu_place(dev_ctx.GetPlace()) ||
platform::is_npu_place(dev_ctx.GetPlace()) ||
platform::is_custom_place(dev_ctx.GetPlace())) {
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) || \
Expand Down
3 changes: 1 addition & 2 deletions paddle/fluid/imperative/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -166,5 +166,4 @@ cc_library(
var_type_traits
layer
math_function
phi_tensor
${MLU_DEPS})
phi_tensor)
2 changes: 0 additions & 2 deletions paddle/fluid/imperative/amp_auto_cast.cc
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,6 @@ OpSupportedInfos(const std::string& place,
{"CPU", &platform::is_cpu_place},
{"XPU", &platform::is_xpu_place},
{"NPU", &platform::is_npu_place},
{"MLU", &platform::is_mlu_place},
};
PADDLE_ENFORCE_NE(is_target_place.count(query_place),
0,
Expand Down Expand Up @@ -245,7 +244,6 @@ inline bool NeedCast(const std::shared_ptr<VarType>& var) {
if (paddle::platform::is_gpu_place(place) ||
paddle::platform::is_cuda_pinned_place(place) ||
paddle::platform::is_xpu_place(place) ||
paddle::platform::is_mlu_place(place) ||
paddle::platform::is_custom_place(place) ||
paddle::platform::is_npu_place(place) ||
paddle::platform::is_npu_pinned_place(place)) {
Expand Down
3 changes: 0 additions & 3 deletions paddle/fluid/imperative/reducer.cc
Original file line number Diff line number Diff line change
Expand Up @@ -66,9 +66,6 @@ void Group::DivNRanks(const platform::DeviceContext &context, int64_t nranks) {
#ifdef PADDLE_WITH_XPU_BKCL
// TODO(liuyuhui) support xpu about div nranks in the future
#endif
} else if (platform::is_mlu_place(tensor->place())) {
// TODO(zhangna)
VLOG(4) << "divnrank for mlu not support yet";
}
}

Expand Down
3 changes: 1 addition & 2 deletions paddle/fluid/operators/collective/c_comm_init_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -61,8 +61,7 @@ class CCommInitOp : public framework::OperatorBase {
#endif

PADDLE_ENFORCE_EQ(
platform::is_gpu_place(place) || platform::is_xpu_place(place) ||
platform::is_mlu_place(place),
platform::is_gpu_place(place) || platform::is_xpu_place(place),
true,
platform::errors::PreconditionNotMet(
"CCommInitOp can run on gpu or xpu or mlu place only."));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,6 @@ phi::KernelKey GetReduceExpectedKernelType(
PADDLE_ENFORCE_EQ(
platform::is_gpu_place(ctx.GetPlace()) ||
platform::is_npu_place(ctx.GetPlace()) ||
platform::is_mlu_place(ctx.GetPlace()) ||
platform::is_xpu_place(ctx.GetPlace()) ||
platform::is_custom_place(ctx.GetPlace()),
true,
Expand Down
6 changes: 2 additions & 4 deletions paddle/fluid/operators/interpolate_op.h
Original file line number Diff line number Diff line change
Expand Up @@ -41,8 +41,7 @@ inline std::vector<int> get_new_shape(
"The shape of dimension tensor should be [1],"
"but received d%.",
tensor->dims()));
if (platform::is_gpu_place(tensor->place()) ||
platform::is_mlu_place(tensor->place())) {
if (platform::is_gpu_place(tensor->place())) {
phi::DenseTensor temp;
paddle::framework::TensorCopySync(*tensor, platform::CPUPlace(), &temp);
vec_new_shape.push_back(static_cast<int32_t>(*temp.data<int32_t>()));
Expand All @@ -60,8 +59,7 @@ inline std::vector<T> get_new_data_from_tensor(
std::vector<T> vec_new_data;
auto* new_data = new_data_tensor->data<T>();
phi::DenseTensor cpu_starts_tensor;
if (platform::is_gpu_place(new_data_tensor->place()) ||
platform::is_mlu_place(new_data_tensor->place())) {
if (platform::is_gpu_place(new_data_tensor->place())) {
paddle::framework::TensorCopySync(
*new_data_tensor, platform::CPUPlace(), &cpu_starts_tensor);
new_data = cpu_starts_tensor.data<T>();
Expand Down
3 changes: 0 additions & 3 deletions paddle/fluid/operators/reader/buffered_reader.cc
Original file line number Diff line number Diff line change
Expand Up @@ -99,7 +99,6 @@ BufferedReader::BufferedReader(
cpu_buffer_.resize(buffer_size);
cuda_buffer_.resize(buffer_size);
npu_buffer_.resize(buffer_size);
mlu_buffer_.resize(buffer_size);
xpu_buffer_.resize(buffer_size);
custom_device_buffer_.resize(buffer_size);
ReadTillBufferFullAsync();
Expand Down Expand Up @@ -387,8 +386,6 @@ void BufferedReader::ReadNextImpl(paddle::framework::LoDTensorArray *out) {
*out = std::move(cuda_buffer_[i]);
} else if (platform::is_npu_place(place_)) {
*out = std::move(npu_buffer_[i]);
} else if (platform::is_mlu_place(place_)) {
*out = std::move(mlu_buffer_[i]);
} else if (platform::is_xpu_place(place_)) {
*out = std::move(xpu_buffer_[i]);
} else if (platform::is_custom_place(place_)) {
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/operators/reader/buffered_reader.h
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,6 @@ class BufferedReader : public framework::DecoratedReader {
std::vector<TensorVec> cpu_buffer_;
std::vector<TensorVec> cuda_buffer_;
std::vector<TensorVec> npu_buffer_;
std::vector<TensorVec> mlu_buffer_;
std::vector<TensorVec> xpu_buffer_;
std::vector<TensorVec> custom_device_buffer_;
size_t prev_pos_{-1UL};
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/operators/reduce_ops/reduce_op.h
Original file line number Diff line number Diff line change
Expand Up @@ -625,7 +625,6 @@ class ReduceBaseOp : public framework::OperatorWithKernel {
PADDLE_ENFORCE_EQ(
platform::is_gpu_place(ctx.GetPlace()) ||
platform::is_npu_place(ctx.GetPlace()) ||
platform::is_mlu_place(ctx.GetPlace()) ||
platform::is_xpu_place(ctx.GetPlace()) ||
platform::is_custom_place(ctx.GetPlace()),
true,
Expand Down
2 changes: 0 additions & 2 deletions paddle/fluid/operators/softmax_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,6 @@ class SoftmaxOp : public framework::OperatorWithKernel {
platform::is_gpu_place(ctx.GetPlace()) ||
platform::is_npu_place(ctx.GetPlace()) ||
platform::is_xpu_place(ctx.GetPlace()) ||
platform::is_mlu_place(ctx.GetPlace()) ||
platform::is_custom_place(ctx.GetPlace()),
true,
platform::errors::InvalidArgument(
Expand Down Expand Up @@ -131,7 +130,6 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
if (!(platform::is_gpu_place(ctx.GetPlace()) ||
platform::is_npu_place(ctx.GetPlace()) ||
platform::is_xpu_place(ctx.GetPlace()) ||
platform::is_mlu_place(ctx.GetPlace()) ||
platform::is_custom_place(ctx.GetPlace())))
PADDLE_THROW(platform::errors::InvalidArgument(
"float16 can only be used on GPU/NPU/XPU/MLU and custom place"));
Expand Down
2 changes: 0 additions & 2 deletions paddle/fluid/operators/tile_op_functor.h
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,6 @@ inline std::vector<int> get_repeat_times(
phi::DenseTensor cpu_repeat_tensor;
if (platform::is_gpu_place(repeat_tensor->place()) ||
platform::is_xpu_place(repeat_tensor->place()) ||
platform::is_mlu_place(repeat_tensor->place()) ||
platform::is_npu_place(repeat_tensor->place())) {
paddle::framework::TensorCopySync(
*repeat_tensor, platform::CPUPlace(), &cpu_repeat_tensor);
Expand All @@ -50,7 +49,6 @@ inline std::vector<int> get_repeat_times(
auto tensor = list_repeat_times_tensor[i];
if (platform::is_gpu_place(tensor->place()) ||
platform::is_xpu_place(tensor->place()) ||
platform::is_mlu_place(tensor->place()) ||
platform::is_npu_place(tensor->place())) {
phi::DenseTensor temp;
paddle::framework::TensorCopySync(*tensor, platform::CPUPlace(), &temp);
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/platform/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,6 @@ cc_library(
dlpack
cudnn_workspace_helper
${XPU_CTX_DEPS}
${MLU_CTX_DEPS}
phi_backends
phi_device_context
generator
Expand Down
2 changes: 0 additions & 2 deletions paddle/fluid/platform/device_context.cc
Original file line number Diff line number Diff line change
Expand Up @@ -47,8 +47,6 @@ DeviceType Place2DeviceType(const platform::Place& place) {
return platform::DeviceType::IPU;
} else if (platform::is_npu_place(place)) {
return platform::DeviceType::NPU;
} else if (platform::is_mlu_place(place)) {
return platform::DeviceType::MLU;
} else if (platform::is_custom_place(place)) {
return platform::DeviceType::CUSTOM_DEVICE;
} else {
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/platform/device_context.h
Original file line number Diff line number Diff line change
Expand Up @@ -109,7 +109,6 @@ constexpr DeviceType kCUDA = DeviceType::CUDA;
constexpr DeviceType kXPU = DeviceType::XPU;
constexpr DeviceType kNPU = DeviceType::NPU;
constexpr DeviceType kIPU = DeviceType::IPU;
constexpr DeviceType kMLU = DeviceType::MLU;
constexpr DeviceType kCUSTOM_DEVICE = DeviceType::CUSTOM_DEVICE;

using DeviceContext = phi::DeviceContext;
Expand Down
18 changes: 0 additions & 18 deletions paddle/fluid/platform/monitor.cc
Original file line number Diff line number Diff line change
Expand Up @@ -46,21 +46,3 @@ DEFINE_INT_STATUS(STAT_npu4_mem_size)
DEFINE_INT_STATUS(STAT_npu5_mem_size)
DEFINE_INT_STATUS(STAT_npu6_mem_size)
DEFINE_INT_STATUS(STAT_npu7_mem_size)

// For MLU
DEFINE_INT_STATUS(STAT_mlu0_mem_size)
DEFINE_INT_STATUS(STAT_mlu1_mem_size)
DEFINE_INT_STATUS(STAT_mlu2_mem_size)
DEFINE_INT_STATUS(STAT_mlu3_mem_size)
DEFINE_INT_STATUS(STAT_mlu4_mem_size)
DEFINE_INT_STATUS(STAT_mlu5_mem_size)
DEFINE_INT_STATUS(STAT_mlu6_mem_size)
DEFINE_INT_STATUS(STAT_mlu7_mem_size)
DEFINE_INT_STATUS(STAT_mlu8_mem_size)
DEFINE_INT_STATUS(STAT_mlu9_mem_size)
DEFINE_INT_STATUS(STAT_mlu10_mem_size)
DEFINE_INT_STATUS(STAT_mlu11_mem_size)
DEFINE_INT_STATUS(STAT_mlu12_mem_size)
DEFINE_INT_STATUS(STAT_mlu13_mem_size)
DEFINE_INT_STATUS(STAT_mlu14_mem_size)
DEFINE_INT_STATUS(STAT_mlu15_mem_size)
18 changes: 0 additions & 18 deletions paddle/fluid/platform/monitor.h
Original file line number Diff line number Diff line change
Expand Up @@ -200,21 +200,3 @@ class StatRegistry {
USE_INT_STAT(STAT_npu5_mem_size); \
USE_INT_STAT(STAT_npu6_mem_size); \
USE_INT_STAT(STAT_npu7_mem_size)

#define USE_MLU_MEM_STAT \
USE_INT_STAT(STAT_mlu0_mem_size); \
USE_INT_STAT(STAT_mlu1_mem_size); \
USE_INT_STAT(STAT_mlu2_mem_size); \
USE_INT_STAT(STAT_mlu3_mem_size); \
USE_INT_STAT(STAT_mlu4_mem_size); \
USE_INT_STAT(STAT_mlu5_mem_size); \
USE_INT_STAT(STAT_mlu6_mem_size); \
USE_INT_STAT(STAT_mlu7_mem_size); \
USE_INT_STAT(STAT_mlu8_mem_size); \
USE_INT_STAT(STAT_mlu9_mem_size); \
USE_INT_STAT(STAT_mlu10_mem_size); \
USE_INT_STAT(STAT_mlu11_mem_size); \
USE_INT_STAT(STAT_mlu12_mem_size); \
USE_INT_STAT(STAT_mlu13_mem_size); \
USE_INT_STAT(STAT_mlu14_mem_size); \
USE_INT_STAT(STAT_mlu15_mem_size)
6 changes: 0 additions & 6 deletions paddle/fluid/platform/place.cc
Original file line number Diff line number Diff line change
Expand Up @@ -33,10 +33,6 @@ bool is_xpu_place(const Place &p) {
return p.GetType() == phi::AllocationType::XPU;
}

bool is_mlu_place(const Place &p) {
return p.GetType() == phi::AllocationType::MLU;
}

bool is_npu_place(const Place &p) {
return p.GetType() == phi::AllocationType::NPU;
}
Expand Down Expand Up @@ -77,8 +73,6 @@ bool is_same_place(const Place &p1, const Place &p2) {
return true;
} else if (is_xpu_place(p1)) {
return p1 == p2;
} else if (is_mlu_place(p1)) {
return p1 == p2;
} else if (is_npu_place(p1)) {
return p1 == p2;
} else if (is_ipu_place(p1)) {
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/platform/place.h
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ class PlaceHelper {
bool is_gpu_place(const Place &);
bool is_xpu_place(const Place &);
bool is_npu_place(const Place &);
bool is_mlu_place(const Place &);
bool is_ipu_place(const Place &);
bool is_cpu_place(const Place &);
bool is_cuda_pinned_place(const Place &);
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/platform/profiler/chrometracing_logger.cc
Original file line number Diff line number Diff line change
Expand Up @@ -257,7 +257,6 @@ void ChromeTracingLogger::LogHostTraceEventNode(
case TracerEventType::UserDefined:
case TracerEventType::OperatorInner:
case TracerEventType::Communication:
case TracerEventType::MluRuntime:
case TracerEventType::NumTypes:
default:
output_file_stream_ << string_format(
Expand Down
1 change: 0 additions & 1 deletion paddle/fluid/platform/profiler/profiler.h
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,6 @@ namespace platform {

static constexpr uint32_t kProfileCPUOptionBit = 0;
static constexpr uint32_t kProfileGPUOptionBit = 1;
static constexpr uint32_t kProfileMLUOptionBit = 2;
static constexpr uint32_t kProfileCustomDeviceOptionBit = 3;

void SynchronizeDevice();
Expand Down
3 changes: 1 addition & 2 deletions paddle/fluid/platform/profiler/utils.cc
Original file line number Diff line number Diff line change
Expand Up @@ -165,8 +165,7 @@ const char* StringTracerEventType(TracerEventType type) {
"Optimization",
"Communication",
"PythonOp",
"PythonUserDefined",
"MluRuntime"};
"PythonUserDefined"};
return categary_name_[static_cast<int>(type)];
}

Expand Down
3 changes: 0 additions & 3 deletions paddle/fluid/pybind/place.cc
Original file line number Diff line number Diff line change
Expand Up @@ -656,16 +656,13 @@ void BindPlace(pybind11::module &m) { // NOLINT
[](platform::Place &self) {
return platform::is_cuda_pinned_place(self);
})
.def("is_mlu_place",
[](platform::Place &self) { return platform::is_mlu_place(self); })
.def(
"is_custom_place",
[](platform::Place &self) { return platform::is_custom_place(self); })
.def("gpu_device_id", [](platform::Place &self) { return self.device; })
.def("xpu_device_id", [](platform::Place &self) { return self.device; })
.def("npu_device_id", [](platform::Place &self) { return self.device; })
.def("ipu_device_id", [](platform::Place &self) { return self.device; })
.def("mlu_device_id", [](platform::Place &self) { return self.device; })
.def("custom_device_id",
[](platform::Place &self) { return self.device; })
.def("set_place",
Expand Down
2 changes: 1 addition & 1 deletion paddle/fluid/pybind/tensor.cc
Original file line number Diff line number Diff line change
Expand Up @@ -401,7 +401,7 @@ void BindTensor(pybind11::module &m) { // NOLINT
Args:
lod (numpy.ndarray): The data to set.
place (CPUPlace|CUDAPlace|XPUPlace|IPUPlace|CUDAPinnedPlace|NPUPlace|MLUPlace): The place where the
place (CPUPlace|CUDAPlace|XPUPlace|IPUPlace|CUDAPinnedPlace|NPUPlace): The place where the
Tensor is to be set.
zero_copy (bool, optional): Whether to share memory with the input numpy array.
This parameter only works with CPUPlace. Default: False.
Expand Down
Loading

0 comments on commit 4d5a3ad

Please sign in to comment.