From 490557c4067d376911d616ea80c3e70351f219f8 Mon Sep 17 00:00:00 2001 From: co63oc Date: Thu, 31 Oct 2024 11:49:25 +0800 Subject: [PATCH] remove use_calc_stream param [fluid_ops] c_split (#68969) * Fix * Fix * test position --- paddle/fluid/operators/collective/c_split_op.cc | 4 ---- paddle/fluid/operators/ops_signature/c_split_sig.cc | 5 +---- paddle/phi/kernels/c_split_kernel.h | 2 -- paddle/phi/kernels/cpu/c_split_kernel.cc | 2 -- paddle/phi/kernels/gpu/c_split_kernel.cu | 2 -- paddle/phi/kernels/xpu/c_split_kernel.cc | 2 -- paddle/phi/ops/yaml/inconsistent/static_ops.yaml | 3 ++- paddle/phi/ops/yaml/op_version.yaml | 8 ++++++++ 8 files changed, 11 insertions(+), 17 deletions(-) diff --git a/paddle/fluid/operators/collective/c_split_op.cc b/paddle/fluid/operators/collective/c_split_op.cc index c6b9704fe740db..68d5a59223a2b5 100644 --- a/paddle/fluid/operators/collective/c_split_op.cc +++ b/paddle/fluid/operators/collective/c_split_op.cc @@ -103,10 +103,6 @@ class CSplitOpMaker : public framework::OpProtoAndCheckerMaker { AddAttr("rank", "(int default 0) rank id.").SetDefault(0); AddAttr("nranks", "(int default 1) number of ranks.").SetDefault(1); AddAttr("ring_id", "(int default 0) ring id.").SetDefault(0); - AddAttr( - "use_calc_stream", - "(bool default false) eject CUDA operations to calculation stream.") - .SetDefault(false); AddAttr("use_model_parallel", "(bool default false) use this op with model parallel.") .SetDefault(true); diff --git a/paddle/fluid/operators/ops_signature/c_split_sig.cc b/paddle/fluid/operators/ops_signature/c_split_sig.cc index 53cd79c755ab58..fa5e40ecbb5c66 100644 --- a/paddle/fluid/operators/ops_signature/c_split_sig.cc +++ b/paddle/fluid/operators/ops_signature/c_split_sig.cc @@ -19,10 +19,7 @@ namespace phi { KernelSignature CSplitOpArgumentMapping( const ArgumentMappingContext& ctx UNUSED) { return KernelSignature( - "c_split", - {"X"}, - {"rank", "nranks", "ring_id", "use_calc_stream", "use_model_parallel"}, - {"Out"}); + "c_split", {"X"}, {"rank", "nranks", "use_model_parallel"}, {"Out"}); } } // namespace phi diff --git a/paddle/phi/kernels/c_split_kernel.h b/paddle/phi/kernels/c_split_kernel.h index 6ec945e123a814..850a31615a4f3e 100644 --- a/paddle/phi/kernels/c_split_kernel.h +++ b/paddle/phi/kernels/c_split_kernel.h @@ -23,8 +23,6 @@ void CSplitKernel(const Context& ctx, const DenseTensor& x, int rank, int nranks, - int ring_id, - bool use_calc_stream, bool use_model_parallel, DenseTensor* out); diff --git a/paddle/phi/kernels/cpu/c_split_kernel.cc b/paddle/phi/kernels/cpu/c_split_kernel.cc index 59fee201f9258e..1b7175a877d787 100644 --- a/paddle/phi/kernels/cpu/c_split_kernel.cc +++ b/paddle/phi/kernels/cpu/c_split_kernel.cc @@ -23,8 +23,6 @@ void CSplitKernel(const Context& ctx, const DenseTensor& x, int rank, int nranks, - int ring_id, - bool use_calc_stream, bool use_model_parallel, DenseTensor* out) { PADDLE_THROW(common::errors::Unavailable( diff --git a/paddle/phi/kernels/gpu/c_split_kernel.cu b/paddle/phi/kernels/gpu/c_split_kernel.cu index a178bfb77f8542..b9796e784d04f9 100644 --- a/paddle/phi/kernels/gpu/c_split_kernel.cu +++ b/paddle/phi/kernels/gpu/c_split_kernel.cu @@ -54,8 +54,6 @@ void CSplitKernel(const Context& ctx, const DenseTensor& x, int rank, int nranks, - int ring_id, - bool use_calc_stream, bool use_model_parallel, DenseTensor* out) { auto place = ctx.GetPlace(); diff --git a/paddle/phi/kernels/xpu/c_split_kernel.cc b/paddle/phi/kernels/xpu/c_split_kernel.cc index 4bfb34fb28c938..49146ea7778573 100644 --- a/paddle/phi/kernels/xpu/c_split_kernel.cc +++ b/paddle/phi/kernels/xpu/c_split_kernel.cc @@ -24,8 +24,6 @@ void CSplitKernel(const Context& dev_ctx, const DenseTensor& x, int rank, int nranks, - int ring_id, - bool use_calc_stream, bool use_model_parallel, DenseTensor* out) { using XPUType = typename XPUTypeTrait::Type; diff --git a/paddle/phi/ops/yaml/inconsistent/static_ops.yaml b/paddle/phi/ops/yaml/inconsistent/static_ops.yaml index 51539c0d781ba6..5703fc213181c2 100644 --- a/paddle/phi/ops/yaml/inconsistent/static_ops.yaml +++ b/paddle/phi/ops/yaml/inconsistent/static_ops.yaml @@ -167,13 +167,14 @@ inplace : (x -> out) - op : c_split - args : (Tensor x, int rank = 0, int nranks = 1, int ring_id = 0, bool use_calc_stream = false, bool use_model_parallel = true) + args : (Tensor x, int rank = 0, int nranks = 1, int ring_id = 0, bool use_model_parallel = true) output : Tensor(out) infer_meta : func : CSplitInferMeta param : [x, nranks] kernel : func : c_split + param: [x, rank, nranks, use_model_parallel] - op : coalesce_tensor_ args : (Tensor[] input, DataType dtype, bool copy_data = false, bool set_constant = false, bool persist_output = false, float constant = 0.0, bool use_align = true, int align_size = -1, int size_of_dtype = -1, int64_t[] concated_shapes = {}, int64_t[] concated_ranks = {}) diff --git a/paddle/phi/ops/yaml/op_version.yaml b/paddle/phi/ops/yaml/op_version.yaml index 578e1c056daf10..29fb04ab9187de 100644 --- a/paddle/phi/ops/yaml/op_version.yaml +++ b/paddle/phi/ops/yaml/op_version.yaml @@ -89,6 +89,14 @@ - add_input : ValueTensor comment : In order to support multi-tag task. +- op : c_split + version : + - checkpoint : Upgrade c_split delete 1 attribute[use_calc_stream]. + action : + - delete_attr : use_calc_stream + comment : eject CUDA operations to calculation stream. + default : false + - op : clip version : - checkpoint : Upgrade clip add a new input [Min]