diff --git a/paddle/fluid/framework/operator.cc b/paddle/fluid/framework/operator.cc index 01d8a3771b100..fed4541ee9f2c 100644 --- a/paddle/fluid/framework/operator.cc +++ b/paddle/fluid/framework/operator.cc @@ -1146,7 +1146,7 @@ void OperatorWithKernel::RunImpl(const Scope& scope, if (FLAGS_run_pt_kernel && pten::KernelFactory::Instance().ContainsKernel(type_.c_str())) { if (pt_kernel_signature_.get() == nullptr || pt_kernel_.get() == nullptr) { - ChoosePtKernel(exe_ctx); + ChoosePtenKernel(exe_ctx); } run_pt_kernel_ = pt_kernel_->IsValid(); } @@ -1192,7 +1192,7 @@ void OperatorWithKernel::RunImpl(const Scope& scope, platform::RecordEvent record_event("compute", platform::EventRole::kInnerOp); if (run_pt_kernel_) { - auto op_kernel_ctx = BuildPtKernelContext(*runtime_ctx, *dev_ctx); + auto op_kernel_ctx = BuildPtenKernelContext(*runtime_ctx, *dev_ctx); (*pt_kernel_)(&op_kernel_ctx); } else { (*kernel_func_)( @@ -1282,26 +1282,26 @@ OpKernelType OperatorWithKernel::InnerGetExpectedKernelType( return expected_kernel_key; } -void OperatorWithKernel::ChoosePtKernel(const ExecutionContext& ctx) const { +void OperatorWithKernel::ChoosePtenKernel(const ExecutionContext& ctx) const { pt_kernel_signature_.reset( - new KernelSignature(this->GetExpectedPtKernelArgs(ctx))); + new KernelSignature(this->GetExpectedPtenKernelArgs(ctx))); VLOG(1) << KernelSignatureToString(*pt_kernel_signature_.get()); kernel_type_.reset(new OpKernelType(InnerGetExpectedKernelType(ctx))); auto pt_kernel_name = pten::KernelName(pt_kernel_signature_->first); - auto pt_kernel_key = TransOpKernelTypeToPtKernelKey(*kernel_type_.get()); + auto pt_kernel_key = TransOpKernelTypeToPtenKernelKey(*kernel_type_.get()); pt_kernel_.reset( new pten::Kernel(pten::KernelFactory::Instance().SelectKernel( pt_kernel_name, pt_kernel_key))); if (pt_kernel_->IsValid()) { - VLOG(1) << "Static mode ChoosePtKernel - kernel name: " << pt_kernel_name + VLOG(1) << "Static mode ChoosePtenKernel - kernel name: " << pt_kernel_name << " | kernel key: " << pt_kernel_key << " | kernel: " << *pt_kernel_; } else { - VLOG(1) << "Static mode ChoosePtKernel - kernel `" << pt_kernel_name + VLOG(1) << "Static mode ChoosePtenKernel - kernel `" << pt_kernel_name << "` not found."; } } @@ -1774,7 +1774,7 @@ OpKernelType OperatorWithKernel::GetKernelTypeForVar( tensor.layout()); } -KernelSignature OperatorWithKernel::GetExpectedPtKernelArgs( +KernelSignature OperatorWithKernel::GetExpectedPtenKernelArgs( const ExecutionContext& ctx) const { if (KernelSignatureMap::Instance().Has(Type())) { return *(KernelSignatureMap::Instance().GetNullable(Type())); @@ -1786,7 +1786,7 @@ KernelSignature OperatorWithKernel::GetExpectedPtKernelArgs( } } -pten::KernelContext OperatorWithKernel::BuildPtKernelContext( +pten::KernelContext OperatorWithKernel::BuildPtenKernelContext( const RuntimeContext& ctx, const platform::DeviceContext& dev_ctx) const { VLOG(1) << RuntimeContextDebugString(ctx); @@ -1834,7 +1834,7 @@ pten::KernelContext OperatorWithKernel::BuildPtKernelContext( std::vector> tmp_inputs; for (auto var : ins_vector) { - auto pt_in = framework::InputVariableToPtTensor(*var, in_def); + auto pt_in = framework::InputVariableToPtenTensor(*var, in_def); tmp_inputs.emplace_back(pt_in); } op_kernel_ctx.EmplaceBackInputs(tmp_inputs); @@ -1846,7 +1846,7 @@ pten::KernelContext OperatorWithKernel::BuildPtKernelContext( std::vector> tmp_outputs; for (auto var : outs_vector) { - auto pt_out = framework::OutputVariableToPtTensor(var, out_def); + auto pt_out = framework::OutputVariableToPtenTensor(var, out_def); tmp_outputs.emplace_back(pt_out); } op_kernel_ctx.EmplaceBackOutputs(tmp_outputs); diff --git a/paddle/fluid/framework/operator.h b/paddle/fluid/framework/operator.h index 29c60877b8116..224974001c469 100644 --- a/paddle/fluid/framework/operator.h +++ b/paddle/fluid/framework/operator.h @@ -537,9 +537,9 @@ class OperatorWithKernel : public OperatorBase { * output arguments registered in the original OpMaker do not match in some * cases, so we use map to record the arguments required by the kernel. * When selecting Kernel during Op execution, select the arguments of the - * original Op according to the GetExpectedPtKernelArgs returned arguments. + * original Op according to the GetExpectedPtenKernelArgs returned arguments. */ - virtual KernelSignature GetExpectedPtKernelArgs( + virtual KernelSignature GetExpectedPtenKernelArgs( const ExecutionContext& ctx) const; private: @@ -583,9 +583,9 @@ class OperatorWithKernel : public OperatorBase { const std::string& name) const; /* member functions for adapting to pten lib */ - void ChoosePtKernel(const ExecutionContext& ctx) const; + void ChoosePtenKernel(const ExecutionContext& ctx) const; - pten::KernelContext BuildPtKernelContext( + pten::KernelContext BuildPtenKernelContext( const RuntimeContext& ctx, const platform::DeviceContext& dev_ctx) const; protected: diff --git a/paddle/fluid/framework/pten_utils.cc b/paddle/fluid/framework/pten_utils.cc index fbe9a4759bbf1..e0e43db139065 100644 --- a/paddle/fluid/framework/pten_utils.cc +++ b/paddle/fluid/framework/pten_utils.cc @@ -66,8 +66,8 @@ std::shared_ptr MakeTensorImpl( const LoDTensor& tensor, const platform::Place& place, proto::VarType::Type type) { return MakeTensorImpl( - tensor, pten::TransToPtBackend(place), pten::TransToPtDataType(type), - pten::TransToPtDataLayout(tensor.layout())); + tensor, pten::TransToPtenBackend(place), pten::TransToPtenDataType(type), + pten::TransToPtenDataLayout(tensor.layout())); } template <> @@ -75,8 +75,8 @@ std::shared_ptr MakeTensorImpl( const Tensor& tensor, const platform::Place& place, proto::VarType::Type type) { return MakeTensorImpl( - tensor, pten::TransToPtBackend(place), pten::TransToPtDataType(type), - pten::TransToPtDataLayout(tensor.layout())); + tensor, pten::TransToPtenBackend(place), pten::TransToPtenDataType(type), + pten::TransToPtenDataLayout(tensor.layout())); } template <> @@ -93,7 +93,7 @@ void ShareTensorImpl(pten::DenseTensor* tensor_impl, pten::TransToProtoVarType(tensor_impl->data_type())); } -std::shared_ptr InputVariableToPtTensor( +std::shared_ptr InputVariableToPtenTensor( const framework::Variable& variable, const pten::TensorArgDef& arg_def) { auto expected_place = pten::TransToFluidPlace(arg_def.backend); @@ -138,7 +138,7 @@ std::shared_ptr InputVariableToPtTensor( return nullptr; } -std::shared_ptr OutputVariableToPtTensor( +std::shared_ptr OutputVariableToPtenTensor( framework::Variable* variable, const pten::TensorArgDef& arg_def) { // mutable_data before run kernel, to avoid share output form // KernelContext to original tensor @@ -170,7 +170,8 @@ std::shared_ptr OutputVariableToPtTensor( return nullptr; } -OpKernelType TransPtKernelKeyToOpKernelType(const pten::KernelKey& kernel_key) { +OpKernelType TransPtenKernelKeyToOpKernelType( + const pten::KernelKey& kernel_key) { proto::VarType::Type data_type = pten::TransToProtoVarType(kernel_key.dtype()); platform::Place place = pten::TransToFluidPlace(kernel_key.backend()); @@ -187,9 +188,9 @@ OpKernelType TransPtKernelKeyToOpKernelType(const pten::KernelKey& kernel_key) { return OpKernelType(data_type, place, data_layout, library_type); } -pten::KernelKey TransOpKernelTypeToPtKernelKey( +pten::KernelKey TransOpKernelTypeToPtenKernelKey( const OpKernelType& kernel_type) { - pten::Backend backend = pten::TransToPtBackend(kernel_type.place_); + pten::Backend backend = pten::TransToPtenBackend(kernel_type.place_); if (kernel_type.library_type_ == LibraryType::kMKLDNN) { backend = pten::Backend::MKLDNN; } else if (kernel_type.library_type_ == LibraryType::kCUDNN) { @@ -198,9 +199,9 @@ pten::KernelKey TransOpKernelTypeToPtKernelKey( // do } paddle::experimental::DataLayout layout = - pten::TransToPtDataLayout(kernel_type.data_layout_); + pten::TransToPtenDataLayout(kernel_type.data_layout_); paddle::experimental::DataType dtype = - pten::TransToPtDataType(kernel_type.data_type_); + pten::TransToPtenDataType(kernel_type.data_type_); return pten::KernelKey(backend, layout, dtype); } @@ -215,16 +216,17 @@ KernelArgsNameMakerByOpProto::GetInputArgsNames() { auto& in = op_proto_->inputs()[i]; auto& in_name = in.name(); if ((in.has_extra() && in.extra()) || (in.has_quant() && in.quant())) { - VLOG(1) << "Parse PtKernel input: skip extra & quant input - " << in_name; + VLOG(1) << "Parse PtenKernel input: skip extra & quant input - " + << in_name; continue; } // If contains dispensable input, we should override the - // GetExpectedPtKernelArgs method self + // GetExpectedPtenKernelArgs method self if (in.has_dispensable() && in.dispensable()) { - VLOG(1) << "Parse PtKernel input: skip dispensable input - " << in_name; + VLOG(1) << "Parse PtenKernel input: skip dispensable input - " << in_name; continue; } - VLOG(1) << "Parse PtKernel input: " << in_name; + VLOG(1) << "Parse PtenKernel input: " << in_name; input_names_.emplace_back(in_name); } return input_names_; @@ -236,7 +238,7 @@ KernelArgsNameMakerByOpProto::GetOutputArgsNames() { auto& out = op_proto_->outputs()[i]; auto& out_name = out.name(); // TODO(chenweihang): outputs also need skip some cases - VLOG(1) << "Parse PtKernel output: " << out_name; + VLOG(1) << "Parse PtenKernel output: " << out_name; output_names_.emplace_back(out_name); } return output_names_; @@ -250,16 +252,17 @@ KernelArgsNameMakerByOpProto::GetAttrsArgsNames() { if (attr_name == "use_mkldnn" || attr_name == "op_role" || attr_name == "op_role_var" || attr_name == "op_namescope" || attr_name == "op_callstack" || attr_name == "op_device") { - VLOG(1) << "Parse PtKernel attribute: skip needless attr - " << attr_name; + VLOG(1) << "Parse PtenKernel attribute: skip needless attr - " + << attr_name; continue; } if ((attr.has_extra() && attr.extra()) || (attr.has_quant() && attr.quant())) { - VLOG(1) << "Parse PtKernel attribute: skip extra & quant attr - " + VLOG(1) << "Parse PtenKernel attribute: skip extra & quant attr - " << attr_name; continue; } - VLOG(1) << "Parse PtKernel attribute: " << attr_name; + VLOG(1) << "Parse PtenKernel attribute: " << attr_name; attr_names_.emplace_back(attr_name); } diff --git a/paddle/fluid/framework/pten_utils.h b/paddle/fluid/framework/pten_utils.h index 14dbe933195be..263101657ceb9 100644 --- a/paddle/fluid/framework/pten_utils.h +++ b/paddle/fluid/framework/pten_utils.h @@ -33,37 +33,39 @@ namespace framework { /* tensor translate */ -template -std::shared_ptr MakeTensorImpl( +template +std::shared_ptr MakeTensorImpl( const VariableT& tensor, pten::Backend backend, paddle::experimental::DataType dtype, paddle::experimental::DataLayout layout); -template -std::shared_ptr MakeTensorImpl(const LoDTensor& tensor, - const platform::Place& place, - proto::VarType::Type type); +template +std::shared_ptr MakeTensorImpl(const LoDTensor& tensor, + const platform::Place& place, + proto::VarType::Type type); -template -std::shared_ptr MakeTensorImpl(const Tensor& tensor, - const platform::Place& place, - proto::VarType::Type type); +template +std::shared_ptr MakeTensorImpl(const Tensor& tensor, + const platform::Place& place, + proto::VarType::Type type); -template -void ShareTensorImpl(PtTensorImplT* tensor_impl, LoDTensor* out); +template +void ShareTensorImpl(PtenTensorImplT* tensor_impl, LoDTensor* out); -template -void ShareTensorImpl(PtTensorImplT* tensor_impl, Tensor* out); +template +void ShareTensorImpl(PtenTensorImplT* tensor_impl, Tensor* out); -std::shared_ptr InputVariableToPtTensor( +std::shared_ptr InputVariableToPtenTensor( const framework::Variable& variable, const pten::TensorArgDef& arg_def); -std::shared_ptr OutputVariableToPtTensor( +std::shared_ptr OutputVariableToPtenTensor( framework::Variable* variable, const pten::TensorArgDef& arg_def); /* Kernel Key translate */ -OpKernelType TransPtKernelKeyToOpKernelType(const pten::KernelKey& kernel_key); -pten::KernelKey TransOpKernelTypeToPtKernelKey(const OpKernelType& kernel_type); +OpKernelType TransPtenKernelKeyToOpKernelType( + const pten::KernelKey& kernel_key); +pten::KernelKey TransOpKernelTypeToPtenKernelKey( + const OpKernelType& kernel_type); /* Kernel Args parse */ diff --git a/paddle/fluid/framework/pten_utils_test.cc b/paddle/fluid/framework/pten_utils_test.cc index 3ba2da3df0580..b3f0e516a4781 100644 --- a/paddle/fluid/framework/pten_utils_test.cc +++ b/paddle/fluid/framework/pten_utils_test.cc @@ -41,7 +41,7 @@ TEST(TcmptUtils, MakeTensor) { ASSERT_EQ(dense_x->data_type(), pten::DataType::FLOAT32); } -TEST(TcmptUtils, VarToPtTensor) { +TEST(TcmptUtils, VarToPtenTensor) { // 1. create Variable Variable v; auto selected_rows = v.GetMutable(); @@ -57,7 +57,7 @@ TEST(TcmptUtils, VarToPtTensor) { auto tensor_def = pten::TensorArgDef(expect_backend, pten::DataLayout::NCHW, pten::DataType::INT32); // 2. test API - auto tensor_x = InputVariableToPtTensor(v, tensor_def); + auto tensor_x = InputVariableToPtenTensor(v, tensor_def); // 3. check result ASSERT_EQ(tensor_x->backend(), expect_backend); ASSERT_EQ(tensor_x->data_type(), pten::DataType::INT32); diff --git a/paddle/fluid/imperative/prepared_operator.cc b/paddle/fluid/imperative/prepared_operator.cc index 97d893babae18..749f4ec76a75c 100644 --- a/paddle/fluid/imperative/prepared_operator.cc +++ b/paddle/fluid/imperative/prepared_operator.cc @@ -153,12 +153,12 @@ PreparedOp PrepareImpl(const NameVarMap& ins, if (FLAGS_run_pt_kernel && pten::KernelFactory::Instance().ContainsKernel(op.Type().c_str())) { - auto pt_kernel_signature = op.GetExpectedPtKernelArgs(dygraph_exe_ctx); + auto pt_kernel_signature = op.GetExpectedPtenKernelArgs(dygraph_exe_ctx); VLOG(1) << framework::KernelSignatureToString(pt_kernel_signature); auto pt_kernel_name = pten::KernelName(pt_kernel_signature.first); - auto pt_kernel_key = TransOpKernelTypeToPtKernelKey(expected_kernel_key); + auto pt_kernel_key = TransOpKernelTypeToPtenKernelKey(expected_kernel_key); auto pt_kernel = pten::KernelFactory::Instance().SelectKernel( pt_kernel_name, pt_kernel_key); @@ -171,7 +171,7 @@ PreparedOp PrepareImpl(const NameVarMap& ins, return PreparedOp(op, ctx, expected_kernel_key, pt_kernel_signature, pt_kernel, dev_ctx); } else { - VLOG(1) << "Dynamic mode ChoosePtKernel - kernel `" << pt_kernel_name + VLOG(1) << "Dynamic mode ChoosePtenKernel - kernel `" << pt_kernel_name << "` not found."; } } @@ -243,7 +243,7 @@ PreparedOp PreparedOp::Prepare(const NameVarMap& ins, } template -static pten::KernelContext BuildDygraphPtKernelContext( +static pten::KernelContext BuildDygraphPtenKernelContext( const framework::KernelSignature& pt_kernel_signature, const pten::Kernel& pt_kernel, const NameVarMap& ins, const NameVarMap& outs, const framework::AttributeMap& attrs, @@ -292,7 +292,7 @@ static pten::KernelContext BuildDygraphPtKernelContext( for (auto var : ins_vector) { const auto& variable = var->Var(); - auto pt_in = framework::InputVariableToPtTensor(variable, in_def); + auto pt_in = framework::InputVariableToPtenTensor(variable, in_def); tmp_inputs.emplace_back(pt_in); } op_kernel_ctx.EmplaceBackInputs(tmp_inputs); @@ -306,7 +306,7 @@ static pten::KernelContext BuildDygraphPtKernelContext( for (auto var : outs_vector) { auto* variable = var->MutableVar(); - auto pt_out = framework::OutputVariableToPtTensor(variable, out_def); + auto pt_out = framework::OutputVariableToPtenTensor(variable, out_def); tmp_outputs.emplace_back(pt_out); } op_kernel_ctx.EmplaceBackOutputs(tmp_outputs); @@ -401,7 +401,7 @@ static void PreparedOpRunPtImpl( static_cast(op).InferShape( &infer_shape_ctx); - auto op_kernel_ctx = BuildDygraphPtKernelContext( + auto op_kernel_ctx = BuildDygraphPtenKernelContext( pt_kernel_signature, pt_kernel, ins, outs, attrs, default_attrs, *dev_ctx); diff --git a/paddle/fluid/operators/fill_any_like_op.cc b/paddle/fluid/operators/fill_any_like_op.cc index b46a1c3c89b6a..494341694b72e 100644 --- a/paddle/fluid/operators/fill_any_like_op.cc +++ b/paddle/fluid/operators/fill_any_like_op.cc @@ -48,7 +48,7 @@ class FillAnyLikeOp : public framework::OperatorWithKernel { tensor.layout()); } - framework::KernelSignature GetExpectedPtKernelArgs( + framework::KernelSignature GetExpectedPtenKernelArgs( const framework::ExecutionContext &ctx) const override { return std::make_pair( "fill_any_like", diff --git a/paddle/fluid/operators/scale_op.cc b/paddle/fluid/operators/scale_op.cc index 329a649a5a34d..e2ae1ef8eca31 100644 --- a/paddle/fluid/operators/scale_op.cc +++ b/paddle/fluid/operators/scale_op.cc @@ -71,7 +71,7 @@ class ScaleOp : public framework::OperatorWithKernel { return framework::OpKernelType(input_data_type, ctx.GetPlace()); } - framework::KernelSignature GetExpectedPtKernelArgs( + framework::KernelSignature GetExpectedPtenKernelArgs( const framework::ExecutionContext &ctx) const override { if (ctx.HasInput("ScaleTensor")) { return std::make_pair( diff --git a/paddle/pten/core/convert_utils.cc b/paddle/pten/core/convert_utils.cc index 74d87101d7175..32f2497dd18a5 100644 --- a/paddle/pten/core/convert_utils.cc +++ b/paddle/pten/core/convert_utils.cc @@ -20,7 +20,7 @@ limitations under the License. */ namespace pten { // TODO(chenweihang): Add other place trans cases later -Backend TransToPtBackend(const paddle::platform::Place& place) { +Backend TransToPtenBackend(const paddle::platform::Place& place) { if (paddle::platform::is_cpu_place(place)) { return Backend::CPU; } else if (paddle::platform::is_gpu_place(place)) { @@ -30,7 +30,7 @@ Backend TransToPtBackend(const paddle::platform::Place& place) { } } -paddle::experimental::DataType TransToPtDataType( +paddle::experimental::DataType TransToPtenDataType( const paddle::framework::proto::VarType::Type& dtype) { // Set the order of case branches according to the frequency with // the data type is used @@ -64,7 +64,7 @@ paddle::experimental::DataType TransToPtDataType( } } -DataLayout TransToPtDataLayout(const paddle::framework::DataLayout& layout) { +DataLayout TransToPtenDataLayout(const paddle::framework::DataLayout& layout) { switch (layout) { case paddle::framework::DataLayout::kNHWC: return DataLayout::NHWC; diff --git a/paddle/pten/core/convert_utils.h b/paddle/pten/core/convert_utils.h index e97d2a8c73210..aa79cb240dd04 100644 --- a/paddle/pten/core/convert_utils.h +++ b/paddle/pten/core/convert_utils.h @@ -30,10 +30,10 @@ namespace pten { using DataType = paddle::experimental::DataType; using DataLayout = paddle::experimental::DataLayout; -Backend TransToPtBackend(const paddle::platform::Place& place); -DataType TransToPtDataType( +Backend TransToPtenBackend(const paddle::platform::Place& place); +DataType TransToPtenDataType( const paddle::framework::proto::VarType::Type& dtype); -DataLayout TransToPtDataLayout(const paddle::framework::DataLayout& layout); +DataLayout TransToPtenDataLayout(const paddle::framework::DataLayout& layout); paddle::platform::Place TransToFluidPlace(const Backend& backend); paddle::framework::proto::VarType::Type TransToProtoVarType( diff --git a/paddle/pten/kernels/cuda/math.cu b/paddle/pten/kernels/cuda/math.cu index b9230dbf47a1f..1306cdc8017e6 100644 --- a/paddle/pten/kernels/cuda/math.cu +++ b/paddle/pten/kernels/cuda/math.cu @@ -77,7 +77,7 @@ void Mean(const CUDAContext& dev_ctx, const DenseTensor& x, DenseTensor* out) { pten::DenseTensor tmp( TensorMeta(paddle::framework::make_ddim( {static_cast(temp_storage_bytes)}), - pten::TransToPtBackend(dev_ctx.GetPlace()), + pten::TransToPtenBackend(dev_ctx.GetPlace()), x.data_type(), x.layout()), TensorStatus());