未验证 提交 30a02d27 编写于 作者: R Ruibin Cheung 提交者: GitHub

[clang-tidy] enable modernize-use-equals-default (#55983)

上级 4d094b0c
...@@ -181,7 +181,7 @@ modernize-redundant-void-arg, ...@@ -181,7 +181,7 @@ modernize-redundant-void-arg,
-modernize-unary-static-assert, -modernize-unary-static-assert,
-modernize-use-bool-literals, -modernize-use-bool-literals,
modernize-use-emplace, modernize-use-emplace,
-modernize-use-equals-default, modernize-use-equals-default,
-modernize-use-equals-delete, -modernize-use-equals-delete,
-modernize-use-noexcept, -modernize-use-noexcept,
modernize-use-nullptr, modernize-use-nullptr,
......
...@@ -34,7 +34,7 @@ DistTensorSpec::DistTensorSpec(const DistTensorSpec& spec) { ...@@ -34,7 +34,7 @@ DistTensorSpec::DistTensorSpec(const DistTensorSpec& spec) {
dist_attr_.copy_from(spec.dist_attr()); dist_attr_.copy_from(spec.dist_attr());
} }
DistTensorSpec::~DistTensorSpec() {} DistTensorSpec::~DistTensorSpec() = default;
DistTensorSpec::DistTensorSpec(const Tensor& tensor) { DistTensorSpec::DistTensorSpec(const Tensor& tensor) {
shape_ = tensor.shape(); shape_ = tensor.shape();
......
...@@ -47,7 +47,7 @@ ProcessGroupNCCL::NCCLTask::NCCLTask(const Place& place, ...@@ -47,7 +47,7 @@ ProcessGroupNCCL::NCCLTask::NCCLTask(const Place& place,
comm_event_(place), comm_event_(place),
task_place_(place) {} task_place_(place) {}
ProcessGroupNCCL::NCCLTask::~NCCLTask() {} ProcessGroupNCCL::NCCLTask::~NCCLTask() = default;
bool ProcessGroupNCCL::NCCLTask::IsCompleted() { return comm_event_.Query(); } bool ProcessGroupNCCL::NCCLTask::IsCompleted() { return comm_event_.Query(); }
......
...@@ -48,7 +48,7 @@ FleetExecutor::FleetExecutor(const FleetExecutorDesc& exe_desc) ...@@ -48,7 +48,7 @@ FleetExecutor::FleetExecutor(const FleetExecutorDesc& exe_desc)
InitMessageBus(); InitMessageBus();
} }
FleetExecutor::~FleetExecutor() { FleetExecutor::~FleetExecutor() { // NOLINT
for (const auto& carrier_id : carrier_ids_) { for (const auto& carrier_id : carrier_ids_) {
GlobalMap<std::string, Carrier>::Get(carrier_id)->Release(); GlobalMap<std::string, Carrier>::Get(carrier_id)->Release();
} }
......
...@@ -24,7 +24,7 @@ namespace distributed { ...@@ -24,7 +24,7 @@ namespace distributed {
Interceptor::Interceptor(int64_t interceptor_id, TaskNode* node) Interceptor::Interceptor(int64_t interceptor_id, TaskNode* node)
: interceptor_id_(interceptor_id), node_(node) {} : interceptor_id_(interceptor_id), node_(node) {}
Interceptor::~Interceptor() { Interceptor::~Interceptor() { // NOLINT
// FIXME(wangxi): throw in stop function // FIXME(wangxi): throw in stop function
// std::lock_guard<std::mutex> lock(mutex_); // std::lock_guard<std::mutex> lock(mutex_);
// PADDLE_ENFORCE_EQ(messages_.empty(), true, // PADDLE_ENFORCE_EQ(messages_.empty(), true,
......
...@@ -38,7 +38,7 @@ inline double GetCurrentUS() { ...@@ -38,7 +38,7 @@ inline double GetCurrentUS() {
return 1e+6 * time.tv_sec + time.tv_usec; return 1e+6 * time.tv_sec + time.tv_usec;
} }
Communicator::Communicator() {} Communicator::Communicator() = default;
void Communicator::InitGFlag(const std::string &gflags) { void Communicator::InitGFlag(const std::string &gflags) {
VLOG(3) << "Init With Gflags:" << gflags; VLOG(3) << "Init With Gflags:" << gflags;
......
...@@ -1205,7 +1205,7 @@ Node *GraphShard::find_node(uint64_t id) { ...@@ -1205,7 +1205,7 @@ Node *GraphShard::find_node(uint64_t id) {
return iter == node_location.end() ? nullptr : bucket[iter->second]; return iter == node_location.end() ? nullptr : bucket[iter->second];
} }
GraphTable::~GraphTable() { GraphTable::~GraphTable() { // NOLINT
#ifdef PADDLE_WITH_GPU_GRAPH #ifdef PADDLE_WITH_GPU_GRAPH
clear_graph(); clear_graph();
#endif #endif
......
...@@ -1999,7 +1999,7 @@ void PaddleBoxDataFeed::PutToFeedVec(const std::vector<Record*>& ins_vec) { ...@@ -1999,7 +1999,7 @@ void PaddleBoxDataFeed::PutToFeedVec(const std::vector<Record*>& ins_vec) {
#endif #endif
} }
SlotRecordInMemoryDataFeed::~SlotRecordInMemoryDataFeed() { SlotRecordInMemoryDataFeed::~SlotRecordInMemoryDataFeed() { // NOLINT
#if defined(PADDLE_WITH_CUDA) && defined(PADDLE_WITH_HETERPS) #if defined(PADDLE_WITH_CUDA) && defined(PADDLE_WITH_HETERPS)
stop_token_.store(true); stop_token_.store(true);
for (auto& thread : pack_threads_) { for (auto& thread : pack_threads_) {
......
...@@ -1821,7 +1821,7 @@ class MultiSlotInMemoryDataFeed : public InMemoryDataFeed<Record> { ...@@ -1821,7 +1821,7 @@ class MultiSlotInMemoryDataFeed : public InMemoryDataFeed<Record> {
class SlotRecordInMemoryDataFeed : public InMemoryDataFeed<SlotRecord> { class SlotRecordInMemoryDataFeed : public InMemoryDataFeed<SlotRecord> {
public: public:
SlotRecordInMemoryDataFeed() {} SlotRecordInMemoryDataFeed() = default;
virtual ~SlotRecordInMemoryDataFeed(); virtual ~SlotRecordInMemoryDataFeed();
void Init(const DataFeedDesc& data_feed_desc) override; void Init(const DataFeedDesc& data_feed_desc) override;
void LoadIntoMemory() override; void LoadIntoMemory() override;
......
...@@ -37,7 +37,7 @@ FetchAsyncOpHandle::FetchAsyncOpHandle(ir::Node *node, ...@@ -37,7 +37,7 @@ FetchAsyncOpHandle::FetchAsyncOpHandle(ir::Node *node,
local_exec_scopes_(local_exec_scopes), local_exec_scopes_(local_exec_scopes),
return_merged_(return_merged) {} return_merged_(return_merged) {}
FetchAsyncOpHandle::~FetchAsyncOpHandle() {} FetchAsyncOpHandle::~FetchAsyncOpHandle() = default;
void FetchAsyncOpHandle::RecordWaitEventOnCtx( void FetchAsyncOpHandle::RecordWaitEventOnCtx(
platform::DeviceContext *waited_ctx) { platform::DeviceContext *waited_ctx) {
......
...@@ -35,7 +35,7 @@ FetchOpHandle::FetchOpHandle(ir::Node *node, ...@@ -35,7 +35,7 @@ FetchOpHandle::FetchOpHandle(ir::Node *node,
local_exec_scopes_(local_exec_scopes), local_exec_scopes_(local_exec_scopes),
return_merged_(return_merged) {} return_merged_(return_merged) {}
FetchOpHandle::~FetchOpHandle() {} FetchOpHandle::~FetchOpHandle() = default;
void FetchOpHandle::RecordWaitEventOnCtx(platform::DeviceContext *waited_ctx) { void FetchOpHandle::RecordWaitEventOnCtx(platform::DeviceContext *waited_ctx) {
PADDLE_THROW(platform::errors::PermissionDenied( PADDLE_THROW(platform::errors::PermissionDenied(
......
...@@ -27,7 +27,7 @@ namespace paddle { ...@@ -27,7 +27,7 @@ namespace paddle {
namespace framework { namespace framework {
namespace details { namespace details {
struct DebugTools { struct DebugTools {
DebugTools() {} DebugTools() = default;
std::string path = ""; std::string path = "";
int stack_limit = 1; int stack_limit = 1;
}; };
......
...@@ -30,7 +30,7 @@ std::string OpHandleBase::DebugString() const { ...@@ -30,7 +30,7 @@ std::string OpHandleBase::DebugString() const {
return ss.str(); return ss.str();
} }
OpHandleBase::~OpHandleBase() PADDLE_MAY_THROW { OpHandleBase::~OpHandleBase() PADDLE_MAY_THROW { // NOLINT
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
for (auto &ev : events_) { for (auto &ev : events_) {
if (ev.second) { if (ev.second) {
......
...@@ -40,7 +40,7 @@ ScaleLossGradOpHandle::ScaleLossGradOpHandle(ir::Node *node, ...@@ -40,7 +40,7 @@ ScaleLossGradOpHandle::ScaleLossGradOpHandle(ir::Node *node,
this->SetDeviceContext(place_, dev_ctx); this->SetDeviceContext(place_, dev_ctx);
} }
ScaleLossGradOpHandle::~ScaleLossGradOpHandle() {} ScaleLossGradOpHandle::~ScaleLossGradOpHandle() = default;
struct ScaleLossGradFunctor { struct ScaleLossGradFunctor {
float coeff_; float coeff_;
......
...@@ -19,7 +19,7 @@ ...@@ -19,7 +19,7 @@
namespace paddle { namespace paddle {
namespace framework { namespace framework {
namespace details { namespace details {
SSAGraphExecutor::~SSAGraphExecutor() {} SSAGraphExecutor::~SSAGraphExecutor() = default;
void ClearFetchOp(ir::Graph* graph, std::vector<OpHandleBase*>* fetch_ops) { void ClearFetchOp(ir::Graph* graph, std::vector<OpHandleBase*>* fetch_ops) {
if (fetch_ops->empty()) return; if (fetch_ops->empty()) return;
......
...@@ -18,7 +18,7 @@ namespace paddle { ...@@ -18,7 +18,7 @@ namespace paddle {
namespace framework { namespace framework {
namespace details { namespace details {
VarHandleBase::~VarHandleBase() {} VarHandleBase::~VarHandleBase() = default;
VarHandle::~VarHandle() { VLOG(4) << "deleting var handle " << DebugString(); } VarHandle::~VarHandle() { VLOG(4) << "deleting var handle " << DebugString(); }
......
...@@ -24,7 +24,7 @@ namespace paddle { ...@@ -24,7 +24,7 @@ namespace paddle {
namespace framework { namespace framework {
namespace ir { namespace ir {
AdaptivePool2dConvertGlobalPass::AdaptivePool2dConvertGlobalPass() { AdaptivePool2dConvertGlobalPass::AdaptivePool2dConvertGlobalPass() { // NOLINT
AddOpCompat(OpCompat("pool2d")) AddOpCompat(OpCompat("pool2d"))
.AddInput("X") .AddInput("X")
.IsTensor() .IsTensor()
......
...@@ -32,7 +32,7 @@ class Graph; ...@@ -32,7 +32,7 @@ class Graph;
class AdaptivePool2dConvertGlobalPass : public FusePassBase { class AdaptivePool2dConvertGlobalPass : public FusePassBase {
public: public:
AdaptivePool2dConvertGlobalPass(); AdaptivePool2dConvertGlobalPass();
virtual ~AdaptivePool2dConvertGlobalPass() {} virtual ~AdaptivePool2dConvertGlobalPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -51,7 +51,7 @@ struct ConstantFolding : public PatternBase { ...@@ -51,7 +51,7 @@ struct ConstantFolding : public PatternBase {
}; };
} // namespace patterns } // namespace patterns
ConstantFoldingPass::ConstantFoldingPass() {} ConstantFoldingPass::ConstantFoldingPass() = default;
void ConstantFoldingPass::ApplyImpl(ir::Graph *graph) const { void ConstantFoldingPass::ApplyImpl(ir::Graph *graph) const {
PADDLE_ENFORCE_NOT_NULL( PADDLE_ENFORCE_NOT_NULL(
......
...@@ -756,7 +756,7 @@ void ConvEltwiseAddBNFusePass::ApplyImpl(ir::Graph* graph) const { ...@@ -756,7 +756,7 @@ void ConvEltwiseAddBNFusePass::ApplyImpl(ir::Graph* graph) const {
AddStatis(found_conv_bn_count); AddStatis(found_conv_bn_count);
} }
ConvTransposeBNFusePass::ConvTransposeBNFusePass() { ConvTransposeBNFusePass::ConvTransposeBNFusePass() { // NOLINT
AddOpCompat(OpCompat("conv2d_transpose")) AddOpCompat(OpCompat("conv2d_transpose"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
...@@ -800,7 +800,8 @@ ConvTransposeBNFusePass::ConvTransposeBNFusePass() { ...@@ -800,7 +800,8 @@ ConvTransposeBNFusePass::ConvTransposeBNFusePass() {
.End(); .End();
} }
ConvTransposeEltwiseAddBNFusePass::ConvTransposeEltwiseAddBNFusePass() { ConvTransposeEltwiseAddBNFusePass::
ConvTransposeEltwiseAddBNFusePass() { // NOLINT
AddOpCompat(OpCompat("conv2d_transpose")) AddOpCompat(OpCompat("conv2d_transpose"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
...@@ -844,7 +845,7 @@ ConvTransposeEltwiseAddBNFusePass::ConvTransposeEltwiseAddBNFusePass() { ...@@ -844,7 +845,7 @@ ConvTransposeEltwiseAddBNFusePass::ConvTransposeEltwiseAddBNFusePass() {
.End(); .End();
} }
DepthwiseConvBNFusePass::DepthwiseConvBNFusePass() { DepthwiseConvBNFusePass::DepthwiseConvBNFusePass() { // NOLINT
AddOpCompat(OpCompat("depthwise_conv2d")) AddOpCompat(OpCompat("depthwise_conv2d"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
......
...@@ -30,7 +30,7 @@ using platform::MemEvent; ...@@ -30,7 +30,7 @@ using platform::MemEvent;
const double CostData::NOT_MEASURED = -1; const double CostData::NOT_MEASURED = -1;
CostData::~CostData() { CostData::~CostData() { // NOLINT
// TODO(zhhsplendid): when we save a copy of program/graph, we should delete // TODO(zhhsplendid): when we save a copy of program/graph, we should delete
// here. // here.
} }
......
...@@ -56,7 +56,7 @@ static const std::initializer_list<std::string> rnn_variable_names{ ...@@ -56,7 +56,7 @@ static const std::initializer_list<std::string> rnn_variable_names{
class ComputePropagateScalesMkldnnPassTest : public testing::Test { class ComputePropagateScalesMkldnnPassTest : public testing::Test {
public: public:
ComputePropagateScalesMkldnnPassTest() { ComputePropagateScalesMkldnnPassTest() { // NOLINT
pass.reset(new ComputePropagateScalesMkldnnPass()); pass.reset(new ComputePropagateScalesMkldnnPass());
} }
......
...@@ -32,7 +32,7 @@ class Graph; ...@@ -32,7 +32,7 @@ class Graph;
PADDLE_ENFORCE_NOT_NULL( \ PADDLE_ENFORCE_NOT_NULL( \
id, platform::errors::InvalidArgument("Subgraph has no node %s.", #id)); id, platform::errors::InvalidArgument("Subgraph has no node %s.", #id));
DepthwiseConvMKLDNNPass::DepthwiseConvMKLDNNPass() { DepthwiseConvMKLDNNPass::DepthwiseConvMKLDNNPass() { // NOLINT
AddOpCompat(OpCompat("depthwise_conv2d")) AddOpCompat(OpCompat("depthwise_conv2d"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
......
...@@ -25,7 +25,7 @@ class Graph; ...@@ -25,7 +25,7 @@ class Graph;
class DepthwiseConvMKLDNNPass : public FusePassBase { class DepthwiseConvMKLDNNPass : public FusePassBase {
public: public:
DepthwiseConvMKLDNNPass(); DepthwiseConvMKLDNNPass();
virtual ~DepthwiseConvMKLDNNPass() {} virtual ~DepthwiseConvMKLDNNPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -23,7 +23,7 @@ namespace paddle { ...@@ -23,7 +23,7 @@ namespace paddle {
namespace framework { namespace framework {
namespace ir { namespace ir {
Int8ScaleCalculationMkldnnPass::Int8ScaleCalculationMkldnnPass() { Int8ScaleCalculationMkldnnPass::Int8ScaleCalculationMkldnnPass() { // NOLINT
AddOpCompat(OpCompat("conv2d")) AddOpCompat(OpCompat("conv2d"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
......
...@@ -27,7 +27,7 @@ class Graph; ...@@ -27,7 +27,7 @@ class Graph;
class Int8ScaleCalculationMkldnnPass : public FusePassBase { class Int8ScaleCalculationMkldnnPass : public FusePassBase {
public: public:
Int8ScaleCalculationMkldnnPass(); Int8ScaleCalculationMkldnnPass();
virtual ~Int8ScaleCalculationMkldnnPass() {} virtual ~Int8ScaleCalculationMkldnnPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -76,7 +76,7 @@ void QuantizeConvInput(Scope* scope, ...@@ -76,7 +76,7 @@ void QuantizeConvInput(Scope* scope,
} // namespace } // namespace
ParamsQuantizationMkldnnPass::ParamsQuantizationMkldnnPass() { ParamsQuantizationMkldnnPass::ParamsQuantizationMkldnnPass() { // NOLINT
AddOpCompat(OpCompat("fused_conv2d")) AddOpCompat(OpCompat("fused_conv2d"))
.AddInput("Input") .AddInput("Input")
.IsTensor() .IsTensor()
......
...@@ -27,7 +27,7 @@ class Graph; ...@@ -27,7 +27,7 @@ class Graph;
class ParamsQuantizationMkldnnPass : public FusePassBase { class ParamsQuantizationMkldnnPass : public FusePassBase {
public: public:
ParamsQuantizationMkldnnPass(); ParamsQuantizationMkldnnPass();
virtual ~ParamsQuantizationMkldnnPass() {} virtual ~ParamsQuantizationMkldnnPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -65,7 +65,7 @@ struct TestScope { ...@@ -65,7 +65,7 @@ struct TestScope {
}; };
struct ProgramStrategy { struct ProgramStrategy {
virtual ~ProgramStrategy() {} virtual ~ProgramStrategy() = default;
std::unique_ptr<Graph> CreateGraph() { std::unique_ptr<Graph> CreateGraph() {
CreateProgram(); CreateProgram();
......
...@@ -170,7 +170,8 @@ void FuseQuantTranspose2DequantOneDNNPass::ApplyImpl(Graph *graph) const { ...@@ -170,7 +170,8 @@ void FuseQuantTranspose2DequantOneDNNPass::ApplyImpl(Graph *graph) const {
FuseTranspose2Dequantize(graph, "transpose2"); FuseTranspose2Dequantize(graph, "transpose2");
} }
FuseQuantTranspose2DequantOneDNNPass::FuseQuantTranspose2DequantOneDNNPass() { FuseQuantTranspose2DequantOneDNNPass::
FuseQuantTranspose2DequantOneDNNPass() { // NOLINT
AddOpCompat(OpCompat("transpose2")) AddOpCompat(OpCompat("transpose2"))
.AddInput("X") .AddInput("X")
.IsTensor() .IsTensor()
......
...@@ -23,7 +23,7 @@ namespace ir { ...@@ -23,7 +23,7 @@ namespace ir {
class FuseQuantTranspose2DequantOneDNNPass : public FusePassBase { class FuseQuantTranspose2DequantOneDNNPass : public FusePassBase {
public: public:
virtual ~FuseQuantTranspose2DequantOneDNNPass() {} virtual ~FuseQuantTranspose2DequantOneDNNPass() = default;
FuseQuantTranspose2DequantOneDNNPass(); FuseQuantTranspose2DequantOneDNNPass();
protected: protected:
......
...@@ -31,7 +31,7 @@ namespace ir { ...@@ -31,7 +31,7 @@ namespace ir {
GET_IR_NODE(reshape2_op); \ GET_IR_NODE(reshape2_op); \
GET_IR_NODE(reshape2_out); GET_IR_NODE(reshape2_out);
ShuffleChannelMKLDNNDetectPass::ShuffleChannelMKLDNNDetectPass() { ShuffleChannelMKLDNNDetectPass::ShuffleChannelMKLDNNDetectPass() { // NOLINT
AddOpCompat(OpCompat("reshape2")) AddOpCompat(OpCompat("reshape2"))
.AddInput("X") .AddInput("X")
.IsTensor() .IsTensor()
......
...@@ -27,7 +27,7 @@ class Graph; ...@@ -27,7 +27,7 @@ class Graph;
class ShuffleChannelMKLDNNDetectPass : public FusePassBase { class ShuffleChannelMKLDNNDetectPass : public FusePassBase {
public: public:
ShuffleChannelMKLDNNDetectPass(); ShuffleChannelMKLDNNDetectPass();
virtual ~ShuffleChannelMKLDNNDetectPass() {} virtual ~ShuffleChannelMKLDNNDetectPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -37,7 +37,7 @@ namespace paddle { ...@@ -37,7 +37,7 @@ namespace paddle {
namespace framework { namespace framework {
namespace ir { namespace ir {
class Node; class Node;
ReverseRollFusePass::ReverseRollFusePass() { ReverseRollFusePass::ReverseRollFusePass() { // NOLINT
AddOpCompat(OpCompat("reshape2")) AddOpCompat(OpCompat("reshape2"))
.AddInput("X") .AddInput("X")
.IsTensor() .IsTensor()
......
...@@ -54,7 +54,7 @@ namespace ir { ...@@ -54,7 +54,7 @@ namespace ir {
class ReverseRollFusePass : public FusePassBase { class ReverseRollFusePass : public FusePassBase {
public: public:
ReverseRollFusePass(); ReverseRollFusePass();
virtual ~ReverseRollFusePass() {} virtual ~ReverseRollFusePass() = default;
protected: protected:
void ApplyImpl(ir::Graph *graph) const override; void ApplyImpl(ir::Graph *graph) const override;
......
...@@ -31,7 +31,7 @@ namespace ir { ...@@ -31,7 +31,7 @@ namespace ir {
GET_IR_NODE(reshape2_op); \ GET_IR_NODE(reshape2_op); \
GET_IR_NODE(reshape2_out); GET_IR_NODE(reshape2_out);
ShuffleChannelDetectPass::ShuffleChannelDetectPass() { ShuffleChannelDetectPass::ShuffleChannelDetectPass() { // NOLINT
AddOpCompat(OpCompat("reshape2")) AddOpCompat(OpCompat("reshape2"))
.AddInput("X") .AddInput("X")
.IsTensor() .IsTensor()
......
...@@ -27,7 +27,7 @@ class Graph; ...@@ -27,7 +27,7 @@ class Graph;
class ShuffleChannelDetectPass : public FusePassBase { class ShuffleChannelDetectPass : public FusePassBase {
public: public:
ShuffleChannelDetectPass(); ShuffleChannelDetectPass();
virtual ~ShuffleChannelDetectPass() {} virtual ~ShuffleChannelDetectPass() = default;
protected: protected:
void ApplyImpl(ir::Graph* graph) const override; void ApplyImpl(ir::Graph* graph) const override;
......
...@@ -67,7 +67,7 @@ SigmoidElementmulFusePattern::SigmoidElementmulFusePattern( ...@@ -67,7 +67,7 @@ SigmoidElementmulFusePattern::SigmoidElementmulFusePattern(
} // namespace patterns } // namespace patterns
SigmoidElementmulFusePass::SigmoidElementmulFusePass() {} SigmoidElementmulFusePass::SigmoidElementmulFusePass() = default;
void SigmoidElementmulFusePass::ApplyImpl(ir::Graph* graph) const { void SigmoidElementmulFusePass::ApplyImpl(ir::Graph* graph) const {
PADDLE_ENFORCE_NOT_NULL( PADDLE_ENFORCE_NOT_NULL(
......
...@@ -28,7 +28,7 @@ namespace ir { ...@@ -28,7 +28,7 @@ namespace ir {
class Node; class Node;
TrtMapOpsToMatrixMultiplyPass::TrtMapOpsToMatrixMultiplyPass() {} TrtMapOpsToMatrixMultiplyPass::TrtMapOpsToMatrixMultiplyPass() = default;
void TrtMapOpsToMatrixMultiplyPass::ApplyImpl(ir::Graph* graph) const { void TrtMapOpsToMatrixMultiplyPass::ApplyImpl(ir::Graph* graph) const {
PADDLE_ENFORCE_NOT_NULL( PADDLE_ENFORCE_NOT_NULL(
......
...@@ -149,7 +149,7 @@ struct YoloBoxPattern : public PatternBase { ...@@ -149,7 +149,7 @@ struct YoloBoxPattern : public PatternBase {
}; };
} // namespace patterns } // namespace patterns
YoloBoxFusePass::YoloBoxFusePass() {} YoloBoxFusePass::YoloBoxFusePass() = default;
void YoloBoxFusePass::ApplyImpl(ir::Graph* graph) const { void YoloBoxFusePass::ApplyImpl(ir::Graph* graph) const {
PADDLE_ENFORCE_NOT_NULL( PADDLE_ENFORCE_NOT_NULL(
......
...@@ -27,7 +27,7 @@ InterpreterCoreNoEventGarbageCollector:: ...@@ -27,7 +27,7 @@ InterpreterCoreNoEventGarbageCollector::
} }
InterpreterCoreNoEventGarbageCollector:: InterpreterCoreNoEventGarbageCollector::
~InterpreterCoreNoEventGarbageCollector() { ~InterpreterCoreNoEventGarbageCollector() { // NOLINT
queue_.reset(nullptr); queue_.reset(nullptr);
} }
......
...@@ -37,7 +37,7 @@ VariableScope::VariableScope(Scope* scope) { ...@@ -37,7 +37,7 @@ VariableScope::VariableScope(Scope* scope) {
"You have passed a nullptr to construct VariableScope.")); "You have passed a nullptr to construct VariableScope."));
} }
VariableScope::~VariableScope() {} VariableScope::~VariableScope() = default;
Scope* VariableScope::GetMutableScope() const { return scope_; } Scope* VariableScope::GetMutableScope() const { return scope_; }
......
...@@ -40,7 +40,7 @@ class KernelArgsNameMakerByOpProto : public KernelArgsNameMaker { ...@@ -40,7 +40,7 @@ class KernelArgsNameMakerByOpProto : public KernelArgsNameMaker {
platform::errors::InvalidArgument("Op proto cannot be nullptr.")); platform::errors::InvalidArgument("Op proto cannot be nullptr."));
} }
~KernelArgsNameMakerByOpProto() override {} ~KernelArgsNameMakerByOpProto() override = default;
const paddle::small_vector<const char*>& GetInputArgsNames() override; const paddle::small_vector<const char*>& GetInputArgsNames() override;
const paddle::small_vector<const char*>& GetOutputArgsNames() override; const paddle::small_vector<const char*>& GetOutputArgsNames() override;
......
...@@ -187,7 +187,7 @@ void ProgramProcessor::AddDepToBlockOp(const BlockDesc &block) { ...@@ -187,7 +187,7 @@ void ProgramProcessor::AddDepToBlockOp(const BlockDesc &block) {
} }
} }
ProgramProcessor::ProgramProcessor() {} ProgramProcessor::ProgramProcessor() = default;
} // namespace framework } // namespace framework
} // namespace paddle } // namespace paddle
...@@ -72,7 +72,7 @@ void ReaderBase::Start() { ...@@ -72,7 +72,7 @@ void ReaderBase::Start() {
} }
} }
ReaderBase::~ReaderBase() {} ReaderBase::~ReaderBase() = default;
DecoratedReader::~DecoratedReader() { DecoratedReader::~DecoratedReader() {
VLOG(1) << "~DecoratedReader"; VLOG(1) << "~DecoratedReader";
......
...@@ -114,7 +114,7 @@ struct VarIdToTypeIndexMapHolder { ...@@ -114,7 +114,7 @@ struct VarIdToTypeIndexMapHolder {
} }
private: private:
VarIdToTypeIndexMapHolder() { VarIdToTypeIndexMapHolder() { // NOLINT
VarIdToTypeIndexMapInitializer::Init(&id_to_type_map_, &type_to_id_map_); VarIdToTypeIndexMapInitializer::Init(&id_to_type_map_, &type_to_id_map_);
} }
......
...@@ -131,7 +131,9 @@ AutoCastGuard::AutoCastGuard(std::shared_ptr<Tracer> tracer, AmpLevel level) ...@@ -131,7 +131,9 @@ AutoCastGuard::AutoCastGuard(std::shared_ptr<Tracer> tracer, AmpLevel level)
} }
} }
AutoCastGuard::~AutoCastGuard() { tracer_->SetAmpLevel(pre_amp_level_); } AutoCastGuard::~AutoCastGuard() { // NOLINT
tracer_->SetAmpLevel(pre_amp_level_);
}
AmpOperators::AmpOperators() AmpOperators::AmpOperators()
: allow_ops_(new std::unordered_set<std::string>()), : allow_ops_(new std::unordered_set<std::string>()),
...@@ -163,7 +165,7 @@ AmpOperators::AmpOperators() ...@@ -163,7 +165,7 @@ AmpOperators::AmpOperators()
<< unsupported_bf16_ops_->size(); << unsupported_bf16_ops_->size();
} }
AmpOperators::~AmpOperators() {} AmpOperators::~AmpOperators() = default;
AmpOperators& AmpOperators::Instance() { AmpOperators& AmpOperators::Instance() {
static AmpOperators instance; static AmpOperators instance;
......
...@@ -23,7 +23,7 @@ namespace paddle { ...@@ -23,7 +23,7 @@ namespace paddle {
namespace inference { namespace inference {
namespace analysis { namespace analysis {
Analyzer::Analyzer() {} Analyzer::Analyzer() = default;
void Analyzer::Run(Argument *argument) { RunAnalysis(argument); } void Analyzer::Run(Argument *argument) { RunAnalysis(argument); }
......
...@@ -27,7 +27,7 @@ namespace paddle { ...@@ -27,7 +27,7 @@ namespace paddle {
namespace inference { namespace inference {
namespace analysis { namespace analysis {
PassRegistry::PassRegistry() { PassRegistry::PassRegistry() { // NOLINT
// Register manually to avoid the trivial `USE_OP` like macro for easier use // Register manually to avoid the trivial `USE_OP` like macro for easier use
// and link. // and link.
passes_.emplace("ir_analysis_pass", passes_.emplace("ir_analysis_pass",
......
...@@ -50,7 +50,7 @@ class EigenGpuStreamDevice : public Eigen::StreamInterface { ...@@ -50,7 +50,7 @@ class EigenGpuStreamDevice : public Eigen::StreamInterface {
EigenGpuStreamDevice() : scratch_(nullptr), semaphore_(nullptr) { EigenGpuStreamDevice() : scratch_(nullptr), semaphore_(nullptr) {
Eigen::initializeDeviceProp(); Eigen::initializeDeviceProp();
} }
~EigenGpuStreamDevice() override {} ~EigenGpuStreamDevice() override = default;
void Reinitialize(gpuStream_t cuda_stream, void Reinitialize(gpuStream_t cuda_stream,
phi::Allocator* allocator, phi::Allocator* allocator,
......
...@@ -29,7 +29,7 @@ namespace tensorrt { ...@@ -29,7 +29,7 @@ namespace tensorrt {
class ActivationOpConverter : public OpConverter { class ActivationOpConverter : public OpConverter {
public: public:
ActivationOpConverter() {} ActivationOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -21,7 +21,7 @@ namespace tensorrt { ...@@ -21,7 +21,7 @@ namespace tensorrt {
class ElementwiseTensorOpConverter : public OpConverter { class ElementwiseTensorOpConverter : public OpConverter {
public: public:
ElementwiseTensorOpConverter() {} ElementwiseTensorOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
...@@ -325,7 +325,7 @@ class ElementwiseTensorModOpConverter : public ElementwiseTensorOpConverter { ...@@ -325,7 +325,7 @@ class ElementwiseTensorModOpConverter : public ElementwiseTensorOpConverter {
// https://github.com/PaddlePaddle/Paddle/blob/release/2.4/python/paddle/tensor/math.py#L420 // https://github.com/PaddlePaddle/Paddle/blob/release/2.4/python/paddle/tensor/math.py#L420
class PowOpConverter : public OpConverter { class PowOpConverter : public OpConverter {
public: public:
PowOpConverter() {} PowOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -21,7 +21,7 @@ namespace tensorrt { ...@@ -21,7 +21,7 @@ namespace tensorrt {
class EqualOpConverter : public OpConverter { class EqualOpConverter : public OpConverter {
public: public:
EqualOpConverter() {} EqualOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
...@@ -74,7 +74,7 @@ class EqualOpConverter : public OpConverter { ...@@ -74,7 +74,7 @@ class EqualOpConverter : public OpConverter {
class NotEqualOpConverter : public OpConverter { class NotEqualOpConverter : public OpConverter {
public: public:
NotEqualOpConverter() {} NotEqualOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -20,7 +20,7 @@ namespace tensorrt { ...@@ -20,7 +20,7 @@ namespace tensorrt {
class SquareOpConverter : public OpConverter { class SquareOpConverter : public OpConverter {
public: public:
SquareOpConverter() {} SquareOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -29,7 +29,7 @@ namespace tensorrt { ...@@ -29,7 +29,7 @@ namespace tensorrt {
class TopKOpConverter : public OpConverter { class TopKOpConverter : public OpConverter {
public: public:
TopKOpConverter() {} TopKOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -29,7 +29,7 @@ namespace tensorrt { ...@@ -29,7 +29,7 @@ namespace tensorrt {
class UnaryOpConverter : public OpConverter { class UnaryOpConverter : public OpConverter {
public: public:
UnaryOpConverter() {} UnaryOpConverter() = default;
void operator()(const framework::proto::OpDesc& op, void operator()(const framework::proto::OpDesc& op,
const framework::Scope& scope, const framework::Scope& scope,
bool test_mode) override { bool test_mode) override {
......
...@@ -23,7 +23,7 @@ namespace tensorrt { ...@@ -23,7 +23,7 @@ namespace tensorrt {
class ExprWrapper { class ExprWrapper {
public: public:
ExprWrapper() {} ExprWrapper() = default;
ExprWrapper(const nvinfer1::IDimensionExpr* expr, ExprWrapper(const nvinfer1::IDimensionExpr* expr,
nvinfer1::IExprBuilder* expr_builder) { nvinfer1::IExprBuilder* expr_builder) {
this->expr = expr; this->expr = expr;
......
...@@ -35,7 +35,7 @@ namespace tensorrt { ...@@ -35,7 +35,7 @@ namespace tensorrt {
// Just tell by the op_types. // Just tell by the op_types.
struct SimpleOpTypeSetTeller : public Teller { struct SimpleOpTypeSetTeller : public Teller {
SimpleOpTypeSetTeller() { SimpleOpTypeSetTeller() { // NOLINT
#if IS_TRT_VERSION_GE(7130) #if IS_TRT_VERSION_GE(7130)
// use TensorRT plugin // use TensorRT plugin
teller_set.insert("group_norm"); teller_set.insert("group_norm");
...@@ -3083,7 +3083,7 @@ struct SimpleOpTypeSetTeller : public Teller { ...@@ -3083,7 +3083,7 @@ struct SimpleOpTypeSetTeller : public Teller {
struct GenericPluginTeller : public Teller { struct GenericPluginTeller : public Teller {
public: public:
GenericPluginTeller() {} GenericPluginTeller() = default;
bool operator()(const framework::OpDesc& desc, bool operator()(const framework::OpDesc& desc,
bool use_no_calib_int8 = false, bool use_no_calib_int8 = false,
bool with_dynamic_shape = false) override { bool with_dynamic_shape = false) override {
...@@ -3125,7 +3125,7 @@ struct GenericPluginTeller : public Teller { ...@@ -3125,7 +3125,7 @@ struct GenericPluginTeller : public Teller {
struct CustomPluginTeller : public Teller { struct CustomPluginTeller : public Teller {
public: public:
CustomPluginTeller() {} CustomPluginTeller() = default;
bool operator()(const framework::OpDesc& desc, bool operator()(const framework::OpDesc& desc,
bool use_no_calib_int8 = false, bool use_no_calib_int8 = false,
bool with_dynamic_shape = false) override { bool with_dynamic_shape = false) override {
...@@ -3178,7 +3178,7 @@ bool OpTeller::Tell(const framework::ir::Node* node, ...@@ -3178,7 +3178,7 @@ bool OpTeller::Tell(const framework::ir::Node* node,
return false; return false;
} }
OpTeller::OpTeller() { OpTeller::OpTeller() { // NOLINT
tellers_.emplace_back(new tensorrt::SimpleOpTypeSetTeller); tellers_.emplace_back(new tensorrt::SimpleOpTypeSetTeller);
tellers_.emplace_back(new tensorrt::GenericPluginTeller); tellers_.emplace_back(new tensorrt::GenericPluginTeller);
tellers_.emplace_back(new tensorrt::CustomPluginTeller); tellers_.emplace_back(new tensorrt::CustomPluginTeller);
......
...@@ -405,7 +405,7 @@ char const* EmbLayerNormPlugin::getPluginNamespace() const noexcept { ...@@ -405,7 +405,7 @@ char const* EmbLayerNormPlugin::getPluginNamespace() const noexcept {
return mNamespace.c_str(); return mNamespace.c_str();
} }
EmbLayerNormPluginCreator::EmbLayerNormPluginCreator() {} EmbLayerNormPluginCreator::EmbLayerNormPluginCreator() = default;
char const* EmbLayerNormPluginCreator::getPluginName() const noexcept { char const* EmbLayerNormPluginCreator::getPluginName() const noexcept {
return EMB_LAYER_NORM_NAME; return EMB_LAYER_NORM_NAME;
......
...@@ -771,7 +771,7 @@ char const* EmbLayerNormVarSeqlenPluginBase::getPluginNamespace() ...@@ -771,7 +771,7 @@ char const* EmbLayerNormVarSeqlenPluginBase::getPluginNamespace()
} }
EmbLayerNormVarSeqlenPluginBaseCreator:: EmbLayerNormVarSeqlenPluginBaseCreator::
EmbLayerNormVarSeqlenPluginBaseCreator() {} EmbLayerNormVarSeqlenPluginBaseCreator() = default;
char const* EmbLayerNormVarSeqlenPluginBaseCreator::getPluginName() char const* EmbLayerNormVarSeqlenPluginBaseCreator::getPluginName()
const noexcept { const noexcept {
......
...@@ -405,7 +405,7 @@ void RoiAlignPluginDynamic::serialize(void* buffer) const TRT_NOEXCEPT { ...@@ -405,7 +405,7 @@ void RoiAlignPluginDynamic::serialize(void* buffer) const TRT_NOEXCEPT {
void RoiAlignPluginDynamic::destroy() TRT_NOEXCEPT {} void RoiAlignPluginDynamic::destroy() TRT_NOEXCEPT {}
RoiAlignPluginDynamicCreator::RoiAlignPluginDynamicCreator() {} RoiAlignPluginDynamicCreator::RoiAlignPluginDynamicCreator() = default;
void RoiAlignPluginDynamicCreator::setPluginNamespace(const char* lib_namespace) void RoiAlignPluginDynamicCreator::setPluginNamespace(const char* lib_namespace)
TRT_NOEXCEPT { TRT_NOEXCEPT {
......
...@@ -36,7 +36,7 @@ StackPluginDynamic::StackPluginDynamic(void const* serial_data, ...@@ -36,7 +36,7 @@ StackPluginDynamic::StackPluginDynamic(void const* serial_data,
DeserializeValue(&serial_data, &serial_length, &with_fp16_); DeserializeValue(&serial_data, &serial_length, &with_fp16_);
} }
StackPluginDynamic::~StackPluginDynamic() {} StackPluginDynamic::~StackPluginDynamic() = default;
nvinfer1::IPluginV2DynamicExt* StackPluginDynamic::clone() const TRT_NOEXCEPT { nvinfer1::IPluginV2DynamicExt* StackPluginDynamic::clone() const TRT_NOEXCEPT {
return new StackPluginDynamic(axis_, num_stack_, with_fp16_); return new StackPluginDynamic(axis_, num_stack_, with_fp16_);
...@@ -230,7 +230,7 @@ int StackPluginDynamic::enqueue(const nvinfer1::PluginTensorDesc* input_desc, ...@@ -230,7 +230,7 @@ int StackPluginDynamic::enqueue(const nvinfer1::PluginTensorDesc* input_desc,
return cudaGetLastError() != cudaSuccess; return cudaGetLastError() != cudaSuccess;
} }
StackPluginDynamicCreator::StackPluginDynamicCreator() {} StackPluginDynamicCreator::StackPluginDynamicCreator() = default;
const char* StackPluginDynamicCreator::getPluginName() const TRT_NOEXCEPT { const char* StackPluginDynamicCreator::getPluginName() const TRT_NOEXCEPT {
return "stack_plugin"; return "stack_plugin";
......
...@@ -437,7 +437,7 @@ nvinfer1::IPluginV2Ext* YoloBoxPlugin::clone() const TRT_NOEXCEPT { ...@@ -437,7 +437,7 @@ nvinfer1::IPluginV2Ext* YoloBoxPlugin::clone() const TRT_NOEXCEPT {
input_w_); input_w_);
} }
YoloBoxPluginCreator::YoloBoxPluginCreator() {} YoloBoxPluginCreator::YoloBoxPluginCreator() = default;
void YoloBoxPluginCreator::setPluginNamespace(const char* lib_namespace) void YoloBoxPluginCreator::setPluginNamespace(const char* lib_namespace)
TRT_NOEXCEPT { TRT_NOEXCEPT {
......
...@@ -34,7 +34,7 @@ class AttributeVisitor { ...@@ -34,7 +34,7 @@ class AttributeVisitor {
public: public:
ir::IrContext* ctx; ir::IrContext* ctx;
AttributeVisitor() { ctx = ir::IrContext::Instance(); } AttributeVisitor() { ctx = ir::IrContext::Instance(); }
~AttributeVisitor() {} ~AttributeVisitor() = default;
public: public:
virtual ir::Attribute operator()(int i) { virtual ir::Attribute operator()(int i) {
......
...@@ -127,7 +127,7 @@ class CUDAGraphAllocator ...@@ -127,7 +127,7 @@ class CUDAGraphAllocator
: underlying_allocator_(allocator) {} : underlying_allocator_(allocator) {}
public: public:
~CUDAGraphAllocator() override {} ~CUDAGraphAllocator() override = default;
static std::shared_ptr<Allocator> Create( static std::shared_ptr<Allocator> Create(
const std::shared_ptr<Allocator>& allocator) { const std::shared_ptr<Allocator>& allocator) {
...@@ -1272,7 +1272,7 @@ AllocatorFacadePrivate::AllocatorMap AllocatorFacadePrivate::system_allocators_; ...@@ -1272,7 +1272,7 @@ AllocatorFacadePrivate::AllocatorMap AllocatorFacadePrivate::system_allocators_;
AllocatorFacade::AllocatorFacade() : m_(new AllocatorFacadePrivate()) {} AllocatorFacade::AllocatorFacade() : m_(new AllocatorFacadePrivate()) {}
// delete m_ may cause core dump when the destructor of python in conflict with // delete m_ may cause core dump when the destructor of python in conflict with
// cpp. // cpp.
AllocatorFacade::~AllocatorFacade() {} AllocatorFacade::~AllocatorFacade() = default;
AllocatorFacade& AllocatorFacade::Instance() { AllocatorFacade& AllocatorFacade::Instance() {
static AllocatorFacade* instance = new AllocatorFacade; static AllocatorFacade* instance = new AllocatorFacade;
......
...@@ -38,7 +38,7 @@ namespace operators { ...@@ -38,7 +38,7 @@ namespace operators {
class CCommInitAllInferShape : public framework::InferShapeBase { class CCommInitAllInferShape : public framework::InferShapeBase {
public: public:
~CCommInitAllInferShape() override {} ~CCommInitAllInferShape() override = default;
void operator()(framework::InferShapeContext* ctx) const override{}; void operator()(framework::InferShapeContext* ctx) const override{};
}; };
......
...@@ -38,7 +38,7 @@ namespace operators { ...@@ -38,7 +38,7 @@ namespace operators {
class CCommInitMultiTrainerInferShape : public framework::InferShapeBase { class CCommInitMultiTrainerInferShape : public framework::InferShapeBase {
public: public:
~CCommInitMultiTrainerInferShape() override {} ~CCommInitMultiTrainerInferShape() override = default;
void operator()(framework::InferShapeContext* ctx) const override{}; void operator()(framework::InferShapeContext* ctx) const override{};
}; };
......
...@@ -44,7 +44,7 @@ static __device__ __forceinline__ double RealSqrt(double x) { return sqrt(x); } ...@@ -44,7 +44,7 @@ static __device__ __forceinline__ double RealSqrt(double x) { return sqrt(x); }
template <typename T> template <typename T>
struct PairForLayerNorm { struct PairForLayerNorm {
__device__ __forceinline__ PairForLayerNorm() {} __device__ __forceinline__ PairForLayerNorm() = default;
__device__ __forceinline__ PairForLayerNorm(const T& first, const T& second) __device__ __forceinline__ PairForLayerNorm(const T& first, const T& second)
: first_(first), second_(second) {} : first_(first), second_(second) {}
......
...@@ -113,7 +113,7 @@ class BeamSearchFunctor<phi::CPUContext, T> { ...@@ -113,7 +113,7 @@ class BeamSearchFunctor<phi::CPUContext, T> {
* The basic items help to sort. * The basic items help to sort.
*/ */
struct Item { struct Item {
Item() {} Item() = default;
Item(size_t offset, size_t id, float score) Item(size_t offset, size_t id, float score)
: offset(offset), id(id), score(score) {} : offset(offset), id(id), score(score) {}
// offset in the higher lod level. // offset in the higher lod level.
......
...@@ -21,7 +21,7 @@ namespace operators { ...@@ -21,7 +21,7 @@ namespace operators {
namespace math { namespace math {
struct Triple { struct Triple {
__device__ __forceinline__ Triple() {} __device__ __forceinline__ Triple() = default;
__device__ __forceinline__ Triple(int o, int i, float s) __device__ __forceinline__ Triple(int o, int i, float s)
: offset(o), id(i), score(s) {} : offset(o), id(i), score(s) {}
......
...@@ -22,7 +22,7 @@ namespace paddle { ...@@ -22,7 +22,7 @@ namespace paddle {
namespace operators { namespace operators {
namespace math { namespace math {
Sampler::~Sampler() {} Sampler::~Sampler() = default;
UniformSampler::UniformSampler(int64_t range, unsigned int seed) UniformSampler::UniformSampler(int64_t range, unsigned int seed)
: Sampler(range, seed), inv_range_(1.0 / (range + 1)) { : Sampler(range, seed), inv_range_(1.0 / (range + 1)) {
......
...@@ -36,7 +36,9 @@ void PyReader::ReadNext(paddle::framework::LoDTensorArray* out) { ...@@ -36,7 +36,9 @@ void PyReader::ReadNext(paddle::framework::LoDTensorArray* out) {
if (!success) out->clear(); if (!success) out->clear();
} }
PyReader::~PyReader() { queue_->Close(); } PyReader::~PyReader() { // NOLINT
queue_->Close();
}
void PyReader::Shutdown() { queue_->Close(); } void PyReader::Shutdown() { queue_->Close(); }
......
...@@ -47,7 +47,7 @@ std::unordered_map<uint32_t, uint64_t> CreateThreadIdMapping() { ...@@ -47,7 +47,7 @@ std::unordered_map<uint32_t, uint64_t> CreateThreadIdMapping() {
} }
} // namespace details } // namespace details
CudaTracer::CudaTracer() {} CudaTracer::CudaTracer() = default;
void CudaTracer::PrepareTracing() { void CudaTracer::PrepareTracing() {
PADDLE_ENFORCE_EQ( PADDLE_ENFORCE_EQ(
......
...@@ -32,7 +32,7 @@ CustomTracer::CustomTracer(const std::string& dev_type) : dev_type_(dev_type) { ...@@ -32,7 +32,7 @@ CustomTracer::CustomTracer(const std::string& dev_type) : dev_type_(dev_type) {
#endif #endif
} }
CustomTracer::~CustomTracer() { CustomTracer::~CustomTracer() { // NOLINT
#ifdef PADDLE_WITH_CUSTOM_DEVICE #ifdef PADDLE_WITH_CUSTOM_DEVICE
phi::DeviceManager::ProfilerFinalize(dev_type_, &collector_, context_); phi::DeviceManager::ProfilerFinalize(dev_type_, &collector_, context_);
#endif #endif
......
...@@ -29,7 +29,7 @@ namespace ir { ...@@ -29,7 +29,7 @@ namespace ir {
// AbstractType, TypeStorage, AbstractAttribute, AttributeStorage, Dialect. // AbstractType, TypeStorage, AbstractAttribute, AttributeStorage, Dialect.
class IrContextImpl { class IrContextImpl {
public: public:
IrContextImpl() {} IrContextImpl() = default;
~IrContextImpl() { ~IrContextImpl() {
std::lock_guard<ir::SpinLock> guard(destructor_lock_); std::lock_guard<ir::SpinLock> guard(destructor_lock_);
......
...@@ -66,7 +66,7 @@ struct ParametricStorageManager { ...@@ -66,7 +66,7 @@ struct ParametricStorageManager {
std::function<void(StorageBase *)> destroy_; std::function<void(StorageBase *)> destroy_;
}; };
StorageManager::StorageManager() {} StorageManager::StorageManager() = default;
StorageManager::~StorageManager() = default; StorageManager::~StorageManager() = default;
......
...@@ -53,7 +53,7 @@ using IntArray = experimental::IntArray; ...@@ -53,7 +53,7 @@ using IntArray = experimental::IntArray;
class AbstractAutogradMeta { class AbstractAutogradMeta {
public: public:
// No AbstractAutogradMeta should be created // No AbstractAutogradMeta should be created
virtual ~AbstractAutogradMeta() {} virtual ~AbstractAutogradMeta() = default;
}; };
/** /**
......
...@@ -403,12 +403,7 @@ void Tensor::reset() { ...@@ -403,12 +403,7 @@ void Tensor::reset() {
/* Part 6: Operator overloading */ /* Part 6: Operator overloading */
Tensor &Tensor::operator=(const Tensor &x) & { Tensor &Tensor::operator=(const Tensor &x) & = default;
impl_ = x.impl_;
autograd_meta_ = x.autograd_meta_;
name_ = x.name_;
return *this;
}
Tensor &Tensor::operator=(Tensor &&x) & { Tensor &Tensor::operator=(Tensor &&x) & {
impl_ = std::move(x.impl_); impl_ = std::move(x.impl_);
......
...@@ -109,7 +109,7 @@ namespace phi { ...@@ -109,7 +109,7 @@ namespace phi {
namespace dynload { namespace dynload {
struct PathNode { struct PathNode {
PathNode() {} PathNode() = default;
std::string path = ""; std::string path = "";
}; };
......
...@@ -66,7 +66,7 @@ class EigenGpuStreamDevice : public Eigen::StreamInterface { ...@@ -66,7 +66,7 @@ class EigenGpuStreamDevice : public Eigen::StreamInterface {
EigenGpuStreamDevice() : scratch_(nullptr), semaphore_(nullptr) { EigenGpuStreamDevice() : scratch_(nullptr), semaphore_(nullptr) {
Eigen::initializeDeviceProp(); Eigen::initializeDeviceProp();
} }
~EigenGpuStreamDevice() override {} ~EigenGpuStreamDevice() override = default;
void Reinitialize(gpuStream_t cuda_stream, void Reinitialize(gpuStream_t cuda_stream,
Allocator* allocator, Allocator* allocator,
......
...@@ -95,7 +95,7 @@ struct OneDNNContext::Impl { ...@@ -95,7 +95,7 @@ struct OneDNNContext::Impl {
p_mutex_.reset(new std::mutex()); p_mutex_.reset(new std::mutex());
} }
~Impl() {} ~Impl() = default;
void ResetBlobMap(void* ptr) { void ResetBlobMap(void* ptr) {
VLOG(4) << OneDNNContext::tls().get_curr_exec() << " " << ptr; VLOG(4) << OneDNNContext::tls().get_curr_exec() << " " << ptr;
......
...@@ -27,7 +27,7 @@ template <class T> ...@@ -27,7 +27,7 @@ template <class T>
class Point_ { class Point_ {
public: public:
// default constructor // default constructor
Point_() {} Point_() = default;
Point_(T _x, T _y) {} Point_(T _x, T _y) {}
Point_(const Point_& pt UNUSED) {} Point_(const Point_& pt UNUSED) {}
......
...@@ -53,7 +53,7 @@ void CreateLstmGrad(phi::funcs::LstmMetaGrad<T>* lstm_grad) { ...@@ -53,7 +53,7 @@ void CreateLstmGrad(phi::funcs::LstmMetaGrad<T>* lstm_grad) {
template <typename T> template <typename T>
struct GradCell { struct GradCell {
virtual ~GradCell() {} virtual ~GradCell() = default;
virtual void operator()(const CPUContext& dev_ctx UNUSED, virtual void operator()(const CPUContext& dev_ctx UNUSED,
DenseTensor* gate_tensor UNUSED, DenseTensor* gate_tensor UNUSED,
DenseTensor* state_tensor UNUSED, DenseTensor* state_tensor UNUSED,
...@@ -355,7 +355,7 @@ struct LSTMGradCell : GradCell<T> { ...@@ -355,7 +355,7 @@ struct LSTMGradCell : GradCell<T> {
template <typename T, typename GradCellType> template <typename T, typename GradCellType>
struct GradLayer { struct GradLayer {
explicit GradLayer(const GradCellType& cell) : cell_(cell) {} explicit GradLayer(const GradCellType& cell) : cell_(cell) {}
virtual ~GradLayer() {} virtual ~GradLayer() = default;
void run_rnn_grad_function( void run_rnn_grad_function(
const CPUContext& dev_ctx, const CPUContext& dev_ctx,
const DenseTensor* input, const DenseTensor* input,
...@@ -690,7 +690,7 @@ struct SingleGradLayer : GradLayer<T, GradCellType> { ...@@ -690,7 +690,7 @@ struct SingleGradLayer : GradLayer<T, GradCellType> {
// explicit SingleGradLayer(GradCellType& cell) : cell_(cell) {} // explicit SingleGradLayer(GradCellType& cell) : cell_(cell) {}
explicit SingleGradLayer(const GradCellType& cell) explicit SingleGradLayer(const GradCellType& cell)
: GradLayer<T, GradCellType>(cell) {} : GradLayer<T, GradCellType>(cell) {}
~SingleGradLayer() override {} ~SingleGradLayer() override = default;
void operator()(const CPUContext& dev_ctx, void operator()(const CPUContext& dev_ctx,
const DenseTensor* input, const DenseTensor* input,
const DenseTensor* output, const DenseTensor* output,
...@@ -802,7 +802,7 @@ template <typename T, typename GradCellType> ...@@ -802,7 +802,7 @@ template <typename T, typename GradCellType>
struct BidirGradLayer : GradLayer<T, GradCellType> { struct BidirGradLayer : GradLayer<T, GradCellType> {
explicit BidirGradLayer(const GradCellType& cell) explicit BidirGradLayer(const GradCellType& cell)
: GradLayer<T, GradCellType>(cell) {} : GradLayer<T, GradCellType>(cell) {}
~BidirGradLayer() override {} ~BidirGradLayer() override = default;
void operator()(const CPUContext& dev_ctx, void operator()(const CPUContext& dev_ctx,
const DenseTensor* input, const DenseTensor* input,
const DenseTensor* output, const DenseTensor* output,
......
...@@ -34,7 +34,7 @@ namespace phi { ...@@ -34,7 +34,7 @@ namespace phi {
template <typename T> template <typename T>
struct Cell { struct Cell {
virtual ~Cell() {} virtual ~Cell() = default;
virtual void operator()(const CPUContext* dev_ctx UNUSED, virtual void operator()(const CPUContext* dev_ctx UNUSED,
DenseTensor* input UNUSED, DenseTensor* input UNUSED,
const DenseTensor* weight_hh UNUSED, const DenseTensor* weight_hh UNUSED,
...@@ -208,7 +208,7 @@ struct LSTMCell : Cell<T> { ...@@ -208,7 +208,7 @@ struct LSTMCell : Cell<T> {
template <typename T, typename CellType> template <typename T, typename CellType>
struct Layer { struct Layer {
explicit Layer(const CellType& cell) : cell_(cell) {} explicit Layer(const CellType& cell) : cell_(cell) {}
virtual ~Layer() {} virtual ~Layer() = default;
void preprocess(const CPUContext& dev_ctx, void preprocess(const CPUContext& dev_ctx,
const DenseTensor& input, const DenseTensor& input,
const DenseTensor& weight, const DenseTensor& weight,
......
...@@ -65,7 +65,7 @@ struct PointerWrapper { ...@@ -65,7 +65,7 @@ struct PointerWrapper {
const void* ins_addr[Size]; const void* ins_addr[Size];
__device__ inline const void* operator[](int i) const { return ins_addr[i]; } __device__ inline const void* operator[](int i) const { return ins_addr[i]; }
PointerWrapper() {} PointerWrapper() = default;
PointerWrapper(const phi::GPUContext& ctx, PointerWrapper(const phi::GPUContext& ctx,
const std::vector<phi::DenseTensor>& ins, const std::vector<phi::DenseTensor>& ins,
const T** pre_alloced_host_ptr) { const T** pre_alloced_host_ptr) {
...@@ -84,7 +84,7 @@ template <typename T, int Size> ...@@ -84,7 +84,7 @@ template <typename T, int Size>
struct PADDLE_ALIGN(256) AlignedPointerWrapper struct PADDLE_ALIGN(256) AlignedPointerWrapper
: public PointerWrapper<T, Size> { : public PointerWrapper<T, Size> {
public: public:
AlignedPointerWrapper() {} AlignedPointerWrapper() = default;
AlignedPointerWrapper(const phi::GPUContext& ctx, AlignedPointerWrapper(const phi::GPUContext& ctx,
const std::vector<phi::DenseTensor>& ins, const std::vector<phi::DenseTensor>& ins,
const T** pre_alloced_host_ptr) { const T** pre_alloced_host_ptr) {
...@@ -98,7 +98,7 @@ struct PointerToPointer { ...@@ -98,7 +98,7 @@ struct PointerToPointer {
void** ins_addr{nullptr}; void** ins_addr{nullptr};
__device__ inline const void* operator[](int i) const { return ins_addr[i]; } __device__ inline const void* operator[](int i) const { return ins_addr[i]; }
PointerToPointer() {} PointerToPointer() = default;
PointerToPointer(const phi::GPUContext& ctx, PointerToPointer(const phi::GPUContext& ctx,
const std::vector<phi::DenseTensor>& ins, const std::vector<phi::DenseTensor>& ins,
const T** pre_alloced_host_ptr, const T** pre_alloced_host_ptr,
...@@ -186,9 +186,7 @@ struct PointerToPointerAndCol { ...@@ -186,9 +186,7 @@ struct PointerToPointerAndCol {
template <int MovSize> template <int MovSize>
struct alignas(MovSize) Packed { struct alignas(MovSize) Packed {
__device__ Packed() { __device__ Packed() = default;
// do nothing
}
union { union {
char buf[MovSize]; char buf[MovSize];
}; };
...@@ -621,7 +619,7 @@ struct PointerAndColArray ...@@ -621,7 +619,7 @@ struct PointerAndColArray
public: public:
funcs::ValueArray<IndexT, Size> val_array; funcs::ValueArray<IndexT, Size> val_array;
PointerAndColArray() {} PointerAndColArray() = default;
PointerAndColArray(const phi::GPUContext& ctx, PointerAndColArray(const phi::GPUContext& ctx,
const int out_col_num, const int out_col_num,
IndexT* out_cols, IndexT* out_cols,
......
...@@ -32,7 +32,7 @@ namespace phi { ...@@ -32,7 +32,7 @@ namespace phi {
template <typename MT, typename InT, typename OutT> template <typename MT, typename InT, typename OutT>
struct MaskedSelectGradFunctor { struct MaskedSelectGradFunctor {
HOSTDEVICE MaskedSelectGradFunctor() {} HOSTDEVICE MaskedSelectGradFunctor() = default;
HOSTDEVICE inline void operator()(OutT* out, HOSTDEVICE inline void operator()(OutT* out,
const MT* mask, const MT* mask,
......
...@@ -30,7 +30,7 @@ namespace phi { ...@@ -30,7 +30,7 @@ namespace phi {
template <typename MT, typename InT, typename OutT> template <typename MT, typename InT, typename OutT>
struct MaskedSelectFunctor { struct MaskedSelectFunctor {
HOSTDEVICE MaskedSelectFunctor() {} HOSTDEVICE MaskedSelectFunctor() = default;
HOSTDEVICE inline void operator()(OutT* out, HOSTDEVICE inline void operator()(OutT* out,
const MT* mask, const MT* mask,
......
...@@ -63,7 +63,7 @@ __device__ __forceinline__ double inline_pow(double base, double exponent) { ...@@ -63,7 +63,7 @@ __device__ __forceinline__ double inline_pow(double base, double exponent) {
template <typename T> template <typename T>
struct NonzeroFunctor { struct NonzeroFunctor {
HOSTDEVICE explicit inline NonzeroFunctor() {} HOSTDEVICE explicit inline NonzeroFunctor() = default;
HOSTDEVICE inline T operator()(const T x) const { HOSTDEVICE inline T operator()(const T x) const {
return static_cast<T>(static_cast<double>(x) != 0); return static_cast<T>(static_cast<double>(x) != 0);
} }
...@@ -71,7 +71,7 @@ struct NonzeroFunctor { ...@@ -71,7 +71,7 @@ struct NonzeroFunctor {
template <typename T> template <typename T>
struct AbsFunctor { struct AbsFunctor {
HOSTDEVICE explicit inline AbsFunctor() {} HOSTDEVICE explicit inline AbsFunctor() = default;
HOSTDEVICE inline T operator()(const T x) const { HOSTDEVICE inline T operator()(const T x) const {
return static_cast<T>(inline_abs(x)); return static_cast<T>(inline_abs(x));
} }
......
...@@ -177,9 +177,7 @@ typename std::enable_if<HasCanPackAs<T>::value == false, bool>::type CanPackAs( ...@@ -177,9 +177,7 @@ typename std::enable_if<HasCanPackAs<T>::value == false, bool>::type CanPackAs(
template <typename T, int N> template <typename T, int N>
struct alignas(sizeof(T) * N) Pack { struct alignas(sizeof(T) * N) Pack {
__device__ Pack() { __device__ Pack() = default;
// do nothing
}
T elem[N]; T elem[N];
}; };
......
...@@ -375,7 +375,7 @@ class CudnnBNAddReluTester { ...@@ -375,7 +375,7 @@ class CudnnBNAddReluTester {
SetUp(); SetUp();
} }
~CudnnBNAddReluTester() {} ~CudnnBNAddReluTester() = default;
void CheckForward(float diff, bool is_relative_atol = false) { void CheckForward(float diff, bool is_relative_atol = false) {
LOG(INFO) << "[CheckForward, diff=" << diff LOG(INFO) << "[CheckForward, diff=" << diff
......
...@@ -235,7 +235,7 @@ class CudnnNormConvolutionTester { ...@@ -235,7 +235,7 @@ class CudnnNormConvolutionTester {
SetUp(); SetUp();
} }
~CudnnNormConvolutionTester() {} ~CudnnNormConvolutionTester() = default;
void CheckForward(float diff, bool is_relative_atol = false) { void CheckForward(float diff, bool is_relative_atol = false) {
phi::GPUContext *ctx = static_cast<phi::GPUContext *>( phi::GPUContext *ctx = static_cast<phi::GPUContext *>(
......
...@@ -89,7 +89,7 @@ struct TestFusedDropoutActBias { ...@@ -89,7 +89,7 @@ struct TestFusedDropoutActBias {
ctx = reinterpret_cast<phi::GPUContext *>(devicectx); ctx = reinterpret_cast<phi::GPUContext *>(devicectx);
} }
~TestFusedDropoutActBias() {} ~TestFusedDropoutActBias() = default;
void SetUp() { void SetUp() {
const int n = rows * cols; const int n = rows * cols;
......
...@@ -95,7 +95,7 @@ struct TestFusedLayernormResidualDropoutBias { ...@@ -95,7 +95,7 @@ struct TestFusedLayernormResidualDropoutBias {
ctx = reinterpret_cast<phi::GPUContext *>(devicectx); ctx = reinterpret_cast<phi::GPUContext *>(devicectx);
} }
~TestFusedLayernormResidualDropoutBias() {} ~TestFusedLayernormResidualDropoutBias() = default;
void SetUp() { void SetUp() {
using U = LayerNormParamType<T>; using U = LayerNormParamType<T>;
......
...@@ -35,7 +35,7 @@ struct DataRecord { ...@@ -35,7 +35,7 @@ struct DataRecord {
size_t batch_size{1}; size_t batch_size{1};
size_t num_samples; // total number of samples size_t num_samples; // total number of samples
DataRecord() { DataRecord() { // NOLINT
turns = new std::vector<std::vector< turns = new std::vector<std::vector<
int64_t>>[FLAGS_max_turn_num]; // turns data : FLAGS_max_turn_num int64_t>>[FLAGS_max_turn_num]; // turns data : FLAGS_max_turn_num
turns_mask = new std::vector<std::vector< turns_mask = new std::vector<std::vector<
...@@ -48,7 +48,7 @@ struct DataRecord { ...@@ -48,7 +48,7 @@ struct DataRecord {
Load(path); Load(path);
} }
~DataRecord() { ~DataRecord() { // NOLINT
delete[] turns; delete[] turns;
delete[] turns_mask; delete[] turns_mask;
} }
......
...@@ -50,7 +50,7 @@ class DemoPredictor : public PaddlePredictor { ...@@ -50,7 +50,7 @@ class DemoPredictor : public PaddlePredictor {
return nullptr; return nullptr;
} }
~DemoPredictor() override {} ~DemoPredictor() override = default;
}; };
template <> template <>
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册