未验证 提交 39ffef0d 编写于 作者: P PuQing 提交者: GitHub

[PHI decoupling] move "flags.h" from fluid to phi (#48696)

上级 dd57860d
......@@ -20,7 +20,7 @@
#include "paddle/fluid/distributed/store/tcp_utils.h"
#include "paddle/fluid/platform/enforce.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
namespace paddle {
namespace distributed {
......
......@@ -25,7 +25,7 @@
#include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h"
#include "paddle/fluid/eager/tests/test_utils.h"
#include "paddle/fluid/imperative/tracer.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
#ifdef WITH_GPERFTOOLS
#include "gperftools/profiler.h"
......
......@@ -24,7 +24,7 @@
#include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h"
#include "paddle/fluid/eager/tests/test_utils.h"
#include "paddle/fluid/imperative/tracer.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
#ifdef WITH_GPERFTOOLS
#include "gperftools/profiler.h"
......
......@@ -24,9 +24,9 @@
#include <vector>
#include "glog/logging.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/os_info.h"
#include "paddle/fluid/platform/profiler/utils.h"
#include "paddle/phi/core/flags.h"
DECLARE_bool(use_stream_safe_cuda_allocator);
PADDLE_DEFINE_EXPORTED_string(static_executor_perfstat_filepath,
......
......@@ -14,7 +14,7 @@
#include "paddle/fluid/imperative/flags.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_uint64(dygraph_debug,
0,
......
......@@ -21,7 +21,7 @@
#include <mutex> // NOLINT
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_string(
tracer_profile_fname,
......
......@@ -1422,8 +1422,7 @@ CreatePaddlePredictor<AnalysisConfig, PaddleEngineKind::kAnalysis>(
}
// support set flags from enviorment.
const platform::ExportedFlagInfoMap &env_map =
platform::GetExportedFlagInfoMap();
const phi::ExportedFlagInfoMap &env_map = phi::GetExportedFlagInfoMap();
std::ostringstream os;
os << "--tryfromenv=";
for (auto &pair : env_map) {
......
......@@ -18,8 +18,8 @@
#include <mutex> // NOLINT
#include "paddle/fluid/memory/allocation/aligned_allocator.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/profiler/event_tracing.h"
#include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_READONLY_bool(
free_idle_chunk,
......
......@@ -17,7 +17,7 @@ limitations under the License. */
#include "paddle/fluid/framework/new_executor/standalone_executor.h"
#include "paddle/fluid/operators/assign_op.h"
#include "paddle/fluid/operators/controlflow/control_flow_op_helper.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/kernels/funcs/math_function.h"
#ifdef PADDLE_WITH_MKLDNN
......
......@@ -25,7 +25,6 @@ limitations under the License. */
#include "paddle/fluid/platform/mkldnn_helper.h"
#endif
#include "paddle/fluid/framework/infershape_utils.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/infermeta/binary.h"
namespace paddle {
......
......@@ -21,7 +21,6 @@ limitations under the License. */
#include "paddle/fluid/framework/infershape_utils.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/op_version_registry.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/core/infermeta_utils.h"
#include "paddle/phi/infermeta/backward.h"
#include "paddle/phi/infermeta/binary.h"
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include <vector>
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
namespace paddle {
namespace operators {
......@@ -113,7 +113,7 @@ class ConvInceptionFusionOpMaker : public framework::OpProtoAndCheckerMaker {
"allocated/freed each time the operator runs, larger "
"workspace size can increase performance but also requires "
"better hardware. This size should be chosen carefully.")
.SetDefault(platform::GetDefaultConvWorkspaceSizeLimitMB());
.SetDefault(phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB());
AddComment(R"DOC(
)DOC");
}
......
......@@ -31,10 +31,6 @@ if(WITH_PYTHON)
endif()
endif()
cc_library(
flags
SRCS flags.cc
DEPS gflags)
cc_library(
denormal
SRCS denormal.cc
......@@ -178,11 +174,6 @@ if(WITH_GLOO)
DEPS framework_proto gloo_wrapper enforce)
endif()
cc_library(
cudnn_workspace_helper
SRCS cudnn_workspace_helper.cc
DEPS)
# separate init from device_context to avoid cycle dependencies
cc_library(
init
......
......@@ -32,7 +32,7 @@ limitations under the License. */
#include <algorithm>
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
DECLARE_double(fraction_of_cpu_memory_to_use);
DECLARE_uint64(initial_cpu_memory_in_mb);
......
......@@ -24,7 +24,6 @@ limitations under the License. */
#include "paddle/fluid/memory/memory.h"
#include "paddle/fluid/platform/cuda_device_guard.h"
#include "paddle/fluid/platform/enforce.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/lock_guard_ptr.h"
#include "paddle/fluid/platform/macros.h"
#include "paddle/fluid/platform/monitor.h"
......@@ -32,6 +31,7 @@ limitations under the License. */
#include "paddle/fluid/platform/profiler/mem_tracing.h"
#include "paddle/fluid/string/split.h"
#include "paddle/phi/backends/gpu/gpu_info.h"
#include "paddle/phi/core/flags.h"
#ifdef PADDLE_WITH_HIP
#include "paddle/fluid/platform/dynload/miopen.h"
......
......@@ -101,7 +101,7 @@ limitations under the License. */
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_types.h"
#endif
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
namespace phi {
class ErrorSummary;
......
......@@ -17,9 +17,9 @@
#include "glog/logging.h"
#include "paddle/fluid/framework/op_proto_maker.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/profiler/common_event.h"
#include "paddle/fluid/platform/profiler/host_event_recorder.h"
#include "paddle/phi/core/flags.h"
// Used to filter events, works like glog VLOG(level).
// RecordEvent will works if host_trace_level >= level.
......
......@@ -252,7 +252,7 @@ static void RegisterGlobalVarGetterSetter() {
REGISTER_PUBLIC_GLOBAL_VAR(FLAGS_rpc_prefetch_thread_num);
#endif
const auto &flag_map = platform::GetExportedFlagInfoMap();
const auto &flag_map = phi::GetExportedFlagInfoMap();
for (const auto &pair : flag_map) {
const std::string &name = pair.second.name;
bool is_writable = pair.second.is_writable;
......
......@@ -22,7 +22,7 @@ def map_code_template(attrs_str, attrs_checker_str):
return f"""// This file is generated by paddle/phi/api/yaml/generator/ops_extra_info_gen.py
#include "paddle/fluid/operators/ops_extra_info.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
namespace paddle {{
namespace operators {{
......
......@@ -202,7 +202,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv2d_fusion
extra :
......@@ -211,7 +211,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv2d_transpose
backward : conv2d_transpose_grad
......@@ -219,7 +219,7 @@
attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, bool force_fp32_output = false,
str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : conv3d
backward : conv3d_grad
......@@ -227,12 +227,12 @@
attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
bool use_addto = false, bool fuse_residual_connection = false, bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv3d_transpose
backward : conv3d_transpose_grad
extra :
attrs : [bool use_cudnn = true, bool use_mkldnn = false, int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()]
attrs : [bool use_cudnn = true, bool use_mkldnn = false, int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : cos
backward : cos_grad, cos_double_grad, cos_triple_grad
......@@ -273,7 +273,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : depthwise_conv2d_transpose
backward : depthwise_conv2d_transpose_grad
......@@ -281,7 +281,7 @@
attrs : [bool is_test = false, bool use_cudnn = false, bool use_mkldnn = false, bool force_fp32_output = false,
str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()]
int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : dequantize_linear
extra :
......
add_subdirectory(dynload)
add_subdirectory(gpu)
set(BACKENDS_SRCS all_context.cc cpu/cpu_context.cc)
set(BACKENDS_DEPS enforce place flags eigen3 phi_device_context)
......
cc_library(cudnn_workspace_helper SRCS cudnn_workspace_helper.cc)
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
......@@ -12,13 +12,14 @@
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include <cstdlib>
#include <string>
namespace paddle {
namespace platform {
namespace phi {
namespace backends {
namespace gpu {
static int GetDefaultConvWorkspaceSizeLimitMBImpl() {
const char *env_str = std::getenv("FLAGS_conv_workspace_size_limit");
......@@ -30,6 +31,6 @@ int GetDefaultConvWorkspaceSizeLimitMB() {
static auto workspace_size = GetDefaultConvWorkspaceSizeLimitMBImpl();
return workspace_size;
}
} // namespace platform
} // namespace paddle
} // namespace gpu
} // namespace backends
} // namespace phi
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
......@@ -14,12 +14,13 @@
#pragma once
namespace paddle {
namespace platform {
namespace phi {
namespace backends {
namespace gpu {
static constexpr int kDefaultConvWorkspaceSizeLimitMB = 512;
int GetDefaultConvWorkspaceSizeLimitMB();
} // namespace platform
} // namespace paddle
} // namespace gpu
} // namespace backends
} // namespace phi
......@@ -21,7 +21,7 @@ limitations under the License. */
// TODO(wilber): The phi computing library requires a component to manage
// flags.
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_string(
selected_xpus,
......
......@@ -5,6 +5,11 @@ if(WITH_GPU)
proto_library(external_error_proto SRCS external_error.proto)
endif()
cc_library(
flags
SRCS flags.cc
DEPS gflags)
cc_library(errors SRCS errors.cc)
set(phi_enforce_deps errors flags)
if(WITH_GPU)
......
......@@ -13,13 +13,12 @@
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/core/flags.h"
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#endif
namespace paddle {
namespace platform {
namespace phi {
const ExportedFlagInfoMap &GetExportedFlagInfoMap() {
return *GetMutableExportedFlagInfoMap();
......@@ -30,8 +29,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap() {
return &g_exported_flag_info_map;
}
} // namespace platform
} // namespace paddle
} // namespace phi
PADDLE_DEFINE_EXPORTED_int32(inner_op_parallelism,
0,
......@@ -261,8 +259,9 @@ PADDLE_DEFINE_EXPORTED_bool(
* increased.
* Users need to balance memory and speed.
*/
PADDLE_DEFINE_EXPORTED_int64(conv_workspace_size_limit,
paddle::platform::kDefaultConvWorkspaceSizeLimitMB,
PADDLE_DEFINE_EXPORTED_int64(
conv_workspace_size_limit,
phi::backends::gpu::kDefaultConvWorkspaceSizeLimitMB,
"cuDNN convolution workspace limit in MB unit.");
/**
......
......@@ -20,12 +20,11 @@
#include <type_traits>
#include "gflags/gflags.h"
#include "paddle/fluid/platform/macros.h"
#include "paddle/phi/core/macros.h"
#include "paddle/utils/variant.h"
namespace paddle {
namespace platform {
namespace phi {
struct FlagInfo {
using ValueType =
......@@ -51,7 +50,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap();
static_assert(std::is_same<FlagDeclaredType, ::std::string>::value || \
std::is_arithmetic<FlagDeclaredType>::value, \
"FLAGS should be std::string or arithmetic type"); \
auto *instance = ::paddle::platform::GetMutableExportedFlagInfoMap(); \
auto *instance = ::phi::GetMutableExportedFlagInfoMap(); \
auto &info = (*instance)[#__name]; \
info.name = #__name; \
info.value_ptr = &(FLAGS_##__name); \
......@@ -96,5 +95,4 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap();
__PADDLE_DEFINE_EXPORTED_FLAG( \
name, true, ::std::string, string, default_value, doc)
} // namespace platform
} // namespace paddle
} // namespace phi
......@@ -14,11 +14,11 @@
#include "paddle/fluid/operators/layout_utils.h"
#include "paddle/fluid/operators/norm_utils.cu.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/backends/gpu/gpu_dnn.h"
#include "paddle/phi/common/layout.h"
#include "paddle/phi/core/enforce.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/core/kernel_registry.h"
#include "paddle/phi/kernels/batch_norm_kernel.h"
#include "paddle/phi/kernels/empty_kernel.h"
......
......@@ -22,11 +22,11 @@ namespace cub = hipcub;
#include "paddle/fluid/operators/layout_utils.h"
#include "paddle/fluid/operators/norm_utils.cu.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/backends/gpu/gpu_dnn.h"
#include "paddle/phi/common/layout.h"
#include "paddle/phi/core/enforce.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/core/kernel_registry.h"
#include "paddle/phi/kernels/batch_norm_kernel.h"
#include "paddle/phi/kernels/funcs/batch_norm_utils.h"
......
......@@ -14,10 +14,10 @@
#pragma once
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/common/amp_type_traits.h"
#include "paddle/phi/common/place.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/kernels/funcs/aligned_vector.h"
DECLARE_bool(use_fast_math);
......
......@@ -23,8 +23,8 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/bfloat16.h"
#include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h"
......
......@@ -24,8 +24,8 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/bfloat16.h"
#include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h"
......
......@@ -23,9 +23,9 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/dynload/cudnn.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h"
#include "paddle/phi/kernels/funcs/batch_norm_utils.h"
......
......@@ -17,8 +17,8 @@ limitations under the License. */
#include "paddle/fluid/framework/phi_utils.h"
#include "paddle/fluid/memory/allocation/allocator_strategy.h"
#include "paddle/fluid/platform/device/npu/npu_info.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/init.h"
#include "paddle/phi/core/flags.h"
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
DECLARE_bool(enable_gpu_memory_usage_log);
......@@ -43,7 +43,7 @@ int main(int argc, char** argv) {
}
#endif
const auto& flag_map = paddle::platform::GetExportedFlagInfoMap();
const auto& flag_map = phi::GetExportedFlagInfoMap();
for (const auto& pair : flag_map) {
const std::string& name = pair.second.name;
// NOTE(zhiqiu): some names may not linked in some tests, so add to
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册