未验证 提交 39ffef0d 编写于 作者: P PuQing 提交者: GitHub

[PHI decoupling] move "flags.h" from fluid to phi (#48696)

上级 dd57860d
...@@ -20,7 +20,7 @@ ...@@ -20,7 +20,7 @@
#include "paddle/fluid/distributed/store/tcp_utils.h" #include "paddle/fluid/distributed/store/tcp_utils.h"
#include "paddle/fluid/platform/enforce.h" #include "paddle/fluid/platform/enforce.h"
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
namespace paddle { namespace paddle {
namespace distributed { namespace distributed {
......
...@@ -25,7 +25,7 @@ ...@@ -25,7 +25,7 @@
#include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h" #include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h"
#include "paddle/fluid/eager/tests/test_utils.h" #include "paddle/fluid/eager/tests/test_utils.h"
#include "paddle/fluid/imperative/tracer.h" #include "paddle/fluid/imperative/tracer.h"
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
#ifdef WITH_GPERFTOOLS #ifdef WITH_GPERFTOOLS
#include "gperftools/profiler.h" #include "gperftools/profiler.h"
......
...@@ -24,7 +24,7 @@ ...@@ -24,7 +24,7 @@
#include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h" #include "paddle/fluid/eager/tests/performance_tests/benchmark_utils.h"
#include "paddle/fluid/eager/tests/test_utils.h" #include "paddle/fluid/eager/tests/test_utils.h"
#include "paddle/fluid/imperative/tracer.h" #include "paddle/fluid/imperative/tracer.h"
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
#ifdef WITH_GPERFTOOLS #ifdef WITH_GPERFTOOLS
#include "gperftools/profiler.h" #include "gperftools/profiler.h"
......
...@@ -24,9 +24,9 @@ ...@@ -24,9 +24,9 @@
#include <vector> #include <vector>
#include "glog/logging.h" #include "glog/logging.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/os_info.h" #include "paddle/fluid/platform/os_info.h"
#include "paddle/fluid/platform/profiler/utils.h" #include "paddle/fluid/platform/profiler/utils.h"
#include "paddle/phi/core/flags.h"
DECLARE_bool(use_stream_safe_cuda_allocator); DECLARE_bool(use_stream_safe_cuda_allocator);
PADDLE_DEFINE_EXPORTED_string(static_executor_perfstat_filepath, PADDLE_DEFINE_EXPORTED_string(static_executor_perfstat_filepath,
......
...@@ -14,7 +14,7 @@ ...@@ -14,7 +14,7 @@
#include "paddle/fluid/imperative/flags.h" #include "paddle/fluid/imperative/flags.h"
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_uint64(dygraph_debug, PADDLE_DEFINE_EXPORTED_uint64(dygraph_debug,
0, 0,
......
...@@ -21,7 +21,7 @@ ...@@ -21,7 +21,7 @@
#include <mutex> // NOLINT #include <mutex> // NOLINT
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_string( PADDLE_DEFINE_EXPORTED_string(
tracer_profile_fname, tracer_profile_fname,
......
...@@ -1422,8 +1422,7 @@ CreatePaddlePredictor<AnalysisConfig, PaddleEngineKind::kAnalysis>( ...@@ -1422,8 +1422,7 @@ CreatePaddlePredictor<AnalysisConfig, PaddleEngineKind::kAnalysis>(
} }
// support set flags from enviorment. // support set flags from enviorment.
const platform::ExportedFlagInfoMap &env_map = const phi::ExportedFlagInfoMap &env_map = phi::GetExportedFlagInfoMap();
platform::GetExportedFlagInfoMap();
std::ostringstream os; std::ostringstream os;
os << "--tryfromenv="; os << "--tryfromenv=";
for (auto &pair : env_map) { for (auto &pair : env_map) {
......
...@@ -18,8 +18,8 @@ ...@@ -18,8 +18,8 @@
#include <mutex> // NOLINT #include <mutex> // NOLINT
#include "paddle/fluid/memory/allocation/aligned_allocator.h" #include "paddle/fluid/memory/allocation/aligned_allocator.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/profiler/event_tracing.h" #include "paddle/fluid/platform/profiler/event_tracing.h"
#include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_READONLY_bool( PADDLE_DEFINE_EXPORTED_READONLY_bool(
free_idle_chunk, free_idle_chunk,
......
...@@ -17,7 +17,7 @@ limitations under the License. */ ...@@ -17,7 +17,7 @@ limitations under the License. */
#include "paddle/fluid/framework/new_executor/standalone_executor.h" #include "paddle/fluid/framework/new_executor/standalone_executor.h"
#include "paddle/fluid/operators/assign_op.h" #include "paddle/fluid/operators/assign_op.h"
#include "paddle/fluid/operators/controlflow/control_flow_op_helper.h" #include "paddle/fluid/operators/controlflow/control_flow_op_helper.h"
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
#include "paddle/phi/kernels/funcs/math_function.h" #include "paddle/phi/kernels/funcs/math_function.h"
#ifdef PADDLE_WITH_MKLDNN #ifdef PADDLE_WITH_MKLDNN
......
...@@ -25,7 +25,6 @@ limitations under the License. */ ...@@ -25,7 +25,6 @@ limitations under the License. */
#include "paddle/fluid/platform/mkldnn_helper.h" #include "paddle/fluid/platform/mkldnn_helper.h"
#endif #endif
#include "paddle/fluid/framework/infershape_utils.h" #include "paddle/fluid/framework/infershape_utils.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/infermeta/binary.h" #include "paddle/phi/infermeta/binary.h"
namespace paddle { namespace paddle {
......
...@@ -21,7 +21,6 @@ limitations under the License. */ ...@@ -21,7 +21,6 @@ limitations under the License. */
#include "paddle/fluid/framework/infershape_utils.h" #include "paddle/fluid/framework/infershape_utils.h"
#include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/op_version_registry.h" #include "paddle/fluid/framework/op_version_registry.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/phi/core/infermeta_utils.h" #include "paddle/phi/core/infermeta_utils.h"
#include "paddle/phi/infermeta/backward.h" #include "paddle/phi/infermeta/backward.h"
#include "paddle/phi/infermeta/binary.h" #include "paddle/phi/infermeta/binary.h"
......
...@@ -16,7 +16,7 @@ limitations under the License. */ ...@@ -16,7 +16,7 @@ limitations under the License. */
#include <vector> #include <vector>
#include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h" #include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
namespace paddle { namespace paddle {
namespace operators { namespace operators {
...@@ -113,7 +113,7 @@ class ConvInceptionFusionOpMaker : public framework::OpProtoAndCheckerMaker { ...@@ -113,7 +113,7 @@ class ConvInceptionFusionOpMaker : public framework::OpProtoAndCheckerMaker {
"allocated/freed each time the operator runs, larger " "allocated/freed each time the operator runs, larger "
"workspace size can increase performance but also requires " "workspace size can increase performance but also requires "
"better hardware. This size should be chosen carefully.") "better hardware. This size should be chosen carefully.")
.SetDefault(platform::GetDefaultConvWorkspaceSizeLimitMB()); .SetDefault(phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB());
AddComment(R"DOC( AddComment(R"DOC(
)DOC"); )DOC");
} }
......
...@@ -31,10 +31,6 @@ if(WITH_PYTHON) ...@@ -31,10 +31,6 @@ if(WITH_PYTHON)
endif() endif()
endif() endif()
cc_library(
flags
SRCS flags.cc
DEPS gflags)
cc_library( cc_library(
denormal denormal
SRCS denormal.cc SRCS denormal.cc
...@@ -178,11 +174,6 @@ if(WITH_GLOO) ...@@ -178,11 +174,6 @@ if(WITH_GLOO)
DEPS framework_proto gloo_wrapper enforce) DEPS framework_proto gloo_wrapper enforce)
endif() endif()
cc_library(
cudnn_workspace_helper
SRCS cudnn_workspace_helper.cc
DEPS)
# separate init from device_context to avoid cycle dependencies # separate init from device_context to avoid cycle dependencies
cc_library( cc_library(
init init
......
...@@ -32,7 +32,7 @@ limitations under the License. */ ...@@ -32,7 +32,7 @@ limitations under the License. */
#include <algorithm> #include <algorithm>
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
DECLARE_double(fraction_of_cpu_memory_to_use); DECLARE_double(fraction_of_cpu_memory_to_use);
DECLARE_uint64(initial_cpu_memory_in_mb); DECLARE_uint64(initial_cpu_memory_in_mb);
......
...@@ -24,7 +24,6 @@ limitations under the License. */ ...@@ -24,7 +24,6 @@ limitations under the License. */
#include "paddle/fluid/memory/memory.h" #include "paddle/fluid/memory/memory.h"
#include "paddle/fluid/platform/cuda_device_guard.h" #include "paddle/fluid/platform/cuda_device_guard.h"
#include "paddle/fluid/platform/enforce.h" #include "paddle/fluid/platform/enforce.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/lock_guard_ptr.h" #include "paddle/fluid/platform/lock_guard_ptr.h"
#include "paddle/fluid/platform/macros.h" #include "paddle/fluid/platform/macros.h"
#include "paddle/fluid/platform/monitor.h" #include "paddle/fluid/platform/monitor.h"
...@@ -32,6 +31,7 @@ limitations under the License. */ ...@@ -32,6 +31,7 @@ limitations under the License. */
#include "paddle/fluid/platform/profiler/mem_tracing.h" #include "paddle/fluid/platform/profiler/mem_tracing.h"
#include "paddle/fluid/string/split.h" #include "paddle/fluid/string/split.h"
#include "paddle/phi/backends/gpu/gpu_info.h" #include "paddle/phi/backends/gpu/gpu_info.h"
#include "paddle/phi/core/flags.h"
#ifdef PADDLE_WITH_HIP #ifdef PADDLE_WITH_HIP
#include "paddle/fluid/platform/dynload/miopen.h" #include "paddle/fluid/platform/dynload/miopen.h"
......
...@@ -101,7 +101,7 @@ limitations under the License. */ ...@@ -101,7 +101,7 @@ limitations under the License. */
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_types.h" #include "paddle/fluid/platform/device/gpu/gpu_types.h"
#endif #endif
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
namespace phi { namespace phi {
class ErrorSummary; class ErrorSummary;
......
...@@ -17,9 +17,9 @@ ...@@ -17,9 +17,9 @@
#include "glog/logging.h" #include "glog/logging.h"
#include "paddle/fluid/framework/op_proto_maker.h" #include "paddle/fluid/framework/op_proto_maker.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/profiler/common_event.h" #include "paddle/fluid/platform/profiler/common_event.h"
#include "paddle/fluid/platform/profiler/host_event_recorder.h" #include "paddle/fluid/platform/profiler/host_event_recorder.h"
#include "paddle/phi/core/flags.h"
// Used to filter events, works like glog VLOG(level). // Used to filter events, works like glog VLOG(level).
// RecordEvent will works if host_trace_level >= level. // RecordEvent will works if host_trace_level >= level.
......
...@@ -252,7 +252,7 @@ static void RegisterGlobalVarGetterSetter() { ...@@ -252,7 +252,7 @@ static void RegisterGlobalVarGetterSetter() {
REGISTER_PUBLIC_GLOBAL_VAR(FLAGS_rpc_prefetch_thread_num); REGISTER_PUBLIC_GLOBAL_VAR(FLAGS_rpc_prefetch_thread_num);
#endif #endif
const auto &flag_map = platform::GetExportedFlagInfoMap(); const auto &flag_map = phi::GetExportedFlagInfoMap();
for (const auto &pair : flag_map) { for (const auto &pair : flag_map) {
const std::string &name = pair.second.name; const std::string &name = pair.second.name;
bool is_writable = pair.second.is_writable; bool is_writable = pair.second.is_writable;
......
...@@ -22,7 +22,7 @@ def map_code_template(attrs_str, attrs_checker_str): ...@@ -22,7 +22,7 @@ def map_code_template(attrs_str, attrs_checker_str):
return f"""// This file is generated by paddle/phi/api/yaml/generator/ops_extra_info_gen.py return f"""// This file is generated by paddle/phi/api/yaml/generator/ops_extra_info_gen.py
#include "paddle/fluid/operators/ops_extra_info.h" #include "paddle/fluid/operators/ops_extra_info.h"
#include "paddle/fluid/platform/cudnn_workspace_helper.h" #include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
namespace paddle {{ namespace paddle {{
namespace operators {{ namespace operators {{
......
...@@ -202,7 +202,7 @@ ...@@ -202,7 +202,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f, bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false, float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv2d_fusion - op : conv2d_fusion
extra : extra :
...@@ -211,7 +211,7 @@ ...@@ -211,7 +211,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f, bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false, float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv2d_transpose - op : conv2d_transpose
backward : conv2d_transpose_grad backward : conv2d_transpose_grad
...@@ -219,7 +219,7 @@ ...@@ -219,7 +219,7 @@
attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, bool force_fp32_output = false, attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, bool force_fp32_output = false,
str mkldnn_data_type = "float32", bool fuse_relu = false, str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : conv3d - op : conv3d
backward : conv3d_grad backward : conv3d_grad
...@@ -227,12 +227,12 @@ ...@@ -227,12 +227,12 @@
attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, str mkldnn_data_type = "float32", bool fuse_relu = false, attrs : [bool is_test = false, bool use_cudnn = true, bool use_mkldnn = false, str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
bool use_addto = false, bool fuse_residual_connection = false, bool force_fp32_output = false, bool use_addto = false, bool fuse_residual_connection = false, bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : conv3d_transpose - op : conv3d_transpose
backward : conv3d_transpose_grad backward : conv3d_transpose_grad
extra : extra :
attrs : [bool use_cudnn = true, bool use_mkldnn = false, int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()] attrs : [bool use_cudnn = true, bool use_mkldnn = false, int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : cos - op : cos
backward : cos_grad, cos_double_grad, cos_triple_grad backward : cos_grad, cos_double_grad, cos_triple_grad
...@@ -273,7 +273,7 @@ ...@@ -273,7 +273,7 @@
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, bool use_addto = false,
bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f, bool fuse_residual_connection = false, float Scale_in = 1.0f, float Scale_out = 1.0f,
float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false, float Scale_in_eltwise = 1.0f, 'float[] Scale_weights = {1.0f}', bool force_fp32_output = false,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB(), bool exhaustive_search = false]
- op : depthwise_conv2d_transpose - op : depthwise_conv2d_transpose
backward : depthwise_conv2d_transpose_grad backward : depthwise_conv2d_transpose_grad
...@@ -281,7 +281,7 @@ ...@@ -281,7 +281,7 @@
attrs : [bool is_test = false, bool use_cudnn = false, bool use_mkldnn = false, bool force_fp32_output = false, attrs : [bool is_test = false, bool use_cudnn = false, bool use_mkldnn = false, bool force_fp32_output = false,
str mkldnn_data_type = "float32", bool fuse_relu = false, str mkldnn_data_type = "float32", bool fuse_relu = false,
str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f, str fuse_activation = "", float fuse_alpha = 0.0f, float fuse_beta = 0.0f,
int workspace_size_MB = platform::GetDefaultConvWorkspaceSizeLimitMB()] int workspace_size_MB = phi::backends::gpu::GetDefaultConvWorkspaceSizeLimitMB()]
- op : dequantize_linear - op : dequantize_linear
extra : extra :
......
add_subdirectory(dynload) add_subdirectory(dynload)
add_subdirectory(gpu)
set(BACKENDS_SRCS all_context.cc cpu/cpu_context.cc) set(BACKENDS_SRCS all_context.cc cpu/cpu_context.cc)
set(BACKENDS_DEPS enforce place flags eigen3 phi_device_context) set(BACKENDS_DEPS enforce place flags eigen3 phi_device_context)
......
cc_library(cudnn_workspace_helper SRCS cudnn_workspace_helper.cc)
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
// //
// Licensed under the Apache License, Version 2.0 (the "License"); // Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License. // you may not use this file except in compliance with the License.
...@@ -12,13 +12,14 @@ ...@@ -12,13 +12,14 @@
// See the License for the specific language governing permissions and // See the License for the specific language governing permissions and
// limitations under the License. // limitations under the License.
#include "paddle/fluid/platform/cudnn_workspace_helper.h" #include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include <cstdlib> #include <cstdlib>
#include <string> #include <string>
namespace paddle { namespace phi {
namespace platform { namespace backends {
namespace gpu {
static int GetDefaultConvWorkspaceSizeLimitMBImpl() { static int GetDefaultConvWorkspaceSizeLimitMBImpl() {
const char *env_str = std::getenv("FLAGS_conv_workspace_size_limit"); const char *env_str = std::getenv("FLAGS_conv_workspace_size_limit");
...@@ -30,6 +31,6 @@ int GetDefaultConvWorkspaceSizeLimitMB() { ...@@ -30,6 +31,6 @@ int GetDefaultConvWorkspaceSizeLimitMB() {
static auto workspace_size = GetDefaultConvWorkspaceSizeLimitMBImpl(); static auto workspace_size = GetDefaultConvWorkspaceSizeLimitMBImpl();
return workspace_size; return workspace_size;
} }
} // namespace gpu
} // namespace platform } // namespace backends
} // namespace paddle } // namespace phi
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
// //
// Licensed under the Apache License, Version 2.0 (the "License"); // Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License. // you may not use this file except in compliance with the License.
...@@ -14,12 +14,13 @@ ...@@ -14,12 +14,13 @@
#pragma once #pragma once
namespace paddle { namespace phi {
namespace platform { namespace backends {
namespace gpu {
static constexpr int kDefaultConvWorkspaceSizeLimitMB = 512; static constexpr int kDefaultConvWorkspaceSizeLimitMB = 512;
int GetDefaultConvWorkspaceSizeLimitMB(); int GetDefaultConvWorkspaceSizeLimitMB();
} // namespace gpu
} // namespace platform } // namespace backends
} // namespace paddle } // namespace phi
...@@ -21,7 +21,7 @@ limitations under the License. */ ...@@ -21,7 +21,7 @@ limitations under the License. */
// TODO(wilber): The phi computing library requires a component to manage // TODO(wilber): The phi computing library requires a component to manage
// flags. // flags.
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
PADDLE_DEFINE_EXPORTED_string( PADDLE_DEFINE_EXPORTED_string(
selected_xpus, selected_xpus,
......
...@@ -5,6 +5,11 @@ if(WITH_GPU) ...@@ -5,6 +5,11 @@ if(WITH_GPU)
proto_library(external_error_proto SRCS external_error.proto) proto_library(external_error_proto SRCS external_error.proto)
endif() endif()
cc_library(
flags
SRCS flags.cc
DEPS gflags)
cc_library(errors SRCS errors.cc) cc_library(errors SRCS errors.cc)
set(phi_enforce_deps errors flags) set(phi_enforce_deps errors flags)
if(WITH_GPU) if(WITH_GPU)
......
...@@ -13,13 +13,12 @@ ...@@ -13,13 +13,12 @@
// See the License for the specific language governing permissions and // See the License for the specific language governing permissions and
// limitations under the License. // limitations under the License.
#include "paddle/fluid/platform/flags.h" #include "paddle/phi/core/flags.h"
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/cudnn_workspace_helper.h" #include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#endif #endif
namespace paddle { namespace phi {
namespace platform {
const ExportedFlagInfoMap &GetExportedFlagInfoMap() { const ExportedFlagInfoMap &GetExportedFlagInfoMap() {
return *GetMutableExportedFlagInfoMap(); return *GetMutableExportedFlagInfoMap();
...@@ -30,8 +29,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap() { ...@@ -30,8 +29,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap() {
return &g_exported_flag_info_map; return &g_exported_flag_info_map;
} }
} // namespace platform } // namespace phi
} // namespace paddle
PADDLE_DEFINE_EXPORTED_int32(inner_op_parallelism, PADDLE_DEFINE_EXPORTED_int32(inner_op_parallelism,
0, 0,
...@@ -261,8 +259,9 @@ PADDLE_DEFINE_EXPORTED_bool( ...@@ -261,8 +259,9 @@ PADDLE_DEFINE_EXPORTED_bool(
* increased. * increased.
* Users need to balance memory and speed. * Users need to balance memory and speed.
*/ */
PADDLE_DEFINE_EXPORTED_int64(conv_workspace_size_limit, PADDLE_DEFINE_EXPORTED_int64(
paddle::platform::kDefaultConvWorkspaceSizeLimitMB, conv_workspace_size_limit,
phi::backends::gpu::kDefaultConvWorkspaceSizeLimitMB,
"cuDNN convolution workspace limit in MB unit."); "cuDNN convolution workspace limit in MB unit.");
/** /**
......
...@@ -20,12 +20,11 @@ ...@@ -20,12 +20,11 @@
#include <type_traits> #include <type_traits>
#include "gflags/gflags.h" #include "gflags/gflags.h"
#include "paddle/fluid/platform/macros.h" #include "paddle/phi/core/macros.h"
#include "paddle/utils/variant.h" #include "paddle/utils/variant.h"
namespace paddle { namespace phi {
namespace platform {
struct FlagInfo { struct FlagInfo {
using ValueType = using ValueType =
...@@ -51,7 +50,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap(); ...@@ -51,7 +50,7 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap();
static_assert(std::is_same<FlagDeclaredType, ::std::string>::value || \ static_assert(std::is_same<FlagDeclaredType, ::std::string>::value || \
std::is_arithmetic<FlagDeclaredType>::value, \ std::is_arithmetic<FlagDeclaredType>::value, \
"FLAGS should be std::string or arithmetic type"); \ "FLAGS should be std::string or arithmetic type"); \
auto *instance = ::paddle::platform::GetMutableExportedFlagInfoMap(); \ auto *instance = ::phi::GetMutableExportedFlagInfoMap(); \
auto &info = (*instance)[#__name]; \ auto &info = (*instance)[#__name]; \
info.name = #__name; \ info.name = #__name; \
info.value_ptr = &(FLAGS_##__name); \ info.value_ptr = &(FLAGS_##__name); \
...@@ -96,5 +95,4 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap(); ...@@ -96,5 +95,4 @@ ExportedFlagInfoMap *GetMutableExportedFlagInfoMap();
__PADDLE_DEFINE_EXPORTED_FLAG( \ __PADDLE_DEFINE_EXPORTED_FLAG( \
name, true, ::std::string, string, default_value, doc) name, true, ::std::string, string, default_value, doc)
} // namespace platform } // namespace phi
} // namespace paddle
...@@ -14,11 +14,11 @@ ...@@ -14,11 +14,11 @@
#include "paddle/fluid/operators/layout_utils.h" #include "paddle/fluid/operators/layout_utils.h"
#include "paddle/fluid/operators/norm_utils.cu.h" #include "paddle/fluid/operators/norm_utils.cu.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h" #include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/backends/gpu/gpu_dnn.h" #include "paddle/phi/backends/gpu/gpu_dnn.h"
#include "paddle/phi/common/layout.h" #include "paddle/phi/common/layout.h"
#include "paddle/phi/core/enforce.h" #include "paddle/phi/core/enforce.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/core/kernel_registry.h" #include "paddle/phi/core/kernel_registry.h"
#include "paddle/phi/kernels/batch_norm_kernel.h" #include "paddle/phi/kernels/batch_norm_kernel.h"
#include "paddle/phi/kernels/empty_kernel.h" #include "paddle/phi/kernels/empty_kernel.h"
......
...@@ -22,11 +22,11 @@ namespace cub = hipcub; ...@@ -22,11 +22,11 @@ namespace cub = hipcub;
#include "paddle/fluid/operators/layout_utils.h" #include "paddle/fluid/operators/layout_utils.h"
#include "paddle/fluid/operators/norm_utils.cu.h" #include "paddle/fluid/operators/norm_utils.cu.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h" #include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/backends/gpu/gpu_dnn.h" #include "paddle/phi/backends/gpu/gpu_dnn.h"
#include "paddle/phi/common/layout.h" #include "paddle/phi/common/layout.h"
#include "paddle/phi/core/enforce.h" #include "paddle/phi/core/enforce.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/core/kernel_registry.h" #include "paddle/phi/core/kernel_registry.h"
#include "paddle/phi/kernels/batch_norm_kernel.h" #include "paddle/phi/kernels/batch_norm_kernel.h"
#include "paddle/phi/kernels/funcs/batch_norm_utils.h" #include "paddle/phi/kernels/funcs/batch_norm_utils.h"
......
...@@ -14,10 +14,10 @@ ...@@ -14,10 +14,10 @@
#pragma once #pragma once
#include "paddle/fluid/platform/flags.h"
#include "paddle/phi/backends/gpu/gpu_context.h" #include "paddle/phi/backends/gpu/gpu_context.h"
#include "paddle/phi/common/amp_type_traits.h" #include "paddle/phi/common/amp_type_traits.h"
#include "paddle/phi/common/place.h" #include "paddle/phi/common/place.h"
#include "paddle/phi/core/flags.h"
#include "paddle/phi/kernels/funcs/aligned_vector.h" #include "paddle/phi/kernels/funcs/aligned_vector.h"
DECLARE_bool(use_fast_math); DECLARE_bool(use_fast_math);
......
...@@ -23,8 +23,8 @@ ...@@ -23,8 +23,8 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h" #include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif #endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h" #include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/bfloat16.h" #include "paddle/phi/common/bfloat16.h"
#include "paddle/phi/common/float16.h" #include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h" #include "paddle/phi/kernels/cpu/conv_util.h"
......
...@@ -24,8 +24,8 @@ ...@@ -24,8 +24,8 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h" #include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif #endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h" #include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/bfloat16.h" #include "paddle/phi/common/bfloat16.h"
#include "paddle/phi/common/float16.h" #include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h" #include "paddle/phi/kernels/cpu/conv_util.h"
......
...@@ -23,9 +23,9 @@ ...@@ -23,9 +23,9 @@
#include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h" #include "paddle/phi/kernels/gpudnn/conv_cudnn_v7.h"
#endif #endif
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
#include "paddle/fluid/platform/profiler.h" #include "paddle/fluid/platform/profiler.h"
#include "paddle/phi/backends/dynload/cudnn.h" #include "paddle/phi/backends/dynload/cudnn.h"
#include "paddle/phi/backends/gpu/cuda/cudnn_workspace_helper.h"
#include "paddle/phi/common/float16.h" #include "paddle/phi/common/float16.h"
#include "paddle/phi/kernels/cpu/conv_util.h" #include "paddle/phi/kernels/cpu/conv_util.h"
#include "paddle/phi/kernels/funcs/batch_norm_utils.h" #include "paddle/phi/kernels/funcs/batch_norm_utils.h"
......
...@@ -17,8 +17,8 @@ limitations under the License. */ ...@@ -17,8 +17,8 @@ limitations under the License. */
#include "paddle/fluid/framework/phi_utils.h" #include "paddle/fluid/framework/phi_utils.h"
#include "paddle/fluid/memory/allocation/allocator_strategy.h" #include "paddle/fluid/memory/allocation/allocator_strategy.h"
#include "paddle/fluid/platform/device/npu/npu_info.h" #include "paddle/fluid/platform/device/npu/npu_info.h"
#include "paddle/fluid/platform/flags.h"
#include "paddle/fluid/platform/init.h" #include "paddle/fluid/platform/init.h"
#include "paddle/phi/core/flags.h"
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
DECLARE_bool(enable_gpu_memory_usage_log); DECLARE_bool(enable_gpu_memory_usage_log);
...@@ -43,7 +43,7 @@ int main(int argc, char** argv) { ...@@ -43,7 +43,7 @@ int main(int argc, char** argv) {
} }
#endif #endif
const auto& flag_map = paddle::platform::GetExportedFlagInfoMap(); const auto& flag_map = phi::GetExportedFlagInfoMap();
for (const auto& pair : flag_map) { for (const auto& pair : flag_map) {
const std::string& name = pair.second.name; const std::string& name = pair.second.name;
// NOTE(zhiqiu): some names may not linked in some tests, so add to // NOTE(zhiqiu): some names may not linked in some tests, so add to
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册