提交 0574915e 编写于 作者: W wangmeng28

Merge remote-tracking branch 'upstream/develop' into factorization_machine_layer

...@@ -65,16 +65,12 @@ void Scope::DropKids() { ...@@ -65,16 +65,12 @@ void Scope::DropKids() {
kids_.clear(); kids_.clear();
} }
std::once_flag feed_variable_flag;
framework::Scope& GetGlobalScope() { framework::Scope& GetGlobalScope() {
static std::unique_ptr<framework::Scope> g_scope{nullptr}; static framework::Scope* g_scope = nullptr;
std::call_once(feed_variable_flag, [&]() { if (g_scope == nullptr) {
g_scope.reset(new framework::Scope()); g_scope = new framework::Scope();
g_scope->Var("feed_value"); }
g_scope->Var("fetch_value"); return *g_scope;
});
return *(g_scope.get());
} }
} // namespace framework } // namespace framework
......
...@@ -14,11 +14,6 @@ limitations under the License. */ ...@@ -14,11 +14,6 @@ limitations under the License. */
#include "paddle/memory/memory.h" #include "paddle/memory/memory.h"
#include <algorithm> // for transform
#include <cstring> // for memcpy
#include <memory> // for unique_ptr
#include <mutex> // for call_once
#include "glog/logging.h" #include "glog/logging.h"
#include "paddle/memory/detail/buddy_allocator.h" #include "paddle/memory/detail/buddy_allocator.h"
...@@ -32,19 +27,14 @@ namespace memory { ...@@ -32,19 +27,14 @@ namespace memory {
using BuddyAllocator = detail::BuddyAllocator; using BuddyAllocator = detail::BuddyAllocator;
std::once_flag cpu_allocator_flag;
std::once_flag gpu_allocator_flag;
BuddyAllocator* GetCPUBuddyAllocator() { BuddyAllocator* GetCPUBuddyAllocator() {
static std::unique_ptr<BuddyAllocator> a{nullptr}; static detail::BuddyAllocator* a = nullptr;
if (a == nullptr) {
std::call_once(cpu_allocator_flag, [&]() { a = new detail::BuddyAllocator(new detail::CPUAllocator,
a.reset(new BuddyAllocator(new detail::CPUAllocator, platform::CpuMinChunkSize(),
platform::CpuMinChunkSize(), platform::CpuMaxChunkSize());
platform::CpuMaxChunkSize())); }
}); return a;
return a.get();
} }
template <> template <>
...@@ -65,35 +55,24 @@ size_t Used<platform::CPUPlace>(platform::CPUPlace place) { ...@@ -65,35 +55,24 @@ size_t Used<platform::CPUPlace>(platform::CPUPlace place) {
#ifdef PADDLE_WITH_CUDA #ifdef PADDLE_WITH_CUDA
BuddyAllocator* GetGPUBuddyAllocator(int gpu_id) { BuddyAllocator* GetGPUBuddyAllocator(int gpu_id) {
using BuddyAllocVec = std::vector<BuddyAllocator*>; static BuddyAllocator** as = NULL;
static std::unique_ptr<BuddyAllocVec, void (*)(BuddyAllocVec * p)> as{ if (as == NULL) {
new BuddyAllocVec, [](BuddyAllocVec* p) {
std::for_each(p->begin(), p->end(),
[](BuddyAllocator* p) { delete p; });
}};
// GPU buddy allocators
auto& allocators = *as.get();
// GPU buddy allocator initialization
std::call_once(gpu_allocator_flag, [&]() {
int gpu_num = platform::GetCUDADeviceCount(); int gpu_num = platform::GetCUDADeviceCount();
allocators.reserve(gpu_num); as = new BuddyAllocator*[gpu_num];
for (int gpu = 0; gpu < gpu_num; gpu++) { for (int gpu = 0; gpu < gpu_num; gpu++) {
platform::SetDeviceId(gpu); platform::SetDeviceId(gpu);
allocators.emplace_back(new BuddyAllocator(new detail::GPUAllocator, as[gpu] = new BuddyAllocator(new detail::GPUAllocator,
platform::GpuMinChunkSize(), platform::GpuMinChunkSize(),
platform::GpuMaxChunkSize())); platform::GpuMaxChunkSize());
} }
VLOG(3) << "\n\nNOTE: each GPU device use " VLOG(3) << "\n\nNOTE: each GPU device use "
<< FLAGS_fraction_of_gpu_memory_to_use * 100 << "% of GPU memory.\n" << FLAGS_fraction_of_gpu_memory_to_use * 100 << "% of GPU memory.\n"
<< "You can set environment variable '" << "You can set environment variable '"
<< platform::kEnvFractionGpuMemoryToUse << platform::kEnvFractionGpuMemoryToUse
<< "' to change the fraction of GPU usage.\n\n"; << "' to change the fraction of GPU usage.\n\n";
}); }
platform::SetDeviceId(gpu_id); platform::SetDeviceId(gpu_id);
return allocators[gpu_id]; return as[gpu_id];
} }
template <> template <>
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册