Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
05114693
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
05114693
编写于
11月 04, 2020
作者:
W
Wilber
提交者:
GitHub
11月 04, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Inference] Memory modification for ShrinkMemory. (#28355)
上级
95b18683
变更
19
隐藏空白更改
内联
并排
Showing
19 changed file
with
223 addition
and
2 deletion
+223
-2
paddle/fluid/memory/allocation/CMakeLists.txt
paddle/fluid/memory/allocation/CMakeLists.txt
+1
-0
paddle/fluid/memory/allocation/allocator.h
paddle/fluid/memory/allocation/allocator.h
+3
-0
paddle/fluid/memory/allocation/allocator_facade.cc
paddle/fluid/memory/allocation/allocator_facade.cc
+5
-0
paddle/fluid/memory/allocation/allocator_facade.h
paddle/fluid/memory/allocation/allocator_facade.h
+3
-0
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator.h
.../fluid/memory/allocation/auto_growth_best_fit_allocator.h
+3
-0
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator_test.cc
.../memory/allocation/auto_growth_best_fit_allocator_test.cc
+1
-0
paddle/fluid/memory/allocation/naive_best_fit_allocator.cc
paddle/fluid/memory/allocation/naive_best_fit_allocator.cc
+51
-0
paddle/fluid/memory/allocation/naive_best_fit_allocator.h
paddle/fluid/memory/allocation/naive_best_fit_allocator.h
+1
-0
paddle/fluid/memory/allocation/naive_best_fit_allocator_test.cc
.../fluid/memory/allocation/naive_best_fit_allocator_test.cc
+74
-0
paddle/fluid/memory/allocation/retry_allocator.h
paddle/fluid/memory/allocation/retry_allocator.h
+3
-0
paddle/fluid/memory/allocation/retry_allocator_test.cc
paddle/fluid/memory/allocation/retry_allocator_test.cc
+2
-0
paddle/fluid/memory/allocation/thread_local_allocator.cc
paddle/fluid/memory/allocation/thread_local_allocator.cc
+2
-0
paddle/fluid/memory/allocation/thread_local_allocator.h
paddle/fluid/memory/allocation/thread_local_allocator.h
+4
-0
paddle/fluid/memory/allocation/thread_local_allocator_test.cc
...le/fluid/memory/allocation/thread_local_allocator_test.cc
+1
-0
paddle/fluid/memory/detail/buddy_allocator.cc
paddle/fluid/memory/detail/buddy_allocator.cc
+39
-2
paddle/fluid/memory/detail/buddy_allocator.h
paddle/fluid/memory/detail/buddy_allocator.h
+7
-0
paddle/fluid/memory/detail/buddy_allocator_test.cc
paddle/fluid/memory/detail/buddy_allocator_test.cc
+17
-0
paddle/fluid/memory/malloc.cc
paddle/fluid/memory/malloc.cc
+4
-0
paddle/fluid/memory/malloc.h
paddle/fluid/memory/malloc.h
+2
-0
未找到文件。
paddle/fluid/memory/allocation/CMakeLists.txt
浏览文件 @
05114693
...
...
@@ -4,6 +4,7 @@ cc_library(locked_allocator SRCS locked_allocator.cc DEPS allocator)
cc_library
(
buffered_allocator SRCS buffered_allocator.cc DEPS allocator
)
cc_library
(
best_fit_allocator SRCS best_fit_allocator.cc DEPS allocator
)
cc_library
(
naive_best_fit_allocator SRCS naive_best_fit_allocator.cc DEPS allocator buddy_allocator profiler
)
cc_test
(
naive_best_fit_allocator_test SRCS naive_best_fit_allocator_test.cc DEPS naive_best_fit_allocator
)
cc_test
(
buffered_allocator_test SRCS buffered_allocator_test.cc DEPS locked_allocator buffered_allocator cpu_allocator best_fit_allocator
)
if
(
WITH_MKLDNN
)
...
...
paddle/fluid/memory/allocation/allocator.h
浏览文件 @
05114693
...
...
@@ -178,12 +178,15 @@ class Allocator {
FreeImpl
(
allocation
);
}
inline
void
Release
(
const
platform
::
Place
&
place
)
{
ReleaseImpl
(
place
);
}
// True if the `Allocate` is thread safe.
virtual
bool
IsAllocThreadSafe
()
const
;
protected:
virtual
Allocation
*
AllocateImpl
(
size_t
size
)
=
0
;
virtual
void
FreeImpl
(
Allocation
*
allocation
);
virtual
void
ReleaseImpl
(
const
platform
::
Place
&
place
)
{}
};
using
AllocationDeleter
=
Allocator
::
AllocationDeleter
;
...
...
paddle/fluid/memory/allocation/allocator_facade.cc
浏览文件 @
05114693
...
...
@@ -287,6 +287,11 @@ AllocationPtr AllocatorFacade::Alloc(const platform::Place& place,
return
m_
->
GetAllocator
(
place
,
size
)
->
Allocate
(
size
);
}
void
AllocatorFacade
::
Release
(
const
platform
::
Place
&
place
)
{
m_
->
GetAllocator
(
place
,
/* A non-zero num to choose allocator_ */
1
)
->
Release
(
place
);
}
}
// namespace allocation
}
// namespace memory
}
// namespace paddle
paddle/fluid/memory/allocation/allocator_facade.h
浏览文件 @
05114693
...
...
@@ -44,6 +44,9 @@ class AllocatorFacade {
// Allocate a unique allocation.
AllocationPtr
Alloc
(
const
platform
::
Place
&
place
,
size_t
size
);
// Release unused memory pool.
void
Release
(
const
platform
::
Place
&
place
);
// TODO(yy): Allocate a Copy-On-Write allocation?
private:
AllocatorFacade
();
...
...
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator.h
浏览文件 @
05114693
...
...
@@ -39,6 +39,9 @@ class AutoGrowthBestFitAllocator : public Allocator {
void
FreeImpl
(
Allocation
*
allocation
)
override
;
// Release the memory block which is not used in pool.
void
ReleaseImpl
(
const
platform
::
Place
&
place
)
override
{
FreeIdleChunks
();
}
private:
void
FreeIdleChunks
();
...
...
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator_test.cc
浏览文件 @
05114693
...
...
@@ -65,6 +65,7 @@ static void TestFreeIdleChunk(bool free_idle_chunk,
}
else
{
ASSERT_EQ
(
recorded_allocator
->
AllocatedSize
(),
memory_size
+
alignment
);
}
ag_allocator
->
Release
(
platform
::
CPUPlace
());
}
}
...
...
paddle/fluid/memory/allocation/naive_best_fit_allocator.cc
浏览文件 @
05114693
...
...
@@ -53,6 +53,9 @@ void *Alloc(const Place &place, size_t size);
template
<
typename
Place
>
void
Free
(
const
Place
&
place
,
void
*
p
,
size_t
size
);
template
<
typename
Place
>
void
Release
(
const
Place
&
place
);
template
<
typename
Place
>
size_t
Used
(
const
Place
&
place
);
...
...
@@ -99,6 +102,11 @@ void Free<platform::CPUPlace>(const platform::CPUPlace &place, void *p,
GetCPUBuddyAllocator
()
->
Free
(
p
);
}
template
<
>
void
Release
<
platform
::
CPUPlace
>
(
const
platform
::
CPUPlace
&
place
)
{
GetCPUBuddyAllocator
()
->
Release
();
}
template
<
>
size_t
Used
<
platform
::
CPUPlace
>
(
const
platform
::
CPUPlace
&
place
)
{
return
GetCPUBuddyAllocator
()
->
Used
();
...
...
@@ -186,6 +194,17 @@ void Free<platform::XPUPlace>(const platform::XPUPlace &place, void *p,
#endif
}
template
<
>
void
Release
<
platform
::
XPUPlace
>
(
const
platform
::
XPUPlace
&
place
)
{
#ifdef PADDLE_WITH_XPU
PADDLE_THROW
(
platform
::
errors
::
PermissionDenied
(
"Release XPU pool is not supported."
));
#else
PADDLE_THROW
(
platform
::
errors
::
PermissionDenied
(
"'XPUPlace' is not supported."
));
#endif
}
template
<
>
size_t
Used
<
platform
::
XPUPlace
>
(
const
platform
::
XPUPlace
&
place
)
{
#ifdef PADDLE_WITH_XPU
...
...
@@ -313,6 +332,16 @@ void Free<platform::CUDAPlace>(const platform::CUDAPlace &place, void *p,
#endif
}
template
<
>
void
Release
<
platform
::
CUDAPlace
>
(
const
platform
::
CUDAPlace
&
place
)
{
#ifdef PADDLE_WITH_CUDA
GetGPUBuddyAllocator
(
place
.
device
)
->
Release
();
#else
PADDLE_THROW
(
platform
::
errors
::
PermissionDenied
(
"'CUDAPlace' is not supported in CPU only device."
));
#endif
}
#ifdef PADDLE_WITH_CUDA
BuddyAllocator
*
GetCUDAPinnedBuddyAllocator
()
{
static
std
::
once_flag
init_flag
;
...
...
@@ -371,6 +400,17 @@ void Free<platform::CUDAPinnedPlace>(const platform::CUDAPinnedPlace &place,
#endif
}
template
<
>
void
Release
<
platform
::
CUDAPinnedPlace
>
(
const
platform
::
CUDAPinnedPlace
&
place
)
{
#ifdef PADDLE_WITH_CUDA
GetCUDAPinnedBuddyAllocator
()
->
Release
();
#else
PADDLE_THROW
(
platform
::
errors
::
PermissionDenied
(
"'CUDAPinnedPlace' is not supported in CPU only device."
));
#endif
}
struct
AllocVisitor
:
public
boost
::
static_visitor
<
void
*>
{
inline
explicit
AllocVisitor
(
size_t
size
)
:
size_
(
size
)
{}
...
...
@@ -397,6 +437,13 @@ struct FreeVisitor : public boost::static_visitor<void> {
size_t
size_
;
};
struct
ReleaseVisitor
:
public
boost
::
static_visitor
<
void
>
{
template
<
typename
Place
>
inline
void
operator
()(
const
Place
&
place
)
const
{
Release
<
Place
>
(
place
);
}
};
size_t
Usage
::
operator
()(
const
platform
::
CPUPlace
&
cpu
)
const
{
return
Used
(
cpu
);
}
...
...
@@ -439,6 +486,10 @@ void NaiveBestFitAllocator::FreeImpl(Allocation *allocation) {
delete
allocation
;
}
void
NaiveBestFitAllocator
::
ReleaseImpl
(
const
platform
::
Place
&
place
)
{
boost
::
apply_visitor
(
legacy
::
ReleaseVisitor
(),
place
);
}
}
// namespace allocation
}
// namespace memory
}
// namespace paddle
paddle/fluid/memory/allocation/naive_best_fit_allocator.h
浏览文件 @
05114693
...
...
@@ -35,6 +35,7 @@ class NaiveBestFitAllocator : public Allocator {
protected:
Allocation
*
AllocateImpl
(
size_t
size
)
override
;
void
FreeImpl
(
Allocation
*
allocation
)
override
;
void
ReleaseImpl
(
const
platform
::
Place
&
place
)
override
;
private:
platform
::
Place
place_
;
...
...
paddle/fluid/memory/allocation/naive_best_fit_allocator_test.cc
0 → 100644
浏览文件 @
05114693
// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/memory/allocation/naive_best_fit_allocator.h"
#include <algorithm>
#include <chrono> // NOLINT
#include <condition_variable> // NOLINT
#include <mutex> // NOLINT
#include <string>
#include <thread> // NOLINT
#include <vector>
#include "gtest/gtest.h"
namespace
paddle
{
namespace
memory
{
namespace
allocation
{
TEST
(
NaiveBestFitAllocatorTest
,
CpuAlloc
)
{
NaiveBestFitAllocator
alloc
{
platform
::
CPUPlace
()};
{
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
}
alloc
.
Release
(
platform
::
CPUPlace
());
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
alloc
.
Release
(
platform
::
CPUPlace
());
}
#ifdef PADDLE_WITH_CUDA
TEST
(
NaiveBestFitAllocatorTest
,
GpuAlloc
)
{
NaiveBestFitAllocator
alloc
{
platform
::
CUDAPlace
(
0
)};
{
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
}
alloc
.
Release
(
platform
::
CUDAPlace
(
0
));
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
alloc
.
Release
(
platform
::
CUDAPlace
(
0
));
}
TEST
(
NaiveBestFitAllocatorTest
,
CudaPinnedAlloc
)
{
NaiveBestFitAllocator
alloc
{
platform
::
CUDAPinnedPlace
()};
{
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
}
alloc
.
Release
(
platform
::
CUDAPinnedPlace
());
size_t
size
=
(
1
<<
20
);
auto
allocation
=
alloc
.
Allocate
(
size
);
alloc
.
Release
(
platform
::
CUDAPinnedPlace
());
}
#endif
}
// namespace allocation
}
// namespace memory
}
// namespace paddle
paddle/fluid/memory/allocation/retry_allocator.h
浏览文件 @
05114693
...
...
@@ -47,6 +47,9 @@ class RetryAllocator : public Allocator {
protected:
void
FreeImpl
(
Allocation
*
allocation
)
override
;
Allocation
*
AllocateImpl
(
size_t
size
)
override
;
void
ReleaseImpl
(
const
platform
::
Place
&
place
)
override
{
underlying_allocator_
->
Release
(
place
);
}
private:
std
::
shared_ptr
<
Allocator
>
underlying_allocator_
;
...
...
paddle/fluid/memory/allocation/retry_allocator_test.cc
浏览文件 @
05114693
...
...
@@ -96,6 +96,7 @@ TEST(RetryAllocator, RetryAllocator) {
bool
is_all_equal
=
std
::
all_of
(
addresses
.
begin
(),
addresses
.
end
(),
[
val
](
void
*
p
)
{
return
p
==
val
;
});
ASSERT_TRUE
(
is_all_equal
);
allocator
->
Release
(
platform
::
CPUPlace
());
}
}
...
...
@@ -135,6 +136,7 @@ TEST(RetryAllocator, RetryAllocatorLastAllocFailure) {
auto
allocation
=
allocator
.
Allocate
(
allocate_size
);
ASSERT_TRUE
(
false
);
allocation
.
reset
();
allocator
.
Release
(
p
);
}
catch
(
BadAlloc
&
ex
)
{
ASSERT_TRUE
(
std
::
string
(
ex
.
what
()).
find
(
"Cannot allocate"
)
!=
std
::
string
::
npos
);
...
...
paddle/fluid/memory/allocation/thread_local_allocator.cc
浏览文件 @
05114693
...
...
@@ -72,6 +72,8 @@ void ThreadLocalAllocatorImpl::FreeImpl(ThreadLocalAllocation* allocation) {
delete
allocation
;
}
void
ThreadLocalAllocatorImpl
::
ReleaseImpl
()
{
buddy_allocator_
->
Release
();
}
}
// namespace allocation
}
// namespace memory
}
// namespace paddle
paddle/fluid/memory/allocation/thread_local_allocator.h
浏览文件 @
05114693
...
...
@@ -52,6 +52,7 @@ class ThreadLocalAllocatorImpl
explicit
ThreadLocalAllocatorImpl
(
const
platform
::
Place
&
p
);
ThreadLocalAllocation
*
AllocateImpl
(
size_t
size
);
void
FreeImpl
(
ThreadLocalAllocation
*
allocation
);
void
ReleaseImpl
();
private:
std
::
unique_ptr
<
memory
::
detail
::
BuddyAllocator
>
buddy_allocator_
;
...
...
@@ -91,6 +92,9 @@ class ThreadLocalCUDAAllocator : public Allocator {
auto
allocator_impl
=
tl_allocation
->
GetAllocator
();
allocator_impl
->
FreeImpl
(
tl_allocation
);
}
void
ReleaseImpl
(
const
platform
::
Place
&
p
)
override
{
return
ThreadLocalCUDAAllocatorPool
::
Instance
().
Get
(
gpu_id_
)
->
ReleaseImpl
();
}
private:
int
gpu_id_
;
...
...
paddle/fluid/memory/allocation/thread_local_allocator_test.cc
浏览文件 @
05114693
...
...
@@ -62,6 +62,7 @@ TEST(ThreadLocalAllocator, cross_scope_release) {
auto
tl_allocator_impl
=
ThreadLocalCUDAAllocatorPool
::
Instance
().
Get
(
devices
[
j
]);
allocator_addresses
[
j
][
i
]
=
tl_allocator_impl
.
get
();
memory
::
Release
(
platform
::
CUDAPlace
(
devices
[
j
]));
}
});
}
...
...
paddle/fluid/memory/detail/buddy_allocator.cc
浏览文件 @
05114693
...
...
@@ -39,9 +39,10 @@ BuddyAllocator::~BuddyAllocator() {
while
(
!
pool_
.
empty
())
{
auto
block
=
static_cast
<
MemoryBlock
*>
(
std
::
get
<
2
>
(
*
pool_
.
begin
()));
auto
desc
=
cache_
.
LoadDesc
(
block
);
VLOG
(
10
)
<<
"Free from block ("
<<
block
<<
", "
<<
desc
->
get_size
()
<<
")"
;
VLOG
(
10
)
<<
"Free from block ("
<<
block
<<
", "
<<
desc
->
get_total_size
()
<<
")"
;
system_allocator_
->
Free
(
block
,
desc
->
get_size
(),
desc
->
get_index
());
system_allocator_
->
Free
(
block
,
desc
->
get_
total_
size
(),
desc
->
get_index
());
cache_
.
Invalidate
(
block
);
pool_
.
erase
(
pool_
.
begin
());
}
...
...
@@ -161,6 +162,39 @@ void BuddyAllocator::Free(void* p) {
IndexSizeAddress
(
desc
->
get_index
(),
desc
->
get_total_size
(),
block
));
}
void
BuddyAllocator
::
Release
()
{
std
::
lock_guard
<
std
::
mutex
>
lock
(
mutex_
);
int
num
=
0
;
uint64_t
bytes
=
0
;
bool
del_flag
=
false
;
for
(
auto
iter
=
pool_
.
begin
();
iter
!=
pool_
.
end
();)
{
auto
remain_size
=
std
::
get
<
1
>
(
*
iter
);
auto
remain_ptr
=
std
::
get
<
2
>
(
*
iter
);
for
(
auto
&
chunk
:
chunks_
)
{
auto
init_size
=
std
::
get
<
1
>
(
chunk
);
auto
init_ptr
=
std
::
get
<
2
>
(
chunk
);
if
(
init_size
==
remain_size
&&
init_ptr
==
remain_ptr
)
{
++
num
;
bytes
+=
init_size
;
total_free_
-=
init_size
;
auto
block
=
static_cast
<
MemoryBlock
*>
(
std
::
get
<
2
>
(
chunk
));
system_allocator_
->
Free
(
init_ptr
,
init_size
,
std
::
get
<
0
>
(
chunk
));
cache_
.
Invalidate
(
block
);
del_flag
=
true
;
break
;
}
}
if
(
del_flag
)
{
iter
=
pool_
.
erase
(
iter
);
}
else
{
iter
++
;
}
}
VLOG
(
10
)
<<
"Release "
<<
num
<<
" chunk, Free "
<<
bytes
<<
" bytes."
;
}
size_t
BuddyAllocator
::
Used
()
{
return
total_used_
;
}
size_t
BuddyAllocator
::
GetMinChunkSize
()
{
return
min_chunk_size_
;
}
size_t
BuddyAllocator
::
GetMaxChunkSize
()
{
return
max_chunk_size_
;
}
...
...
@@ -213,6 +247,9 @@ BuddyAllocator::PoolSet::iterator BuddyAllocator::RefillPool(
total_free_
+=
allocate_bytes
;
// record the chunk.
chunks_
.
insert
(
IndexSizeAddress
(
index
,
allocate_bytes
,
p
));
// dump the block into pool
return
pool_
.
insert
(
IndexSizeAddress
(
index
,
allocate_bytes
,
p
)).
first
;
}
...
...
paddle/fluid/memory/detail/buddy_allocator.h
浏览文件 @
05114693
...
...
@@ -40,6 +40,8 @@ class BuddyAllocator {
public:
void
*
Alloc
(
size_t
unaligned_size
);
void
Free
(
void
*
ptr
);
// Release the unused memory pool, a real free operation for the OS.
void
Release
();
size_t
Used
();
size_t
GetMinChunkSize
();
size_t
GetMaxChunkSize
();
...
...
@@ -92,6 +94,11 @@ class BuddyAllocator {
*/
PoolSet
pool_
;
/**
* \brief Record the allocated chunks when Refill pool.
*/
PoolSet
chunks_
;
private:
/*! Unify the metadata format between GPU and CPU allocations */
MetadataCache
cache_
;
...
...
paddle/fluid/memory/detail/buddy_allocator_test.cc
浏览文件 @
05114693
...
...
@@ -305,6 +305,23 @@ TEST(BuddyAllocator, SpeedAna) {
std
::
cerr
<<
"time cost "
<<
diff
.
count
()
<<
std
::
endl
;
}
TEST
(
BuddyAllocator
,
Release
)
{
// In a 8 GB machine, the pool size will be about 800 MB
FLAGS_fraction_of_gpu_memory_to_use
=
0.1
;
FLAGS_initial_gpu_memory_in_mb
=
0
;
FLAGS_reallocate_gpu_memory_in_mb
=
0
;
BuddyAllocator
buddy_allocator
(
std
::
unique_ptr
<
SystemAllocator
>
(
new
GPUAllocator
(
TEST_GPU_ID
)),
platform
::
GpuMinChunkSize
(),
platform
::
GpuMaxChunkSize
());
// Less than pool size
TestBuddyAllocator
(
&
buddy_allocator
,
10
);
TestBuddyAllocator
(
&
buddy_allocator
,
10
<<
10
);
TestBuddyAllocator
(
&
buddy_allocator
,
50
<<
20
);
buddy_allocator
.
Release
();
}
#endif
}
// namespace detail
...
...
paddle/fluid/memory/malloc.cc
浏览文件 @
05114693
...
...
@@ -31,5 +31,9 @@ AllocationPtr Alloc(const platform::Place &place, size_t size) {
return
allocation
::
AllocatorFacade
::
Instance
().
Alloc
(
place
,
size
);
}
void
Release
(
const
platform
::
Place
&
place
)
{
return
allocation
::
AllocatorFacade
::
Instance
().
Release
(
place
);
}
}
// namespace memory
}
// namespace paddle
paddle/fluid/memory/malloc.h
浏览文件 @
05114693
...
...
@@ -38,5 +38,7 @@ extern AllocationPtr Alloc(const platform::Place& place, size_t size);
extern
AllocationPtr
Alloc
(
const
platform
::
DeviceContext
&
dev_ctx
,
size_t
size
);
extern
void
Release
(
const
platform
::
Place
&
place
);
}
// namespace memory
}
// namespace paddle
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录