未验证 提交 3f6229c6 编写于 作者: F From00 提交者: GitHub

Replace shared_ptr with unique_ptr in base_ptr_test (#38530)

上级 b1e73347
...@@ -35,7 +35,7 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test { ...@@ -35,7 +35,7 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test {
void OneByOneAllocTest() { void OneByOneAllocTest() {
for (size_t i = 0; i < alloc_times_; ++i) { for (size_t i = 0; i < alloc_times_; ++i) {
size_t size = dis_(random_engine_); size_t size = dis_(random_engine_);
std::shared_ptr<Allocation> allocation = AllocShared(place_, size); AllocationPtr allocation = Alloc(place_, size);
void* base_ptr = allocation->base_ptr(); void* base_ptr = allocation->base_ptr();
void* system_ptr = void* system_ptr =
...@@ -47,21 +47,21 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test { ...@@ -47,21 +47,21 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test {
} }
void BatchByBatchAllocTest() { void BatchByBatchAllocTest() {
std::vector<std::shared_ptr<Allocation>> allocations; std::vector<AllocationPtr> allocations;
allocations.reserve(batch_size_); allocations.reserve(batch_size_);
size_t batch_num = alloc_times_ / batch_size_; size_t batch_num = alloc_times_ / batch_size_;
for (size_t i = 0; i < batch_num; ++i) { for (size_t i = 0; i < batch_num; ++i) {
for (size_t j = 0; j < batch_size_; ++j) { for (size_t j = 0; j < batch_size_; ++j) {
size_t size = dis_(random_engine_); size_t size = dis_(random_engine_);
std::shared_ptr<Allocation> allocation = AllocShared(place_, size); AllocationPtr allocation = Alloc(place_, size);
void* base_ptr = allocation->base_ptr(); void* base_ptr = allocation->base_ptr();
void* system_ptr = void* system_ptr =
platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId()); platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId());
EXPECT_EQ(base_ptr, system_ptr); EXPECT_EQ(base_ptr, system_ptr);
allocations.emplace_back(allocation); allocations.emplace_back(std::move(allocation));
} }
allocations.clear(); allocations.clear();
} }
...@@ -70,19 +70,19 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test { ...@@ -70,19 +70,19 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test {
} }
void ContinuousAllocTest() { void ContinuousAllocTest() {
std::vector<std::shared_ptr<Allocation>> allocations; std::vector<AllocationPtr> allocations;
allocations.reserve(alloc_times_); allocations.reserve(alloc_times_);
for (size_t i = 0; i < alloc_times_; ++i) { for (size_t i = 0; i < alloc_times_; ++i) {
size_t size = dis_(random_engine_); size_t size = dis_(random_engine_);
std::shared_ptr<Allocation> allocation = AllocShared(place_, size); AllocationPtr allocation = Alloc(place_, size);
void* base_ptr = allocation->base_ptr(); void* base_ptr = allocation->base_ptr();
void* system_ptr = void* system_ptr =
platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId()); platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId());
EXPECT_EQ(base_ptr, system_ptr); EXPECT_EQ(base_ptr, system_ptr);
allocations.emplace_back(allocation); allocations.emplace_back(std::move(allocation));
} }
allocations.clear(); allocations.clear();
...@@ -90,7 +90,7 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test { ...@@ -90,7 +90,7 @@ class CUDAAllocatoionBasePtrTest : public ::testing::Test {
} }
void ZeroSizeAllocTest() { void ZeroSizeAllocTest() {
std::shared_ptr<Allocation> allocation = AllocShared(place_, 0); AllocationPtr allocation = Alloc(place_, 0);
void* base_ptr = allocation->base_ptr(); void* base_ptr = allocation->base_ptr();
void* system_ptr = void* system_ptr =
platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId()); platform::GetGpuBasePtr(allocation->ptr(), place_.GetDeviceId());
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册