Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
ce46ef22
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 2 年 前同步成功
通知
2325
Star
20933
Fork
5424
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
ce46ef22
编写于
6月 20, 2019
作者:
H
hong19860320
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
ARM cpu_info refine
test=develop
上级
251255bc
变更
3
展开全部
隐藏空白更改
内联
并排
Showing
3 changed file
with
773 addition
and
724 deletion
+773
-724
paddle/fluid/lite/core/context.h
paddle/fluid/lite/core/context.h
+4
-5
paddle/fluid/lite/core/cpu_info.cc
paddle/fluid/lite/core/cpu_info.cc
+726
-658
paddle/fluid/lite/core/cpu_info.h
paddle/fluid/lite/core/cpu_info.h
+43
-61
未找到文件。
paddle/fluid/lite/core/context.h
浏览文件 @
ce46ef22
...
@@ -67,7 +67,7 @@ class Context<TargetType::kARM> {
...
@@ -67,7 +67,7 @@ class Context<TargetType::kARM> {
ARMContext
&
operator
=
(
const
ARMContext
&
ctx
)
{}
ARMContext
&
operator
=
(
const
ARMContext
&
ctx
)
{}
// NOTE: InitOnce should only be used by ContextScheduler
// NOTE: InitOnce should only be used by ContextScheduler
void
InitOnce
()
{}
void
InitOnce
()
{
DeviceInfo
::
Init
();
}
void
CopyShared
(
const
ARMContext
*
ctx
)
{}
void
CopyShared
(
const
ARMContext
*
ctx
)
{}
...
@@ -78,20 +78,19 @@ class Context<TargetType::kARM> {
...
@@ -78,20 +78,19 @@ class Context<TargetType::kARM> {
return
DeviceInfo
::
Global
().
SetCache
(
l1size
,
l2size
,
l3size
);
return
DeviceInfo
::
Global
().
SetCache
(
l1size
,
l2size
,
l3size
);
}
}
void
SetArch
(
ARMArch
arch
)
{
return
DeviceInfo
::
Global
().
SetArch
(
arch
);
}
void
SetArch
(
ARMArch
arch
)
{
return
DeviceInfo
::
Global
().
SetArch
(
arch
);
}
void
BindDev
()
{
return
DeviceInfo
::
Global
().
BindDev
();
}
PowerMode
mode
()
const
{
return
DeviceInfo
::
Global
().
mode
();
}
PowerMode
mode
()
const
{
return
DeviceInfo
::
Global
().
mode
();
}
int
threads
()
const
{
return
DeviceInfo
::
Global
().
threads
();
}
int
threads
()
const
{
return
DeviceInfo
::
Global
().
threads
();
}
ARMArch
arch
()
const
{
return
DeviceInfo
::
Global
().
arch
();
}
ARMArch
arch
()
const
{
return
DeviceInfo
::
Global
().
arch
();
}
int
l1_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l1_cache_size
();
}
int
l2_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l2_cache_size
();
}
int
l3_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l3_cache_size
();
}
template
<
typename
T
>
template
<
typename
T
>
T
*
workspace_data
()
{
T
*
workspace_data
()
{
return
DeviceInfo
::
Global
().
workspace_data
<
T
>
();
return
DeviceInfo
::
Global
().
workspace_data
<
T
>
();
}
}
int
l1_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l1_cache_size
();
}
int
l2_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l2_cache_size
();
}
int
l3_cache_size
()
const
{
return
DeviceInfo
::
Global
().
l3_cache_size
();
}
bool
ExtendWorkspace
(
DDimLite
dims
)
{
bool
ExtendWorkspace
(
DDimLite
dims
)
{
return
DeviceInfo
::
Global
().
ExtendWorkspace
(
dims
);
return
DeviceInfo
::
Global
().
ExtendWorkspace
(
dims
);
}
}
...
...
paddle/fluid/lite/core/cpu_info.cc
浏览文件 @
ce46ef22
此差异已折叠。
点击以展开。
paddle/fluid/lite/core/cpu_info.h
浏览文件 @
ce46ef22
...
@@ -14,6 +14,7 @@
...
@@ -14,6 +14,7 @@
#pragma once
#pragma once
#include <cstdarg>
#include <string>
#include <string>
#include <vector>
#include <vector>
#include "paddle/fluid/lite/core/lite_tensor.h"
#include "paddle/fluid/lite/core/lite_tensor.h"
...
@@ -47,92 +48,73 @@ typedef enum {
...
@@ -47,92 +48,73 @@ typedef enum {
class
DeviceInfo
{
class
DeviceInfo
{
public:
public:
int
idx_
;
int
max_freq_
;
int
min_freq_
;
int
generate_arch_
;
int
compute_core_num_
;
int
max_memory_
;
int
sharemem_size_
;
std
::
string
device_name_
;
std
::
string
compute_ability_
;
std
::
vector
<
int
>
L1_cache_
;
std
::
vector
<
int
>
L2_cache_
;
std
::
vector
<
int
>
L3_cache_
;
std
::
vector
<
int
>
core_ids_
;
std
::
vector
<
int
>
big_core_ids_
;
std
::
vector
<
int
>
little_core_ids_
;
std
::
vector
<
int
>
cluster_ids_
;
std
::
vector
<
ARMArch
>
archs_
;
ARMArch
arch_
;
// LITE_POWER_HIGH stands for using big cores,
// LITE_POWER_LOW stands for using small core,
// LITE_POWER_FULL stands for using all cores
PowerMode
mode_
;
std
::
vector
<
int
>
active_ids_
;
TensorLite
workspace_
;
int64_t
count_
{
0
};
static
DeviceInfo
&
Global
()
{
static
DeviceInfo
&
Global
()
{
static
auto
*
x
=
new
DeviceInfo
;
static
auto
*
x
=
new
DeviceInfo
;
return
*
x
;
return
*
x
;
}
}
static
void
Init
()
{
static
int
Init
()
{
auto
&
info
=
Global
();
static
int
ret
=
Global
().
Setup
();
InitInternal
(
&
info
)
;
return
ret
;
}
}
void
SetRunMode
(
PowerMode
mode
,
int
threads
);
int
Setup
();
void
SetRunMode
(
PowerMode
mode
,
int
thread_num
);
void
SetCache
(
int
l1size
,
int
l2size
,
int
l3size
);
void
SetCache
(
int
l1size
,
int
l2size
,
int
l3size
);
void
SetArch
(
ARMArch
arch
)
{
arch_
=
arch
;
}
void
SetArch
(
ARMArch
arch
)
{
arch_
=
arch
;
}
void
BindDev
();
PowerMode
mode
()
const
{
return
mode_
;
}
PowerMode
mode
()
const
{
return
mode_
;
}
int
threads
()
const
{
return
active_ids_
.
size
();
}
int
threads
()
const
{
return
active_ids_
.
size
();
}
ARMArch
arch
()
const
{
return
arch_
;
}
ARMArch
arch
()
const
{
return
arch_
;
}
int
l1_cache_size
()
const
{
return
L1_cache_
[
active_ids_
[
0
]];
}
int
l2_cache_size
()
const
{
return
L2_cache_
[
active_ids_
[
0
]];
}
int
l3_cache_size
()
const
{
return
L3_cache_
[
active_ids_
[
0
]];
}
template
<
typename
T
>
template
<
typename
T
>
T
*
workspace_data
()
{
T
*
workspace_data
()
{
return
workspace_
.
mutable_data
<
T
>
();
return
workspace_
.
mutable_data
<
T
>
();
}
}
int
l1_cache_size
()
const
{
return
L1_cache_
[
active_ids_
[
0
]];
}
int
l2_cache_size
()
const
{
return
L2_cache_
[
active_ids_
[
0
]];
}
int
l3_cache_size
()
const
{
return
L3_cache_
[
active_ids_
[
0
]];
}
bool
ExtendWorkspace
(
DDimLite
dims
);
bool
ExtendWorkspace
(
DDimLite
dims
);
private:
private:
DeviceInfo
()
=
default
;
int
core_num_
;
static
void
InitInternal
(
DeviceInfo
*
dev
);
std
::
vector
<
int
>
max_freqs_
;
};
std
::
vector
<
int
>
min_freqs_
;
int
mem_size_
;
size_t
arm_get_meminfo
();
std
::
string
dev_name_
;
int
arm_get_cpucount
();
void
arm_get_cpu_arch
(
std
::
vector
<
ARMArch
>*
archs
);
bool
get_cpu_info_from_name
(
DeviceInfo
*
cpu_info
,
std
::
string
hardware_name
);
#ifdef LITE_WITH_LINUX
std
::
vector
<
int
>
L1_cache_
;
std
::
vector
<
int
>
L2_cache_
;
void
set_default_cache
(
DeviceInfo
*
dev
);
std
::
vector
<
int
>
L3_cache_
;
std
::
vector
<
int
>
core_ids_
;
std
::
string
arm_get_cpu_name
();
std
::
vector
<
int
>
big_core_ids_
;
std
::
vector
<
int
>
little_core_ids_
;
std
::
vector
<
int
>
cluster_ids_
;
std
::
vector
<
ARMArch
>
archs_
;
int
get_max_freq_khz
(
int
cpuid
);
ARMArch
arch_
;
// LITE_POWER_HIGH stands for using big cores,
// LITE_POWER_LOW stands for using small core,
// LITE_POWER_FULL stands for using all cores
PowerMode
mode_
;
std
::
vector
<
int
>
active_ids_
;
TensorLite
workspace_
;
int64_t
count_
{
0
};
int
arm_sort_cpuid_by_max_frequency
(
int
cpu_count
,
std
::
vector
<
int
>*
cpuids
,
void
SetCacheInfo
(
int
cache_id
,
int
argc
,
...);
const
std
::
vector
<
int
>&
cpu_freq
,
void
SetArchInfo
(
int
argc
,
...);
std
::
vector
<
int
>*
cluster_ids
);
bool
SetCPUInfoByName
();
int
check_online
(
const
std
::
vector
<
int
>&
core_ids
);
void
SetCPUInfoByProb
();
int
set_sched_affinity
(
const
std
::
vector
<
int
>&
cpuids
);
void
RequestPowerFullMode
(
const
int
thread_num
);
void
RequestPowerHighMode
(
const
int
thread_num
);
void
RequestPowerLowMode
(
const
int
thread_num
);
void
RequestPowerNoBindMode
(
const
int
thread_num
);
void
RequestPowerRandHighMode
(
const
int
shift_num
,
const
int
thread_num
);
void
RequestPowerRandLowMode
(
const
int
shift_num
,
const
int
thread_num
);
#endif // LITE_WITH_LINUX
DeviceInfo
()
=
default
;
};
#endif // LITE_WITH_ARM
#endif // LITE_WITH_ARM
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录