// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // Parts of the following code in this file refs to // https://github.com/Tencent/ncnn/blob/master/src/cpu.cpp // Tencent is pleased to support the open source community by making ncnn // available. // // Copyright (C) 2017 THL A29 Limited, a Tencent company. All rights reserved. // // Licensed under the BSD 3-Clause License (the "License"); you may not use this // file except in compliance with the License. You may obtain a copy of the // License at // // https://opensource.org/licenses/BSD-3-Clause // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, WITHOUT // WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the // License for the specific language governing permissions and limitations under // the License. #ifdef LITE_WITH_LINUX #include #include #endif #ifdef LITE_WITH_ANDROID #include #endif #if __APPLE__ #include "TargetConditionals.h" #if LITE_WITH_IPHONE #include #include #include #endif // LITE_WITH_IPHONE #endif // __APPLE__ #ifdef ARM_WITH_OMP #include #endif #include #include #include "lite/core/device_info.h" namespace paddle { namespace lite { #if ((defined LITE_WITH_ARM) || (defined LITE_WITH_MLU)) thread_local lite_api::PowerMode DeviceInfo::mode_; thread_local ARMArch DeviceInfo::arch_; thread_local int DeviceInfo::mem_size_; thread_local std::vector DeviceInfo::active_ids_; thread_local TensorLite DeviceInfo::workspace_; thread_local int64_t DeviceInfo::count_ = 0; #ifdef LITE_WITH_MLU thread_local cnmlCoreVersion_t DeviceInfo::mlu_core_version_{CNML_MLU270}; thread_local int DeviceInfo::mlu_core_number_{1}; thread_local bool DeviceInfo::use_first_conv_{false}; thread_local std::vector DeviceInfo::mean_vec_; thread_local std::vector DeviceInfo::std_vec_; thread_local DataLayoutType DeviceInfo::input_layout_{DATALAYOUT(kNCHW)}; #endif #ifdef TARGET_IOS const int DEFAULT_L1_CACHE_SIZE = 64 * 1024; const int DEFAULT_L2_CACHE_SIZE = 2048 * 1024; const int DEFAULT_L3_CACHE_SIZE = 0; #else const int DEFAULT_L1_CACHE_SIZE = 32 * 1024; const int DEFAULT_L2_CACHE_SIZE = 512 * 1024; const int DEFAULT_L3_CACHE_SIZE = 0; #endif int get_cpu_num() { #ifdef LITE_WITH_LINUX // get cpu count from /sys/devices/system/cpu/cpunum/uevent int max_cpu_num = 128; int cpu_num = 0; for (int i = 0; i < max_cpu_num; ++i) { char path[256]; snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/uevent", i); FILE* fp = fopen(path, "rb"); if (!fp) { break; } cpu_num++; fclose(fp); } if (cpu_num < 1) { cpu_num = 1; } return cpu_num; #elif defined(TARGET_IOS) int cpu_num = 0; size_t len = sizeof(cpu_num); sysctlbyname("hw.ncpu", &cpu_num, &len, NULL, 0); if (cpu_num < 1) { cpu_num = 1; } return cpu_num; #else return 1; #endif } size_t get_mem_size() { #ifdef LITE_WITH_LINUX // get cpu count from /proc/cpuinfo FILE* fp = fopen("/proc/meminfo", "rb"); if (!fp) { return 1; } size_t memsize = 0; char line[1024]; while (!feof(fp)) { char* s = fgets(line, 1024, fp); if (!s) { break; } sscanf(s, "MemTotal: %d kB", &memsize); } fclose(fp); return memsize; #elif defined(TARGET_IOS) // to be implemented printf("not implemented, set to default 4GB\n"); return 4096 * 1024; #endif return 0; } void get_cpu_arch(std::vector* archs, const int cpu_num) { archs->resize(cpu_num); for (int i = 0; i < cpu_num; ++i) { archs->at(i) = kARMArch_UNKOWN; } #ifdef LITE_WITH_LINUX //! get CPU ARCH FILE* fp = fopen("/proc/cpuinfo", "rb"); if (!fp) { return; } int cpu_idx = 0; char line[1024]; while (!feof(fp)) { char* s = fgets(line, 1024, fp); if (!s) { break; } if (strstr(line, "part") != NULL) { ARMArch arch_type = kARMArch_UNKOWN; int arch_id = 0; sscanf(s, "CPU part\t: %x", &arch_id); switch (arch_id) { case 0xd03: arch_type = kA53; break; case 0xd05: arch_type = kA55; break; case 0xd07: arch_type = kA57; break; case 0xd08: arch_type = kA72; break; case 0xd09: arch_type = kA73; break; case 0xd0a: arch_type = kA75; break; case 0xd40: arch_type = kA76; break; case 0x804: // 855 arch_type = kA76; break; case 0x805: // 855 arch_type = kA55; break; case 0x802: // 845 arch_type = kA75; break; case 0x803: // 845 arch_type = kA55; break; case 0x801: // 835 arch_type = kA73; break; case 0x800: // 835 arch_type = kA73; break; case 0x205: // 820 arch_type = kA72; break; default: LOG(ERROR) << "Unknow cpu arch: " << arch_id; } archs->at(cpu_idx) = arch_type; cpu_idx++; } } fclose(fp); for (; cpu_idx > 0 && cpu_idx < cpu_num; ++cpu_idx) { archs->at(cpu_idx) = archs->at(cpu_idx - 1); } #elif defined(TARGET_IOS) for (int i = 0; i < cpu_num; ++i) { archs->at(i) = kAPPLE; } #endif } #ifdef LITE_WITH_LINUX std::string get_cpu_name() { std::string cpu_name = ""; FILE* fp = fopen("/proc/cpuinfo", "rb"); if (!fp) { return ""; } char line[1024]; bool first_model_name = true; while (!feof(fp)) { char* s = fgets(line, 1024, fp); if (!s) { break; } if (strstr(line, "Hardware") != NULL) { cpu_name += std::string(line); } if (strstr(line, "model name") != NULL && first_model_name) { cpu_name += std::string(line); first_model_name = false; } } #ifdef LITE_WITH_ANDROID // cpu name concat board name, platform name and chip name char board_name[128]; char platform_name[128]; char chip_name[128]; __system_property_get("ro.product.board", board_name); __system_property_get("ro.board.platform", platform_name); __system_property_get("ro.chipname", chip_name); cpu_name = cpu_name + "_" + board_name + "_" + platform_name + "_" + chip_name; #endif std::transform(cpu_name.begin(), cpu_name.end(), cpu_name.begin(), ::toupper); fclose(fp); return cpu_name; } int get_min_freq_khz(int cpuid) { // first try, for all possible cpu char path[256]; snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/cpufreq/cpuinfo_max_freq", cpuid); FILE* fp = fopen(path, "rb"); if (!fp) { return -1; } int min_freq_khz = -1; fscanf(fp, "%d", &min_freq_khz); fclose(fp); return min_freq_khz; } int get_max_freq_khz(int cpuid) { // first try, for all possible cpu char path[256]; snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpufreq/stats/cpu%d/time_in_state", cpuid); FILE* fp = fopen(path, "rb"); if (!fp) { // second try, for online cpu snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/cpufreq/stats/time_in_state", cpuid); fp = fopen(path, "rb"); } int max_freq_khz = 0; if (fp) { while (!feof(fp)) { int freq_khz = 0; int nscan = fscanf(fp, "%d %*d", &freq_khz); if (nscan != 1) { break; } if (freq_khz > max_freq_khz) { max_freq_khz = freq_khz; } } } if (max_freq_khz == 0 || !fp) { // third try, for online cpu snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/cpufreq/cpuinfo_max_freq", cpuid); fp = fopen(path, "rb"); if (!fp) { return -1; } int max_freq_khz = -1; fscanf(fp, "%d", &max_freq_khz); fclose(fp); return max_freq_khz; } fclose(fp); return max_freq_khz; } void sort_cpuid_by_max_freq(const std::vector& max_freqs, std::vector* cpu_ids, std::vector* cluster_ids) { int cpu_num = max_freqs.size(); if (cpu_num == 0) { return; } cpu_ids->resize(cpu_num); cluster_ids->resize(cpu_num); for (int i = 0; i < cpu_num; i++) { cpu_ids->at(i) = i; } // sort cpuid as big core first // simple bubble sort for (int i = 0; i < cpu_num; i++) { for (int j = i + 1; j < cpu_num; j++) { if (max_freqs[i] < max_freqs[j]) { // swap int tmp = cpu_ids->at(i); cpu_ids->at(i) = cpu_ids->at(j); cpu_ids->at(j) = tmp; } } } // SMP int mid_max_freq = (max_freqs[cpu_ids->at(0)] + max_freqs[cpu_ids->at(cpu_num - 1)]) / 2; for (int i = 0; i < cpu_num; i++) { cpu_ids->at(i) = i; if (max_freqs[i] >= mid_max_freq) { cluster_ids->at(i) = 0; } else { cluster_ids->at(i) = 1; } } } void get_cpu_cache_size(int cpu_id, int* l1_cache_size, int* l2_cache_size, int* l3_cache_size) { int max_cache_idx_num = 10; *l1_cache_size = DEFAULT_L1_CACHE_SIZE; *l2_cache_size = DEFAULT_L2_CACHE_SIZE; *l3_cache_size = DEFAULT_L3_CACHE_SIZE; for (int i = 0; i < max_cache_idx_num; i++) { char path[256]; snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/cache/index%d/level", cpu_id, i); FILE* fp = fopen(path, "rb"); if (fp) { int level = -1; fscanf(fp, "%d", &level); fclose(fp); snprintf(path, sizeof(path), "/sys/devices/system/cpu/cpu%d/cache/index%d/size", cpu_id, i); fp = fopen(path, "rb"); if (fp) { int size = -1; fscanf(fp, "%d", &size); fclose(fp); if (size >= 0) { if (level == 1) { *l1_cache_size = size * 1024; } else if (level == 2) { *l2_cache_size = size * 1024; } else if (level == 3) { *l3_cache_size = size * 1024; } } } } } } bool check_cpu_online(const std::vector& cpu_ids) { if (cpu_ids.size() == 0) { return false; } char path[256]; bool all_online = true; for (int i = 0; i < cpu_ids.size(); ++i) { snprintf( path, sizeof(path), "/sys/devices/system/cpu/cpu%d/online", cpu_ids[i]); FILE* fp = fopen(path, "rb"); int is_online = 0; if (fp) { fscanf(fp, "%d", &is_online); fclose(fp); } else { LOG(ERROR) << "Failed to query the online statue of CPU id:" << cpu_ids[i]; } if (is_online == 0) { all_online = false; LOG(ERROR) << "CPU id:" << cpu_ids[i] << " is offine"; } } return all_online; } int set_sched_affinity(const std::vector& cpu_ids) { // #define CPU_SETSIZE 1024 // #define __NCPUBITS (8 * sizeof (unsigned long)) // typedef struct // { // unsigned long __bits[CPU_SETSIZE / __NCPUBITS]; // } cpu_set_t; // set affinity for thread #ifdef __GLIBC__ pid_t pid = syscall(SYS_gettid); #else pid_t pid = gettid(); #endif cpu_set_t mask; CPU_ZERO(&mask); for (int i = 0; i < cpu_ids.size(); ++i) { CPU_SET(cpu_ids[i], &mask); } int syscallret = syscall(__NR_sched_setaffinity, pid, sizeof(mask), &mask); if (syscallret) { return -1; } return 0; } bool bind_threads(const std::vector cpu_ids) { #ifdef ARM_WITH_OMP int thread_num = cpu_ids.size(); omp_set_num_threads(thread_num); std::vector ssarets; for (int i = 0; i < thread_num; ++i) { ssarets.push_back(0); } #pragma omp parallel for for (int i = 0; i < thread_num; i++) { ssarets[i] = set_sched_affinity(cpu_ids); } for (int i = 0; i < thread_num; i++) { if (ssarets[i] != 0) { LOG(ERROR) << "Set cpu affinity failed, core id: " << cpu_ids[i]; return false; } } #else // ARM_WITH_OMP std::vector first_cpu_id; first_cpu_id.push_back(cpu_ids[0]); int ssaret = set_sched_affinity(first_cpu_id); if (ssaret != 0) { LOG(ERROR) << "Set cpu affinity failed, core id: " << cpu_ids[0]; return false; } #endif // ARM_WITH_OMP return true; } #endif // LITE_WITH_LINUX void DeviceInfo::SetDotInfo(int argc, ...) { va_list arg_ptr; va_start(arg_ptr, argc); dot_.resize(core_num_); if (argc == 1) { bool flag = va_arg(arg_ptr, int) > 0; for (int i = 0; i < core_num_; ++i) { dot_[i] = flag; } } else { bool flag_big_core = va_arg(arg_ptr, int) > 0; bool flag_little_core = va_arg(arg_ptr, int) > 0; int big_core_num = big_core_ids_.size(); int little_core_num = little_core_ids_.size(); for (int i = 0; i < big_core_num; ++i) { dot_[big_core_ids_[i]] = flag_big_core; } for (int i = 0; i < little_core_num; ++i) { dot_[little_core_ids_[i]] = flag_little_core; } } va_end(arg_ptr); } void DeviceInfo::SetFP16Info(int argc, ...) { va_list arg_ptr; va_start(arg_ptr, argc); fp16_.resize(core_num_); if (argc == 1) { bool flag = va_arg(arg_ptr, int) > 0; for (int i = 0; i < core_num_; ++i) { fp16_[i] = flag; } } else { bool flag_big_core = va_arg(arg_ptr, int) > 0; bool flag_little_core = va_arg(arg_ptr, int) > 0; int big_core_num = big_core_ids_.size(); int little_core_num = little_core_ids_.size(); for (int i = 0; i < big_core_num; ++i) { fp16_[big_core_ids_[i]] = flag_big_core; } for (int i = 0; i < little_core_num; ++i) { fp16_[little_core_ids_[i]] = flag_little_core; } } va_end(arg_ptr); } void DeviceInfo::SetFP32Info(int argc, ...) { va_list arg_ptr; va_start(arg_ptr, argc); fp32_.resize(core_num_); if (argc == 1) { bool flag = va_arg(arg_ptr, int) > 0; for (int i = 0; i < core_num_; ++i) { fp32_[i] = flag; } } else { bool flag_big_core = va_arg(arg_ptr, int) > 0; bool flag_little_core = va_arg(arg_ptr, int) > 0; int big_core_num = big_core_ids_.size(); int little_core_num = little_core_ids_.size(); for (int i = 0; i < big_core_num; ++i) { fp32_[big_core_ids_[i]] = flag_big_core; } for (int i = 0; i < little_core_num; ++i) { fp32_[little_core_ids_[i]] = flag_little_core; } } va_end(arg_ptr); } // cache_id : 0 -> L1, 1 -> L2, 2 -> L3 void DeviceInfo::SetCacheInfo(int cache_id, int argc, ...) { va_list arg_ptr; va_start(arg_ptr, argc); std::vector* cache; switch (cache_id) { case 0: cache = &L1_cache_; break; case 1: cache = &L2_cache_; break; case 2: cache = &L3_cache_; break; default: break; } cache->resize(core_num_); if (argc == 1) { int cache_size = va_arg(arg_ptr, int); for (int i = 0; i < core_num_; ++i) { (*cache)[i] = cache_size; } } else { int big_core_num = big_core_ids_.size(); int little_core_num = little_core_ids_.size(); int big_core_cache_size = va_arg(arg_ptr, int); int little_core_cache_size = va_arg(arg_ptr, int); for (int i = 0; i < big_core_num; ++i) { (*cache)[big_core_ids_[i]] = big_core_cache_size; } for (int i = 0; i < little_core_num; ++i) { (*cache)[little_core_ids_[i]] = little_core_cache_size; } } va_end(arg_ptr); } void DeviceInfo::SetArchInfo(int argc, ...) { va_list arg_ptr; va_start(arg_ptr, argc); archs_.resize(core_num_); if (argc == 1) { ARMArch arch = (ARMArch)va_arg(arg_ptr, int); for (int i = 0; i < core_num_; ++i) { archs_[i] = arch; } } else { ARMArch big_core_arch = (ARMArch)va_arg(arg_ptr, int); ARMArch little_core_arch = (ARMArch)va_arg(arg_ptr, int); int big_core_num = big_core_ids_.size(); int little_core_num = little_core_ids_.size(); for (int i = 0; i < big_core_num; ++i) { archs_[big_core_ids_[i]] = big_core_arch; } for (int i = 0; i < little_core_num; ++i) { archs_[little_core_ids_[i]] = little_core_arch; } } va_end(arg_ptr); } bool DeviceInfo::SetCPUInfoByName() { /* Snapdragon */ if (dev_name_.find("SM8150") != std::string::npos) { // 855 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA76, kA55); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 2, 256 * 1024, 128 * 1024); SetCacheInfo(2, 1, 2048 * 1024); SetFP16Info(1, 1); SetDotInfo(1, 1); return true; } else if (dev_name_.find("SDM845") != std::string::npos) { // 845 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA75, kA55); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 2, 256 * 1024, 128 * 1024); SetCacheInfo(2, 1, 2048 * 1024); SetFP16Info(1, 1); return true; } else if (dev_name_.find("SDM710") != std::string::npos) { // 710 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {6, 7}; little_core_ids_ = {0, 1, 2, 3, 4, 5}; cluster_ids_ = {1, 1, 1, 1, 1, 1, 0, 0}; SetArchInfo(2, kA75, kA55); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 2, 256 * 1024, 128 * 1024); SetCacheInfo(2, 1, 1024 * 1024); return true; } else if (dev_name_.find("MSM8998") != std::string::npos) { // 835 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA73, kA53); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 2, 1024 * 1024, /*real cache size is 2M, while that will get bad performace on conv3x3s1 or gemm, set to 1M or 512K*/ 1024 * 1024); return true; } else if (dev_name_.find("MSM8996") != std::string::npos) { // 820 core_num_ = 4; core_ids_ = {0, 1, 2, 3}; big_core_ids_ = {2, 3}; little_core_ids_ = {0, 1}; cluster_ids_ = {1, 1, 0, 0}; SetArchInfo(1, kA72); SetCacheInfo(0, 1, 24 * 1024); SetCacheInfo(1, 2, 1024 * 1024, 512 * 1024); return true; } else if (dev_name_.find("SDM660") != std::string::npos || dev_name_.find("SDM636") != std::string::npos) { // 660, 636 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(1, kA73); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 1, 1024 * 1024); return true; } else if (dev_name_.find("MSM8976") != std::string::npos) { // 652,653 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA72, kA53); SetCacheInfo(0, 1, 32 * 1024); SetCacheInfo(1, 2, 1024 * 1024, 512 * 1024); return true; } else if (dev_name_.find("MSM8953") != std::string::npos) { // 625 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; little_core_ids_ = {}; cluster_ids_ = {0, 0, 0, 0, 0, 0, 0, 0}; SetArchInfo(1, kA53); SetCacheInfo(0, 1, 32 * 1024); SetCacheInfo(1, 1, 1024 * 1024); return true; } else if (dev_name_.find("MSM8939") != std::string::npos) { // 615 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {0, 1, 2, 3}; little_core_ids_ = {4, 5, 6, 7}; cluster_ids_ = {0, 0, 0, 0, 1, 1, 1, 1}; SetArchInfo(1, kA53); SetCacheInfo(0, 1, 32 * 1024); SetCacheInfo(1, 2, 512 * 1024, 256 * 1024); return true; /* MediaTek */ } else if (dev_name_.find("MT6797") != std::string::npos) { // X20/X23/X25/X27 core_num_ = 10; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7, 8, 9}; big_core_ids_ = {8, 9}; little_core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; cluster_ids_ = {1, 1, 1, 1, 1, 1, 1, 1, 0, 0}; SetArchInfo(2, kA72, kA53); SetCacheInfo(0, 1, 32 * 1024); SetCacheInfo(1, 2, 1024 * 1024, 512 * 1024); return true; } else if (dev_name_.find("MT6799") != std::string::npos) { // X30 core_num_ = 10; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7, 8, 9}; big_core_ids_ = {8, 9}; little_core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; cluster_ids_ = {1, 1, 1, 1, 1, 1, 1, 1, 0, 0}; SetArchInfo(2, kA73, kA53); return true; } else if (dev_name_.find("MT6795") != std::string::npos || dev_name_.find("MT6762") != std::string::npos || dev_name_.find("MT6755T") != std::string::npos || dev_name_.find("MT6755S") != std::string::npos || dev_name_.find("MT6753") != std::string::npos || dev_name_.find("MT6752") != std::string::npos || dev_name_.find("MT6750") != std::string::npos) { // X10, P22, P15/P18, MT6753, MT6752/MT6752M, MT6750 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; little_core_ids_ = {}; cluster_ids_ = {0, 0, 0, 0, 0, 0, 0, 0}; SetArchInfo(1, kA53); return true; } else if (dev_name_.find("MT6758") != std::string::npos || dev_name_.find("MT6757") != std::string::npos || dev_name_.find("MT6763") != std::string::npos || dev_name_.find("MT6755M") != std::string::npos || dev_name_.find("MT6755") != std::string::npos) { // P30, P20/P25, P23, P10 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(1, kA53); return true; } else if (dev_name_.find("MT6771") != std::string::npos) { // P60 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA73, kA53); return true; } else if (dev_name_.find("MT6765") != std::string::npos || dev_name_.find("MT6739") != std::string::npos || dev_name_.find("MT6738") != std::string::npos || dev_name_.find("MT6737") != std::string::npos) { // A22, MT6739, MT6738, MT6767 core_num_ = 4; core_ids_ = {0, 1, 2, 3}; big_core_ids_ = {0, 1, 2, 3}; little_core_ids_ = {}; cluster_ids_ = {0, 0, 0, 0}; SetArchInfo(1, kA53); return true; } else if (dev_name_.find("KIRIN980") != std::string::npos || dev_name_.find("KIRIN990") != std::string::npos) { // Kirin 980, Kirin 990 core_num_ = 8; core_ids_ = {0, 1, 2, 3, 4, 5, 6, 7}; big_core_ids_ = {4, 5, 6, 7}; little_core_ids_ = {0, 1, 2, 3}; cluster_ids_ = {1, 1, 1, 1, 0, 0, 0, 0}; SetArchInfo(2, kA76, kA55); SetCacheInfo(0, 2, 64 * 1024, 32 * 1024); SetCacheInfo(1, 2, 512 * 1024, 128 * 1024); SetCacheInfo(2, 1, 4096 * 1024); SetFP16Info(1, 1); SetDotInfo(1, 1); return true; } else if (dev_name_.find("FT2000PLUS") != std::string::npos) { core_num_ = 64; core_ids_.resize(core_num_); big_core_ids_.resize(core_num_); cluster_ids_.resize(core_num_); for (int i = 0; i < core_num_; ++i) { core_ids_[i] = i; big_core_ids_[i] = i; cluster_ids_[i] = 0; } little_core_ids_ = {}; SetCacheInfo(0, 1, 64 * 1024); SetCacheInfo(1, 1, 32 * 1024 * 1024); SetCacheInfo(2, 1, 128 * 1024 * 1024); return true; } return false; } void DeviceInfo::SetCPUInfoByProb() { #ifdef LITE_WITH_LINUX // get big.LITTLE cores by sorting CPU frequency sort_cpuid_by_max_freq(max_freqs_, &core_ids_, &cluster_ids_); big_core_ids_.clear(); little_core_ids_.clear(); for (int i = 0; i < cluster_ids_.size(); ++i) { if (cluster_ids_[i] == 0) { big_core_ids_.push_back(core_ids_[i]); } else { little_core_ids_.push_back(core_ids_[i]); } } // get l1, l2, l3 cache size for each core for (int i = 0; i < core_num_; i++) { get_cpu_cache_size(i, &(L1_cache_[i]), &(L2_cache_[i]), &(L3_cache_[i])); } #endif // LITE_WITH_LINUX } void DeviceInfo::RequestPowerFullMode(int thread_num) { int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); active_ids_.clear(); for (int i = 0; i < thread_num; ++i) { if (i < big_core_size) { active_ids_.push_back(big_core_ids_[i]); } else if (i < big_core_size + little_core_size) { active_ids_.push_back(little_core_ids_[i - big_core_size]); } } mode_ = lite_api::PowerMode::LITE_POWER_FULL; } void DeviceInfo::RequestPowerHighMode(int thread_num) { int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); active_ids_.clear(); if (big_core_size > 0) { mode_ = lite_api::PowerMode::LITE_POWER_HIGH; if (thread_num > big_core_size) { LOG(ERROR) << "Request thread num: " << thread_num << ", exceed the big cores size: " << big_core_size << ", truncate thread num to " << big_core_size; active_ids_ = big_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back(big_core_ids_[big_core_size - 1 - i]); } } } else { mode_ = lite_api::PowerMode::LITE_POWER_LOW; LOG(ERROR) << "HIGH POWER MODE is not support, switch to little cores."; if (thread_num > little_core_size) { active_ids_ = little_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back(little_core_ids_[i]); } } } } void DeviceInfo::RequestPowerLowMode(int thread_num) { int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); active_ids_.clear(); if (little_core_size > 0) { mode_ = lite_api::PowerMode::LITE_POWER_LOW; if (thread_num > little_core_size) { LOG(WARNING) << "Request thread num: " << thread_num << ", exceed the little cores size: " << little_core_size << ", truncate thread num to " << little_core_size; active_ids_ = little_core_ids_; } else { for (int i = 0; i < thread_num; i++) { active_ids_.push_back(little_core_ids_[i]); } } } else { mode_ = lite_api::PowerMode::LITE_POWER_HIGH; LOG(WARNING) << "LOW POWER MODE is not support, switch to big cores"; if (thread_num > big_core_size) { active_ids_ = big_core_ids_; } else { for (int i = 0; i < thread_num; i++) { active_ids_.push_back(big_core_ids_[i]); } } } } void DeviceInfo::RequestPowerNoBindMode(int thread_num) { active_ids_.clear(); if (thread_num > core_ids_.size()) { active_ids_ = core_ids_; } else { active_ids_.resize(thread_num); for (uint32_t i = 0; i < thread_num; ++i) { if (i < big_core_ids_.size()) { active_ids_[i] = big_core_ids_[i]; } else { active_ids_[i] = little_core_ids_[i - big_core_ids_.size()]; } } } mode_ = lite_api::PowerMode::LITE_POWER_NO_BIND; } void DeviceInfo::RequestPowerRandHighMode(int shift_num, int thread_num) { int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); active_ids_.clear(); if (big_core_size > 0) { mode_ = lite_api::PowerMode::LITE_POWER_RAND_HIGH; if (thread_num > big_core_size) { LOG(WARNING) << "Request thread num: " << thread_num << ", exceed the big cores size: " << big_core_size << ", truncate thread num to " << big_core_size; active_ids_ = big_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back(big_core_ids_[(i + shift_num) % big_core_size]); } } } else { mode_ = lite_api::PowerMode::LITE_POWER_LOW; LOG(WARNING) << "HIGH POWER MODE is not support, switch to little cores."; if (thread_num > little_core_size) { active_ids_ = little_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back(little_core_ids_[i]); } } } } void DeviceInfo::RequestPowerRandLowMode(int shift_num, int thread_num) { int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); active_ids_.clear(); if (little_core_size > 0) { mode_ = lite_api::PowerMode::LITE_POWER_RAND_LOW; if (thread_num > little_core_size) { LOG(WARNING) << "Request thread num: " << thread_num << ", exceed the little cores size: " << little_core_size << ", truncate thread num to " << little_core_size; active_ids_ = little_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back( little_core_ids_[(i + shift_num) % little_core_size]); } } } else { mode_ = lite_api::PowerMode::LITE_POWER_HIGH; LOG(WARNING) << "LOW POWER MODE is not support, switch to big cores."; if (thread_num > big_core_size) { active_ids_ = big_core_ids_; } else { for (int i = 0; i < thread_num; ++i) { active_ids_.push_back(big_core_ids_[i]); } } } } int DeviceInfo::Setup() { core_num_ = get_cpu_num(); mem_size_ = get_mem_size(); get_cpu_arch(&archs_, core_num_); // set defalut CPU info SetCacheInfo(0, 1, DEFAULT_L1_CACHE_SIZE); SetCacheInfo(1, 1, DEFAULT_L2_CACHE_SIZE); SetCacheInfo(2, 1, DEFAULT_L3_CACHE_SIZE); SetFP32Info(1, 1); SetFP16Info(1, 0); SetDotInfo(1, 0); max_freqs_.resize(core_num_); min_freqs_.resize(core_num_); #ifdef LITE_WITH_LINUX // get max&min freq for (int i = 0; i < core_num_; ++i) { int max_freq = get_max_freq_khz(i); int min_freq = get_min_freq_khz(i); max_freqs_[i] = max_freq / 1000; min_freqs_[i] = min_freq / 1000; } // get cache size and big.LITTLE core ids dev_name_ = get_cpu_name(); if (!SetCPUInfoByName()) { SetCPUInfoByProb(); } #else #ifdef TARGET_IOS dev_name_ = "Apple"; #else dev_name_ = "Unknown"; #endif core_ids_.resize(core_num_); cluster_ids_.resize(core_num_); big_core_ids_.resize(core_num_); for (int i = 0; i < core_num_; ++i) { max_freqs_[i] = 1000000; min_freqs_[i] = 1000000; cluster_ids_[i] = 0; core_ids_[i] = i; big_core_ids_[i] = i; } #endif // output info LOG(INFO) << "ARM multiprocessors name: " << dev_name_; LOG(INFO) << "ARM multiprocessors number: " << core_num_; for (int i = 0; i < core_num_; ++i) { LOG(INFO) << "ARM multiprocessors ID: " << core_ids_[i] << ", max freq: " << max_freqs_[i] << ", min freq: " << min_freqs_[i] << ", cluster ID: " << cluster_ids_[core_ids_[i]] << ", CPU ARCH: A" << static_cast(archs_[i]); } LOG(INFO) << "L1 DataCache size is: "; for (int i = 0; i < core_num_; ++i) { LOG(INFO) << L1_cache_[i] / 1024 << " KB"; } LOG(INFO) << "L2 Cache size is: "; for (int i = 0; i < core_num_; ++i) { LOG(INFO) << L2_cache_[i] / 1024 << " KB"; } LOG(INFO) << "L3 Cache size is: "; for (int i = 0; i < core_num_; ++i) { LOG(INFO) << L3_cache_[i] / 1024 << " KB"; } LOG(INFO) << "Total memory: " << mem_size_ << "KB"; // set default run mode SetRunMode(lite_api::PowerMode::LITE_POWER_NO_BIND, 1); // use single thread by default return 0; } #ifdef LITE_WITH_MLU void DeviceInfo::SetMLURunMode(lite_api::MLUCoreVersion core_version, int core_number, bool use_first_conv, const std::vector& mean_vec, const std::vector& std_vec, DataLayoutType input_layout) { switch (core_version) { case (lite_api::MLUCoreVersion::MLU_220): mlu_core_version_ = CNML_MLU220; break; case (lite_api::MLUCoreVersion::MLU_270): mlu_core_version_ = CNML_MLU270; break; default: mlu_core_version_ = CNML_MLU270; break; } mlu_core_number_ = core_number; use_first_conv_ = use_first_conv; mean_vec_ = mean_vec; std_vec_ = std_vec; input_layout_ = input_layout; } cnmlCoreVersion_t DeviceInfo::MLUCoreVersion() { return mlu_core_version_; } int DeviceInfo::MLUCoreNumber() { return mlu_core_number_; } bool DeviceInfo::UseFirstConv() { return use_first_conv_; } const std::vector& DeviceInfo::MeanVec() const { return mean_vec_; } const std::vector& DeviceInfo::StdVec() const { return std_vec_; } DataLayoutType DeviceInfo::InputLayout() const { return input_layout_; } #endif // LITE_WITH_MLU void DeviceInfo::SetRunMode(lite_api::PowerMode mode, int thread_num) { #ifdef ARM_WITH_OMP thread_num = std::min(thread_num, core_num_); #else thread_num = 1; // force thread_num to 1 if OpenMP is disabled #endif #ifdef LITE_WITH_LINUX int big_core_size = big_core_ids_.size(); int little_core_size = little_core_ids_.size(); int big_little_core_size = big_core_size + little_core_size; thread_num = std::min(thread_num, big_little_core_size); count_++; int shift_num = (count_ / 10) % big_core_size; switch (mode) { case lite_api::LITE_POWER_FULL: RequestPowerFullMode(thread_num); break; case lite_api::LITE_POWER_HIGH: RequestPowerHighMode(thread_num); break; case lite_api::LITE_POWER_LOW: RequestPowerLowMode(thread_num); break; case lite_api::LITE_POWER_NO_BIND: RequestPowerNoBindMode(thread_num); break; case lite_api::LITE_POWER_RAND_HIGH: RequestPowerRandHighMode(shift_num, thread_num); break; case lite_api::LITE_POWER_RAND_LOW: RequestPowerRandLowMode(shift_num, thread_num); break; default: LOG(FATAL) << "Unsupported power mode: " << static_cast(mode); break; } if (active_ids_.empty()) { active_ids_.push_back(0); } #ifdef ARM_WITH_OMP omp_set_num_threads(active_ids_.size()); #endif if (mode_ != lite_api::LITE_POWER_NO_BIND) { if (check_cpu_online(active_ids_)) { bind_threads(active_ids_); } else { LOG(WARNING) << "Some cores are offline, switch to NO BIND MODE"; mode_ = lite_api::LITE_POWER_NO_BIND; } } #else // LITE_WITH_LINUX // only LITE_POWER_NO_BIND is supported in other OS RequestPowerNoBindMode(thread_num); #ifdef ARM_WITH_OMP omp_set_num_threads(active_ids_.size()); #endif #endif // LITE_WITH_LINUX //! alloc memory for sgemm in this context workspace_.Resize({llc_size()}); workspace_.mutable_data(); arch_ = archs_[active_ids_[0]]; } void DeviceInfo::SetCache(int l1size, int l2size, int l3size) { SetCacheInfo(0, 1, l1size); SetCacheInfo(1, 1, l2size); SetCacheInfo(2, 1, l3size); workspace_.Resize({llc_size()}); workspace_.mutable_data(); } bool DeviceInfo::ExtendWorkspace(size_t size) { workspace_.Resize( {static_cast(size + static_cast(llc_size()))}); return workspace_.mutable_data() != nullptr; } #endif // LITE_WITH_ARM #ifdef LITE_WITH_MLU void SetMluDevice(int device_id) { LOG(INFO) << "Set mlu device " << device_id; cnrtDev_t dev_handle; CNRT_CALL(cnrtGetDeviceHandle(&dev_handle, device_id)); CNRT_CALL(cnrtSetCurrentDevice(dev_handle)); } void Device::Init() { SetMluDevice(idx_); GetInfo(); CreateQueue(); } void Device::GetInfo() {} void Device::CreateQueue() { exec_queue_.clear(); io_queue_.clear(); for (size_t i = 0; i < max_queue_; ++i) { cnrtQueue_t exec_queue; cnrtQueue_t io_queue; cnrtCreateQueue(&exec_queue); cnrtCreateQueue(&io_queue); exec_queue_.push_back(exec_queue); io_queue_.push_back(io_queue); cnrtCreateQueue(&exec_queue); exec_queue_.push_back(exec_queue); } } #endif // LITE_WITH_MLU #ifdef LITE_WITH_CUDA void Device::Init() { GetInfo(); CreateStream(); } void Device::GetInfo() { cudaGetDeviceProperties(&device_prop_, idx_); cudaRuntimeGetVersion(&runtime_version_); sm_version_ = (device_prop_.major << 8 | device_prop_.minor); has_hmma_ = (sm_version_ == 0x0700 || sm_version_ == 0x0702 || sm_version_ == 0x0705); has_fp16_ = (sm_version_ == 0x0602 || sm_version_ == 0x0600 || sm_version_ == 0x0503 || has_hmma_); has_imma_ = (sm_version_ == 0x0702 || sm_version_ == 0x0705); has_int8_ = (sm_version_ == 0x0601 || sm_version_ == 0x0700 || has_imma_); } void Device::CreateStream() { exec_stream_.clear(); io_stream_.clear(); for (int i = 0; i < max_stream_; i++) { cudaStream_t exec_stream; cudaStream_t io_stream; cudaStreamCreate(&exec_stream); cudaStreamCreate(&io_stream); exec_stream_.push_back(exec_stream); io_stream_.push_back(io_stream); } } #endif } // namespace lite } // namespace paddle