cpu_info.h 3.5 KB
Newer Older
T
tensor-tang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once

#include <string>
#include <vector>
19
#include "paddle/fluid/lite/core/lite_tensor.h"
T
tensor-tang 已提交
20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69
#include "paddle/fluid/lite/utils/cp_logging.h"

namespace paddle {
namespace lite {

#ifdef LITE_WITH_ARM

typedef enum {
  LITE_POWER_HIGH = 0,
  LITE_POWER_LOW = 1,
  LITE_POWER_FULL = 2,
  LITE_POWER_NO_BIND = 3,
  LITE_POWER_RAND_HIGH = 4,
  LITE_POWER_RAND_LOW = 5
} PowerMode;

typedef enum {
  kAPPLE = 0,
  kA53 = 53,
  kA55 = 55,
  kA57 = 57,
  kA72 = 72,
  kA73 = 73,
  kA75 = 75,
  kA76 = 76,
  kARMArch_UNKOWN = -1
} ARMArch;

class DeviceInfo {
 public:
  int idx_;
  int max_freq_;
  int min_freq_;
  int generate_arch_;
  int compute_core_num_;
  int max_memory_;
  int sharemem_size_;

  std::string device_name_;
  std::string compute_ability_;

  std::vector<int> L1_cache_;
  std::vector<int> L2_cache_;
  std::vector<int> L3_cache_;
  std::vector<int> core_ids_;
  std::vector<int> big_core_ids_;
  std::vector<int> little_core_ids_;
  std::vector<int> cluster_ids_;
  std::vector<ARMArch> archs_;

70 71 72 73 74 75 76 77 78
  ARMArch arch_;
  // LITE_POWER_HIGH stands for using big cores,
  // LITE_POWER_LOW stands for using small core,
  // LITE_POWER_FULL stands for using all cores
  PowerMode mode_;
  std::vector<int> active_ids_;
  TensorLite workspace_;
  int64_t count_{0};

T
tensor-tang 已提交
79 80 81 82 83
  static DeviceInfo& Global() {
    static auto* x = new DeviceInfo;
    return *x;
  }

84
  static void Init() {
T
tensor-tang 已提交
85
    auto& info = Global();
86
    InitInternal(&info);
T
tensor-tang 已提交
87 88
  }

89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107
  void SetRunMode(PowerMode mode, int threads);
  void SetCache(int l1size, int l2size, int l3size);
  void SetArch(ARMArch arch) { arch_ = arch; }
  void BindDev();

  PowerMode mode() const { return mode_; }
  int threads() const { return active_ids_.size(); }
  ARMArch arch() const { return arch_; }

  template <typename T>
  T* workspace_data() {
    return workspace_.mutable_data<T>();
  }

  int l1_cache_size() const { return L1_cache_[active_ids_[0]]; }
  int l2_cache_size() const { return L2_cache_[active_ids_[0]]; }
  int l3_cache_size() const { return L3_cache_[active_ids_[0]]; }
  bool ExtendWorkspace(DDimLite dims);

T
tensor-tang 已提交
108 109
 private:
  DeviceInfo() = default;
110
  static void InitInternal(DeviceInfo* dev);
T
tensor-tang 已提交
111 112 113 114 115 116 117 118 119 120
};

size_t arm_get_meminfo();

int arm_get_cpucount();

void arm_get_cpu_arch(std::vector<ARMArch>* archs);

bool get_cpu_info_from_name(DeviceInfo* cpu_info, std::string hardware_name);

121
#ifdef LITE_WITH_LINUX
T
tensor-tang 已提交
122 123 124 125 126 127 128 129 130 131 132 133 134

void set_default_cache(DeviceInfo* dev);

std::string arm_get_cpu_name();

int get_max_freq_khz(int cpuid);

int arm_sort_cpuid_by_max_frequency(int cpu_count, std::vector<int>* cpuids,
                                    const std::vector<int>& cpu_freq,
                                    std::vector<int>* cluster_ids);
int check_online(const std::vector<int>& core_ids);
int set_sched_affinity(const std::vector<int>& cpuids);

135
#endif  // LITE_WITH_LINUX
T
tensor-tang 已提交
136 137 138 139 140

#endif  // LITE_WITH_ARM

}  // namespace lite
}  // namespace paddle