CMakeLists.txt 5.2 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11
set(kernel_declare_file
    ${PADDLE_BINARY_DIR}/paddle/phi/kernels/declarations.h.tmp
    CACHE INTERNAL "declarations.h file")
set(kernel_declare_file_final
    ${PADDLE_BINARY_DIR}/paddle/phi/kernels/declarations.h)
file(
  WRITE ${kernel_declare_file}
  "// Generated by the paddle/phi/kernels/CMakeLists.txt.  DO NOT EDIT!\n\n#pragma once\n\n"
)
file(APPEND ${kernel_declare_file}
     "#include \"paddle/phi/core/kernel_registry.h\"\n\n")
12 13 14
set(kernel_declare_file_prune
    ${PADDLE_BINARY_DIR}/paddle/phi/kernels/declarations.h.prune
    CACHE INTERNAL "declarations.h file")
15

16
# phi functors and functions called by kernels
C
Chen Weihang 已提交
17
add_subdirectory(funcs)
C
Chen Weihang 已提交
18

19 20 21
# kernel autotune
add_subdirectory(autotune)

22 23
# phi depends all phi kernel targets
set_property(GLOBAL PROPERTY PHI_KERNELS "")
24

25
# [ 1. Common kernel compilation dependencies ]
26 27
set(COMMON_KERNEL_DEPS
    dense_tensor
28
    string_tensor
29 30
    sparse_coo_tensor
    sparse_csr_tensor
31
    tensor_array
32 33 34 35 36
    kernel_context
    kernel_factory
    arg_map_context
    convert_utils
    lod_utils
37
    custom_kernel
38
    string_infermeta
39
    phi_tensor_utils)
40 41 42 43 44 45 46 47 48
set(COMMON_KERNEL_DEPS
    ${COMMON_KERNEL_DEPS}
    eigen_function
    blas
    math_function
    im2col
    vol2col
    concat_and_split_functor
    selected_rows_functor)
49
# remove this dep after removing fluid deps on tensor creation
50
set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} phi_api_utils lod_utils)
Z
zhangkaihuo 已提交
51 52
set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} infermeta infermeta_utils
                       sparse_infermeta)
53
set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} switch_autotune)
54

55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70
set(COMMON_KERNEL_DEPS
    ${COMMON_KERNEL_DEPS}
    threadpool
    jit_kernel_helper
    softmax
    cross_entropy
    matrix_bit_code
    lapack_function
    lstm_compute
    gru_compute
    deformable_conv_functor
    matrix_reduce
    segment_pooling
    pooling
    maxouting
    matrix_inverse
71
    matrix_solve
72
    phi_dynload_warpctc
H
Hui Zhang 已提交
73
    phi_dynload_warprnnt
74
    sequence_padding
F
Feiyu Chan 已提交
75
    sequence_scale
76
    fft
77
    phi_data_layout_transform
78
    gpc
79 80
    utf8proc
    gather_scatter_functor)
81

82 83 84 85
if(WITH_FLASHATTN)
  set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} phi_dynload_flashattn)
endif()

W
Wen Sun 已提交
86
set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} process_group)
L
LiYuRio 已提交
87
if(WITH_NCCL OR WITH_RCCL)
88 89 90 91 92
  set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} process_group_nccl
                         nccl_comm_context)
endif()
if(WITH_GLOO)
  set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} gloo_comm_context)
L
LiYuRio 已提交
93
endif()
94
set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} processgroup_comm_utils)
95 96 97
if(WITH_CUDNN_FRONTEND)
  set(COMMON_KERNEL_DEPS ${COMMON_KERNEL_DEPS} cudnn-frontend)
endif()
98
copy_if_different(${kernel_declare_file} ${kernel_declare_file_final})
99 100 101 102 103 104 105 106 107 108 109 110 111 112

file(GLOB kernel_h "*.h" "selected_rows/*.h" "sparse/*.h" "strings/*.h")
file(GLOB kernel_impl_h "impl/*.h" "selected_rows/impl/*.h")
file(GLOB kernel_primitive_h "primitive/*.h")

file(
  GLOB
  kernel_cu
  "gpu/*.cu"
  "gpu/*.cu.cc"
  "gpudnn/*.cu"
  "kps/*.cu"
  "selected_rows/gpu/*.cu"
  "sparse/gpu/*.cu"
113 114
  "strings/gpu/*.cu"
  "fusion/gpu/*.cu")
115

116
if(WITH_CUTLASS)
117
  file(GLOB cutlass_cu "fusion/cutlass/conv2d/*.cu" "fusion/cutlass/*.cu")
118 119 120
  list(APPEND kernel_cu ${cutlass_cu})
endif()

121 122 123 124 125 126 127 128 129 130 131 132
if(WITH_MKLDNN)
  file(
    GLOB
    kernel_cc
    "*.cc"
    "cpu/*.cc"
    "selected_rows/*.cc"
    "selected_rows/cpu/*.cc"
    "sparse/*.cc"
    "sparse/cpu/*.cc"
    "strings/*.cc"
    "strings/cpu/*.cc"
133 134
    "onednn/*.cc"
    "fusion/*.cc"
135
    "fusion/onednn/*.cc"
136
    "fusion/cpu/*.cc")
137 138 139 140 141 142 143 144 145 146 147
else()
  file(
    GLOB
    kernel_cc
    "*.cc"
    "cpu/*.cc"
    "selected_rows/*.cc"
    "selected_rows/cpu/*.cc"
    "sparse/*.cc"
    "sparse/cpu/*.cc"
    "strings/*.cc"
148 149 150
    "strings/cpu/*.cc"
    "fusion/*.cc"
    "fusion/cpu/*.cc")
151 152
endif()

153
file(GLOB kernel_xpu "xpu/*.cc" "selected_rows/xpu/*.cc" "fusion/xpu/*.cc")
154 155 156

add_library(phi_cpu ${kernel_cc})
kernel_declare("${kernel_cc}")
157 158 159 160 161 162
if(WITH_MKLDNN)
  target_link_libraries(phi_cpu ${COMMON_KERNEL_DEPS}
                        get_kerneltype_forvar_utils)
else()
  target_link_libraries(phi_cpu ${COMMON_KERNEL_DEPS})
endif()
163 164

set(ADD_PHI_KERNELS phi_cpu)
165 166 167 168 169 170 171 172 173

if(WITH_GPU OR WITH_ROCM)
  if(WITH_GPU)
    add_library(phi_gpu ${kernel_cu})
  elseif(WITH_ROCM)
    hip_add_library(phi_gpu STATIC ${kernel_cu})
  endif()
  kernel_declare("${kernel_cu}")
  target_link_libraries(phi_gpu ${COMMON_KERNEL_DEPS})
174
  set(ADD_PHI_KERNELS ${ADD_PHI_KERNELS} phi_gpu)
175 176 177 178
endif()

if(WITH_XPU)
  if(WITH_XPU_KP)
L
Leo Chen 已提交
179 180 181 182 183 184 185 186 187 188
    file(COPY ${CMAKE_CURRENT_SOURCE_DIR}/kps/
         DESTINATION ${CMAKE_CURRENT_BINARY_DIR}/kps/)
    file(GLOB kernel_xpu_kps "${CMAKE_CURRENT_BINARY_DIR}/kps/*.cu")
    foreach(kernel ${kernel_xpu_kps})
      get_filename_component(name ${kernel} NAME_WE)
      file(RENAME ${kernel} "${CMAKE_CURRENT_BINARY_DIR}/kps/${name}.kps")
    endforeach()
    file(GLOB kernel_xpu_kps "${CMAKE_CURRENT_BINARY_DIR}/kps/*.kps")
    xpu_add_library(phi_xpu STATIC ${kernel_xpu} ${kernel_xpu_kps} DEPENDS
                    ${COMMON_KERNEL_DEPS})
189 190 191
  else()
    add_library(phi_xpu ${kernel_xpu})
  endif()
192 193
  kernel_declare("${kernel_xpu}")
  kernel_declare("${kernel_xpu_kps}")
194
  target_link_libraries(phi_xpu ${COMMON_KERNEL_DEPS})
195 196 197 198
  set(ADD_PHI_KERNELS ${ADD_PHI_KERNELS} phi_xpu)
endif()

set_property(GLOBAL PROPERTY PHI_KERNELS ${ADD_PHI_KERNELS})
199 200 201 202

if(NOT "${KERNEL_LIST}" STREQUAL "")
  prune_declaration_h()
endif()