名称 最后提交 最后更新
..
mkldnn disable copying of datatype when sharing buffer between two tensors. (#37247)
CMakeLists.txt [NPU] cherry-pick basic NPU components/allocator/operator/executor supports from ascendrc (#32144)
elementwise_add_op.cc Support elementwise_add triple grad Kernel (#36508)
elementwise_add_op.cu Add the transformop parameter in TensorReduceFunctorImpl (#38135)
elementwise_add_op.h [Pten]Move math to new directory and change 「math」 to 「math_kernel」 (#38604)
elementwise_add_op_npu.cc [NPU] reorganization for device API abstraction (#37110)
elementwise_add_op_xpu.cc update elementwise api in kunlun (#35021)
elementwise_div_op.cc [PTen]Elementwise_div Kernel Refactor (#37418)
elementwise_div_op.cu Remove useless headers for some grad ops (#38732)
elementwise_div_op.h Remove useless headers for some grad ops (#38732)
elementwise_div_op_npu.cc [NPU] reorganization for device API abstraction (#37110)
elementwise_div_op_xpu.cc update elementwise api in kunlun (#35021)
elementwise_floordiv_op.cc add the op version check for the elementwise ops, test=op_version (#30010)
elementwise_floordiv_op.cu Unify the functor definition of elementwise add, sub, mul, div, floordiv, max, min. (#35684)
elementwise_floordiv_op.h Unify the functor definition of elementwise add, sub, mul, div, floordiv, max, min. (#35684)
elementwise_floordiv_op_npu.cc [NPU] reorganization for device API abstraction (#37110)
elementwise_floordiv_op_xpu.cc update elementwise api in kunlun (#35021)
elementwise_functor.h optimize elementwise_mul_grad using new interfaces (#37728)
elementwise_max_op.cc Add fmax and fmin operators (#37826)
elementwise_max_op.cu Add fmax and fmin operators (#37826)
elementwise_max_op.h Add fmax and fmin operators (#37826)
elementwise_max_op_npu.cc [NPU] reorganization for device API abstraction (#37110)
elementwise_max_op_xpu.cc update elementwise api in kunlun (#35021)
elementwise_min_op.cc Add fmax and fmin operators (#37826)
elementwise_min_op.cu Support FP16 for more ops (#38123)
elementwise_min_op.h
elementwise_min_op_npu.cc
elementwise_min_op_xpu.cc
elementwise_mod_op.cc
elementwise_mod_op.cu
elementwise_mod_op.h
elementwise_mod_op_npu.cc
elementwise_mul_op.cc
elementwise_mul_op.cu
elementwise_mul_op.h
elementwise_mul_op_npu.cc
elementwise_mul_op_xpu.cc
elementwise_npu.h
elementwise_op.h
elementwise_op_broadcast.cu.h
elementwise_op_function.h
elementwise_op_impl.cu.h
elementwise_op_npu_test.cc
elementwise_pow_op.cc
elementwise_pow_op.cu
elementwise_pow_op.h
elementwise_pow_op_npu.cc
elementwise_pow_op_xpu.cc
elementwise_sub_op.cc
elementwise_sub_op.cu
elementwise_sub_op.h
elementwise_sub_op_npu.cc
elementwise_sub_op_xpu.cc
elementwise_xpu.h
test_elementwise_add_grad_grad.cc
test_elementwise_add_op_inplace.cc
test_elementwise_div_grad_grad.cc
test_elementwise_op_grad_grad.h
unity_build_rule.cmake

项目简介

PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)

🚀 Github 镜像仓库 🚀

源项目地址

https://github.com/paddlepaddle/paddle

deep-learningdistributed-trainingefficiencymachine-learningneural-networkpaddlepaddlepythonscalability

发行版本 60

PaddlePaddle 2.5.0 Release Note

全部发行版

贡献者 246

全部贡献者

开发语言

  • C++ 49.8 %
  • Python 41.0 %
  • Cuda 7.0 %
  • CMake 1.1 %
  • Shell 0.6 %
反馈
建议
客服 返回
顶部