1. 18 10月, 2017 6 次提交
    • M
      MatMul operator (#4856) · 16489827
      Markus Kliegl 提交于
      * initial matmul operator
      
      Similar to np.matmul, but also has transpose_X and transpose_Y flags,
      and only supports tensors from rank 1 to 3 inclusive.
      
      For GPU, uses cublas?gemmStridedBatched. For CPU, uses
      cblas_?gemm_batch if available via MKL; otherwise a simple serial
      implementation that loops over the batch dimension is employed for now.
      16489827
    • D
      Merge pull request #4571 from sidgoyal78/momentum_op · fd96914d
      dzhwinter 提交于
      Add momentum operator
      fd96914d
    • Y
      Remove private data members in OpRegister (#4871) · 5d67677c
      Yu Yang 提交于
      5d67677c
    • A
    • Q
      fix mac compile (#4870) · 46f7854e
      Qiao Longfei 提交于
      46f7854e
    • Q
      Impl optimizer (#4734) · df0946eb
      Qiao Longfei 提交于
      * init parameter base class
      
      * optimize the Comments of optimizer
      
      * basic implimentation of optimizer
      
      * add test_optimizer
      
      * add no_grad_set to interface
      
      * update optimizer.py
      
      * python code can run
      
      * fix some problem
      
      * add sync_with_cpp to Python Program and Block
      
      * sync vars and ops in block from cpp
      
      * optimize code and add some comment
      
      * add more check for sync
      
      * update optimizer with return value of Backward
      
      * rm unused code
      
      * infer shape when create gradient vairiable
      
      * update test_optimizer
      
      * update test_program.py
      
      * update backward test
      
      * follow comment
      df0946eb
  2. 17 10月, 2017 34 次提交