test_gemm_accuracy.cpp 3.7 KB
Newer Older
Z
zhaojiaying01 已提交
1 2 3 4 5 6 7 8 9 10 11
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

Y
yanantao78 已提交
12 13
#include <cstdlib>
#include <ctime>
Y
yanantao78 已提交
14
#include <iostream>
Z
zhaojiaying01 已提交
15
#include "../test_helper.h"
Z
zhaojiaying01 已提交
16
#include "common/log.h"
Z
zhaojiaying01 已提交
17
#include "memory/t_malloc.h"
18
#include "operators/math/gemm/cblas.h"
Z
zhaojiaying01 已提交
19

Z
zhaojiaying01 已提交
20 21
#define a(i, j) a[(i)*lda + (j)]
#define b(i, j) b[(i)*ldb + (j)]
Y
yanantao78 已提交
22
#define c(i, j) c[(i)*ldc + (j)]
Z
zhaojiaying01 已提交
23 24
#define c1(i, j) c1[(i)*ldc + (j)]

Z
ZhenWang 已提交
25
void print_matrix(int m, int n, int ldc, float *c) {
Y
yanantao78 已提交
26 27 28 29
  for (int i = 0; i < m; ++i) {
    std::cout << c(i, 0);
    for (int j = 1; j < n; ++j) {
      std::cout << " | " << c(i, j);
Y
yanantao78 已提交
30 31
    }
    std::cout << std::endl;
Y
yanantao78 已提交
32 33
  }
  std::cout << std::endl;
Y
yanantao78 已提交
34 35
}

36 37 38 39 40 41
int do_sgemm(int m, int n, int k, int pr) {
  const float alpha = 1.f;
  const float beta = 0.f;
  const int lda = k;
  const int ldb = n;
  const int ldc = n;
Z
zhaojiaying01 已提交
42

Y
yanantao78 已提交
43 44 45 46 47 48 49 50
  float *a =
      static_cast<float *>(paddle_mobile::memory::Alloc(sizeof(float) * m * k));
  float *b =
      static_cast<float *>(paddle_mobile::memory::Alloc(sizeof(float) * k * n));
  float *c =
      static_cast<float *>(paddle_mobile::memory::Alloc(sizeof(float) * m * n));
  float *c1 =
      static_cast<float *>(paddle_mobile::memory::Alloc(sizeof(float) * m * n));
Y
do trim  
yanantao78 已提交
51

52 53 54 55 56
  std::mt19937 rng(111);
  std::uniform_real_distribution<double> uniform_dist(0, 1);
  const float lower = -10.f;
  const float upper = 10.f;

Z
zhaojiaying01 已提交
57
  for (int i = 0; i < m * k; ++i) {
58
    a[i] = static_cast<float>(uniform_dist(rng) * (upper - lower) + lower);
Z
zhaojiaying01 已提交
59 60
  }
  for (int i = 0; i < k * n; ++i) {
61
    b[i] = static_cast<float>(uniform_dist(rng) * (upper - lower) + lower);
Z
zhaojiaying01 已提交
62
  }
63
  memcpy(c, c1, sizeof(float) * m * n);
Y
do trim  
yanantao78 已提交
64

Y
yanantao78 已提交
65 66 67 68 69 70
  for (int i = 0; i < m; ++i) {
    for (int j = 0; j < n; ++j) {
      float r = 0;
      for (int p = 0; p < k; p++) {
        r += a(i, p) * b(p, j);
      }
71
      c1(i, j) = alpha * r;
Z
zhaojiaying01 已提交
72 73
    }
  }
Y
do trim  
yanantao78 已提交
74

75 76 77 78 79
  std::cout << "run cblas_sgemm..." << std::endl;
  paddle_mobile::operators::math::cblas_sgemm(false, false, m, n, k, alpha, a,
                                              lda, b, ldb, 0.f, c, ldc);

  std::cout << "compare results..." << std::endl;
Z
zhaojiaying01 已提交
80
  for (int i = 0; i < m * n; ++i) {
81 82 83 84
    if (abs(c[i] - c1[i]) >= 1e-2) {
      std::cout << "c[" << i << "] != c1[" << i << "]: " << c[i] << " vs "
                << c1[i] << std::endl;
      exit(1);
Z
zhaojiaying01 已提交
85 86
    }
  }
Y
do trim  
yanantao78 已提交
87

Y
yanantao78 已提交
88 89
  if (pr > 0) {
    std::cout << "A:" << std::endl;
Z
ZhenWang 已提交
90
    print_matrix(m, k, lda, a);
Y
yanantao78 已提交
91
    std::cout << "B:" << std::endl;
Z
ZhenWang 已提交
92
    print_matrix(k, n, ldb, b);
Y
yanantao78 已提交
93
    std::cout << "C:" << std::endl;
Z
ZhenWang 已提交
94
    print_matrix(m, n, ldc, c);
Y
yanantao78 已提交
95
    std::cout << "C1:" << std::endl;
Z
ZhenWang 已提交
96
    print_matrix(m, n, ldc, c1);
Y
yanantao78 已提交
97
  }
Y
do trim  
yanantao78 已提交
98

Y
yanantao78 已提交
99 100 101 102
  paddle_mobile::memory::Free(a);
  paddle_mobile::memory::Free(b);
  paddle_mobile::memory::Free(c);
  paddle_mobile::memory::Free(c1);
Y
do trim  
yanantao78 已提交
103

Y
yanantao78 已提交
104 105 106
  return 0;
}

107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129
int main(int argc, char *argv[]) {
  do_sgemm(1, 1, 1, 1);

  do_sgemm(9, 9, 1, 1);
  do_sgemm(999, 99, 1, 0);
  do_sgemm(999, 1, 1, 0);
  do_sgemm(1, 9, 9, 1);
  do_sgemm(1, 99, 999, 0);
  do_sgemm(1, 1, 999, 0);

  do_sgemm(9, 9, 9, 1);
  do_sgemm(10, 6, 12, 1);
  do_sgemm(512, 256, 384, 0);
  do_sgemm(1366, 768, 256, 0);
  do_sgemm(1255, 755, 333, 0);
  do_sgemm(555, 777, 999, 0);

  do_sgemm(10, 6, 12, 1);
  do_sgemm(512, 256, 384, 0);
  do_sgemm(1366, 768, 256, 0);
  do_sgemm(1255, 755, 333, 0);
  do_sgemm(555, 777, 999, 0);

Z
zhaojiaying01 已提交
130 131
  return 0;
}