diff --git a/lite/backends/arm/math/gemm_prepacked_int8.cc b/lite/backends/arm/math/gemm_prepacked_int8.cc index 86ee6e4635def6379486a73a9db5954479338089..24991750a33e53699b284d2a440e5df3f99f8cf6 100644 --- a/lite/backends/arm/math/gemm_prepacked_int8.cc +++ b/lite/backends/arm/math/gemm_prepacked_int8.cc @@ -2026,7 +2026,6 @@ inline void gemm_sdot_int8_kernel(const int8_t* a_ptr, "vadd.f32 q3, q11, q3\n" /* r21, add offset */ \ "vadd.f32 q4, q12, q4\n" /* r30, add offset */ \ "vadd.f32 q5, q13, q5\n" /* r31, add offset */ \ - /*"vld1.32 {d12-d13}, [%[vmax]]\n"*/ /* set q4 = -127 \n"*/ \ "vmov.f32 q6, #-127.0\n" \ "vcge.f32 q7, q8, q6\n" /* @ q8 >= -127 \n */ \ "vcge.f32 q10, q9, q6\n" /* @ q8 >= -127 \n */ \ @@ -2131,7 +2130,6 @@ inline void gemm_int8_kernel(const int8_t* a_ptr, int is_relu, int k, int rem) { - //float vmax[4] = {-127.0, -127.0, -127.0, -127.0}; asm volatile(GEMM_INT8_KERNEL GEMM_INT8_INT8_OUT : [a_ptr] "+r"(a_ptr), [b_ptr] "+r"(b_ptr), @@ -2144,7 +2142,6 @@ inline void gemm_int8_kernel(const int8_t* a_ptr, [alpha] "r"(alpha), [bias] "r"(bias), [rem] "r"(rem), - // [vmax] "r"(vmax), [scale] "r"(scale) : "q0", "q1",