提交 f3d9c641 编写于 作者: Z zhangyang0701

change format

上级 111d156d
...@@ -75,7 +75,8 @@ void convert_to_chw(float **data_in, int channel, int height, int width, ...@@ -75,7 +75,8 @@ void convert_to_chw(float **data_in, int channel, int height, int width,
} }
} }
void dump_stride_float(std::string filename, paddle_mobile::PaddleTensor input_tensor) { void dump_stride_float(std::string filename,
paddle_mobile::PaddleTensor input_tensor) {
auto data_ptr = reinterpret_cast<float *>(input_tensor.data.data()); auto data_ptr = reinterpret_cast<float *>(input_tensor.data.data());
int c = (input_tensor.shape)[1]; int c = (input_tensor.shape)[1];
int h = (input_tensor.shape)[2]; int h = (input_tensor.shape)[2];
...@@ -83,7 +84,7 @@ void dump_stride_float(std::string filename, paddle_mobile::PaddleTensor input_t ...@@ -83,7 +84,7 @@ void dump_stride_float(std::string filename, paddle_mobile::PaddleTensor input_t
int n = (input_tensor.shape)[0]; int n = (input_tensor.shape)[0];
float *data_tmp = float *data_tmp =
reinterpret_cast<float *>(malloc(c * h * w * sizeof(float))); reinterpret_cast<float *>(malloc(c * h * w * sizeof(float)));
//convert_to_chw(&data_ptr, c, h, w, data_tmp); // convert_to_chw(&data_ptr, c, h, w, data_tmp);
std::ofstream out(filename.c_str()); std::ofstream out(filename.c_str());
float result = 0; float result = 0;
int datasize = abs(c * h * w * n); int datasize = abs(c * h * w * n);
...@@ -98,7 +99,8 @@ void dump_stride_float(std::string filename, paddle_mobile::PaddleTensor input_t ...@@ -98,7 +99,8 @@ void dump_stride_float(std::string filename, paddle_mobile::PaddleTensor input_t
out.close(); out.close();
} }
void dump_stride(std::string filename, paddle_mobile::PaddleTensor input_tensor) { void dump_stride(std::string filename,
paddle_mobile::PaddleTensor input_tensor) {
if (input_tensor.dtypeid == typeid(float)) { if (input_tensor.dtypeid == typeid(float)) {
dump_stride_float(filename, input_tensor); dump_stride_float(filename, input_tensor);
} else { } else {
...@@ -135,17 +137,16 @@ PaddleMobileConfig GetConfig1() { ...@@ -135,17 +137,16 @@ PaddleMobileConfig GetConfig1() {
int main() { int main() {
open_device(); open_device();
PaddleMobileConfig config1 = GetConfig1(); PaddleMobileConfig config1 = GetConfig1();
auto predictor1 = auto predictor1 =
CreatePaddlePredictor<PaddleMobileConfig, CreatePaddlePredictor<PaddleMobileConfig,
PaddleEngineKind::kPaddleMobile>(config1); PaddleEngineKind::kPaddleMobile>(config1);
std::cout << "Finishing loading model" << std::endl; std::cout << "Finishing loading model" << std::endl;
for(int i = 0; i < 1; ++i){ for (int i = 0; i < 1; ++i) {
int img_length1 = 144 * 14 * 14; int img_length1 = 144 * 14 * 14;
auto img1 = reinterpret_cast<float *>(fpga_malloc(img_length1 * sizeof(float))); auto img1 =
reinterpret_cast<float *>(fpga_malloc(img_length1 * sizeof(float)));
readStream(g_image1, reinterpret_cast<char *>(img1)); readStream(g_image1, reinterpret_cast<char *>(img1));
std::cout << "Finishing initializing data" << std::endl; std::cout << "Finishing initializing data" << std::endl;
...@@ -170,8 +171,8 @@ for(int i = 0; i < 1; ++i){ ...@@ -170,8 +171,8 @@ for(int i = 0; i < 1; ++i){
std::string dumpName = "marker2_api_fetch_" + std::to_string(fetchNum); std::string dumpName = "marker2_api_fetch_" + std::to_string(fetchNum);
dump_stride(dumpName, v1[fetchNum]); dump_stride(dumpName, v1[fetchNum]);
} }
} }
///////////////////////////////////// /////////////////////////////////////
PaddleMobileConfig config = GetConfig(); PaddleMobileConfig config = GetConfig();
auto predictor = auto predictor =
...@@ -194,14 +195,14 @@ for(int i = 0; i < 1; ++i){ ...@@ -194,14 +195,14 @@ for(int i = 0; i < 1; ++i){
t_img_info.data.Reset(img_info, 3 * sizeof(float)); t_img_info.data.Reset(img_info, 3 * sizeof(float));
t_img.dtypeid = typeid(float); t_img.dtypeid = typeid(float);
//quantize(&img, img_length); // quantize(&img, img_length);
//t_img.dtypeid = typeid(int8_t); // t_img.dtypeid = typeid(int8_t);
t_img.layout = LAYOUT_HWC; t_img.layout = LAYOUT_HWC;
t_img.shape = std::vector<int>({1, 432, 1280, 3}); t_img.shape = std::vector<int>({1, 432, 1280, 3});
t_img.name = "Image information"; t_img.name = "Image information";
t_img.data.Reset(img, img_length * sizeof(float)); t_img.data.Reset(img, img_length * sizeof(float));
//t_img.data.Reset(img, img_length * sizeof(int8_t)); // t_img.data.Reset(img, img_length * sizeof(int8_t));
// for(int i = 0; i < 100; ++i){ // for(int i = 0; i < 100; ++i){
predictor->FeedPaddleTensors({t_img_info, t_img}); predictor->FeedPaddleTensors({t_img_info, t_img});
std::cout << "Finishing feeding data " << std::endl; std::cout << "Finishing feeding data " << std::endl;
......
...@@ -19,8 +19,8 @@ limitations under the License. */ ...@@ -19,8 +19,8 @@ limitations under the License. */
#include <iostream> #include <iostream>
#include "../../src/io/paddle_inference_api.h" #include "../../src/io/paddle_inference_api.h"
using namespace paddle_mobile; //NOLINT using namespace paddle_mobile; // NOLINT
using namespace paddle_mobile::fpga; //NOLINT using namespace paddle_mobile::fpga; // NOLINT
static const char *g_image = "../images/mobilenet_txtdata/1.txt"; static const char *g_image = "../images/mobilenet_txtdata/1.txt";
static const char *g_model = "../models/keycurve_l2_regular4_model/__model__"; static const char *g_model = "../models/keycurve_l2_regular4_model/__model__";
...@@ -119,7 +119,9 @@ PaddleMobileConfig GetConfig() { ...@@ -119,7 +119,9 @@ PaddleMobileConfig GetConfig() {
int main() { int main() {
open_device(); open_device();
PaddleMobileConfig config = GetConfig(); PaddleMobileConfig config = GetConfig();
auto predictor = CreatePaddlePredictor<paddle_mobile::PaddleMobileConfig, PaddleEngineKind::kPaddleMobile>(config); auto predictor =
CreatePaddlePredictor<paddle_mobile::PaddleMobileConfig,
PaddleEngineKind::kPaddleMobile>(config);
std::cout << "Finishing loading model" << std::endl; std::cout << "Finishing loading model" << std::endl;
int img_length = 256 * 416 * 3; int img_length = 256 * 416 * 3;
......
...@@ -19,8 +19,8 @@ limitations under the License. */ ...@@ -19,8 +19,8 @@ limitations under the License. */
#include <iostream> #include <iostream>
#include "../../src/io/paddle_inference_api.h" #include "../../src/io/paddle_inference_api.h"
using namespace paddle_mobile; //NOLINT using namespace paddle_mobile; // NOLINT
using namespace paddle_mobile::fpga; //NOLINT using namespace paddle_mobile::fpga; // NOLINT
static const char *g_image = "../images/yolo_test_txtimg/1.txt"; static const char *g_image = "../images/yolo_test_txtimg/1.txt";
static const char *g_model = "../models/yolo_bn_l2_model/__model__"; static const char *g_model = "../models/yolo_bn_l2_model/__model__";
...@@ -51,8 +51,7 @@ signed char float_to_int8(float fdata) { ...@@ -51,8 +51,7 @@ signed char float_to_int8(float fdata) {
} }
void quantize(float **data_in, int data_size) { void quantize(float **data_in, int data_size) {
float *tmp = *data_in; float *tmp = *data_in;
signed char *tmp_data = signed char *tmp_data = (signed char *)fpga_malloc(data_size * sizeof(char));
(signed char *)fpga_malloc(data_size * sizeof(char));
for (int i = 0; i < data_size; i++) { for (int i = 0; i < data_size; i++) {
tmp_data[i] = float_to_int8((*data_in)[i] + 128); tmp_data[i] = float_to_int8((*data_in)[i] + 128);
} }
...@@ -120,7 +119,9 @@ PaddleMobileConfig GetConfig() { ...@@ -120,7 +119,9 @@ PaddleMobileConfig GetConfig() {
int main() { int main() {
open_device(); open_device();
PaddleMobileConfig config = GetConfig(); PaddleMobileConfig config = GetConfig();
auto predictor = CreatePaddlePredictor<PaddleMobileConfig, PaddleEngineKind::kPaddleMobile>(config); auto predictor =
CreatePaddlePredictor<PaddleMobileConfig,
PaddleEngineKind::kPaddleMobile>(config);
std::cout << "Finishing loading model" << std::endl; std::cout << "Finishing loading model" << std::endl;
int img_length = 256 * 416 * 3; int img_length = 256 * 416 * 3;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册