diff --git a/deploy/jeston/readme.md b/deploy/jeston/readme.md index 21c35f46b68312b639dc2f8505f568698cd2aed2..c1ff8cb721e51e3e3608c03fa91d165c6a31f1f0 100644 --- a/deploy/jeston/readme.md +++ b/deploy/jeston/readme.md @@ -1,14 +1,14 @@ -# Jeston部署PaddleOCR模型 +# Jetson部署PaddleOCR模型 -本节介绍PaddleOCR在Jeston NX、TX2、nano、AGX等系列硬件的部署。 +本节介绍PaddleOCR在Jetson NX、TX2、nano、AGX等系列硬件的部署。 ## 1. 环境准备 -需要准备一台Jeston开发板,如果需要TensorRT预测,需准备好TensorRT环境,建议使用7.1.3版本的TensorRT; +需要准备一台Jetson开发板,如果需要TensorRT预测,需准备好TensorRT环境,建议使用7.1.3版本的TensorRT; -1. jeston安装PaddlePaddle +1. Jetson安装PaddlePaddle PaddlePaddle下载[链接](https://www.paddlepaddle.org.cn/inference/user_guides/download_lib.html#python) 请选择适合的您Jetpack版本、cuda版本、trt版本的安装包。 @@ -33,12 +33,12 @@ cd PaddleOCR pip3 install -r requirements.txt ``` -*注:jeston硬件CPU较差,依赖安装较慢,请耐心等待* +*注:jetson硬件CPU较差,依赖安装较慢,请耐心等待* ## 2. 执行预测 -从[文档](https://github.com/PaddlePaddle/PaddleOCR/blob/dygraph/doc/doc_ch/ppocr_introduction.md#6-%E6%A8%A1%E5%9E%8B%E5%BA%93) 模型库中获取PPOCR模型,下面以PP-OCRv3模型为例,介绍在PPOCR模型在jeston上的使用方式: +从[文档](https://github.com/PaddlePaddle/PaddleOCR/blob/dygraph/doc/doc_ch/ppocr_introduction.md#6-%E6%A8%A1%E5%9E%8B%E5%BA%93) 模型库中获取PPOCR模型,下面以PP-OCRv3模型为例,介绍在PPOCR模型在jetson上的使用方式: 下载并解压PP-OCRv3模型 ``` diff --git a/deploy/jeston/readme_en.md b/deploy/jeston/readme_en.md index b4e183503f1d06d2b4ba8dae7fb99d1321681919..1c81395df55dca381fff1adb20451ef7252be08b 100644 --- a/deploy/jeston/readme_en.md +++ b/deploy/jeston/readme_en.md @@ -1,14 +1,14 @@ -# Jeston Deployment for PaddleOCR +# Jetson Deployment for PaddleOCR -This section introduces the deployment of PaddleOCR on Jeston NX, TX2, nano, AGX and other series of hardware. +This section introduces the deployment of PaddleOCR on Jetson NX, TX2, nano, AGX and other series of hardware. ## 1. Prepare Environment -You need to prepare a Jeston development hardware. If you need TensorRT, you need to prepare the TensorRT environment. It is recommended to use TensorRT version 7.1.3; +You need to prepare a Jetson development hardware. If you need TensorRT, you need to prepare the TensorRT environment. It is recommended to use TensorRT version 7.1.3; -1. Install PaddlePaddle in Jeston +1. Install PaddlePaddle in Jetson The PaddlePaddle download [link](https://www.paddlepaddle.org.cn/inference/user_guides/download_lib.html#python) Please select the appropriate installation package for your Jetpack version, cuda version, and trt version. Here, we download paddlepaddle_gpu-2.3.0rc0-cp36-cp36m-linux_aarch64.whl. @@ -32,11 +32,11 @@ cd PaddleOCR pip3 install -r requirements.txt ``` -*Note: Jeston hardware CPU is poor, dependency installation is slow, please wait patiently* +*Note: Jetson hardware CPU is poor, dependency installation is slow, please wait patiently* ## 2. Perform prediction -Obtain the PPOCR model from the [document](https://github.com/PaddlePaddle/PaddleOCR/blob/dygraph/doc/doc_ch/ppocr_introduction.md#6-%E6%A8%A1%E5%9E%8B%E5%BA%93) model library. The following takes the PP-OCRv3 model as an example to introduce the use of the PPOCR model on jeston: +Obtain the PPOCR model from the [document](https://github.com/PaddlePaddle/PaddleOCR/blob/dygraph/doc/doc_ch/ppocr_introduction.md#6-%E6%A8%A1%E5%9E%8B%E5%BA%93) model library. The following takes the PP-OCRv3 model as an example to introduce the use of the PPOCR model on Jetson: Download and unzip the PP-OCRv3 models. ```