mirror of
https://github.com/PaddlePaddle/PaddleOCR.git
synced 2025-06-03 21:53:39 +08:00
[Windows Inference Library (version 3.0.0-beta1) - MKL](https://www.paddlepaddle.org.cn/inference/master/guides/install/download_lib.html#windows) renames the mkldnn folder to onednn.
English | 简体中文
PP-OCR Deployment
Paddle Deployment Introduction
Paddle provides a variety of deployment schemes to meet the deployment requirements of different scenarios. Please choose according to the actual situation:
PP-OCR Deployment
PP-OCR has supported muti deployment schemes. Click the link to get the specific tutorial.
- Python Inference
- C++ Inference
- Serving (Python/C++)
- Paddle-Lite (ARM CPU/OpenCL ARM GPU)
- Paddle.js
- Jetson Inference
- Paddle2ONNX
If you need the deployment tutorial of academic algorithm models other than PP-OCR, please directly enter the main page of corresponding algorithms, entrance。