You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -36,17 +37,18 @@ MindOCR is an open-source toolbox for OCR development and application based on [
36
37
37
38
## Installation
38
39
39
-
<detailsclosemarkdown>
40
+
<detailsopenmarkdown>
41
+
<summary> Details </summary>
40
42
41
43
#### Prerequisites
42
44
43
45
MindOCR is built on MindSpore AI framework, which supports CPU/GPU/NPU devices.
44
46
MindOCR is compatible with the following framework versions. For details and installation guideline, please refer to the installation links shown below.
- mindspore lite (for inference) [[install](docs/en/inference/environment.md)]
51
+
- mindspore lite (for offline inference) >= 2.2.0[[install](docs/en/inference/environment.md)]
50
52
51
53
52
54
#### Dependency
@@ -126,10 +128,12 @@ python tools/eval.py \
126
128
127
129
For more illustration and usage, please refer to the model training section in [Tutorials](#tutorials).
128
130
129
-
### 3. Model Inference - Quick Guideline
131
+
### 3. Model Offline Inference - Quick Guideline
130
132
131
-
You can do MindSpore Lite inference in MindOCR using **MindOCR models** or **Third-party models** (PaddleOCR, MMOCR, etc.).
132
-
Please refer to [MindOCR Models Inference - Quick Start](docs/en/inference/inference_quickstart.md) or [Third-party Models Inference - Quick Start](docs/en/inference/inference_thirdparty_quickstart.md).
133
+
You can do MindSpore Lite inference in MindOCR using **MindOCR models** or **Third-party models** (PaddleOCR, MMOCR, etc.). Please refer to the following documents
134
+
-[Python/C++ Inference on Ascend 310](docs/en/inference/inference_tutorial.md)
For the detailed performance of the trained models, please refer to [configs](./configs).
188
202
189
203
For details of MindSpore Lite and ACL inference models support, please refer to [MindOCR Models Support List](docs/en/inference/inference_quickstart.md) and [Third-party Models Support List](docs/en/inference/inference_thirdparty_quickstart.md) (PaddleOCR, MMOCR, etc.).
@@ -219,14 +233,49 @@ MindOCR provides a [dataset conversion tool](tools/dataset_converters) to OCR da
We will include more datasets for training and evaluation. This list will be continuously updated.
223
251
252
+
## Frequently Asked Questions
253
+
Frequently asked questions about configuring environment and mindocr, please refer to [FAQ](docs/en/tutorials/frequently_asked_questions.md).
254
+
224
255
## Notes
225
256
226
257
### What is New
258
+
259
+
<detailsclosemarkdown>
260
+
<summary>News</summary>
261
+
227
262
- 2023/12/14
263
+
1. Add new trained models
264
+
-[LayoutXLM SER](configs/kie/vi_layoutxlm) for key information extraction
265
+
-[VI-LayoutXLM SER](configs/kie/layoutlm_series) for key information extraction
266
+
-[PP-OCRv3 DBNet](configs/det/dbnet/db_mobilenetv3_ppocrv3.yaml) for text detection and [PP-OCRv3 SVTR](configs/rec/svtr/svtr_ppocrv3_ch.yaml) for recognition, supporting online inferece and finetuning
267
+
2. Add more benchmark datasets and their results
268
+
-[XFUND](configs/kie/vi_layoutxlm/README_CN.md)
269
+
3. Multiple specifications support for Ascend 910: DBNet ResNet-50, DBNet++ ResNet-50, CRNN VGG7, SVTR-Tiny, FCENet, ABINet
270
+
- 2023/11/28
271
+
1. Add offline inference support for PP-OCRv4
272
+
-[PP-OCRv4 DBNet](deploy/py_infer/src/configs/det/ppocr/ch_PP-OCRv4_det_cml.yaml) for text detection and [PP-OCRv4 CRNN](deploy/py_infer/src/configs/rec/ppocr/ch_PP-OCRv4_rec_distillation.yaml) for text recognition, supporting offline inferece
273
+
2. Fix bugs of third-party models offline inference
274
+
- 2023/11/17
228
275
1. Add new trained models
229
276
-[YOLOv8](configs/layout/yolov8) for layout analysis
277
+
2. Add more benchmark datasets and their results
278
+
-[PublayNet](configs/layout/yolov8/README_CN.md)
230
279
- 2023/07/06
231
280
1. Add new trained models
232
281
-[RobustScanner](configs/rec/robustscanner) for text recognition
@@ -285,13 +334,14 @@ which can be enabled by add "shape_list" to the `eval.dataset.output_columns` li
285
334
- 2023/03/13
286
335
1. Add system test and CI workflow.
287
336
2. Add modelarts adapter to allow training on OpenI platform. To train on OpenI:
288
-
```text
289
-
i) Create a new training task on the openi cloud platform.
290
-
ii) Link the dataset (e.g., ic15_mindocr) on the webpage.
291
-
iii) Add run parameter `config` and write the yaml file path on the website UI interface, e.g., '/home/work/user-job-dir/V0001/configs/rec/test.yaml'
292
-
iv) Add run parameter `enable_modelarts` and set True on the website UI interface.
293
-
v) Fill in other blanks and launch.
294
-
```
337
+
```text
338
+
i) Create a new training task on the openi cloud platform.
339
+
ii) Link the dataset (e.g., ic15_mindocr) on the webpage.
340
+
iii) Add run parameter `config` and write the yaml file path on the website UI interface, e.g., '/home/work/user-job-dir/V0001/configs/rec/test.yaml'
341
+
iv) Add run parameter `enable_modelarts` and set True on the website UI interface.
0 commit comments