commit
a01794954c
43
README.md
43
README.md
|
@ -4,24 +4,36 @@
|
|||
# Kindai-OCR
|
||||
OCR system for recognizing modern Japanese magazines
|
||||
|
||||
# Updates:
|
||||
Kindai V2.0 employed Transformer OCR for text recognition. Transformer OCR was trained on NDL and CODH datasets.
|
||||
|
||||
## About
|
||||
|
||||
This repo contains an OCR sytem for converting modern Japanese images to text.
|
||||
This is a result of [N2I project](http://codh.rois.ac.jp/collaboration/#n2i) for digitization of modern Japanese documents.
|
||||
This repo contains an OCR system for converting modern Japanese images to text. The software has been developed by Dr. Anh Duc Le, while he was working for <a href="http://codh.rois.ac.jp/">ROIS-DS Center for Open Data in the Humanities</a>.
|
||||
|
||||
The system has 2 main modules: text line extraction and text line recognition. The overall architechture is shown in the below figures.
|
||||
![alt text](https://github.com/ducanh841988/Kindai-OCR/blob/master/images/TextlineExtraction.jpg "text line extraction")
|
||||
The system has 2 main modules: text line extraction and text line recognition. The overall architecture is shown in the below figures.
|
||||
|
||||
### Text line extraction
|
||||
We retrain the CRAFT (Character Region Awareness for Text Detection) on 1000 annotated images provided by Center for Research and Development of Higher Education, The University of Tokyo.
|
||||
|
||||
<img src="./images/TextlineExtraction.jpg" height="200" />
|
||||
|
||||
### Text line recognition
|
||||
For Kindai V1.0, we employ the attention-based encoder-decoder on our previous publication. We train the text line recognition on 1000 annotated images and 1600 unannotated images provided by Center for Research and Development of Higher Education, The University of Tokyo and National Institute for Japanese Language and Linguistics, respectively.
|
||||
|
||||
<img src="./images/kindai_v1.jpg" height="200" />
|
||||
|
||||
For Kindai V2.0, we trained a transformer with more data from National Diet Library (NDL) and The Center for Open Data in The Humanities (CODH).
|
||||
[NDL dataset](https://github.com/ndl-lab/pdmocrdataset-part2) contains 3,997 pages, 103,256 lines and [CODH dataset](http://codh.rois.ac.jp/modern-magazine/dataset/) contains 1985 pages and 59,465 lines.
|
||||
|
||||
|
||||
<img src="./images/kindai_v2.png" height="200" />
|
||||
|
||||
For text line extraction, we retrain the CRAFT (Character Region Awareness for Text Detection) on 1000 annotated images provided by Center for Research and Development of Higher Education, The University of Tokyo.
|
||||
![alt text](https://github.com/ducanh841988/Kindai-OCR/blob/master/images/TextlineRecognition.jpg "text line recognition")
|
||||
|
||||
Text line recognition,
|
||||
For Kindai V1.0, we employ the attention-based encoder-decoder on our previous publication. We train the text line recognition on 1000 annotated images and 1600 unannotated images provided by Center for Research and Development of Higher Education, The University of Tokyo and National Institute for Japanese Language and Linguistics, respectively.
|
||||
For Kindai V2.0, we trained a transformer with more data from National Diet Library and The Center for Open Data in The Humanities.
|
||||
|
||||
## Installing Kindai OCR
|
||||
|
||||
Python==3.7.11
|
||||
```Python==3.7.11
|
||||
torch==1.7.0
|
||||
torchvision==0.8.1
|
||||
opencv-python==3.4.2.17
|
||||
|
@ -31,13 +43,13 @@ Polygon3
|
|||
pillow==4.3.0
|
||||
pytorch-lightning==1.3.5
|
||||
einops==0.3.0
|
||||
editdistance==0.5.3
|
||||
editdistance==0.5.3
|
||||
```
|
||||
|
||||
|
||||
## Running Kindai OCR
|
||||
- You should first download the pre_trained models and put them into ./pretrain/ folder.
|
||||
[VGG model](https://drive.google.com/file/d/1_A1dEFKxyiz4Eu1HOCDbjt1OPoEh90qr/view?usp=sharing), [CRAFT model](https://drive.google.com/file/d/1-9xt_jjs4btMrz5wzrU1-kyp2c6etFab/view?usp=sharing), [OCR V1.0 model](https://drive.google.com/file/d/1mibg7D2D5rvPhhenLeXNilSLMBloiexl/view?usp=sharing)
|
||||
[OCR V2.0 model] (https://drive.google.com/file/d/1cq4PwPS2mXXRjOApst2i7n4G3mBSVqpI/view?usp=drive_link)
|
||||
[VGG model](https://drive.google.com/file/d/1_A1dEFKxyiz4Eu1HOCDbjt1OPoEh90qr/view?usp=sharing), [CRAFT model](https://drive.google.com/file/d/1-9xt_jjs4btMrz5wzrU1-kyp2c6etFab/view?usp=sharing), [OCR V1.0 model](https://drive.google.com/file/d/1mibg7D2D5rvPhhenLeXNilSLMBloiexl/view?usp=sharing), [OCR V2.0 model](https://drive.google.com/file/d/1cq4PwPS2mXXRjOApst2i7n4G3mBSVqpI/view?usp=drive_link)
|
||||
- Copy your images into ./data/test/ folder
|
||||
- run the following script to recognize images:
|
||||
`python test_kindai_1.0.py`
|
||||
|
@ -48,9 +60,10 @@ editdistance==0.5.3
|
|||
- using --cuda = True for GPU device and Fasle for CPU device
|
||||
- using --canvas_size ot set image size for text line detection
|
||||
- An example result from our OCR system
|
||||
<img src="https://github.com/ducanh841988/Kindai-OCR/blob/master/data/result1/res_k188701_021_39.jpg" width="700">
|
||||
|
||||
## Running Kindai OCR
|
||||
<img src="./data/result1/res_k188701_021_39.jpg" width="500">
|
||||
|
||||
## Citation
|
||||
If you find Kindai OCR useful in your research, please consider citing:
|
||||
Anh Duc Le, Daichi Mochihashi, Katsuya Masuda, Hideki Mima, and Nam Tuan Ly. 2019. Recognition of Japanese historical text lines by an attention-based encoder-decoder and text line generation. In Proceedings of the 5th International Workshop on Historical Document Imaging and Processing (HIP ’19). Association for Computing Machinery, New York, NY, USA, 37–41. DOI:https://doi.org/10.1145/3352631.3352641
|
||||
|
||||
|
|
Before Width: | Height: | Size: 69 KiB After Width: | Height: | Size: 69 KiB |
Binary file not shown.
After Width: | Height: | Size: 186 KiB |
Loading…
Reference in New Issue