Good news! Our new work exhibits state-of-the-art performances on the DocUNet Benchmark dataset: DocScanner: Robust Document Image Rectification with Progressive Learning
DocTr: Document Image Transformer for Geometric Unwarping and Illumination Correction
ACM MM 2021 Oral
Any questions or discussions are welcomed!
DocTr consists of two main components: a geometric unwarping transformer (GeoTr) and an illumination correction transformer (IllTr).
- For geometric unwarping, we train the GeoTr network using the Doc3D dataset.
- For illumination correction, we train the IllTr network based on the DRIC dataset.
dataset path structure
DATASET_ROOT/
-imgs/
img files
-uvs/
uv files
to train the model
python train.py
to change config, edit
batch_size = 1
num_epochs = 30
in train.py
- Download the pretrained models from Google Drive or Baidu Cloud, and put them to
$ROOT/model_pretrained/
. - Geometric unwarping:
python inference.py
- Geometric unwarping and illumination rectification:
python inference.py --ill_rec True
- We use the same evaluation code as DocUNet Benchmark dataset based on Matlab 2019a.
- Please compare the scores according to your Matlab version.
- Use the rectified images available from Google Drive or Baidu Cloud for reproducing the quantitative performance on the DocUNet Benchmark reported in the paper and further comparison.
If you find this code useful for your research, please use the following BibTeX entry.
@inproceedings{feng2021doctr,
title={DocTr: Document Image Transformer for Geometric Unwarping and Illumination Correction},
author={Feng, Hao and Wang, Yuechen and Zhou, Wengang and Deng, Jiajun and Li, Houqiang},
booktitle={Proceedings of the 29th ACM International Conference on Multimedia},
pages={273--281},
year={2021}
}
@article{feng2021docscanner,
title={DocScanner: Robust Document Image Rectification with Progressive Learning},
author={Feng, Hao and Zhou, Wengang and Deng, Jiajun and Tian, Qi and Li, Houqiang},
journal={arXiv preprint arXiv:2110.14968},
year={2021}
}