Paper
24 January 2011 High recall document content extraction
Chang An, Henry S. Baird
Author Affiliations +
Proceedings Volume 7874, Document Recognition and Retrieval XVIII; 787405 (2011) https://doi.org/10.1117/12.876706
Event: IS&T/SPIE Electronic Imaging, 2011, San Francisco Airport, California, United States
Abstract
We report methodologies for computing high-recall masks for document image content extraction, that is, the location and segmentation of regions containing handwriting, machine-printed text, photographs, blank space, etc. The resulting segmentation is pixel-accurate, which accommodates arbitrary zone shapes (not merely rectangles). We describe experiments showing that iterated classifiers can increase recall of all content types, with little loss of precision. We also introduce two methodological enhancements: (1) a multi-stage voting rule; and (2) a scoring policy that views blank pixels as a "don't care" class with other content classes. These enhancements improve both recall and precision, achieving at least 89% recall and at least 87% precision among three content types: machine-print, handwriting, and photo.
© (2011) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Chang An and Henry S. Baird "High recall document content extraction", Proc. SPIE 7874, Document Recognition and Retrieval XVIII, 787405 (24 January 2011); https://doi.org/10.1117/12.876706
Lens.org Logo
CITATIONS
Cited by 1 scholarly publication.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Image segmentation

Photography

Feature extraction

Image enhancement

Image classification

Eye

Image processing algorithms and systems

Back to Top