Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/3857
Title: Automated form reading
Authors: Chu, Kim-ching
Subjects: Image processing
Optical character recognition devices
Document imaging systems
Hong Kong Polytechnic University -- Dissertations
Issue Date: 2000
Publisher: The Hong Kong Polytechnic University
Abstract: Forms are used extensively to collect and distribute data. The main task of an automated form reader is to locate the date filled in forms and to encode the content into appropriate symbolic descriptions. In this thesis, we aim to develop efficient algorithms for an automated form reading system. The key problems tackled in this thesis are image preprocessing, script determination, fast keywords matching, and printed character recognition. Preprocessing of digital images is a very important step in document analysis. We discuss in this thesis a combined intensity histogram and a local contrast feature for image binarization, horizontal Run Length Smoothing Algorithm (RLSA) followed by 8-neightbouring connection method for page segmentation, the use of simple criteria for text and line extraction, and fast skew estimation and correction using the extracted lines with a backup of an interline cross-correlation method for those forms without lines. Our approach for skew estimation is efficient and effective for those forms containing a lot of lines. The skewed angled can be detected with an error of smaller than 1°. It is very common that the documents contain more than one script. In this thesis, a robust script determination approach is proposed which can cope with different fonts, sizes, styles and darkness of text in document images. Two neural networks are employed. The first neural network is trained to derive a set of 15 masks which are used for extracting 15 features. The coefficients of masks are then quantized for reduced computational complexity. The second neural network is trained with 15 extracted features to perform the script separation. Experimental results show that 97% of the image can be correctly classified.
A Dynamic Recognition Neural Network (DRNN) is proposed in this thesis to perform fast keywords matching. Different sets of features are used to deal with different scripts. For English, projection profiles (x and y) are used while for Chinese, contour features are utilized. Testing on 29 name cards shows that a 90% correct matching rate can be achieved. An algorithm based on the vertical projection and a peak-to-valley function is adopted for segmenting characters. By applying the algorithm on form images with 100 dpi scanning resolution, about 86% of the characters can be correctly segmented. A neural network is then employed to classify the segmented characters into 50 groups. Both intensity features and structure-based features extracted from the skeleton image were utilized. An accuracy of 85% to 87.5% can be achieved when testing on the images with 100 dpi scanning resolution and higher accuracy of 94% to 96.6% can be achieved if the scanning resolution is 150 dpi.
Description: xiii, 80 leaves : ill. ; 31 cm.
PolyU Library Call No.: [THS] LG51 .H577M EIE 2000 Chu
Rights: All rights reserved.
Type: Thesis
URI: http://hdl.handle.net/10397/3857
Appears in Collections:EIE Theses
PolyU Electronic Theses

Files in This Item:
File Description SizeFormat 
b15319878_link.htmFor PolyU Users 161 BHTMLView/Open
b15319878_ir.pdfFor All Users (Non-printable) 2.77 MBAdobe PDFView/Open


All items in the PolyU Institutional Repository are protected by copyright, with all rights reserved, unless otherwise indicated. No item in the PolyU IR may be reproduced for commercial or resale purposes.