IMAGE S EGMENTATION AND TEXT EXTRACTION: APPLICATION TO THE EXTRACTION OF TEXTUAL INFORMATION IN SCENE IMAGES

Similar documents
Scene Text Detection Using Machine Learning Classifiers

OCR For Handwritten Marathi Script

Recognition of Gurmukhi Text from Sign Board Images Captured from Mobile Camera

Segmentation Framework for Multi-Oriented Text Detection and Recognition

An Efficient Character Segmentation Based on VNP Algorithm

EXTRACTING TEXT FROM VIDEO

Automatic Detection of Change in Address Blocks for Reply Forms Processing

Bus Detection and recognition for visually impaired people

Time Stamp Detection and Recognition in Video Frames

SCENE TEXT BINARIZATION AND RECOGNITION

Automatically Algorithm for Physician s Handwritten Segmentation on Prescription

Text Information Extraction And Analysis From Images Using Digital Image Processing Techniques

Text Detection in Indoor/Outdoor Scene Images

Keyword Spotting in Document Images through Word Shape Coding

I. INTRODUCTION. Figure-1 Basic block of text analysis

CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM

A New Approach to Detect and Extract Characters from Off-Line Printed Images and Text

An ICA based Approach for Complex Color Scene Text Binarization

Refinement of digitized documents through recognition of mathematical formulae

A Fast Caption Detection Method for Low Quality Video Images

Part-Based Skew Estimation for Mathematical Expressions

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

Enhanced Image. Improved Dam point Labelling

Extracting Characters From Books Based On The OCR Technology

A new approach to reference point location in fingerprint recognition

Mobile Camera Based Text Detection and Translation

Character Recognition

Text-Tracking Wearable Camera System for the Blind

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation

Robust PDF Table Locator

HANDWRITTEN GURMUKHI CHARACTER RECOGNITION USING WAVELET TRANSFORMS

Hand Written Character Recognition using VNP based Segmentation and Artificial Neural Network

Optical Character Recognition (OCR) for Printed Devnagari Script Using Artificial Neural Network

Face Recognition Technology Based On Image Processing Chen Xin, Yajuan Li, Zhimin Tian

Smith et al. [6] developed a text detection algorithm by using vertical edge. In this method, vertical edges are first detected with a predefined temp

HCR Using K-Means Clustering Algorithm

Texture Segmentation by Windowed Projection

Historical Handwritten Document Image Segmentation Using Background Light Intensity Normalization

AUTOMATIC LOGO EXTRACTION FROM DOCUMENT IMAGES

A Survey of Problems of Overlapped Handwritten Characters in Recognition process for Gurmukhi Script

Determining Document Skew Using Inter-Line Spaces

Vision. OCR and OCV Application Guide OCR and OCV Application Guide 1/14

Image Text Extraction and Recognition using Hybrid Approach of Region Based and Connected Component Methods

Using Adaptive Run Length Smoothing Algorithm for Accurate Text Localization in Images

Extraction Characters from Scene Image based on Shape Properties and Geometric Features

Restoring Warped Document Image Based on Text Line Correction

Text Recognition in Videos using a Recurrent Connectionist Approach

Skew Detection and Correction of Document Image using Hough Transform Method

Layout Segmentation of Scanned Newspaper Documents

An Objective Evaluation Methodology for Handwritten Image Document Binarization Techniques

Connected Component Clustering Based Text Detection with Structure Based Partition and Grouping

A Model-based Line Detection Algorithm in Documents

Extracting Layers and Recognizing Features for Automatic Map Understanding. Yao-Yi Chiang

Text Area Detection from Video Frames

Biometric Security System Using Palm print

Image Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments

ECG782: Multidimensional Digital Signal Processing

An Approach to Detect Text and Caption in Video

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

12/12 A Chinese Words Detection Method in Camera Based Images Qingmin Chen, Yi Zhou, Kai Chen, Li Song, Xiaokang Yang Institute of Image Communication

A Generalized Method to Solve Text-Based CAPTCHAs

II. WORKING OF PROJECT

CITS 4402 Computer Vision

A System for Automatic Extraction of the User Entered Data from Bankchecks

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task

Human Motion Detection and Tracking for Video Surveillance

A New Algorithm for Detecting Text Line in Handwritten Documents

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Handwritten Hindi Numerals Recognition System

N.Priya. Keywords Compass mask, Threshold, Morphological Operators, Statistical Measures, Text extraction

Data Hiding in Binary Text Documents 1. Q. Mei, E. K. Wong, and N. Memon

Text Extraction from Natural Scene Images and Conversion to Audio in Smart Phone Applications

AN EFFICIENT BINARIZATION TECHNIQUE FOR FINGERPRINT IMAGES S. B. SRIDEVI M.Tech., Department of ECE

Localization, Extraction and Recognition of Text in Telugu Document Images

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Character Segmentation and Recognition Algorithm of Text Region in Steel Images

OTCYMIST: Otsu-Canny Minimal Spanning Tree for Born-Digital Images

Arabic Text Extraction from Video Film

Keywords Connected Components, Text-Line Extraction, Trained Dataset.

A Technique for Classification of Printed & Handwritten text

OBJECT SORTING IN MANUFACTURING INDUSTRIES USING IMAGE PROCESSING

Fish species recognition from video using SVM classifier

Recognition-based Segmentation of Nom Characters from Body Text Regions of Stele Images Using Area Voronoi Diagram

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation.

IMAGE ENHANCEMENT in SPATIAL DOMAIN by Intensity Transformations

An Accurate Method for Skew Determination in Document Images

TEVI: Text Extraction for Video Indexing

IMPLEMENTING ON OPTICAL CHARACTER RECOGNITION USING MEDICAL TABLET FOR BLIND PEOPLE

Mobile Camera Based Calculator

Text Enhancement with Asymmetric Filter for Video OCR. Datong Chen, Kim Shearer and Hervé Bourlard

Locating 1-D Bar Codes in DCT-Domain

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

A Statistical approach to line segmentation in handwritten documents

Video Inter-frame Forgery Identification Based on Optical Flow Consistency

Optical Character Recognition Based Speech Synthesis System Using LabVIEW

A Miniature-Based Image Retrieval System

ADAPTIVE DOCUMENT IMAGE THRESHOLDING USING IFOREGROUND AND BACKGROUND CLUSTERING

Handwritten Script Recognition at Block Level

Textural Features for Image Database Retrieval

Segmentation of Kannada Handwritten Characters and Recognition Using Twelve Directional Feature Extraction Techniques

Transcription:

International Seminar on Application of Science Mathematics 2011 ISASM2011 IMAGE S EGMENTATION AND TEXT EXTRACTION: APPLICATION TO THE EXTRACTION OF TEXTUAL INFORMATION IN SCENE IMAGES Danial Md Nor 1, Rosli Omar 2, M. Zarar M.Jenu 3, and Jean-Marc Ogier 4 1,2,3 Faculty of Electrical and Electronics Engineering, Universiti Tun Hussein Onn Malaysia 86400 Parit Raja, Batu Pahat, Johor, MALAYSIA. 4 L3I University of La Rochelle, 17042 La Rochelle cedix 1, FRANCE. 1 danial@uthm.edu.my, 2 roslio@uthm.edu.my, 3 zarar@uthm.edu.my, 4 jean-marc.ogier@univ-lr.fr This paper proposes a solution to the problem of extraction of textual information in presentation scene images. The problem has gained numerous of interest by document analysis and recognition (DAR) community. As an extention in DAR, new research domain, Camera Based Document Analysis and Recognition (CBDAR) has been established which deals with the textual information in scene images taken by low cost hand held devices like digital camera, cell phones, etc. A lot of applications like text translation, reading text for visually impaired and blind person, information retrieval from media document, e-learning, etc., can be built using the techniques developed in CBDAR domain. The proposed approach of extraction of textual information is composed of three steps: image segmentation, text localization and extraction, and Optical Character Recognition. First of all, for pre-processing the resolution of each image is checked for re-sampling to a common resolution format (720 X 540). Then, the final image is converted to grayscale and binarized using Otsu segmentation method for further processing. In addition, looking at the mean horizontal run length of both black and white pixels, the proper segmentation of foreground objects is checked. In the post-processing step, the text localizer validates the candidate text regions proposed by text detector. We have employed a connected component approach for text localization. The extracted text is then has been successfully recognized using ABBYY FineReader for OCR. Keywords : Image Segmentation, Scene text, Textual Information,Text Extraction, OCR Scope: Science Engineering. 1. Introduction The presence of the text data in images and videos containing information that is useful for clearing automatically, indexing, and structuring an image. In the extraction of this information involves the detection, localization, tracking, extraction, enhancement, and recognition of text from the image provided. Nevertheless, because of diferences change the text size, style, orientation, and alignment, and low image contrast and complex background to the problem of automatic text extraction extremely challenging [1]. Although a comprehensive survey of related problems such as face detection, analysis, documents, and indexing of images & videos can be found, the problem of extracting the text information is not well explored. In this paper, we present a generic framework and methodology for automatically extract the text content of the image obtained from slide video recordings. In particular, we use the video lectures as our initial target because it can be the basis for other scenarios such as meetings and conferences. In addition to OCR, we also discuss how the unique information available in text layout. This information could be used for indexing and retrieval.

2. Related Work Development and progress of various approaches to the extraction of text information from the image and video have been proposed for specific application, including page segmentation [2], text color extraction [3], video frame text detection [4] and content-based image or video indexing[5, 6]. However, extensive research, it is not easy to design series general-purpose systems. This is because there many possible sources of variation when extracting text. Shaded from the textured background or, from the low-contrast or complex images, or images with variations in font size, style, color, orientation, and alignment. This variation makes the problem very difficult to draw automatically. Generally text-detection methods can be classified into three categories. The first one consists of connected component-based methods, which assume that the text regions have uniform colors and satisfy certain size, shape, and spatial alignment constraints. However, these methods are not effective when the text have similar colors with background. The second one consists of the texture based methods, which assume that the text regions have special texture. Though these methods are comparatively less sensitive to background colors, they may not differentiate the texts from the text-like backgrounds. The third one consists of the edge-based methods [7]. The text regions are detected under the assumption that the edge of the background and the object regions are sparser than those of the text regions. However, this kind of approaches is not very effective to detect texts with large font size. Chen et al. [4] compared the Support Vector Machines (SVM) based method with the multilayer perceptrons (MLP) based one for text verification over four independent features, namely, the distance map feature, the grayscale spatial derivative feature, the constant gradient variance feature and the DCT coefficients feature. They found that better detection results are obtained by SVM rather than by MLP. Multi-resolution-based text detection methods are often adopted to detect texts in different scales [8]. Texts with different scales will have different features in the each sub-band. Integrating the detected text in different sub-bands achieves better performance. Furthermore, the available redundant temporal information of videos is often used in candidate text region verification and falsely detected text region elimination. Multi-resolution text regions can be detected from the edge maps down-sampled from that of the original images. Moreover, they also made full use of the temporal redundancy to eliminate the falsely detected text regions [9]. The main contribution of this paper lies in the following three aspects: (1) it is a fast OCR method as in Fig.1, (2) it is proposed for feature extraction of textual information, and (3) it is enhancing textual information for its feature extraction. MPEG Video Frame Extraction Image Segmentation Image Classsification Text Extraction OCR n Figure 1. Flowchart of Text Extraction 3. Text Extraction The aim of Optical Character Recognition (OCR) is to classify optical patterns (often contained in a digital image) corresponding to alphanumeric or other characters. The process of OCR involves several steps including segmentation, feature e xtraction, and classification. In principle, any standard OCR software can now be used to recognize the text in the segmented frames. However, a hard look at the properties of the candidate character regions in the segmented frames or image reveals that most OCR software packages will have significant difficulty to

International Seminar on Application of Science Mathematics 2011 ISASM2011 recognize the text.document images are different from natural images because they contain mainly text with a few graphics and images. Due to the very low-resolution of images of those captured using handheld devices, it is hard to extract the complete layout structure (logical or physical) of the documents and even worse to apply standard OCR systems. For this reason, a shallow representation of the low-resolution captured document images is proposed. In case of original electronic documents in the repository, the extraction of the same signature is straightforward; the PDF or PowerPoint form of the original electronic documents is converted into a relatively high-resolution image (TIFF, JPEG, etc.) on which the signature is computed. Finally, the captured document s signature is compared to with all the original electronic documents signatures in order to find a match. 4. Experimental Results 4.1 Image segmentation First of all, the resolution of each document i.e. both the image version of the original electronic documents and the pre-processed captured documents is checked for re-sampling to a common resolution format (720 X 540). If the perceived document image is resolution-wise different than the common resolution, then the re-sampling is required to bring the image up to the common resolution. This is necessary as in this scenario, the geometrical properties of the various visual features are considered during the matching for the identification of the captured document images. Due to poor resolution, it is not feasible to go up to the character level as long as the adjacent characters are overlapped in the captured documents. First, the captured document images are pre-processed for the perspective correction and noise removal Then, the final image is converted to grayscale and binarized using Otsu segmentation method for further processing [Otsu, 1979]. Furthermore, looking at the mean horizontal run length of both black and white pixels the proper segmentation of foreground objects is checked. For example, for the document images having dark background and light foreground, the output of the binarization is reversed i.e. black background (represented as 0 s) and white foreground (represented as 1 s) (Fig.4.1(a)). Black and white (binarization) (a) (b) Figure 4.1: Otsu segmentation: (a) original slide document, (b) output of Otsu segmentation and white pixels in the output of the Otsu segmentation is computed. Normally, the mean horizontal run length of the background pixels is much higher than that of foreground pixels. If the mean horizontal run length of the black pixels is comparatively higher than that of the white pixels in the binary images of the output of Otsu segmentation then the black and white pixels are simply swapped for the required image. Fig.1 illustrates one of such images having dark background and lighter foreground. For this particular image (Fig.41(b)), the mean horizontal length of the black pixels is 32.3 and for white pixels it is 6.1. The image in Fig. 6.1b is

corrected using this black and white run information for perfect segmentation i.e. black foreground and white background. 4.2 Text localization and extraction 4.2.1 Run-Length S moothing Algorithm The method RLSA is applied row-by-row and column-by-column to the above mentioned binary document images representing white pixels by 1 s and black pixels by 0 s. The RLSA transforms a binary sequence x into an output sequence y according to the rules described by Behera as follows [10]: i) 1 s in x are changed to 0 s in y if the number of adjacent 1 s is less than a pre-defined limit, T. ii) 0 s in x are unchanged in y. For example, with T = 5 the sequence x is mapped into y, which is illustrated in Fig. 4.2. Figure 4.2: RLSA algorithm (adapted from [10]) The basic idea of the RLSA is to connect the neighboring black areas when they are separated by less than T pixels. The degree of connectivity depends on T, the distribution of white and black pixels in the document and the dpi (dots per inch) resolution of the document. The two distinct bit-maps are generated using the RLSA in both horizontal and vertical directions. Often, the spacing between the components in the document image tends to differ horizontally and vertically. Therefore, two different thresholds Th and Tv are used for the RLSA in respective horizontal and vertical direction. For the slide documents, these thresholds are tuned as Th = 80 and Tv = 100. The two bit-maps of respective RLSA output in horizontal and vertical directions are combined using a logical AND operator to detect various components in the document images. Additional horizontal smoothing using the RLSA (Ts = 15) produces the final segmentation result. Fig. 4.3 illustrates the RLSA algorithm in horizontal, vertical and combining output of the image in Fig. 4.3 (c). The values of these thresholds have been evaluated and tuned using about a hundred slide images. RLSA horizontally RLSA vertically Intersected Image (a) (b) (c) Figure 4.3: Output of the RLSA: (a) horizontal direction, (b) vertical direction and (c)

combining both the directions, of the binary image in Fig. 4.1 (b) International Seminar on Application of Science Mathematics 2011 ISASM2011 4.3 Optical Character Recognition Binarization is achieved with a gray threshold value derived from Otsu s [38] method. Additional steps are done for a better binary image. Firstly, a horizontal projection profile analysis is done to detect white text on black background. If the text color is deemed to be white, the OT-region is inverted. This is necessary as most OCR software works best with black text on white background. An example of Otsu segmentation is depicted in Fig. 4.4. 2117 16 1617 172673 RLSA additional horizontally 1496 1629 Final Text 3272 4437 10373 16071 34581 85680 2885 3150 35 36 2558 3488 3346 4800 4208 6384 3307 32 32 4720 524 756 (a) (b) Figure 4.4: Otsu segmentation: (a) original slide document, (b) output of Otsu segmentation Finally, a vertical projection profile analysis is done to discard unwanted pixel regions that are less than a calculated threshold. The final binarized image is shown in Fig. 4.5 (a) and its connected components is illustrated in Fig. 4.5 (b). Connected components will be explained later. We used ABBYY FineReader 8.0 (http://finereader.abbyy.com) for OCR. We found that it was sufficient for our purpose. Show text only BBC) (a) (b) Figure 4.5: The extracted text; (a) result for OCR of a binarized frame. (b) connected components.

Frequency of Nodes 4.3.1 Analysis of Extracted Characters ABBYY FineReader is a good OCR software but not for low resolution document. After extracting the individual characters in a document and determining their properties, we compare them to the original set. Letters like "g", "j" & "y", which extend below the baseline are in one group, tall letters, such as "l" and "T" are in another, short characters, like "a" and "c" another, and floating characters, such as "'" and "^" are in the last. Once classified, an extracted character is compared to learned characters which are in the same group. If no good match is found, the extracted character is then compared to the other original characters, regardless of group. Because we found that some characters made it past the original character recognition algorithm, we deemed it necessary to perform additional operations on poorly recognized characters. The mainly observable cause of misrecognition in our original program was linked characters as wide character. An "r" would just barely touch an "i", and the character would be recognized as an "n". To alleviate this problem, we split the character at its most narrow point. This algorithm could possibly cause problems with something like "mi"-- with a poorly scanned "m", the joined character could be broken in the middle of the "m", find an "n", and do something unpredictable with the remnants of the "m" and the "i". Another common cause of misrecognition were split characters. An "r" might be split down the middle, leaving an "l" -like figure on the left, and something incomprehensible on the right. 5. Enhancing Textual Information In addition to OCR which was obtained from the binarized frame, we propose several other unique features that can be used to represent low resolution images. The layout shape signature can be used as a feature vector to measure the distance between two images in a query by example searching system. Here we briefly discuss the connection of two features derived from the results of the extraction of textual information Length of Sentences and the Minimum Spanning Tree of Sentences (MST). 5.1 Length of Sentences Using text as a feature vector will result in dissimilarity between the original text and the generated ones. Hence an algorithm is proposed to classify horizontal words into three categories namely short, medium and long sentences. The sentences are determined from a number of connected components as in Fig 5.1. A connected component is by a set of character. 1000 900 800 700 600 500 400 300 200 100 0 0 10 20 30 40 50 60 70 Number of Character Bounding Box Figure 5.1: The histogram of classified connected components.

198.2524 474.3216 591.7516 141.3542 461.4694 589 57 656.3452 464.097 118.9832 421.8685 266.4132 154.5477 688.2042 595.9211 294.1564 449.5787 336.8813 662.4394 476.8281 105.2616 122.1147 130.0961 273.6421 719.0174 224.0201 68.1836 577.9455 516.3816 721.6682 468.6513 210.5825 721.9259 529.3893 451.7079 541.1562 International Seminar on Application of Science Mathematics 2011 ISASM2011 5.2 Minimum s panning trees Basically a minimum spanning tree is a subset of the edges of the graph, so that there s a path form any node to any other node and that the sum of the weights of the edges is minimum. We consider each sentence as a node. Here s the minimum spanning tree of the example (Fig.5.2) of Prim s Algorithm by Robert C. Prim.The algorithm (greedily) builds the minimal spanning tree by iteratively adding nodes into a working tree: 1. Start with a tree which contains only one node. 2. Identify a node (outside the tree) which is closest to the tree and add the minimum weight edge from that node to some node in the tree and incorporate the additional node as a part of the tree. 3. If there are less then n 1 edges in the tree, go to 2 For the example graph, here s how it would run. The resulting graph will have 8 MST as in Fig. 5.1 (c). Node 9 Node 8 61.0574 Node 6 Node 7 59.4916 Node 5 Node 6 Node 4 Node 5 Node 9 28.5 147.0782 Node 4 52.6308 Node 3 Node 8 70.6771 Node 7 Node 3 210.9342 34.0918 Node 2 Node 1 Node 1 Node 2 (a) (b) (c) 6.0 Conclusions Figure 5.1: Prim s Algorithm; (a) Nodes, (b) Edges,(c) MST Image Segmentation is an important task and requires careful scrutiny. We have presented an innovative and novel framework for the separation of information in an image. This research, which aims to obtain text information including OCR and layout signature, has focused on the captured images from slide show videos. While OCR enables us to search words in the document, layout signature will be used in indexing. The good indexing scheme of an image or frame will be able to produce fast and accurate retrieval. However, the results are very dependent on the quality of OCR images or documents. In general, the proposed algorithm can give a reliable OCR results. This allows keyword searching for the retrieval to be implemented. However, for low-resolution images, we propose a new signature taken from textual information. A new feature from textual information that we obtain as categorized nodes and the MST is unique for an image. In the future, we intend to develop Content-Based Indexing and Retrieval (CBIR) system using the obtained feature vector. We believe that through proper indexing, the proposed CBIR system is able to address the problem of low resolution document. In this paper, our focus is on scene images from the meetings and conferences. After we have an efficient CBIR, it is not difficult to make a similar system for video and image from other scenarios.

References 1. Jung, C., Q. Liu, and J. Kim, A new approach for text segmentation using a stroke filter. Signal Processing, 2008. 88(7): p. 1907-1916. 2. Drivas, D. and A. Amin, Page segmentation and classification utilizing bottom-up approach. In Proc. of ICDAR, August 1995: p. 610-614. 3. Saidane, Z. and C. Garcia, An Automatic Method for Video Character Segmentation. 2008: p. 557 566. 4. Chen, D., J.-M. Odobez, and H. Bourlard, Text detection and recognition in images and video frames. Pattern Recognition, 2004. 37(3): p. 595-608. 5. Al-Tayeche, R. and A. Khalil, CBIR: Content Based Image Retrieval. 2003. 6. Ameen, M.A., Content-Based Image Search Engine. 2004: p. 255-263. 7. Jung, K., K. In Kim, and A. K. Jain, Text information extraction in images and video: a survey. Pattern Recognition, 2004. 37(5): p. 977-997. 8. Basu, S., et al., Text line extraction from multi-skewed handwritten documents. Pattern Recognition, 2007. 40(6): p. 1825-1839. 9. Lee, S. and J. Kim, Complementary combination of holistic and component analysis for recognition of low-resolution video character images. Pattern Recognition Letters, 2008. 29(4): p. 383-391. 10. Behera, A., D. Lalanne, and R. Ingold. Enhancement of layout-based identification of lowresolution documents using geometrical color distribution. in Document Analysis and Recognition, 2005. Proceedings. Eighth International Conference on. 2005.