Text Information Extraction And Analysis From Images Using Digital Image Processing Techniques

Similar documents
Text Extraction from Complex Color Images Using Optical Character Recognition

Text Extraction of Vehicle Number Plate and Document Images Using Discrete Wavelet Transform in MATLAB

Scene Text Detection Using Machine Learning Classifiers

[Pal, 3(1): January, 2014] ISSN: Impact Factor: 1.852

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

International Journal of Scientific & Engineering Research, Volume 6, Issue 8, August ISSN

Nitesh Kumar Singh, Avinash verma, Anurag kumar

Text Enhancement with Asymmetric Filter for Video OCR. Datong Chen, Kim Shearer and Hervé Bourlard

Automatic Video Caption Detection and Extraction in the DCT Compressed Domain

A Text Detection, Localization and Segmentation System for OCR in Images

Recognition of Gurmukhi Text from Sign Board Images Captured from Mobile Camera

An Edge-based Text Region Extraction Algorithm for Indoor Mobile Robot Navigation

I. INTRODUCTION. Figure-1 Basic block of text analysis

EXTRACTING TEXT FROM VIDEO

TEXT DETECTION AND RECOGNITION IN CAMERA BASED IMAGES

IDIAP IDIAP. Martigny ffl Valais ffl Suisse

Segmentation Framework for Multi-Oriented Text Detection and Recognition

Image Text Extraction and Recognition using Hybrid Approach of Region Based and Connected Component Methods

Time Stamp Detection and Recognition in Video Frames

N.Priya. Keywords Compass mask, Threshold, Morphological Operators, Statistical Measures, Text extraction

12/12 A Chinese Words Detection Method in Camera Based Images Qingmin Chen, Yi Zhou, Kai Chen, Li Song, Xiaokang Yang Institute of Image Communication

CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM

Study on the Signboard Region Detection in Natural Image

TEVI: Text Extraction for Video Indexing

International Journal of Electrical, Electronics ISSN No. (Online): and Computer Engineering 3(2): 85-90(2014)

An Efficient Character Segmentation Based on VNP Algorithm

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-7)

OCR For Handwritten Marathi Script

Motion Detection. Final project by. Neta Sokolovsky

HANDWRITTEN GURMUKHI CHARACTER RECOGNITION USING WAVELET TRANSFORMS

Council for Innovative Research Peer Review Research Publishing System Journal: INTERNATIONAL JOURNAL OF COMPUTERS & TECHNOLOGY

Chapter 3 Image Registration. Chapter 3 Image Registration

TEXT EXTRACTION FROM AN IMAGE BY USING DIGITAL IMAGE PROCESSING

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

Review on Text String Detection from Natural Scenes

Keywords: Thresholding, Morphological operations, Image filtering, Adaptive histogram equalization, Ceramic tile.

Bus Detection and recognition for visually impaired people

An Approach to Detect Text and Caption in Video

Using Adaptive Run Length Smoothing Algorithm for Accurate Text Localization in Images

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

Short Survey on Static Hand Gesture Recognition

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

Human Motion Detection and Tracking for Video Surveillance

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Film Line scratch Detection using Neural Network and Morphological Filter

Understanding Tracking and StroMotion of Soccer Ball

ABSTRACT 1. INTRODUCTION 2. RELATED WORK

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

2 Proposed Methodology

IDIAP IDIAP. Martigny ffl Valais ffl Suisse

An ICA based Approach for Complex Color Scene Text Binarization

Classifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao

An Accurate Method for Skew Determination in Document Images

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY /$ IEEE

Effects Of Shadow On Canny Edge Detection through a camera

Janitor Bot - Detecting Light Switches Jiaqi Guo, Haizi Yu December 10, 2010

Counting Particles or Cells Using IMAQ Vision

EDGE BASED REGION GROWING

SCALE INVARIANT TEMPLATE MATCHING

EECS 556 Image Processing W 09. Image enhancement. Smoothing and noise removal Sharpening filters

Artifacts and Textured Region Detection

Mobile Camera Based Text Detection and Translation

Improving License Plate Recognition Rate using Hybrid Algorithms

Image Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments

HCR Using K-Means Clustering Algorithm

A Review of various Text Localization Techniques

A Phase-based Approach for Caption Detection in Videos

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Searching Video Collections:Part I

Automatically Algorithm for Physician s Handwritten Segmentation on Prescription

Extracting and Segmenting Container Name from Container Images

Layout Segmentation of Scanned Newspaper Documents

Research on QR Code Image Pre-processing Algorithm under Complex Background

A Survey of Problems of Overlapped Handwritten Characters in Recognition process for Gurmukhi Script

An Approach for Real Time Moving Object Extraction based on Edge Region Determination

Available online at ScienceDirect. Procedia Computer Science 96 (2016 )

Suspicious Activity Detection of Moving Object in Video Surveillance System

Image-Based Competitive Printed Circuit Board Analysis

Optical Character Recognition (OCR) for Printed Devnagari Script Using Artificial Neural Network

Detection of Edges Using Mathematical Morphological Operators

Keywords Binary Linked Object, Binary silhouette, Fingertip Detection, Hand Gesture Recognition, k-nn algorithm.

EE795: Computer Vision and Intelligent Systems

Color Local Texture Features Based Face Recognition

SECTION 5 IMAGE PROCESSING 2

Automatic Detection of Change in Address Blocks for Reply Forms Processing

Identifying and Reading Visual Code Markers

Midterm Examination CS 534: Computational Photography

Lecture 6: Edge Detection

Locating 1-D Bar Codes in DCT-Domain

Connected Component Clustering Based Text Detection with Structure Based Partition and Grouping

MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK

DIGITAL IMAGE ANALYSIS. Image Classification: Object-based Classification

[10] Industrial DataMatrix barcodes recognition with a random tilt and rotating the camera

Texture Sensitive Image Inpainting after Object Morphing

AUTOMATIC OBJECT DETECTION IN VIDEO SEQUENCES WITH CAMERA IN MOTION. Ninad Thakoor, Jean Gao and Huamei Chen

Robot localization method based on visual features and their geometric relationship

IMAGE S EGMENTATION AND TEXT EXTRACTION: APPLICATION TO THE EXTRACTION OF TEXTUAL INFORMATION IN SCENE IMAGES

Ulrik Söderström 16 Feb Image Processing. Segmentation

Transcription:

Text Information Extraction And Analysis From Images Using Digital Image Processing Techniques Partha Sarathi Giri Department of Electronics and Communication, M.E.M.S, Balasore, Odisha Abstract Text data present in images contain useful information for automatic explanation, indexing, and structuring of images. Extraction of this information involves detection, localization, tracking, extraction, enhancement, and recognition of the text from a given image. However variations of text due to differences in size, style, orientation, and alignment, as well as low image contrast and complex background make the problem of automatic text extraction extremely challenging in the computer vision research area. The proposed methods compare two basic approaches for extracting text region in images: edge-based and connected-component based. The algorithms are implemented and evaluated using a set of images that vary along the dimensions of lighting, scale and orientation. Accuracy, precision and recall rates for each approach are analyzed to determine the success and limitations of each approach. Keywords: Text information extraction; Text detection; Text localization I. INTRODUCTION Text Extraction from image is concerned with extracting the relevant text data from a collection of images. Recent studies in the field of image processing show a great amount of interest in content retrieval from images and videos. This content can be in the form of objects, color, texture, shape as well as the relationships between them. The semantic information provided by an image can be useful for content based image retrieval, as well as for indexing and classification purposes [4, 10]. As stated by Jung, Kim and Jain in [4], text data is particularly interesting, because text can be used to easily and clearly describe the contents of an image. Since the text data can be embedded in an image or video in different font styles, sizes, orientations, colors, and against a complex background, the problem of extracting the candidate text region becomes a challenging one [4]. Current Optical Character Recognition (OCR) techniques can only handle text against a plain monochrome background and cannot extract text from a complex or textured background [7]. As stated text have some common distinctive characteristics in terms of frequency and orientation information, and also spatial cohesion. Spatial cohesion refers to the fact that text characters of the same string appear close to each other and are of similar height, orientation and spacing [7]. Two of the main methods commonly used to determine spatial cohesion are based on edge [1, 2] and connected component [3] features of text characters. Among them, text within an image is of particular interest as (i) it is very useful for describing the contents of an image; (ii) it can be easily extracted compared to other semantic contents, and(iii) it enables applications such as keyword-based image search, automatic video logging, and text-based image indexing. 1.1. Text in images Texts usually have different appearance changes like font, size, style, orientation, alignment, texture, color, contrast, and background. All the changes will make the problem of text extraction become complicated and difficult. Text in images can exhibit many variations with respect to the following properties: Size: Although the text size can vary a lot, assumptions can be made depending on the application domain. Alignment: The characters in the caption text appear in clusters and usually lie horizontally, although sometimes they can appear as non-planar texts as a result of special effects. This does not apply to scene text, which can have numerous perspective distortions. Scene text can be aligned in any direction and can have geometric distortions. 66

Inter-character distance: Characters in a text line have a uniform distance between them. extracted text images can be transformed into plain text using OCR technology. Color: The characters in a text line are subject to have the same or similar colors. This property makes it possible to use a connected component-based approach for text detection. Many of the research informed till date has focused on finding text strings of a single color. However, video images and other complex color documents can contain text strings with more than two colors (polychrome) for effective visualization, i.e., different colors within one word. Motion: The same characters usually exist in consecutive frames in a video with or without movement. This property is used in text tracking and enhancement. Scene text can have arbitrary motion due to camera or object movement. Edge: Most caption and scene texts are designed to be easily read, thereby resulting in strong edges at the boundaries of text and background. Compression: Many digital images are recorded, transferred, and processed in a compressed format. Thus, a faster TIE system can be achieved if one can extract text without decompression. II. TEXT INFORMATION EXTRACTION (TIE) A TIE system receives an input in the form of a still image or a sequence of images. The images can be in gray scale or color, compressed or uncompressed, and the text in the images may move or may not. The problem arises due to TIE system can be divided into the following sub-problems: (i) detection (ii) localization (iii) tracking (iv) extraction and enhancement (v) recognition (OCR) shown in Fig.1.Text detection refers to the determination of the presence of text in a given sequence of images.text localization is the process of determining the location of text in the image and generating bounding boxes around the text.text tracking is performed to reduce the processing time for text localization and to maintain the integrity of position across adjacent frames. Although the precise location of text in an image can be indicated by bounding boxes, the text needs to be segmented from the background to facilitate its recognition. That means, the extracted text image has to be converted into a binary image and enhanced before it is fed into an OCR engine. Text extraction is the stage where the text components are segmented from the background.text Enhancement of the extracted text components is required because the text region usually has lowresolution and is horizontal to noise. Thereafter, the Fig. 1. Architecture of a TIE system. 2.1 Text extraction techniques To implement, test, and compare and contrast two approaches for text region extraction in images, and to discover how the algorithms perform under variations of lighting, orientation, and scale transformations of the text. The algorithms are from Liu and Samarabandu in and Gllavata, Ewerth and Freisleben in. The comparison is based on the accuracy of the results obtained, and precision and recall rates. 2.1.1 Edge based text region extraction Edge-based methods focus on the high contrast between the text and the background. The edges of the text boundary are identified and merged, and then used to filter out the non-text regions. Usually, an edge filter is used for the edge detection, and a smoothing operation or a morphological operator is used for the merging stage. The basic steps of the edge-based text extraction algorithm are explained, and diagrammed in Figure 2.1. (1)Create a Gaussian pyramid by convolving the input image with a Gaussian kernel and successively down-sample each direction by half. (2)Create directional kernels to detect edges at 0, 45, 90 and 135 orientations. (3)Convolve each image in the Gaussian pyramid with each orientation filter.(4)combine the results of step 3 to create the Feature Map.(4)Dilate the resultant image using a sufficiently large structuring element to cluster candidate text regions together.(5)create final output image with text in white pixels against a plain black background. The procedure for extracting a text region from an image can be broadly classified into three basic steps: detection of the text region in the image, localization of 67

the region, and creating the extracted output character image. Figure2 Bloock diagram for edge based text extraction 2.1.1.1 Detection Among the several textual properties in an image, edge-based methods focus on the high contrast between the text and the background. The edges of the text boundary are identified and merged, and then several methods are used to filter out the non-text regions. In this section the region with the possibility of text for a given image is detected. A Gaussian pyramid is created by successively filtering the input image with a Gaussian kernel of size 3x3 and down sampling the image in each direction by half. Down sampling refers to the process whereby an image is resized to a lower resolution from its original resolution. A Gaussian filter of size 3x3 will be used. Each level in the pyramid corresponds to the input image at a different resolution. These images are next convolved with directional filters at different orientation kernels for edge detection in the horizontal (00), vertical (900) and diagonal (450, 1350) directions. 2.1.1.2 Localization The process of localization involves further enhancing the text regions by eliminating non-text regions. One of the properties of text is that usually all characters appear close to each other in the image, thus forming a cluster. By using a morphological dilation operation, these possible text pixels can be clustered together, eliminating pixels that are far from the candidate text regions. Dilation is an operation which expands or enhances the region of interest, using a structural element of the required shape and or size. The process of dilation is carried out using a very large structuring element in order to enhance the regions which lie close to each other. In this algorithm, a structuring element of size [2x2] has been used. Figure 3.3 below shows the result before and after dilation of an original image. The resultant image after dilation may consist of some non-text regions or noise which needs to be eliminated. Figure2.3. (a) Original Image (b) Before dilation (c) After dilation 2.1.1.3 Character extraction The common OCR systems available require the input image to be such that the characters can be easily parsed and recognized. The text and background should be monochrome and background-to-text contrast should be high. Figure2.2 (a) Sample Gaussian pyramid with 4 levels (b) Each resolution image resized to original image size (c) Sample resized image of the pyramid after convolution with 0º kernel After convolving the image with the orientation kernels, a feature map is created. A weighting factor is associated with each pixel to classify it as a candidate or non-candidate for text region. A pixel is a candidate for text if it is highlighted in all of the edge maps created by the directional filters. Thus, the feature map is a combination of all edge maps at different scales and orientations with the highest weighted pixels present in the resultant map. Figure2.4. (a) Original Image (b) Result 2.1.2 Connected Component based text region extraction 2.1.3 Connected Component based methods uses a bottom-up approach by grouping small components into successively larger components until all regions are identified in the image. A geometrical analysis is needed to merge the text components using the spatial arrangement of the components so as to Alter out nontext components and mark the boundaries of the text regions. The basic steps of the connected component 68

based methods text extraction algorithm are explained, and diagrammed in Figure 2.5. (1) Convert the input image to YUV color space. The luminance(y) value is used for further processing. The output is a gray image. (2) Convert the gray image to an edge image.(3) Compute the horizontal and vertical projection profiles of candidate text regions using a histogram with an appropriate threshold value.(4) Use geometric properties of text such as width to height ratio of characters to eliminate possible non-text regions.(5) Binarize the edge image enhancing only the text regions against a plain black background.(6) Create the Gap Image using the gap-filling process and use this as a reference to further eliminate non-text regions from the output. Figure2.5. Basic Block diagram for Connected Component based text extraction 2.1.3.1 Pre-Processing The input image is pre-processed to make easy easier detection of text regions. As proposed in the image is converted to the YUV color space (luminance + chrominance), and only the luminance(y) channel is used for further processing. The Y channel refers to brightness or intensity of the image whereas the U and the V channels refer to the actual color information. Since text present in an image has more contrast with its background, by using only the Y channel, the image can be converted to a grayscale image with only the brightness contrast information present. in each direction. This weight value is the maximum value between the pixel and its neighbors in the left (L), upper (U) and upper-right (UR) directions. The algorithm uses these three neighbor values to detect edges in horizontal, vertical and diagonal directions. The resultant edge image obtained is sharpened in order to increase contrast between the detected edges and its background, making it easier to extract text regions. Figure 2.6(c) shows the sharpened edge image for the Y Channel gray image. 2.1.3.3 Localization The sharpened edge image is considered as the input intensity image for computing the projection profiles, with white candidate text regions against a black background. The vertical projection profile shows the sum of pixels present in each column of the intensity or the sharpened image. Similarly, the horizontal projection profile shows the sum of pixels present in each row of the intensity image. These projection profiles are essentially histograms where each bin is a count of the total number of pixels present in each row or column. Candidate text regions are segmented based on adaptive threshold values, T1 and T2, calculated for the vertical and horizontal projections respectively. Only regions that fall within the threshold limits are considered as candidates for text. The value of threshold T1 is selected to eliminate possible non text regions such as doors, window edges etc. that have a strong vertical orientation. Similarly, the value of threshold T2 is selected to eliminate regions which might be non-text or long edges in the horizontal orientation. Figure2.8. (a) Vertical Projection profile (b) Horizontal Projection profile for Sharpened image Figure2.6 (a) Original image (b) Y channel for test image (c) sharpened edge image for the Y Channel 2.1.3.2 Detection In this process, the connected-component based approach is used to make possible text regions stand out as compared to non-text regions. Every pixel in the edge image is assigned a weight with respect to its neighbors 2.1.3.4 Enhancement and Gap Filling The geometric ratio between the width and the height of the text characters is considered to eliminate possible non-text regions. This ratio value will be defined after experimenting on different kinds of images to get an average value. In this paper, regions with minor to major axis ratio less than 20 are considered as 69

candidate text regions for further processing. Next a gap image will be created which will be used as a reference to refine the localization of the detected text regions. If a pixel in the binary edge image created is surrounded by black (background) pixels in the vertical, horizontal and diagonal directions, this pixel is also substituted with the background value. the accuracy of the system. The results from edge based and connected component based algorithm are shown in Figure-3.1 and Figure-3.2 respectively. Figure2.9 Result obtained by connected component based text detection for test image Figure 2.9(a) Original image (b) Result III. EXPERIMENTAL RESULTS The experimentation of the proposed algorithm was carried out on a data set consisting of different images such as indoor, outdoor, posters etc. These test images vary with respect to scale, lighting and orientation of text in the image. Currently the data set consists of number of images, Hindi character images and variations of 2 indoor poster images as well as outdoor images, taken from a digital camera. The complete lists of test images are summarized in Table1. The performance of each technique has been evaluated based on its precision and recall rates obtained. Precision rate takes into consideration the false positives, which are the non-text regions in the image and have been detected by the algorithm as text regions. Recall rate takes into consideration the false negatives, which are text words in the image, and have not been detected by the algorithm. Thus, precision and recall rates are useful as measures to determine the accuracy of each algorithm in locating correct text regions and eliminating non-text regions. Experiments have been performed to test the proposed system and to measure Figure-3.2 IV. CONCLUSIONS Text locating in natural scene image with complex background is a difficult, challenging, and important problem. In this paper, an accurate text region extraction algorithm based on two methods with grayinformation is presented. The proposed methods work very well on text region in simple images. Texts in natural scene images contain very important information about location information and road signs. One of the further studies is to design the verifying extraction text region by SVM and HMM, and then to design the recognizer system for extraction text regions. V. REFERENCES [1] Xiaoqing Liu and JagathSamarabandu, An Edgebased text region extraction algorithm for Indoor mobile robot navigation, Proceedings of the IEEE, July 2005. [2] Xiaoqing Liu and JagathSamarabandu, Multiscale edge-based Text extraction from Complex images, IEEE, 2006. [3] JulindaGllavata, Ralph Ewerth and Bernd Freisleben, A Robust algorithm for Text detection in images, Proceedings of the 3rd international symposium on Image andsignal Processing and Analysis, 2003. 70

[4] Keechul Jung, KwangIn Kim and Anil K. Jain, Text information extraction in images and video: a survey, The journal of the Pattern Recognition society, 2004. [5] Kongqiao Wang and Jari A. Kangas, Character location in scene images from digital camera, The journal of the Pattern Recognition society, March 2003. [6] K.C. Kim, H.R. Byun, Y.J. Song, Y.W. Choi, S.Y. Chi, K.K. Kim and Y.K Chung, Scene Text Extraction in Natural Scene Images using Hierarchical FeatureCombining and verification, Proceedings of the 17th International Conference on Pattern Recognition (ICPR 04), IEEE. [7] Victor Wu, RaghavanManmatha, and Edward M. Riseman, TextFinder: An Automatic System to Detect and Recognize Text in Images, IEEE Transactions onpattern Analysis and Machine Intelligence, Vol. 21, No. 11, November 1999. [8] Xiaoqing Liu and JagathSamarabandu, A Simple and Fast Text Localization Algorithm for Indoor Mobile Robot Navigation, Proceedings of SPIE- IS&TElectronic Imaging, SPIE Vol. 5672, 2005. [9] Qixiang Ye, Qingming Huang, Wen Gao and Debin Zhao, Fast and Robust text detection in images and video frames, Image and Vision Computing 23, 2005. [10] Rainer Lienhart and Axel Wernicke, Localizing and Segmenting Text in Images and Videos, IEEE Transactions on Circuits and Systems for Video Technology, Vol.12,No.4, April 2002. [11] Qixiang Ye, Wen Gao, Weiqiang Wang and Wei Zeng, A Robust Text Detection Algorithm in Images and Video Frames, IEEE, 2003. [12] http://images.google.com 71