Handwriting Recognition of Diverse Languages

Similar documents
A Study to Recognize Printed Gujarati Characters Using Tesseract OCR

OCR For Handwritten Marathi Script

International Journal of Advance Research in Engineering, Science & Technology

Keywords Connected Components, Text-Line Extraction, Trained Dataset.

Mobile Application with Optical Character Recognition Using Neural Network

Abstract. Problem Statement. Objective. Benefits

Hilditch s Algorithm Based Tamil Character Recognition

LECTURE 6 TEXT PROCESSING

Cursive Handwriting Recognition System Using Feature Extraction and Artificial Neural Network

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation

ISSN: [Mukund* et al., 6(4): April, 2017] Impact Factor: 4.116

Structural Feature Extraction to recognize some of the Offline Isolated Handwritten Gujarati Characters using Decision Tree Classifier

Character Recognition from Google Street View Images

HANDWRITTEN GURMUKHI CHARACTER RECOGNITION USING WAVELET TRANSFORMS

Handwritten Devanagari Character Recognition

Automatic Recognition and Verification of Handwritten Legal and Courtesy Amounts in English Language Present on Bank Cheques

Hand Written Character Recognition using VNP based Segmentation and Artificial Neural Network

Segmentation of Characters of Devanagari Script Documents

Handwritten Devanagari Character Recognition Model Using Neural Network

Extraction and Recognition of Alphanumeric Characters from Vehicle Number Plate

Handwritten Script Recognition at Block Level

An Efficient Character Segmentation Based on VNP Algorithm

Automated Digital Conversion of Hand-Drawn Plots

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

II. WORKING OF PROJECT

INTERNATIONAL RESEARCH JOURNAL OF MULTIDISCIPLINARY STUDIES

Fingerprint Based Gender Classification Using Block-Based DCT

SEVERAL METHODS OF FEATURE EXTRACTION TO HELP IN OPTICAL CHARACTER RECOGNITION

An Improvement Study for Optical Character Recognition by using Inverse SVM in Image Processing Technique

Keywords Binary Linked Object, Binary silhouette, Fingertip Detection, Hand Gesture Recognition, k-nn algorithm.

Comparing Tesseract results with and without Character localization for Smartphone application

Skew Detection and Correction of Document Image using Hough Transform Method

A New Approach to Detect and Extract Characters from Off-Line Printed Images and Text

A Review on Handwritten Character Recognition

FEATURE EXTRACTION TECHNIQUE FOR HANDWRITTEN INDIAN NUMBERS CLASSIFICATION

Handwritten Hindi Character Recognition System Using Edge detection & Neural Network

Fine Classification of Unconstrained Handwritten Persian/Arabic Numerals by Removing Confusion amongst Similar Classes

Neural Network Classifier for Isolated Character Recognition

Finger Print Enhancement Using Minutiae Based Algorithm

A New Technique to Optimize User s Browsing Session using Data Mining

A Multimodal Framework for the Recognition of Ancient Tamil Handwritten Characters in Palm Manuscript Using Boolean Bitmap Pattern of Image Zoning

Indian Multi-Script Full Pin-code String Recognition for Postal Automation

A two-stage approach for segmentation of handwritten Bangla word images

IMPLEMENTING ON OPTICAL CHARACTER RECOGNITION USING MEDICAL TABLET FOR BLIND PEOPLE

A Review on Different Character Segmentation Techniques for Handwritten Gurmukhi Scripts

HCR Using K-Means Clustering Algorithm

Restoring Warped Document Image Based on Text Line Correction

Recognition of Gurmukhi Text from Sign Board Images Captured from Mobile Camera

Writer Recognizer for Offline Text Based on SIFT

MODULE 7 Nearest Neighbour Classifier and its variants LESSON 11. Nearest Neighbour Classifier. Keywords: K Neighbours, Weighted, Nearest Neighbour

Segmentation Based Optical Character Recognition for Handwritten Marathi characters

SEGMENTATION OF CHARACTERS WITHOUT MODIFIERS FROM A PRINTED BANGLA TEXT

A New Technique for Segmentation of Handwritten Numerical Strings of Bangla Language

A Technique for Classification of Printed & Handwritten text

Handwritten Hindi Numerals Recognition System

RESEARCH ON OPTIMIZATION OF IMAGE USING SKELETONIZATION TECHNIQUE WITH ADVANCED ALGORITHM

International Journal of Advance Engineering and Research Development OPTICAL CHARACTER RECOGNITION USING ARTIFICIAL NEURAL NETWORK

A Supervised Method for Multi-keyword Web Crawling on Web Forums

Paper ID: NITETE&TC05 THE HANDWRITTEN DEVNAGARI NUMERALS RECOGNITION USING SUPPORT VECTOR MACHINE

Isolated Handwritten Words Segmentation Techniques in Gurmukhi Script

Keywords Handwritten alphabet recognition, local binary pattern (LBP), feature Descriptor, nearest neighbor classifier.

Automatic License Plate Recognition

Handwritten Marathi Character Recognition on an Android Device

CHAPTER 1 INTRODUCTION

Input sensitive thresholding for ancient Hebrew manuscript

Segmentation of Kannada Handwritten Characters and Recognition Using Twelve Directional Feature Extraction Techniques

A Survey of Problems of Overlapped Handwritten Characters in Recognition process for Gurmukhi Script

Preprocessing of Stream Data using Attribute Selection based on Survival of the Fittest

Toward Part-based Document Image Decoding

On-line handwriting recognition using Chain Code representation

International Journal of Advance Research in Engineering, Science & Technology

AN APPROACH FOR SCANNING BASIC BUSINESS CARD IN ANDROID DEVICE

SCIENCE & TECHNOLOGY

NOVATEUR PUBLICATIONS INTERNATIONAL JOURNAL OF INNOVATIONS IN ENGINEERING RESEARCH AND TECHNOLOGY [IJIERT] ISSN: VOLUME 2, ISSUE 1 JAN-2015

A Generalized Method to Solve Text-Based CAPTCHAs

Object Extraction Using Image Segmentation and Adaptive Constraint Propagation

Skeletonization Algorithm for Numeral Patterns

Skew angle Detection and correction using Radon Transform

A Novel q-parameter Automation in Tsallis Entropy for Image Segmentation

Handwritten English Alphabet Recognition Using Bigram Cost Chengshu (Eric) Li Fall 2015, CS229, Stanford University

I. INTRODUCTION. Figure-1 Basic block of text analysis

A Fast Recognition System for Isolated Printed Characters Using Center of Gravity and Principal Axis

A Hierarchical Pre-processing Model for Offline Handwritten Document Images

Word Slant Estimation using Non-Horizontal Character Parts and Core-Region Information

Layout Segmentation of Scanned Newspaper Documents

NOVATEUR PUBLICATIONS INTERNATIONAL JOURNAL OF INNOVATIONS IN ENGINEERING RESEARCH AND TECHNOLOGY [IJIERT] ISSN: VOLUME 5, ISSUE

Speeding up Online Character Recognition

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Recognition of Unconstrained Malayalam Handwritten Numeral

A Comparative Study of Selected Classification Algorithms of Data Mining

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

TEXT DETECTION AND RECOGNITION FROM IMAGES OF NATURAL SCENE

User Signature Identification and Image Pixel Pattern Verification

Published by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 1

AN EXAMINING FACE RECOGNITION BY LOCAL DIRECTIONAL NUMBER PATTERN (Image Processing)

Segmentation of Isolated and Touching characters in Handwritten Gurumukhi Word using Clustering approach

Handwritten Character Recognition System using Chain code and Correlation Coefficient

3D Object Recognition using Multiclass SVM-KNN

A System towards Indian Postal Automation

OFFLINE SIGNATURE VERIFICATION

Transcription:

Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 6.017 IJCSMC, Vol. 7, Issue. 4, April 2018, pg.109 114 Handwriting Recognition of Diverse Languages Mihir Shah 1, Sahil Mehta 2, Pranav Mody 3, Akhil Sen Roy 4, Sunil P. Khachane 5 1,2,3,4,5 Computer Department, MCT s Rajiv Gandhi Institute of Technology, Andheri West, Mumbai-400053, India 1 shahmn96@gmail.com; 2 mehtasahil31@gmail.com; 3 pranavmodi96@gmail.com; 4 akhil.senroy@gmail.com; 5 sunil.khachane@mctrgit.ac.in Abstract Online Handwriting Recognition for Diverse Languages is a system which is used to recognize digital as well as handwritten inputs. In this paper we will compare two different methods which we used to serve the above purpose. The two methods are K-Nearest Neighbour(KNN) and Tesseract OCR. We begin this paper by introducing the main purpose of this paper which will include a summary of both KNN as well as Tesseract OCR (Optical Character Recognition). After that we will describe in detail the working of both methods, compare them according to their recognition accuracy. This system will also help in Licence plate number recognition, image extraction from images, extracting text from other types of documents, etc. Keywords Optical Character Recognition (OCR), k- Nearest Neighbour (k-nn), Binary Large Objects (BLOBS), Graphical User Interface (GUI), Portable Network Graphics (.png). I. INTRODUCTION This paper includes two different systems which were tested in order to recognize digital texts and more importantly handwritten inputs. The first system which we tested was based on k-nn algorithm. K nearest neighbours is a simple algorithm that stores all available cases and classifies new cases based on a similarity measure (e.g., distance functions). K-Nearest Neighbours is one of the most basic yet essential classification algorithms in Machine Learning. It belongs to the supervised learning domain and finds intense application in pattern recognition, data mining and intrusion detection. Tesseract is an Optical Character Recognition (OCR) engine for various operating systems. It is an open source software which has a very good accuracy and with regular updates it is now able to recognize more than 100 languages. Accuracy of Tesseract OCR varies from 40% to 98%. Further we discuss about both systems in detail. 2018, IJCSMC All Rights Reserved 109

A. ARCHITECTURE: II. DESIGN AND IMPLEMENTATION (K-NN) Fig. 1 Architecture of k-nn There are five different phases in the architecture as follows: 1) Training: A training data set is taken as input and then it is segmented into individual characters. For each segment generated, target values are locked. This classification of training data is stored in.box files. 2) Preprocessing: The input image is first preprocessed i.e. uneven thickness; slant strokes and rotations are removed. This processed input is then passed for segmentation. 3) Segmentation: The preprocessed input is segmented into individual characters. 4) Classification using k-nn: This classification algorithm uses the results of trained data and on the basis of those results, it tries to recognize the input image using the distance formula of k-nn algorithm. 5) Feature Extraction: The classified input is then converted into and text and displayed to the user. B. IMPLEMENTATION: To perform this code, commands are given through the command prompt. The pseudo code for the k-nn is as follows: Classify (X, Y, x) //X: training data, Y: class labels of X, x: unknown sample for i = 1 to m do Compute distance d(x i, x) end for Compute set I containing indices for the k smallest distances d(x i, x). Return majority label for {Y i where i I} While, testing we provided this code with two different images as inputs i.e. Image with Digital text and Image with handwritten text. The k-nn classifier is provided with a training data. After that an input image is been provided and this image undergoes Pre-processing, segmentation, classification using k-nn and feature extraction techniques after which the final output is been provided in.png format as well as on the command prompt. 2018, IJCSMC All Rights Reserved 110

RESULTS: There were multiple input images which were provided to the k-nn based classification system. Some of the example images with their output images are as follows: Fig. 2 Input image for k-nn Fig. 3 Output image for Fig. 2 From the given input image, the accuracy obtained was 57.69%. This is the maximum accuracy which was obtained in the output for all the input images given. Even after training the system the accuracy didn t improve. This system was unsuccessful in recognizing the handwritten inputs. The maximum accuracy obtained in recognizing the handwritten inputs was less than 12%. A. ARCHITECTURE: III. DESIGN AND IMPLEMENTATION (TESSERACT OCR) Fig. 4 Architecture of Tesseract OCR [1] The Tesseract OCR works through various number of steps as follows: 1) Select an input image: This step includes human-machine interaction in which the user is expected to select a image in which the text is to be recognized. This image then goes through various processes after which the text inside the image is finally recognized. 2) Adaptive Thresholding: Adaptive thresholding is a technique which is used to segment an image by setting all pixels whose intensity values are above a threshold value to a foreground value. Eventually the pixels whose intensity values are below the threshold value are set to background value. This process is basically used to convert the input image to a binary image. 3) Connected Component Analysis: The advantage of this analysis is that by inspection of the nesting of outlines, and the number of child and grandchild outlines, it is simple to detect inverse text and recognize it as easily as black-on-white text. 4) Conversion to BLOBS: The Binary image is then converted into BLOBS (Binary Large Objects). BLOBS are divided into two consecutive functions. BLOB extraction and BLOB classification. Separation of different objects are called as BLOB extraction and classification of these different BLOBS is called as 2018, IJCSMC All Rights Reserved 111

BLOB classification. The term LARGE indicates that only the objects with a certain size are of importance. The other Small binary objects are considered as noise. 5) Fine Lines and Words: Outlines are gathered together, purely by nesting, into Blobs. Blobs are organized into text lines, and the lines and regions are analysed for fixed pitch or proportional text. Text lines are broken into words differently according to the kind of character spacing. Fixed pitch text is chopped immediately by character cells. Proportional text is broken into words using definite spaces and fuzzy spaces. [2] 6) RECOGNIZE WORD (PASS 1): In pass 1, an attempt is made to recognize each word. After a satisfactory recognition of the word, this word is passed to an adaptive classifier. Now, the advantage of passing the recognized word to the adaptive classifier is; that the classifier can provide more accuracy in recognition of the text lower down the page. As the recognized words are passed to the classifier the future recognitions of other text images become more and more accurate. 7) RECOGNIZE WORD (PASS 2): The adaptive classifier initially doesn t provide good accuracy in recognition of the text which are at the top of the page since it may have learned something very late to make any contributions in the beginning. Hence, this system runs a second pass over the image. The advantage here is that since the classifier has already recognized and learned a lot from pass 1, it can now make major contributions in providing very good accuracy in recognizing text at the top of the page and also recognize all those words in the entire image which were not properly recognized. 8) A final phase resolves fuzzy spaces and checks alternative hypotheses for the x-height to locate small cap text. [2] B. IMPLEMENTATION: In this system we have also tried to make a GUI (Graphical User Interface) to make it more user friendly. Fig. 5 User Friendly GUI for selecting an image The Fig. 5 shows that this dialog box provides the user with an option to SELECT IMAGE from his device. When the user clicks on Select Image button it opens up the File explorer (For Windows) in the screen. The user then has to select the file, then the image in which the text has to be recognized. The Fig. 5 shows that on clicking the Select Image option various images that the user has stored are seen from his folders. Once the user selects an image from the files, the image s path is then passed to the code. This image is then the input image to the Tesseract OCR. The OCR will then perform all the steps and provide an output in a text format. C. RESULTS: We provided four different inputs i.e. Digital Input, Numbers, Handwritten Alphabets and a handwritten letter. The accuracy was calculated by comparing the output alphabet and the input alphabet or number. Fig. 6 Input Image 1 for Tesseract OCR Fig. 7 Output Image for Input Image 1 2018, IJCSMC All Rights Reserved 112

Fig. 8 Input Image 2 for Tesseract OCR Fig. 9 Output Image for Input Image 2 Fig. 10 Input Image 3 for Tesseract OCR Fig. 11 Output Image for Input Image 3 Fig. 12 Input Image 4 for Tesseract OCR Fig. 13 Output Image for Input Image 4 In Fig. 10 are the alphabets and numbers in English language. These were the handwritten inputs to the tesseract. The alphabets were written with enough spacing and in upper as well as the lower case. Fig. 11 shows the output of the provided input. It was observed that the tesseract had provided the output with a very good accuracy rate. While in Fig. 12 the input was cursive handwritten English language. As seen in Fig. 13 the OCR did recognize a lot of characters with proper spacing but still had a low accuracy rate. D. CALCULATION OF ACCURACY: SR. NO. INPUT IMAGE OUTPUT IMAGE ACCURACY (%) 1 Fig. 6 Fig. 7 98.72 2 Fig. 8 Fig. 9 100 3 Fig. 10 Fig. 11 62.90 4 Fig. 12 Fig. 13 14.65 2018, IJCSMC All Rights Reserved 113

The above table describes in percentage the accuracy rate of all the inputs. IV. CONCLUSION AND FURTHER WORK K-NN: By using this algorithm we can train the data but only up to a certain limit. It has been observed that after training the data again and again and executing the program the accuracy does not increase further. The maximum accuracy obtained here was only 57.69% but it had certain limitations. a) It could not recognize any handwritten inputs which is the main drawback considering the purpose of our project. b) Unfortunately, training the handwritten data also didn t help. TESSERACT OCR: After a unsuccessful attempt of recognizing handwritten input and less accurate digital input using k-nn algorithm, we considered using the Tesseract OCR. Here it was observed that not only the accuracy of recognizing the digital input improved drastically by about 35-40% but the OCR could also recognize well handwritten input which served the main purpose of the project. Our attempts are to retrain the OCR so that the accuracy of recognizing the handwritten inputs also increase eventually. While it is not possible to further train and improve the efficiency of k-nn, our attempts are to improve the efficiency of Tesseract OCR by retraining it. If this training becomes successful, it will become much more accurate in recognizing the handwritten input as well. Further, since OCR provides us with a large range of languages to be recognized, Tesseract OCR can be used for recognition of other languages too. REFERENCES [1] Optical Character Recognition by Open source OCR Tool Tesseract: A Case Study. Authors: Patel, Chirag; Patel, Atul; Patel, Dharmendra. Publication: International Journal of Computer Applications, vol. 55, issue 10, pp. 50-56. Publication Date: 10/2012. Origin: CROSSREF. DOI: 10.5120/8794-2784. [2] An Overview of the Tesseract OCR Engine Ray Smith Google Inc. [3] 1. R. Plamondon and S. N. Srihari, On-line and off-line handwriting recognition: A comprehensive survey, IEEE Trans. Pattern Anal. Mach. Intell., vol. 22, no. 1, pp. 63 84, Jan. 2000. [4] Handwriting Profiling using Generative Adversarial Networks Arna Ghosh and Biswarup Bhattacharya and Somnath Basu Roy Chowdhury Department of Electrical Engineering, Indian Institute of Technology Kharagpur 7 Nov 2016. [5] Multi-Language Online Handwriting Recognition, Daniel Keysers, Thomas Deselaers, Henry A. Rowley, Li-Lun Wang, and Victor Carbune. IEEE Transactions on pattern analysis and machine intelligence, Vol. 39, No. 6, June 2017 [6] Preprocessing for real-time handwritten character recognition, Bartosz Paszkowski, Wojciech Bieniecki and Szymon Grabowski Katedra Informatyki Stosowanej, Politechnika LÃ o dzka wbieniec@kis.p.lodz.pl 2018, IJCSMC All Rights Reserved 114