Available online at ScienceDirect. Procedia Computer Science 45 (2015 )

Similar documents
Fine Classification of Unconstrained Handwritten Persian/Arabic Numerals by Removing Confusion amongst Similar Classes

A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points

A New Approach to Detect and Extract Characters from Off-Line Printed Images and Text

OCR For Handwritten Marathi Script

Available online at ScienceDirect. Procedia Computer Science 85 (2016 )

Segmentation of Arabic handwritten text to lines

Keyword Spotting in Document Images through Word Shape Coding

Isolated Handwritten Words Segmentation Techniques in Gurmukhi Script

Hidden Loop Recovery for Handwriting Recognition

Input sensitive thresholding for ancient Hebrew manuscript

Keywords Connected Components, Text-Line Extraction, Trained Dataset.

Segmentation of Kannada Handwritten Characters and Recognition Using Twelve Directional Feature Extraction Techniques

Automatic discrimination between printed and handwritten text in documents

Understanding Tracking and StroMotion of Soccer Ball

Optical Character Recognition (OCR) for Printed Devnagari Script Using Artificial Neural Network

Available online at ScienceDirect. Procedia Computer Science 59 (2015 )

Available online at ScienceDirect. Procedia Engineering 161 (2016 ) Bohdan Stawiski a, *, Tomasz Kania a

CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS

A Painting Based Technique for Skew Estimation of Scanned Documents

A Technique for Classification of Printed & Handwritten text

II. WORKING OF PROJECT

Three-Dimensional Reconstruction from Projections Based On Incidence Matrices of Patterns

A two-stage approach for segmentation of handwritten Bangla word images

Tracing and Straightening the Baseline in Handwritten Persian/Arabic Text-line: A New Approach Based on Painting-technique

Comparative Study of ROI Extraction of Palmprint

A Feature based on Encoding the Relative Position of a Point in the Character for Online Handwritten Character Recognition

LECTURE 6 TEXT PROCESSING

A Survey of Problems of Overlapped Handwritten Characters in Recognition process for Gurmukhi Script

An Efficient Character Segmentation Based on VNP Algorithm

Layout Segmentation of Scanned Newspaper Documents

Available online at ScienceDirect. Procedia Computer Science 58 (2015 )

Signature Based Document Retrieval using GHT of Background Information

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task

A DIGITAL APPROACH TO HANDWRITTEN DOCUMENTS. B.I.T. - Bureau Ingénieur Tomasi

Topic 4 Image Segmentation

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

A Model-based Line Detection Algorithm in Documents

Segmentation of Isolated and Touching characters in Handwritten Gurumukhi Word using Clustering approach

Recognition of Unconstrained Malayalam Handwritten Numeral

Skew Detection and Correction of Document Image using Hough Transform Method

Identifying Layout Classes for Mathematical Symbols Using Layout Context

Automatically Algorithm for Physician s Handwritten Segmentation on Prescription

Skew Detection Technique for Binary Document Images based on Hough Transform

Handwriting segmentation of unconstrained Oriya text

Short Survey on Static Hand Gesture Recognition

Robust line segmentation for handwritten documents

Data Hiding in Binary Text Documents 1. Q. Mei, E. K. Wong, and N. Memon

Text Extraction from Natural Scene Images and Conversion to Audio in Smart Phone Applications

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation

A Document Image Analysis System on Parallel Processors

Automatic Recognition and Verification of Handwritten Legal and Courtesy Amounts in English Language Present on Bank Cheques

Human Motion Detection and Tracking for Video Surveillance

Available online at ScienceDirect. Energy Procedia 69 (2015 )

AUTOMATIC LOGO EXTRACTION FROM DOCUMENT IMAGES

HCR Using K-Means Clustering Algorithm

A New Algorithm for Detecting Text Line in Handwritten Documents

Handwritten Devanagari Character Recognition Model Using Neural Network

Time Stamp Detection and Recognition in Video Frames

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

Chapter 2. Frequency distribution. Summarizing and Graphing Data

1. INTRODUCTION. AMS Subject Classification. 68U10 Image Processing

Scanner Parameter Estimation Using Bilevel Scans of Star Charts

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Motion Detection Algorithm

Chain Code Histogram based approach

An Efficient QBIR system using Adaptive segmentation and multiple features

OTCYMIST: Otsu-Canny Minimal Spanning Tree for Born-Digital Images

A Statistical approach to line segmentation in handwritten documents

Prepare a stem-and-leaf graph for the following data. In your final display, you should arrange the leaves for each stem in increasing order.

A new approach to reference point location in fingerprint recognition

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

Handwritten Gurumukhi Character Recognition by using Recurrent Neural Network

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

DATA EMBEDDING IN TEXT FOR A COPIER SYSTEM

International Journal of Advance Research in Engineering, Science & Technology

Applied Statistics for the Behavioral Sciences

Learning-Based Candidate Segmentation Scoring for Real-Time Recognition of Online Overlaid Chinese Handwriting

Cursive Handwriting Recognition System Using Feature Extraction and Artificial Neural Network

CS 223B Computer Vision Problem Set 3

Digital Image Processing

Chapter 2 Describing, Exploring, and Comparing Data

Latest development in image feature representation and extraction

Localization, Extraction and Recognition of Text in Telugu Document Images

Available online at ScienceDirect. Procedia Computer Science 56 (2015 )

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme

Classification of Printed Chinese Characters by Using Neural Network

Texture Image Segmentation using FCM

Extracting Characters From Books Based On The OCR Technology

ScienceDirect. Reducing Semantic Gap in Video Retrieval with Fusion: A survey

Measures of Central Tendency. A measure of central tendency is a value used to represent the typical or average value in a data set.

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Color Image Segmentation

Available online at ScienceDirect. Procedia Computer Science 96 (2016 )

Available online at ScienceDirect. Procedia Computer Science 89 (2016 )

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Estimation of Skew Angle in Binary Document Images Using Hough Transform

Estimation of Skew Angle in Binary Document Images Using Hough Transform

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Transcription:

Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 45 (2015 ) 205 214 International Conference on Advanced Computing Technologies and Applications (ICACTA- 2015) Automatic Removal of Handwritten Annotations from Between- Text-Lines and Inside-Text-Line Regions of a Printed Text Document P. Nagabhushan, Rachida Hannane, Abdessamad Elboushaki, Mohammed Javed* Department of Studies in Computer Science, University of Mysore, Mysore-570006, India Abstract Recovering the original printed text document from handwritten annotations, and making it machine readable is still one of the challenging problems in document image analysis, especially when the original document is unavailable. Therefore, our overall aim of this research is to detect and remove any handwritten annotations that may appear in any part of the document, without causing any loss of original printed information. In this paper, we propose two novel methods to remove handwritten annotations that are specifically located in between-text-lines and inside-text-line regions. To remove between-text-line annotations, a two stage algorithm is proposed, which detects the base line of the printed text lines using the analysis of connected components and removes the annotations with the help of statistically computed distance between the text line regions. On the other hand, to remove the inside-text-line annotations, a novel idea of distinguishing between handwritten annotations and machine printed text is proposed, which involves the extraction of three features for the connected components merged at word level from every detected printed text line. As a first distinguishing feature, we compute the density distribution using vertical projection profile; then in the subsequent step, we compute the number of large vertical edges and the major vertical edge as the second and third distinguishing features employing Prewitt edge detection technique. The proposed method is experimented with a dataset of 170 documents having complex handwritten annotations, which results in an overall accuracy of 93.49% in removing handwritten annotations and an accuracy of 96.22% in recovering the original printed text document. 2015 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license Peer-review (http://creativecommons.org/licenses/by-nc-nd/4.0/). under responsibility of scientific committee of International Conference on Advanced Computing Technologies and Applications Peer-review under (ICACTA-2015). responsibility of scientific committee of International Conference on Advanced Computing Technologies and Applications (ICACTA-2015). Keywords: Handwritten annotation removal; Marginal annotation removal; Between-text-line annotations; Inside-text-line annotation. * Corresponding author: Tel: +91-97-4116-1929 E-mail address: javedsolutions[at]gmail.com; mohammed.javed.2013[at]ieee.org 1877-0509 2015 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). Peer-review under responsibility of scientific committee of International Conference on Advanced Computing Technologies and Applications (ICACTA-2015). doi:10.1016/j.procs.2015.03.123

206 P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 1. Introduction Annotating a printed text document refers to the process or the act of adding annotations or writing critical commentary or explanatory notes into the machine printed text documents. Adding annotations at different regions of the printed text document is the habit of many people whenever they read any document. These annotations may be some important observations or corrections marked in their own style; this makes the problem of recovering the original text document very challenging, particularly when the original document is not available. A reader can add different types of annotations in any part of the document; the type of annotations depend on the information being read by the reader. For example, the annotations can be lines underscoring or side scoring to highlight keywords, sentences or part of a paragraph and so on, or even enclosing such a part with a circular, elliptical or contour shapes holding a word or a sentence of interest, or it can be simply a question mark (when the reader does not understand the content), also it can be in the form of more frequent annotations that are handwritten comments (in which the related idea is extrapolated or the missed details are added), or it can be any other type of external remark that can be added anywhere in a document including: in marginal area of the document, in betweentext-line regions, inside-text-line regions crossing over the printed text lines. Based on the regions where annotations appear, we classify them into the following categories: marginal annotations, between-text-line annotations, insidetext-line annotations, and overlapping annotations (see Fig. 1). There are a few attempts reported in the literature related to handwritten annotations in printed text documents 2,4,5,7. However, we understood that the nature of the annotations considered in 2,4,5,7 is predefined in structure and location. The method described in 5 by Mori and Bunke, have achieved high extraction rates by limiting the colors or types of annotations. Although the correction is successful, their method has the limitation on colors of annotations as well as type of documents. Their method does not focus on the extraction of annotations but on the utilization of extracted annotations. Guo and Ma 4 proposed a method that involves the separation of handwritten annotations from machine printed text within a document. Their algorithm is based on the theory of hidden Markov models to distinguish between machine-printed and handwritten materials, and the classification is performed at the word level. Handwritten annotations are not limited to marginal areas and also their approach can deal with document images having handwritten annotations overlaid on machine-printed text. However, their extracted annotations are limited to some predefined characters; their method cannot extract handwritten line drawings which is also a frequently used annotation. Y. Zheng et al. 7 proposed a method to segment and identify handwritten text from machine printed text in a noisy document; their novelty is that they treat noise as a distinguished class and model noise based on selected features. Trained Fisher classifiers are used to identify machine printed text and handwritten text from noise. Lincoln Faria da Silva et al. 2 proposed a method that involves the recognition of the handwritten text and machine printed text in a scanned document image. New features for classification of the handwritten text and machine printed text are proposed as well: Vertical Projection Variance, Major Horizontal Projection Difference, Pixels Distribution, Vertical Edges, and Major Vertical Edge. Unfortunately, their method can work only when the machine printed text and handwritten texts are separated. However, in our proposed method, we explore the feature of vertical Prewitt edge in order to remove the handwritten annotations inside the printed text line regions. To our best knowledge, the proposed methods in the literature do not present a systematic approach for handling the different annotations that appear in printed text documents. Therefore, in order to systematically handle the different annotations coined in this paper, we propose a Handwritten Annotation Removal System (HARS) based on the different types of annotations that a document can have (see Fig. 2). However in this paper, we specifically focus on detecting and removing annotations located in between text lines and inside text line regions. Rest of the paper is organized as follows: In section 2, we detail the proposed methods to remove between-text-line and inside-text-line annotations. In section 3, the experimental results of our proposed methods are reported, and the last section 4 concludes the paper. 2. Proposed Methods The overall proposed system (HARS) for detecting and removing handwritten annotations consists of five major stages (see Fig. 2), where the scanned annotated document is taken as an input and converted to a binary image; this

P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 207 binary image in the first stage is pre-processed to remove noise coming out of the document scanning. The approach used here is described in Peerawit and Kawtrakul 6. Later, a pre-processing to correct any skew present in the document image is applied. We use the method, which is provided by Cao and Li 1. Further, the second stage of removing marginal annotations (See Fig. 2) has been attempted by Abdessamad et al. 3. They use the vertical and horizontal projection profiles in order to detect the marginal boundary and remove the annotations in one stretch. Because of annotations near the marginal area, the marginal boundary detected is approximate. Therefore, they use the concept of connected component analysis to bring back the printed text components cropped during the process of marginal annotation removal. In this research work, we propose two novel methods for removing between-textline annotations and inside-text-line annotations. The proposed methods are presented in the following subsections. Fig. 1: A sample document with handwritten annotations Fig. 2: Proposed Handwritten Annotation Removal System (HARS) 2.1. Removal of Between-text-line Annotations The handwritten annotations that are located between any two adjacent printed text line regions are called between-text-line annotations (see Fig. 3). Fig. 3: Sample of possible annotations in between-text-line regions In this section, the proposed model for removal of between-text-line annotations is developed (see Fig. 4) in which we present a two-stage algorithm for removing these annotations. They are described as follows: 2.1.1 Detection of the sure printed text lines We use the connected component labeling algorithm to determine the average character size of the characters in the document, which is computed statistically. However, the descender-less characters in a printed text line have essentially same base regions, whereas the handwritten annotations located in between-text-lines region have different base regions. Each connected component is represented by the midpoint of its base region (see Fig. 5), and then the vertical projection of these midpoints is obtained (see Fig. 6).

208 P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 \ Fig. 4: Proposed model for removal of between-text-line annotations Each printed text line region contains high density of connected components producing a high peak at its base region. Therefore, the high peaks in the histogram (in Fig. 6) represent the sure printed text lines. To identify these peaks we compute the mean (μ1) of all peaks in the histogram (H1), then select all peaks above the mean (μ1), and label them as a sure printed text lines. If there are any two peaks which are close to each other by a distance less than character size, then the largest peak is considered as the baseline of the text line (see Fig. 7). Fig. 5: Midpoint of the base region in connected components Fig. 6: Vertical projection profile for midpoints of base regions of connected components (H1) 2.1.2 Detection of the probable or missed text lines The reason that some small printed text lines are missed from the previous stage is that, their peaks are below the mean (μ1) obtained from histogram (H1) (see Fig. 6). To retrieve them, the procedure involves the removal of all higher peaks from the histogram (H1), followed by computation of mean (μ2) of all peaks from the remaining histogram (H2) (see Fig. 8), then for every peak P above the mean (μ2) in the histogram (H2), Consider P as base line of text line in the document if: distance[p,r] 2*charactersize R PriBL (1)

P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 209 where PriBL is set of selected peaks in the first iteration. Fig. 7: Baseline detection of sure printed text lines Fig. 8: Computation of the mean of peaks in histogram (H2) Fig. 9 shows a sample document after the recovery of probable missed text lines. After the detection of all the printed text lines in the annotated document, the next stage involves the removal of handwritten annotations between text lines; the method below demonstrates the process of removing of betweentext-line annotations. Step-1: Compute the average character size for every text line in the document. Step-2: For every two consecutive text lines T i and T i+1 remove the annotations between region of T i and region T i+1 of with the following formula T 0.5) * F An T ( F (0.6) * F ) (2) i ( Zi i 1 Zi 1 Zi 1 where : Text line of index i, : Font size of the text line, and An: the annotation. Fig. 9: Baseline detection for probable printed text lines Fig. 10: Removal of Between-Text-Line Annotations During this process, all annotations that appear in between-text-line regions are removed (see Fig. 10). The next stage in the proposed HARS (see Fig. 2) involves the removal of handwritten annotations from inside-text-line regions, which is discussed as follows. 2.2 Removal of Inside-text-line Annotations The handwritten annotations that are located inside the rectangular area in which the entire printed text line fits into are called inside-text-line annotations. We distinguish here two different types of annotations. The first type is overlapping annotations: which overlap with the printed text lines. The other type of annotations is normal annotations located in some free area inside the text line region but not crossing the printed text (see Fig. 11).

210 P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 In this section, a proposed model for removal of free area inside-text-line annotations is developed (see Fig. 12). The different stages involved in the proposed method are: 2.2.1 Merging of connected components at word level Fig. 11: Sample annotations inside-text-line regions The merging algorithm of connected components is used to segment the printed text at word level (see Fig. 13) on which the next operations will be preformed. The algorithm takes as input two neighboring connected components, and checks if they are in the same position and the difference between their heights is less than half character size, if so, then both of them are merged. This procedure will continue until no merging case remains and results in connected components of words from which the different features will be extracted. 2.2.2 Computation of density distribution Fig. 12: Proposed model for removal of inside-text-line annotations The vertical projection profile of the connected components at word level is obtained (see Fig. 14), and the profile is empirically divided in three regions (see Fig. 15) which are : Region1: Height of (0.6) * character size in the top part of the bounding box, called Ascender region. Region2: Height of character size in the middle part of the bounding box, called Middle region. Region3: Height of (0.5) * character size in the bottom part of the bounding box, called Descender region. It is visible that the vertical projection profile is almost similar and consistent for all machine printed words (see Fig. 15). Moreover, in case of printed text, the highest density of black pixels occurs in the middle part of the projection profile; therefore the ascender region and descender region are known to have low pixel densities. However, the density distribution of the profile for handwritten annotations depends on the type of annotation, and it

P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 211 changes from one annotation to another (see Fig. 16). Experimenting with 50 handwritten words, we observed that all three regions densities are all most similar. Based on this evidence, we conclude that the variance in densities of three regions for machine printed text is higher compared to the variance in case of handwritten annotation. Fig. 13: Merging of the characters to get a word Fig. 14: Vertical projection profile of both handwritten annotation and machine printed word Fig. 15: Three density regions in a vertical projection profile of a word Fig. 16: Density distribution in handwritten annotations and machine printed word We compute the variance of three regions densities using the formula: 2 1 2 2 2 2 cc *( DA DM DD ) (3) cc a where D A : Ascender density,d M: Middle density,d D : Descender density, μ cc: The mean of three regions densities 1 Number of black pixel in theregion *( DA DM DD ) and Density D cc a region area 2.2.3 Vertical Edge Detection In a printed text document where black pixels represent foreground (printed text) and white pixels represent the background of the document; the vertical edge is a set of connected pixels that are stretched in vertical direction. A spatial filter mask is based on the directional Prewitt filter (see Fig. 17). Fig. 18 shows the result of applying this mask on a handwritten and on a machine printed word. Vertical edge is also an important feature used here to distinguish the handwritten annotation from machine printed text. This feature also remains invariant to some overlapping annotations inside the connected components. We detect two important features related to vertical edges which are extracted from the connected components of words: number of large vertical edges and the major vertical edge. Fig. 17: Spatial filter masks

212 P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 Fig. 18: Performance of the vertical Prewitt edge Fig. 19: Example of alphabet containing large vertical edges Extracting large vertical edges: Vertical edge which has the height more than (2/3)* is considered as large vertical edge and more than 65% of the English alphabets have more than one vertical edge. (see Fig. 19). 2 length ( LVE) * character size (4) a where LVE is the large vertical edge As observed in Fig. 20, number of large vertical edges in a printed word is greater than the number of large vertical edges in a handwritten word. The following steps show how to compute the large vertical edge parameter: Step-1:Compute the number of the large vertical edge LVE. Step-2:Compute the parameter D LVE and store it as a feature. number of LVE D (5) LVE length(bb) where BB: Bounding Box. Fig. 20: Large vertical edges for handwritten and printed word Fig. 21: Major vertical edge for handwritten and printed text Extracting major vertical edge: Major vertical edge is the one with larger number of pixels in the component. It is also considered as a feature for distinguishing the handwritten annotations from printed words in the free area of the text line regions of a document. In this case we observe that, the difference between the longest edge and the height of its connected component in case of machine printed text is less compared to the case of handwritten annotation; this difference is called major vertical edge (see Fig. 21). The following steps show how to compute the major vertical edge parameter: Step-1:Compute the max or Height of the Major Vertical Edge. Step-2:Compute the parameter D MVE and store it as feature: D MVE where, BB: Bounding Box HMVE Height(BB) (6)

P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 213 2.2.4 Distinguishing between machine printed and handwritten word The Table 1 below summarizes the features used to distinguish between machine printed text and handwritten annotation based on the features described above. The classification rules are learnt in the training phase of 30 documents randomly selected from 170 dataset used here, the result obtained for classification are described in Fig. 22. Table-1: Feature wise comparison between machine printed text and handwritten annotation Features Machine printed text Handwritten annotation Variance of density distribution High Low Number of large vertical edges High Low Major vertical edge High Low 3. Experimental Results Fig. 22: The procedure to remove handwritten annotations obtained experimentally from 30 training documents To our best knowledge, we could not find any standard dataset with complex handwritten annotations, which suits our research requirements. Therefore, we have collected 170 text documents from literary books and scientific articles, and given to heterogeneous group of students for marking random annotations of their choice. They generate the annotations in their own styles and using their own pens with different thickness. Dataset document images with annotations were obtained by using a scanner of 300 dpi. We compute the accuracy of our proposed methods in two ways. 3.1 Accuracy of Removing Handwritten Annotations In order to compute the accuracy of removing between-text-line and inside-text-line annotations, we use the formula B A Accuracy HA 1 (7) A where A is the amount of expected annotations to be removed, and B is the amount of removed annotations. The experimental results show an average removal accuracy of 93.5%. Most of the documents in our dataset have accuracy above 93.49% (see Table 2 ). Table-2: Result of removing the annotations Table-3:Result of recovering original printed documents Accuracy Below 93.49% Above 93.5% No. of Documents (%) 38 62 Accuracy Below 96.22% Above 96.23% No. of Documents (%) 20 80

214 P. Nagabhushan et al. / Procedia Computer Science 45 ( 2015 ) 205 214 3.2 Accuracy of Recovering the Original Document To compute the accuracy of recovering the original document by removing between-text-line and inside-text-line annotations, we use the formula Accuracy OD 1 (8) where is the expected cleaned document without handwritten annotations, and is recovered document after the removal of annotations. Through our experiments, the average accuracy of getting the original document is 96.22% in which most of the documents of our dataset give accuracy above the average (see Table 3). We also compute the correlation coefficient between the original document and the processed document by using the formula 1 n x x y y r i 1 ( )( ) (9) n 1 Sx Sy where the number of pixels is given by n, the pixels of the original document are represented es by xi and the pixels of the processed document are represented by yi. The mean of the xi, yi is denoted by and. The standard deviation of the xi, yi is denoted by sx, sy respectively. However, the overall correlation between the original document and the processed document for 170 images of our database gives an average correlation of 0.9625. Therefore, it is clear that most of the documents that represent 83% from the dataset have a correlation above 0.9625. Further, the individual accuracies of three components shown in HARS (see Fig. 2) is tabulated in Table 4. Table-4: Cumulative accuracy of three components of HARS Accumulative Accuracy (%) Removing Handwritten Annotations Recovering Original Document Marginal 3 91.11 98.34 Between-Text-Line regions 94.14 95.90 Inside-Text-Line regions 91.45 95.60 Finally with the experiments carried out, we observe an overall error of 6.51%, which is due to the presence of overlapping annotations within the printed text line regions. Overlapping annotations have not be tackled in this paper, which could be anticipated as an extension work to the proposed research. 4. Conclusion In this research paper, we have described and discussed the methods to remove both of handwritten annotations located in between-text-lines and inside-text-line regions of a printed text document. The experimental results show that our methods produce consistent and adequate results under reasonable variation of type of annotations. We demonstrate the proposed idea on a dataset of 170 documents which gives an average of 0.9625 in correlation coefficient between the original document and the processed document, and an overall accuracy of 93.49% for removal of handwritten annotations, and 96.22% in case of recovering the original printed text document. However, as the handwritten annotations overlapping with printed text are not yet processed, we got only 6.51% as reduction in the accuracy of removing of the annotations. Therefore, our proposed algorithm removes any annotations inside text line regions except those annotations, which are over the printed text. References 1. Cao Y., Li H.. Skew detection and correction in document images based on straight-line fitting. P R Letters, 24(12):1871 1879, 2003. 2. Da Silva L. F., Conci A., Sanchez A.. Automatic discrimination between printed and handwritten text in documents. Published in Computer Graphics and Image Processing (SIBGRAPI), XXII Brazilian Symposium, pages 261 267, 2009. 3. Elboushaki A., Hannane R., Nagabhushan P., Javed M. Automatic removal of handwritten annotations in marginal area of printed text document. International Conference on ERCICA, Published by Elsevier, 2014. 4. Guo J. K., Ma M. Y. Separating handwritten material from machine printed text using hidden markov models. ICDAR, pages 436 443, 2001. 5. Mori D., Bunke H. Automatic interpretation and execution of manual corrections on text documents. In P. S. P. W. H. Bunke, Editor, Handbook of Character Recognition and Document Image Analysis, World Scientific, Singapore, pages 679 702, 1997. 6. Peerawit W., Kawtrakul A. Marginal noise removal from document images using edge density. In Proc. Fourth Information and Computer Eng. Postgraduate Workshop, 2004. 7. Zheng Y., Li H., Doermann D. The segmentation and identification of handwriting in noisy document images. In Lecture Notes in Computer Science, 5 th International Workshop DAS 2002 Princeton, NJ, USA., pages 95 105, 2002.