Ear Recognition based on 2D Images

Similar documents
Shape Model-Based 3D Ear Detection from Side Face Range Images

Generic Face Alignment Using an Improved Active Shape Model

Ear recognition based on Edge Potential Function

Enhanced Active Shape Models with Global Texture Constraints for Image Analysis

Advanced Ear Biometrics for Human Recognition

Face Alignment Under Various Poses and Expressions

Ear Biometrics Based on Geometrical Method of Feature Extraction

An Adaptive Eigenshape Model

Empirical Evaluation of Advanced Ear Biometrics

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Human pose estimation using Active Shape Models

Application of Geometry Rectification to Deformed Characters Recognition Liqun Wang1, a * and Honghui Fan2

Image retrieval based on region shape similarity

Research Article Ear Recognition Based on Gabor Features and KFDA

Automatic Modelling Image Represented Objects Using a Statistic Based Approach

Task analysis based on observing hands and objects by vision

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Haresh D. Chande #, Zankhana H. Shah *

AAM Based Facial Feature Tracking with Kinect

Intensity-Depth Face Alignment Using Cascade Shape Regression

Graph Matching Iris Image Blocks with Local Binary Pattern

A New Gabor Phase Difference Pattern for Face and Ear Recognition

An Adaptive Threshold LBP Algorithm for Face Recognition

Hand Gesture Extraction by Active Shape Models

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS

Facial Feature Points Tracking Based on AAM with Optical Flow Constrained Initialization

Personal Identification Using Ear Recognition

An algorithm of lips secondary positioning and feature extraction based on YCbCr color space SHEN Xian-geng 1, WU Wei 2

Active Appearance Models

A Robust Two Feature Points Based Depth Estimation Method 1)

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions

Medical Image Segmentation Using Descriptive Image Features

Rotation Invariant Finger Vein Recognition *

Algorithm research of 3D point cloud registration based on iterative closest point 1

Image Enhancement Techniques for Fingerprint Identification

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Active Appearance Models

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Binary Image Processing. Introduction to Computer Vision CSE 152 Lecture 5

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

A Method of weld Edge Extraction in the X-ray Linear Diode Arrays. Real-time imaging

Partial Face Matching between Near Infrared and Visual Images in MBGC Portal Challenge

Recent Advances in Electrical Engineering and Educational Technologies. Human Identification based on Ear Recognition

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

SOME stereo image-matching methods require a user-selected

2 Proposed Methodology

Biometrics Technology: Hand Geometry

Multimodal Belief Fusion for Face and Ear Biometrics

Chapter 3 Image Registration. Chapter 3 Image Registration

UBEAR: A Dataset of Ear Images Captured On-the-move in Uncontrolled Conditions

Recognition of Human Body Movements Trajectory Based on the Three-dimensional Depth Data

Improvement of SURF Feature Image Registration Algorithm Based on Cluster Analysis

Image Based Feature Extraction Technique For Multiple Face Detection and Recognition in Color Images

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Automatic Feature Extraction of Pose-measuring System Based on Geometric Invariants

Learning-based Neuroimage Registration

Hierarchical Shape Statistical Model for Segmentation of Lung Fields in Chest Radiographs

Distance-driven Fusion of Gait and Face for Human Identification in Video

The Elimination Eyelash Iris Recognition Based on Local Median Frequency Gabor Filters

Face recognition using SURF features

Face Alignment Using Active Shape Model And Support Vector Machine

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Biometric Security System Using Palm print

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

manufacturing process.

Segmentation in Noisy Medical Images Using PCA Model Based Particle Filtering

Critique: Efficient Iris Recognition by Characterizing Key Local Variations

Automated Pivot Location for the Cartesian-Polar Hybrid Point Distribution Model

Pose Normalization for Robust Face Recognition Based on Statistical Affine Transformation

An Automatic Face Identification System Using Flexible Appearance Models

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Automatically Building Appearance Models from Image Sequences using Salient Features.

A Novel Extreme Point Selection Algorithm in SIFT

Face View Synthesis Across Large Angles

A QR code identification technology in package auto-sorting system

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK

3D Face Recognition. Anil K. Jain. Dept. of Computer Science & Engineering Michigan State University.

Face Alignment Using Statistical Models and Wavelet Features

New Experiments on ICP-Based 3D Face Recognition and Authentication

Expanding gait identification methods from straight to curved trajectories

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

An adaptive container code character segmentation algorithm Yajie Zhu1, a, Chenglong Liang2, b

MOTION STEREO DOUBLE MATCHING RESTRICTION IN 3D MOVEMENT ANALYSIS

Semi-Automatic Detection of Cervical Vertebrae in X-ray Images Using Generalized Hough Transform

Fusion of Hand Geometry and Palmprint Biometrics

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Designing of Fingerprint Enhancement Based on Curved Region Based Ridge Frequency Estimation

Motion Estimation and Optical Flow Tracking

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

Practical Image and Video Processing Using MATLAB

Coarse-to-Fine Statistical Shape Model by Bayesian Inference

Outline 7/2/201011/6/

EAR RECOGNITION AND OCCLUSION

A 3D Point Cloud Registration Algorithm based on Feature Points

A Novel Image Transform Based on Potential field Source Reverse for Image Analysis

Robust and Accurate Detection of Object Orientation and ID without Color Segmentation

FACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS

Tongue Recognition From Images

Body Trunk Shape Estimation from Silhouettes by Using Homologous Human Body Model

Transcription:

Ear Recognition based on 2D Images Li Yuan, Zhi-chun Mu Abstract-Research of ear recognition and its application is a new subject in the field of biometrics authentication. Ear normalization and alignment is a fundamental module in the ear recognition system. Traditional manually normalization method is not suitable for an automatic recognition system. In this paper, an automatic ear normalization method is proposed based on improved Active Shape Model (ASM). This algorithm is applied on the USTB ear database for ear normalization. Then Full-space Linear Discriminant Analysis (FSLDA) is applied for ear recognition on the normalized ear images with different rotation variations. Experiments are performed on USTB ear image database. Recognition rates show that based on the right ear images, the acceptable head rotation range for ear recognition is between the right rotation of 20 degree to the left rotation of 10 degree. I. INTRODUCTION Earlier research has shown that human ear is one of the representative human biometrics with uniqueness and stability [1]. As an emerging biometrics technology, ear recognition is attracting more and more attention in biometrics recognition [2]. An ear recognition system usually involves ear detection, feature extraction and ear recognition/verification modules. At present, there have already been some existing ear feature extraction methods for 2D intensity images as follows: Alfred lannarelli presented a 12-measurement method, which is not suitable for machine vision because of the difficulty of localizing the anatomical points [1]. Moreno used feature points of outer ear contour and information obtained from ear shape and wrinkles for ear recognition [2]. Burge and Burger used the main curve segments to form Voronoi diagram, and build adjacency graph out from Voronoi diagram. Then a graph matching based algorithm was used for authentication [3]. Hurley treated the ear image as an array of mutually attracting particles that act as the source of Gaussian force field. The original image is described by a set of potential channels and positions of potential wells [4]. Wang used a high order moment-invariant method to extract ear features [5]. Mu presented a long axis based shape and structural feature extraction method, the shape feature is consisted of the curve fitting parameters of the outer ear contour, the structural feature is composed of ratios of between the length of key sections and the length of Manuscript received May 25, 2007. This work is supported by the National Natural Science Foundation of China under the Grant No. 60573058, and the Key Discipline Construction Program of Beijing Municipal Commission of Education No. XK100080537. Li Yuan and Zhi-chun Mu are with the School ofinformation Engineering, University of Science and Technology Beijing, China, (86-10-62334995; e-mail: yuanli64 7, hotmail.com). the long axis [6]. Chang used standard PCA algorithm for ear recognition, and gets the conclusion that ear and face does not have much difference on recognition rate (71.4% and 70.6% respectively on image database from USF) [7]. For ear recognition in 3D, Hui Chen [8] proposed an ICP based 3D ear recognition. Pin Yan [9] proposed an automatic 3D ear recognition system, which includes automatic ear extraction, and ICP based ear recognition. In prior 2D ear recognition, the pre-processing of ear images has been a manual process, which is not applicable for an automatic system. The 2D ear recognition also has the problem caused by rotation variations. So this paper includes two major parts: automatic ear extraction from 2D images using improved active shape model, and the study of the acceptable head rotation range for ear recognition. II. EAR EXTRACTION AND NORMALIZATION In this section, an automatic ear normalization method is proposed based on improved Active Shape Model (ASM). In the offline training step, ASM is used to create the Point Distributed Model using the landmark points on the ear images of the training set. This model is then applied on the test ear images to search the outer ear contour. The final step is to normalize the ear image to standard size and direction according to the long axis of outer ear contour. The long axis is defined as the line crossing through the two points which have the longest distance on the ear contour. After normalization, the long axis of different ear images will be normalized to the same length and same direction. A. ASM based Outer Ear Contour Extraction Using the long axis as the standard for normalization requires accurate location of the outer ear contour. Active Shape Model is a powerful statistical tool for object location with higher accuracy and robustness. It mainly relies on Principle Component Analysis to statistically model the variability in the training set of example shapes, and then iteratively deforms to fit to an example of the object in a new image. The outer ear contour extraction process has the following four steps. 1. Ear Landmarks for Statistical Shape Models A shape model is described by n landmark points that represent the boundary internal features, or even external features. Each training shape XA is represented as a vector: -Xi = (Xil,Xi2,...,Xin,Yil,Iy2,*.*,Yin )T, (i = 1, 2,..., N) (1) 2. Point Distributed Model 978-1-4244-1597-7/07/$25.00 2007 IEEE

ASM creates a Point Distributed Model using the statistical feature of the landmark points [10]. The training samples need to be aligned so that the corresponding points on different samples are comparable. The training shapes are all aligned by scaling, rotation and translation for minimizing the weighted sum of the squared distances between their equivalent landmark points. Here, the similarity transformation is applied using three factors: scaling factor s, rotation factor 0 and translation factor t. As shown in Fig. 1, the shape distribution is more of central tendency after alignment. + 0 006 (a) (b) Fig. 1. Trainings samples before and after alignment: (a) before alignment, (b) before alignment. After alignment, principle component analyses is applied aligned shape. Therefore, an ear shape model can be approximated as follows: X = Tt 68(X+pb) (2) to the where X is the mean shape, p is a set of orthogonal models of shape variation and b is a vector of shape parameters. TtIs is a translation including rotating by 0, scaling by s and translation of (xt, yt ). The shape vector will be different when parameter b changes. Generally, the variation range for b is -3JT7 < bi < 3 T, so that the shape is similar to the ear, where A, is the eigen value of the shape covariance matrix. Fig. 2 shows how the shape model changes when b is different. (d)qz--b+,0 xvt... (e) (0-3 Fig. 2. Different shape models when b changes. 3. Improved ASM for Outer Ear Contour Extraction a contour based gray level matching algorithm is Here, used to make the shape model quickly find the matching feature point. For each landmark point on the training shape, gray level sampling is applied along the normal direction of the boundary. For the jth landmark point on the ith training image, we sample k points along normal direction on each side, so a vector is composed of the 2k+1 points, and its normalized difference is yi. For the jth landmark point of N images, the mean value is y1, the covariance is Sy, thus we get the gray level statistical model for this landmark point. Apply the same method on each landmark point, we get the statistical model of all the landmark points on the target contour. In this paper, an iterative scheme is applied for searching the outer ear contour with the following steps: 1) Initialize the model shape. Firstly, set b = 0 so X = TSJ> X. Then adjust si, 1i and ti to make the initial model Xi as similar to an actual outer ear contour as possible; 2) According to the Gray Level Average Model of each point, we can get its matching point, and then a new shape vector X, is formed; 3) Get the new parameters s, 0, t and b for Xl; 4) Update si, 0i, t, and b of Xi with the new parameters, and a new shape will be formed. 5) If the change ofthe shape is smaller than a threshold or a given iteration number is reached, the algorithm is converged. Otherwise, repeat step (2)-(4). Fig. 3 shows the searching process. In ASM, local texture model matching is conducted under the assumption that the normalized first derivative profile { yj } satisfies a Gaussian distribution. The matching degree of a probe sample y, to the reference model is given by f (ys) = (y -y TS (ys -y1), where y- is the mean of { yj }( 1 < j < N ), and N is the number of images for establishing Point Distribution Model (PDM), SY is the covariance [ 1]. The main idea of finding the matching point is to minimize f (y5) which is equivalent to minimizing the probability that ys comes from the distribution. For most images interpretation task, the landmarks are usually selected in the points that have strong edge information. So the destination points should have also strong edge information. These points tend to lie on the edges

extracted by an edge operator. Thus we can adjust the matching method by adding a weight to the Malahanobis distance function: f (y,) =(c -e)(ys ) y(y Y;(3 wherec is a constant, e is the Sobel edge intensity at the target point. With this improvement, points with strong edge information are more probably chosen as the best candidate. For the ear images, the landmarks on the outer ear contour part have clearly strong edge intensity, while the landmarks on other parts are not necessary with this property. Therefore, we apply the edge constraints to the landmarks on the outer ear contour. Through experiment, we get the statistical value of c, c =260. As illustrated in Fig. 4, most mismatched points on the contour in the standard ASM have been correctly matched by applying the improved matching function. (a) (b) Fig. 5. Process of ear normalization: (a) original image; (b) rotation; (c) scaling, LA =160; (d) ear extraction; (e) normalization to 160X82. C. Ear Normalization Experiment The experiments are performed on the dataset collected by us (USTB dataset) [12]. It contains subsetl and subset2. Subsetl collected ear images of 77 subjects, 4 images for each. The images of each subject are taken under two conditions: illumination variation and orientation variation and individuals were invited to be seated 2m from the camera and change his/her face orientation. The 4 images are 300x 400 pixels in size as shown in Fig. 6. Fig. 6(a) and Fig. 6(d) are the profile view of the head under different lighting condition. Fig. 6(b) and Fig. 6(c) are -30 and +30 rotation respectively. (c) Fig. 4. Search result before and after match function improved (a) Sobel edge extraction, (b) conventional matching function, (c) improved matching function. B. Ear Normalization After the outer ear contour is found using ASM, we choose 30 feature points on the contour. Search the longest distance between two points (x1, yl) and (X2 Y2) among the 30 points on the outer ear contour, connect the two points to form the long axis, the length of the long axis is defined as LA, as shown to be the AB in Fig. 5(a). After normalization, LA for each ear is the same, and the direction of LA is perpendicular to axis X. The normalization process has the following steps: 1) Rotate the image, so that the long axis is perpendicular to axis X, as shown in Fig. 5(b). 2) Through experiment, we found that the mean value ofthe ration between ear height and ear width is 1.95. So in this paper, ear is normalized to be 82 X 160 pixels. So the test image is zoomed to make sure that the long axis is 160, as shown in Fig. 5(c). 3) Extract the ear using four points: two points on the long axis (xl, y ), (x2, Y2), the left most point on the outer ear contour, and the right most point on the outer ear contour, as shown in Fig. 5(d). 4) Normalize the ear to be 82 X 160 pixels, as shown in Fig. 1 5(e). 5) Lighting normalization using the histogram equalization method, as shown in Fig. 5(e). (a) No. 1 (b) No. 2 (c) No. 3 (d) No. 4 Fig. 6. Example images of the USTB ear database. The ear normalization experiments are performed on subsetl. In four experiments, the probe images are No.1 to No. 4 respectively, and the remaining three images for each subject are utilized to establish the PDM. Then, the PDM is utilized by the proposed ASM to search a shape for the one not participating in establishing the PDM. Fig. 7 shows some example images of successful ear normalization. The success rate for the four experiments are 84%, 7500, 86% and 7500. In practical application, the ear images may be taken under lighting variation or with different accessory. The proposed method works well under these conditions. Fig. 8 (a) and (b) show ear normalization under different light conditions. Fig. 8 (c) and (d) show ear normalization with earrings and glasses.

11 shows the ear recognition rate at different rotation angles. When the head turns left, the recognition rates at 5, 10, 15 and 20 degree are over 9000. The recognition rate drops dramatically after the rotation angles of 20 degree. When the head turns right, the recognition rates at 5 and 10 degree are over 900. The recognition rates at 15 and 20 degree are over 80%. The recognition rate drops dramatically after the rotation angles of 20 degree. In the range of 20 degree to 60 degree, the recognition rate ofright rotation is higher than that of left rotation at the same rotation degree. 0.9 _ Fl- 1 1 AE --E± - L {d I _._I -I- -I- -I- -I- -I- n 4~ III. EAR RECOGNITION In this paper, a full-space linear discriminant analysis (FSLDA) algorithm is applied on the USTB ear database [12] which contains ear images with rotation variations. FSLDA makes full use of the discriminant information in the full-space ofthe within-class scatter, namely the discriminant information in the null space and non-null space of the within-class scatter. 0.8 0.7j',, 0.3 0.2 0.1 oi_ o 20 30 40 50 60 70 dimension offeature vector - -Pl-,-- - --r-.-, -P- --"- 80 90 100 I IV. EXPERIMENTAL RESULTS The ear recognition experiments use both subsetl and subset2. USTB ear subset2 is a multimodal image database. It contains profile views of 79 subjects. The subjects in subsetl and subset2 are different. For each subject in subset2, we take the right ear images of right rotation at 0, 5, 10, 15, 20,25, 30, 40, 45 and 60 degrees, and also the right ear images of left rotation at 5, 10. 15. 20. 25. 30. 35. 40 and 45 degrees. Fig. 9 shows some example images. o.i1l 1 0 20 30 40 50 80 70 dimension of feature vector (b) right rotation Fig. 10. Ear recognition rate at different degrees. Fig. 9. E;xample images Irom UST1 ear database The ear images from subset 1 are used for training to get the feature space with the FSLDA mentioned in section 3. The ear images of 0 degree from subset2 are used for the gallery. The ear images of left rotation and right rotation at other degrees from subset2 are used for the probe. Fig.10 (a) shows the recognition rate at different degrees with left rotation. Fig.10 (b) shows the recognition rate at different degrees with right rotation. It can be seen that the recognition drops as the rotation angle increases. When the dimension ofthe feature space is fixed at 80, Fig. 0.7 0.6 1 E..t, 0.4 _ 0.2 0 9747, e 0 96 1 0000 0.9673 0073 09~~~~~~~~)0.9367 "I 7~ 0)X0 0.4664 ` 0. 764 6(46 Wt 0.7975-1 0.4610 01013 0.1 I''' 11-60 -40-20 0 20 40 60 turning angle Fig. 11. The recognition rate at different rotation angles (dimension of feature space= 80).

So in this experiment, the following conclusion can be made: when the right ear images are used for recognition, and if the recognition rate of over 9000 is preferred in practical application, the acceptable head rotation range is between 20 degree of left rotation to 10 degree of right rotation. V. CONCLUSIONS This paper has presented the ear recognition using 2D images. The improved ASM is applied to solve the ear nornalization problem. The main contributions are: (1) automatically search the outer ear contour on the probe ear image; (2) automatic ear extraction based on the long axis of the outer ear contour. Another contribution is the study of sensitivity to rotation using the Full-space Linear Discriminant Analysis (FSLDA). Experiments are performed on USTB ear image database. Recognition rates show that based on the right ear images, the acceptable head rotation range is right rotation of 20 degree and left rotation of 10 degree if we want to get a desirable recognition rate. This result may be helpful for the design of a practical ear recognition system. REFERENCES [1] Afred lannarelli, "Ear Identification." Forensic Identification Series. Paramount Publishing Company, Fremont, California, 1989. [2] Bele Moreno, Angel Aanchez and Jose F.Velez, "Use Outer Ear Images for Personal Identification in Security Applications," in Proc. ofthe 33rdAnnual 1999 International Carnahan Conference, Madrid, 1999, pp. 469-476. [3] Burge. M and Burger. W, "Ear Biometrics in Computer Vision," in: Proc. 15th International Conference ofpattern Recognition, 2000, pp. 822-826. [4] David J. Hurley, Mark S. Nixon and John, N. Carter, "Force field feature extraction for ear biometrics," Computer Vision and Image Understanding, 2005, 98, pp. 491-512. [5] Wang Zhongli, Mu Zhichun, Wang Xiuyan, Mi Hongtao, "Ear recognition based on Moment Invariant Matching", Pattern Recognition andartificial Intelligence (in Chinese), Dec. 2004, Vol 17, No. 4, pp. 502-505. [6] Mu Zhichun, Yuan Li, Xu Zhengguang, "Shape and Structural Feature Based Ear Recognition," Advances in Biometric Person Authentication, Guangzhou, China, Dec. 2004, pp. 663-670. [7] Chang, K., Bowyer. K.W., Sarkar, S, Victor, B, "Comparison and Combination of Ear and Face Images in Appearance-Based Biometrics," IEEE Transactions on Pattern Analysis and Machine Intelligence, September 2003, Vol. 25, No. 9, pp. 1160-1165. [8] Hui. Chen and B. Bhanu, "Human Ear Recognition in 3D," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol.29, no.4, Apr. 2007, pp. 7 18-737. [9] Ping Yan and Kevin W. Bowyer, "Biometric recognition using three-dimensional ear shape," IEEE Transactions on Pattern Analysis and Machine Intelligence, to be published. [10] F. Cootes, C. J. Taylor, D. H. Cooper, J. Graham, "Active shape models-their training and application," Computer Vision and Image Understanding, 1995, vol. 61, no. 1, pp. 38-59. [11] Cootes T F, Taylor C J, Lanitis A, "Multi-Resolution Search with Active Shape Model," in Proc. of the 12th International Conference on Pattern Recognition, 1994, vol.1, pp. 610-612. [12] http://www.ustb.edu.cn/resb/.