Reliable Face Recognition for Intelligent CCTV

Similar documents
Face Detection and Recognition in an Image Sequence using Eigenedginess

Pose Normalization for Robust Face Recognition Based on Statistical Affine Transformation

Criminal Identification System Using Face Detection and Recognition

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Dr. Enrique Cabello Pardos July

Face detection and recognition. Detection Recognition Sally

Mobile Face Recognization

Hand Gesture Extraction by Active Shape Models

Face Recognition using Tensor Analysis. Prahlad R. Enuganti

Face Objects Detection in still images using Viola-Jones Algorithm through MATLAB TOOLS

NOWADAYS, there are many human jobs that can. Face Recognition Performance in Facing Pose Variation

Mouse Pointer Tracking with Eyes

Subject-Oriented Image Classification based on Face Detection and Recognition

A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods

Active Appearance Models

Algorithm for Efficient Attendance Management: Face Recognition based approach

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Applications Video Surveillance (On-line or off-line)

Face Alignment Under Various Poses and Expressions

Illumination invariant face detection

Disguised Face Identification Based Gabor Feature and SVM Classifier

Learning to Recognize Faces in Realistic Conditions

A Study on Similarity Computations in Template Matching Technique for Identity Verification

Image Based Feature Extraction Technique For Multiple Face Detection and Recognition in Color Images

Parallel face Detection and Recognition on GPU

A Survey of Various Face Detection Methods

Face Tracking in Video

Face Recognition-based Automatic Tagging Scheme for SNS

Face Recognition for Mobile Devices

Classifier Case Study: Viola-Jones Face Detector

Viola Jones Face Detection. Shahid Nabi Hiader Raiz Muhammad Murtaz

FACE DETECTION BY HAAR CASCADE CLASSIFIER WITH SIMPLE AND COMPLEX BACKGROUNDS IMAGES USING OPENCV IMPLEMENTATION

Real-Time Model-Based Hand Localization for Unsupervised Palmar Image Acquisition

Recognition problems. Face Recognition and Detection. Readings. What is recognition?

Out-of-Plane Rotated Object Detection using Patch Feature based Classifier

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction

Face Recognition in Video: Adaptive Fusion of Multiple Matchers

Image enhancement for face recognition using color segmentation and Edge detection algorithm

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Image-Based Face Recognition using Global Features

Illumination Invariant Face Alignment Using Multi-band Active Appearance Model

Object and Class Recognition I:

Face Recognition using Tensor Analysis. Prahlad R. Enuganti

Face Detection and Alignment. Prof. Xin Yang HUST

A Hierarchical Face Identification System Based on Facial Components

Image Processing and Image Representations for Face Recognition

On Modeling Variations for Face Authentication

Part-based Face Recognition Using Near Infrared Images

Facial Feature Extraction Based On FPD and GLCM Algorithms

Categorization by Learning and Combining Object Parts

Two-View Face Recognition Using Bayesian Fusion

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

Haresh D. Chande #, Zankhana H. Shah *

Window based detectors

Generic Face Alignment Using an Improved Active Shape Model

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

A robust method for automatic player detection in sport videos

Discriminant Analysis of Haar Features for Accurate Eye Detection

Face detection, validation and tracking. Océane Esposito, Grazina Laurinaviciute, Alexandre Majetniak

Part-based Face Recognition Using Near Infrared Images

An Image Region Selection with Local Binary Pattern based for Face Recognition

Human pose estimation using Active Shape Models

Detecting People in Images: An Edge Density Approach

Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS

A Real Time Facial Expression Classification System Using Local Binary Patterns

Spatial Frequency Domain Methods for Face and Iris Recognition

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Face detection. Bill Freeman, MIT April 5, 2005

Face Recognition Using Active Appearance Models

Face and Nose Detection in Digital Images using Local Binary Patterns

An Efficient Face Detection and Recognition System

Face Recognition Across Non-Uniform Motion Blur, Illumination and Pose

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Gender Classification Technique Based on Facial Features using Neural Network

Recognition of Non-symmetric Faces Using Principal Component Analysis

Digital Vision Face recognition

Real Time Face Detection, Recognition and Tracking System for Human Activity Tracking

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES

The Effect of Facial Components Decomposition on Face Tracking

Dr. K. Nagabhushan Raju Professor, Dept. of Instrumentation Sri Krishnadevaraya University, Anantapuramu, Andhra Pradesh, India

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Face Recognition using Laplacianfaces

Object. Radiance. Viewpoint v

Face Detection on OpenCV using Raspberry Pi

Robust face recognition under the polar coordinate system

Lecture 4 Face Detection and Classification. Lin ZHANG, PhD School of Software Engineering Tongji University Spring 2018

MediaTek Video Face Beautify

Feature Detection and Tracking with Constrained Local Models

Three-Dimensional Face Recognition: A Fishersurface Approach

Age Group Estimation using Face Features Ranjan Jana, Debaleena Datta, Rituparna Saha

Study of Viola-Jones Real Time Face Detector

Vehicle Detection Method using Haar-like Feature on Real Time System

AAM Based Facial Feature Tracking with Kinect

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

Threshold Based Face Detection

Face/Flesh Detection and Face Recognition

Comparison of Different Face Recognition Algorithms

Transcription:

Reliable Face Recognition for Intelligent CCTV Ting Shan, Brian C. Lovell, Shaokang Chen and Abbas Bigdeli National ICT Australia (NICTA) and Intelligent Real-Time Imaging and Sensing Group EMI, School of ITEE, The University of Queensland ABSTRACT: Face Recognition is generating enormous interest due to government concerns about identity management and global terrorist activity. One ambition of Intelligent CCTV is to help prevent terrorism and a key technology is reliable face recognition. With passport quality photographs, current face recognition technologies can approach 95% recognition accuracy. Yet trials show that performance drops to only 10 to 20% when there are significant changes in lighting, pose, and facial expression. We describe research by the authors to address these issues and provide reliable face recognition performance in real-time. Our system has three major components comprising: 1) a Viola-Jones face detection module based on cascaded simple binary features to rapidly detect and locate multiple faces from the input still image or video sequences, 2) a Pose Normalization Module to estimate facial pose and compensate for extreme pose angles, and 3) Adaptive Principal Component Analysis to recognize the normalized faces. Experimental results show that our approach can achieve good recognition rates on face images across a wide range of head poses with different lighting and expressions. Introduction The Prime Minister of Australia, John Howard, returned from London after the July 2005 suicide bombings and said (Howard J 2005), The biggest thing that I have learnt by a country mile out of my visit, particularly to Britain, is the extraordinary value of surveillance cameras. The integrated security camera system deployed in Britain was primarily installed to reduce the incidence of assault and property damage. Yet it was able to help identify and discover the movements of the four suicide bombers within just 24 hours of the bombing. The effectiveness of this particular security system comes about because of the saturation coverage of the British people by CCTV. In Britain, CCTV systems cover cities, public transport and motorways, while in many other countries the coverage is quite haphazard. It was public demand for security in public places that led to this pervasiveness. Moreover, the adoption of centralised digital video databases, largely to reduce management and monitoring costs, has also resulted in an extraordinary co-ordination of the CCTV resources. An emerging testbed for intelligent CCTV is the emerging experimental planetary scale sensor web, IrisNet (Gibbons P B, Karp B, Ke Y, Nath S and Sehan S 2003). IrisNet uses internet connected desktop PCs and inexpensive, off-the-shelf sensors such as webcams, microphones, temperature, and motion sensors deployed globally to provide a wide-area sensor network. IrisNet is deployed as a service on PlanetLab (www.planet-lab.org), a worldwide collaborative network environment for prototyping next generation internet services initiated by Intel Research and Princeton University. Gibbons et al. (Gibbons P B, Karp B, Ke Y, Nath S and Sehan S 2003) 1

envisage a worldwide sensor web in which many users can query, as a single unit, vast quantities of data from thousands or even millions of planetary sensors. One ambition of Intelligent CCTV is to help prevent terrorism rather than just recording the events leading up to an attack and a key technology is reliable face recognition. We have just begun a major project to field trial these technologies in transport centres accessed by large numbers of people on a daily basis. Now we will focus on some of the crucial technologies underpinning such intelligent CCTV services automatically detecting, normalizing and recognizing faces in image and video databases. Face Detection Face detection is a necessary first-step in a face recognition system to locate a face or faces from cluttered backgrounds. Many techniques for face detection have been developed including: feature-based approaches (Govindaraju V 1996, McKenna S, Gong S and Liddell H 1995), template matching (Jeng S H, Liao H Y M, Liu Y T and Chen M Y 1998, Gunn S R and Nixon M S 1994) and image-based approaches (Turk M and Pentland A 1991, Sung K K and Poggio T 1998). In 2001, Viola and Jones (Viola P and Jones M 2001) proposed an image-based face detection system which can achieve remarkably good performance in real-time. Our face detection module is based on the Viola-Jones approach using our own training sets. The cascade face detector uses a sequence of binary classifiers which discard non-face regions and only send likely face candidates to the next level of classifier. Simple classifiers can be constructed which reject the majority of non-face regions at the very early stage of detection, before the use of more complex classifiers with higher discriminative capability. In this way, the more discriminating but more complex classifiers concentrate their processing time on face-like regions. Haar-like wavelet features (Mallat S G 1989) extracted from the image sub-windows are an image representation method which characterises the texture similarities between different regions by computing the sum of pixel values in different regions. It can be computed rapidly from the integral image (Crow F 1984) which maintains a running sum of pixel values in two dimensions. The Adaboost learning algorithm selects the best rectangle features and linearly combines these features into a classifier. Adaboost is a boosting learning algorithm, which can fuse many weak classifiers into a single more precise classifier. The main idea of Adaboost is as follows. At the beginning of training, all training examples are assigned equal weight. During the process of boosting, the weak classifier with the lowest classification error is selected and the weights of the samples which are wrongly classified by the weak classifier increase. The final classifier is a linear combination of the weak classifiers from all rounds, where classifiers with lower classification error have a higher weighting. Figure1 shows examples of face detection using our implementation. 2

Pose Normalization Fig.1: Detection results on several photographs. Pose normalization refers to compensating for the pitch, roll, and yaw motions of the head to allow for non-frontal viewing conditions. It acts as a bridge between the face detection and face recognition modules. In-Plane Rotation A Viola-Jones eye detector is used to locate the eyes, and the face image is rotated to a vertical frontal face image by computing the angle between two eyes and the horizontal baseline (Fig. 2). Before After Fig. 2: A face sample before and after in-plane image rotation 3

Out-of-Plane Rotation Active Appearance Models are a powerful tool to describe deformable object images. Given a collection of training images for a certain object class where the feature points have been manually marked, a shape and texture can be represented by applying PCA to the sample as: x = x + P c g = g + P c where x is the mean shape, g is the mean texture and P s, Pg are matrices describing the shape and texture variations. The parameter c is a vector of model parameters controlling both shape and texture of the model. We assume that this model parameters c is related to the viewing angle, θ, approximately by: c = c0 + c c cos( θ ) + cs sin( θ ) where c 0, cc and cs are vectors which are learned from the training data. Given a new face image with model parameters c, we can estimate orientation and then synthesize new views. Let cres be the residual vector not explained by the rotation model, c res = c ( c 0 + cs cos( θ ) + cs sin( θ )) To calculate the model parameters c (α ) at a new angle, α, we simply use the equation: c ( α ) = c0 + cc cos( α) + cs sin( α) + cres Here α is 0, so the equation will be: c ( 0) = c 0 + c c + c res Then frontal face images can be synthesized (Fig. 3). s g Fig.3: Some synthesized frontal face images from the Feret Face Database. Face Recognition Recent research on face recognition has been focused on diminishing the impact of nuisance factors on face recognition. Many approaches have been proposed for illumination invariant recognition (Yilmaz A and Gokmen M 2000, Gao Y S and Leung M K H 2002) and expression invariant recognition (Beymer D and Poggio T 1995, Black M J, Fleet D J and Yacoob Y 2000). 4

But these methods suffer from the need to have large numbers of example images for training, Chen and Lovell (Chen S K and Lovell B 2004) developed Adaptive Principal Component Analysis (APCA) to compensate for illumination and facial expression variations. It inherits merits from both PCA (eigenfaces) and FLD (Fisher Linear Discriminant or fisherfaces) by warping the face subspace according to the within- and between-class covariance. The method which we may call eigen-fisherfaces consists of four steps: Subspace Projection: Apply PCA to project face images into the face subspace to generate the m-dimensional feature vectors Whitening Transformation: The subspace is whitened according to the eigenvalues of the subspace with a whitening power p 2 p 2 p 2 p cov = diag{ λ, λ, K λ } (1) 1 Filtering the Eigenfaces: Eigen-features are weighted according to the identification-tovariation value ITV with a filtering power q. q q q γ = diag ITV, ITV, K ITV } (2) { 1 2 Optimizing the cost function: Minimize the cost function according to the combination of error rate and the ratio of between-class distance and within-class distance:. M K d (3) jj, k0 OPT =, m d jm, k < d jj k, m [1L m] 0 0 j= 1 k= 1 m, d jm, k0 More details about APCA can be found in [4]. Experiments show this technique performs well under changes in lighting conditions and facial expression (Fig. 4). 2 m m Fig.4: Contours of 95% recognition performance for the original PCA and the proposed APCA method against lighting elevation and azimuth. Tracking A recognition confidence output is calculated from the angle θ which measures how closely the input face image subspace matches the recognized face subspace. Only the face sub-window with high confidence will be registered for tracking (Fig.5). 5

Fig. 5: Some selected frames (from left to right, up to down) showing different recognition confidence levels using different colour rectangles. Conclusion In this paper we describe some technologies underpinning the pattern recognition engine of a system for locating persons using Intelligent CCTV. A fully automated system must be highly robust to nuisance problems such as lighting, expression change, pose, and camera variation. Our system has three components: 1) a Viola-Jones face detection module based on cascaded simple binary features to rapidly detect and locate multiple faces from the input still image or video sequences, 2) view-based Active Appearance Models to estimate facial pose and compensate for extreme pose angles, and 3) Adaptive Principal Component Analysis, or eigen-fisher faces, to recognize the faces since the method is robust to poor lighting and extreme facial expressions. We plan to report on the results of the public field trials at a later date. Acknowlegement This project is supported by a grant from the Australian Government Department of the Prime Minister and Cabinet. National ICT Australia is funded by the Australian Government's Backing Australia's Ability initiative and the Queensland Government, in part through the Australian Research Council. 6

References Beymer D and Poggio T (1995): Face Recognition from One Example View. Proc. Int'l Conf. of Comp. Vision, 500-507 Black M J, Fleet D J and Yacoob Y (2000): Robustly estimating Changes in Image Appearance. In: Computer Vision and Image Understanding, 78(1), 8-31. Chen S K and Lovell B (2004): Illumination and Expression Invariant Face Recognition with One Sample Image per Class. In: 17 th International Conference on Pattern Recognition (ICPR 04) Volume 1 pp. 300-303 Cootes T F, Walker L and Taylor C J (2000): View-Based Active Appearance Models. 4 th International Conference on Automatic Face and Gesture Recognition, pp: 227-232, March. Crow F (1984): Summed-are tables for texture mapping. Proc of SIGGRAPH, Volume 18(3), pages 207-212. Feret (2005): http://www.itl.nist.gov/iad/humanid/feret/ [last visited 23-Nov-2005] Gao Y S and Leung M K H (2002): Face Recognition Using Line Edge Map. In IEEE PAMI. 24(6), June, 764-779 Gibbons P B, Karp B, Ke Y, Nath S and Sehan S (2003), IrisNet: An Architecture for a Worldwide Sensor Web. In: Pervasive Computing, 2(4), 22-23, Oct Dec Govindaraju V (1996) Locating human faces in photographs. In: International Journal of Computer Vision, Volume 19, Issue 2, August, pp: 129-146. Gunn S R and Nixon M S (1994): A dual active contour for head boundary extraction. In: IEE Colloquium on Image Processing for Biometric Measurement, pp: 6/1 6/4, 20 Apr. Horward J (2005): http://smh.com.au/news/national/howard-backs-more-securitycameras/2005/07/24/1122143730105.html [last visited 23-Nov-2005] Jeng S H, Liao H Y M, Liu Y T and Chen M Y (1998): An efficient approach for facial feature detection using geometrical face model. Proc: 13 th International Conference on Pattern Recognition, Volume 3, pp: 426-430, 25-29 Aug Mallat S G (1989): A theory for multi-resolution signal decomposition: The wavelet representation. In: IEEE Transactions on Pattern Analysis and Machine Intelligence, 11(7): 674-693. McKenna S, Gong S and Liddell H (1995) Real-time tracking for an integrated face recognition system. In 2 nd Workshop on Parallel Modelling of Neural Operators, Faro, Portugal, Nov Sung K K and Poggio T (1998): Example-based learning for view-based human face detection. ITEE Transaction on Pattern Analysis and Machine Intelligence. Turk M and Pentland A (1991): Face recognition using eigenfaces. Proc. Computer Vision and Pattern Recognition, pp: 586-591. Yilmaz A and Gokmen M (2000): Eigenhill vs. eigenface and eigenedge. In Procs of International Conference Pattern Recognition, Barcelona, Spain, 827-830 BIOGRAPHY: Ting Shan: PhD student of University of Queensland, his research interest is in face detection, multi-view face recognition and real-time face recognition system. Brian Lovell: Research Leader of the SAFE Sensors group in National ICT Australia and Professor in the School of ITEE, The University of Queensland. His main interest is the analysis of video streams to recognize human activities. Shaokang Chen: Abbas Bigdeli: 7