Face Detection Using Look-Up Table Based Gentle AdaBoost

Similar documents
Subject-Oriented Image Classification based on Face Detection and Recognition

Face Alignment Under Various Poses and Expressions

Object Category Detection: Sliding Windows

Study of Viola-Jones Real Time Face Detector

Recognition problems. Face Recognition and Detection. Readings. What is recognition?

Skin and Face Detection

Postprint.

Color Model Based Real-Time Face Detection with AdaBoost in Color Image

Face and Nose Detection in Digital Images using Local Binary Patterns

FACE DETECTION BY HAAR CASCADE CLASSIFIER WITH SIMPLE AND COMPLEX BACKGROUNDS IMAGES USING OPENCV IMPLEMENTATION

Object Category Detection: Sliding Windows

Object Detection Design challenges

Detection of a Single Hand Shape in the Foreground of Still Images

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction

Real-time Computation of Haar-like features at generic angles for detection algorithms

Haar Wavelets and Edge Orientation Histograms for On Board Pedestrian Detection

Window based detectors

Efficient and Fast Multi-View Face Detection Based on Feature Transformation

Out-of-Plane Rotated Object Detection using Patch Feature based Classifier

Face Tracking in Video

Detecting and Reading Text in Natural Scenes

Face detection, validation and tracking. Océane Esposito, Grazina Laurinaviciute, Alexandre Majetniak

A robust method for automatic player detection in sport videos

Face detection and recognition. Detection Recognition Sally

Learning to Detect Faces. A Large-Scale Application of Machine Learning

Learning a Rare Event Detection Cascade by Direct Feature Selection

Generic Object Class Detection using Feature Maps

Assessment of Building Classifiers for Face Detection


Boosting Object Detection Performance in Crowded Surveillance Videos

Automatic Initialization of the TLD Object Tracker: Milestone Update

Object recognition (part 1)

Viola Jones Face Detection. Shahid Nabi Hiader Raiz Muhammad Murtaz

Fast and Robust Classification using Asymmetric AdaBoost and a Detector Cascade

2 Cascade detection and tracking

Parameter Sensitive Detectors

People detection in complex scene using a cascade of Boosted classifiers based on Haar-like-features

Classifier Case Study: Viola-Jones Face Detector

Progress Report of Final Year Project

Face Detection and Alignment. Prof. Xin Yang HUST

Texture Features in Facial Image Analysis

Object and Class Recognition I:

Detecting Faces in Images. Detecting Faces in Images. Finding faces in an image. Finding faces in an image. Finding faces in an image

Detecting Pedestrians Using Patterns of Motion and Appearance

Face tracking. (In the context of Saya, the android secretary) Anton Podolsky and Valery Frolov

Detecting Pedestrians Using Patterns of Motion and Appearance

Fast Learning for Statistical Face Detection

A Survey of Various Face Detection Methods

Short Paper Boosting Sex Identification Performance

Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation

Gender Classification Technique Based on Facial Features using Neural Network

Fusion-Based Person Detection System

Image Cues Fusion for Object Tracking Based on Particle Filter

Detecting People in Images: An Edge Density Approach

AN HARDWARE ALGORITHM FOR REAL TIME IMAGE IDENTIFICATION 1

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

ii(i,j) = k i,l j efficiently computed in a single image pass using recurrences

Computer Vision Group Prof. Daniel Cremers. 8. Boosting and Bagging

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki

Face Detection on OpenCV using Raspberry Pi

Computer Vision Group Prof. Daniel Cremers. 6. Boosting

Capturing People in Surveillance Video

Class Sep Instructor: Bhiksha Raj

Learning Object Detection from a Small Number of Examples: the Importance of Good Features.

Improving Part based Object Detection by Unsupervised, Online Boosting

Machine Learning for Signal Processing Detecting faces (& other objects) in images

Active learning for visual object recognition

Detection of Multiple, Partially Occluded Humans in a Single Image by Bayesian Combination of Edgelet Part Detectors

Categorization by Learning and Combining Object Parts

Object detection as supervised classification

Recap Image Classification with Bags of Local Features

Disparity Search Range Estimation: Enforcing Temporal Consistency

Object detection. Asbjørn Berge. INF 5300 Advanced Topic: Video Content Analysis. Technology for a better society

Human Detection Based on a Probabilistic Assembly of Robust Part Detectors

Detecting Pedestrians Using Patterns of Motion and Appearance

Face detection. Bill Freeman, MIT April 5, 2005

Multi-angle Face Detection Based on DP-Adaboost

Discriminative Feature Co-occurrence Selection for Object Detection

SKIN COLOUR INFORMATION AND MORPHOLOGY BASED FACE DETECTION TECHNIQUE

Object detection using non-redundant local Binary Patterns

STA 4273H: Statistical Machine Learning

Introduction. How? Rapid Object Detection using a Boosted Cascade of Simple Features. Features. By Paul Viola & Michael Jones

Person Detection in Images using HoG + Gentleboost. Rahul Rajan June 1st July 15th CMU Q Robotics Lab

Traffic Sign Localization and Classification Methods: An Overview

Gradient Feature Selection for Online Boosting

Computer Vision

Visuelle Perzeption für Mensch- Maschine Schnittstellen

Robust Real-Time Face Detection Using Face Certainty Map

Face/Flesh Detection and Face Recognition

Component-based Face Recognition with 3D Morphable Models

Ensemble Methods, Decision Trees

Fast and Robust Face Finding via Local Context

Efficient Detector Adaptation for Object Detection in a Video

Fast Human Detection Using a Cascade of Histograms of Oriented Gradients

A Cascade of Feed-Forward Classifiers for Fast Pedestrian Detection

Part-based Face Recognition Using Near Infrared Images

Towards Practical Facial Feature Detection

Towards Practical Evaluation of Pedestrian Detectors

MediaTek Video Face Beautify

Transcription:

Face Detection Using Look-Up Table Based Gentle AdaBoost Cem Demirkır and Bülent Sankur Boğaziçi University, Electrical-Electronic Engineering Department, 885 Bebek, İstanbul {cemd,sankur}@boun.edu.tr http://busim.ee.boun.edu.tr Abstract. In this work, we propose a face detection method based on the Gentle AdaBoost algorithm which is used for construction of binary tree structured strong classifiers. Gentle AdaBoost algorithm update values are constructed by using the difference of the conditional class probabilities for the given value of Haar features proposed by []. By using this approach, a classifier which can model image classes that have high degree of in-class variations can be constructed and the number of required Haar features can be reduced. Introduction Classical object detection schemes necessitate complex and computationally heavy classifiers for face detection in gray level images. Since the face detector is applied at each location and scale it requires significant computational power. However most of the locations in the searched scene do not contain any face and in fact the odds of finding a face at any one location is very small. Thus most of the non-face blocks can be eliminated with very simple classifiers. To find an efficient alternative to this exhaustive search approach, a rejection based classification approach is used to eliminate rapidly non-face regions in an image. Basedonsucharejectionstrategy,ViolaandJones[] used cascaded non-face rejection stages with low face elimination rate. Their algorithm consisting of simple classifiers are based on easily computable Haar features yield good detection with low false alarm rate. The required number of Haar features using their approach depends on the target false alarm rate. Liehart [6] made performance comparison of boosting algorithms using haar feature based binary-output weak classifiers. Various extensions of this detector structure have been proposed in [2], [3] and [4]. For example Wu [4] has proposed a multi-view face detector using Real AdaBoost confidence-rated Look-Up-Table (LUT) classifiers to detect faces under rotation and pose variations. Our work is in the same line as the Viola- Jones scheme. The contribution consists in the use of a simple real valued Gentle AdaBoost (GAB) algorithm procedure to construct cascaded classifier structure. The GAB approach helps to reduce the required number of Haar features vis-avis the boosting approach that instead uses binary output weak classifiers. Using T. Kanade, A. Jain, and N.K. Ratha (Eds.): AVBPA 25, LNCS 3546, pp. 339 345, 25. c Springer-Verlag Berlin Heidelberg 25

34 Cem Demirkır and Bülent Sankur LUT based confidence values for each Haar feature the information wasted in the binary weak classifier DAB approach can be utilized in the cascaded classifier training. We also used classifier output propagation as proposed in [4] tofurther reduce the number of features. The rest of the paper is organized as follows: in Section 2 we define the GAB procedure. In Section 3, we define the GAB based Haar feature selection and strong classifier construction. In Section 4, we show we can the use of previous classifier output in the following classifier construction. In Section 5 we give experiments and results, and finally in Section 6 the conclusions. 2 GAB Algorithm Boosting is a classification methodology which applies sequentially reweighted versions of the input data to a classifier algorithm, and taking a weighted majority vote of sequence classifiers thereby produced. At each application of the classification algorithm to the reweighted input data, classification algorithm finds an additional classifier f m (x) at stage m. GAB algorithm is a modified version of the Real AdaBoost (RAB) algorithm and it is defined in Figure. The main difference between GAB and RAB is the way it uses the estimates of the weighted class probabilities to update the weak classifier functions, f m (x).in GAB the update is given by f m (x) = (y = x) (y = x), while in the RAB algorithm is given by half the log-ratio f m (x) = 2 [5]. Logratios can be numerically unstable, leading to very large update values, while the update in GAB lies in the range [-,]. This more conservative algorithm has classification performance similar to RAB algorithm, and outperforms it both especially when stability problems arise. log Pw(y= x) (y= x) GAB Algorithm. Start with weights w i =/N, i =,..., N, F (x) =. 2. Repeat for m =, 2,...,M. (a) Estimate f m (x) by weighted least squares fitting of y to x. (b) Update F (x) F (x)+f m (x) (c) Set w i w i exp [ y i.f m (x i)],i =, 2,...,N, and renormalize so that wi =. i 3. Output the classifier sign [F (x)] = sign [ M m= fm (x)]. Fig.. The GAB algorithm allows for the estimator f m (x) to range over real numbers 3 GAB Based Haar Feature Selection In the original algorithm the simple weak classifiers are built by simply comparing the Haar features to a threshold and thereby producing binary outputs. The feature and its threshold are selected to yield the minimum error at every stage.

Face Detection Using Look-Up Table Based Gentle AdaBoost 34 (a) Likelihood values of a typical chosen Haar feature by GAB algorithm for two object classes. The corresponding Haar feature is shown in the upper left of the figure Update Values of a Typical Haar Feature.8.6.4 Update Value, f m (x).2.2.4.6.8.4.3.2...2.3.4 (b) GAB update values f m (x) =(y = x) (y = x) Binary Output for a Simple Threshold Based Weak Classifier for Threshold Value.8.6 Weak Classifier Output.4.2.2.4.6.8.4.3.2...2.3.4 (c) Binary output of the classifier in the case a simple threshold based weak classifier is used Fig. 2. Comparison of GAB update scheme (Fig. a,b) to the simple binary output scheme (Fig. c)

342 Cem Demirkır and Bülent Sankur In contrast, in GAB based feature selection mechanism the weak classifiers are not forced to yield the binary outputs, instead they give the values of the update functions f m (x) at finite number of samples. The comparison of the update values for the binary case and GAB-based case is shown for a typical chosen Haar feature in Figure 2. We use the GAB training algorithm described in Figure to construct a strong stage classifier using confidence values f m (x) for each feature. Under equal prior probabilities for the object classes, the update f m (x) isgiven by f m (x) = (y = x) (y = x) = (x y =)P (y =) (x y = )P (y = ) (x y =)P (y =)+ (x y = )P (y = ) = (x y =) (x y = ) () (x y =)+ (x y = ) where (x y = ±) are the likelihood values computed by using histograms of feature values x for two different object hypothese. Histogram bins are updated by summing the sample weights of the training set. The subscript w denotes the likelihood values with respect to updated sample weights at each boosting round. A stage of Haar feature classifier construction using GAB. Start with weights w i =/2p and /2l where p and l are the number of positive and negatives class samples. 2. Repeat for m =, 2,...,M. (a) For each Haar feature j, f m (x) =(y = x) (y = x) usingonly the feature j values. (b) Choose the best feature confidence [ set of values f m (x) giving the minimum weighted error e m = E w (yi sign[f m(x i )])] for all feature j. (c) Update F (x) F (x)+f m (x) (d) Set w i w i exp [ y i.f m (x i)],i =, 2,...,N, and renormalize so that wi =. i 3. Output the classifier sign [F (x)] = sign [ M m= fm (x)]. Fig. 3. At each iteration AdaBoost finds a set of update values f m (x) which use the values of the feature corresponding to the minimum error GAB algorithm chooses the best Haar feature resulting with the minimum weighted error e m = E w [ (yi sign[f m(x i)])] from among all available features. The chosen update values are accumulated in the classifier output F (x) value. The output of the classifier, F m (x), is thresholded by T m such that the desired target false alarm rate, f, and detection rate, d, of the classifier is achieved. We also used the approach in [4], where the previous stage classifier output,f m (x), is inputted to the next stage classifier. According to this, in the training process,

Face Detection Using Look-Up Table Based Gentle AdaBoost 343 Image Image + F ( Image ) +... ( ) 2 F ( Image ) F 2 ( Image) First Stage Second Stage Fig. 4. Propagating the classifier output to the next stage classifier, the previous stage classifier outputs are used to compute the posterior class probabilities of the object classes, (x y =)and(x y = ), the update values computed from them is used as an update corresponding to the first feature of the classifier classifier output values for the training samples of the next stage is used to prepare the posteriori class probabilities of the object classes. Thereby the first update value f (x) value of the succeeding classifier is computed by using the histogram values resulting from the classifier outputs of the previous stage. In this sense each previous stage conveys its accumulated information to the next stage classifier. A cascaded classifier is produced by this kind of GAB procedure as illustrated in the Fig 4. 4 Experiments Figure 5 illustrates a chosen Haar feature, its corresponding posterior class probabilities as a function of feature values. The rightmost figure shows the classifier output as more and more classifiers are added, abscissa indicates the training sample index. Adding more features, samples of two classes can be separated from each other by simply thresholding the final classifier output F (x). We implemented GAB algorithm for near frontal face detection in cluttered scenes. For training we used about aligned face and 6 non-face images with the size of 24x24 at each stage training. Face images contain near frontal

344 Cem Demirkır and Bülent Sankur Posteriori Class Probabilities for the Given Haar Feature Value Classifier Output Values For Training Samples.9 (y= x) (y=- x).8.8.6 Posteriori Class Probability Value.7.6.5.4.3.2 F(x).4.2 -.2 -.4 -.6. -.8 -.5 -. -.5.5..5-2 3 4 5 6 7 8 Training Sample Index Posteriori Class Probabilities for the Given Haar Feature Value 2 Classifier Output Values For Training Samples.9 (y= x) (y=- x).5 Posteriori Class Probability Value.8.7.6.5.4.3.2 F(x).5 -.5 -. -.5 -.3 -.2 -...2.3-2 2 3 4 5 6 7 8 Training Sample Index Posteriori Class Probabilities for the Given Haar Feature Value 3 Classifier Output Values For Training Samples.9.8 (y= x) (y=- x) 2 Posteriori Class Probability Value.7.6.5.4.3.2 F(x) - -2. -.4 -.3 -.2 -...2.3.4-3 2 3 4 5 6 7 8 Training Sample Index Fig. 5. First column shows the Haar features selected by GAB, second column shows the posterior class probabilities of the object classes, (x y =)and(x y = ), for the corresponding Haar feature, the last column shows classifier output values F (x) values for the training samples after being updated with each f m (x i) faces which are subject to ±5 inplane rotations and in the range of ±3 outof-plane rotations. The number of points to sample the GAB update functions, f m (x i ), used in our work is 64 and the false alarm probability was chosen as 6. These values are stored for each selected feature. In test process, these stored Look-Up Table (LUT) values of the update functions are used to compute the confidence values of each computed feature values. On the same training set, we trained the cascaded structure by using the proposed LUT-based GAB procedure and threshold-based Discrete AdaBoost (DAB) approach. The number of total features produced by GAB based procedure is about 2 percent of the DAB based training case for the same overall target detection and false alarm rate. We tested two methods on the CMU test set containing 53 faces in the 3 images. The number of features and False Alarm (FA)/Detection Rates (DR) are given in Table for the two methods. As seen from Table, not only the LUT-based GAB method performance is higher than the DAB method and also requires much fewer feature as compare to the DAB method, but about one fifth of the number of features the detection performance is even better.

Face Detection Using Look-Up Table Based Gentle AdaBoost 345 Table. Performance and number of feature comparison for the methods. FA : False Alarm, DR : Detection Ratio Method Number of FAs/DR Number of total features LUT-based GAB 5/85.2% 53 DAB 5/8.% 2652 5 Conclusions In this work, we developed a simple and efficient LUT-based GAB training procedure using Haar like features for the near frontal face detection. We tested and compared two methods on the public common test set, CMU test set. This procedure necessitates significantly fewer features with respect to the DAB-based training case for the near frontal face detection problem. We plan to use this approach to implement a rotation invariant multi-view face detection system. References. Viola, P., Jones, M.: Rapid Object Detection Using A Boosted Cascade of Simple Features. IEEE Conference on Computer Vision and Pattern Recognition, (2). 2. Li, S. Z.: Statistical Learning of Multi View Face Detection. ECCV 22, Copenhagen, Denmark. 3. Viola, P., Jones, M.: Fast Multi View Face Detection. Technical Report TR23-96, July 23, Mitsubishi Electric Research Laboratories. 4. Wu, B., Ai, H., Huang, C.: Fast Rotation Invariant Multi-View Face Detection based on Real Adaboost. AFG 4, May 7-9, 24, Seoul, Korea. 5. Friedman, J., Hastie, T., Tibshirani, R.: Additive Logistic regression: a Statitical View of Boosting. Technical Report, Standford University, August 7, 998. 6. R. Liehart, A. Kuranov, V. Pisarevsky, Empirical Analysis of Detection Cascades of Boosted Classifiers for Rapid Object Detection, MRL Tech report, 22.