Pedestrian Detection with Improved LBP and Hog Algorithm

Similar documents
The Population Density of Early Warning System Based On Video Image

An Adaptive Threshold LBP Algorithm for Face Recognition

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

An Object Detection System using Image Reconstruction with PCA

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

Computer Science Faculty, Bandar Lampung University, Bandar Lampung, Indonesia

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach

Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation

Journal of Chemical and Pharmaceutical Research, 2015, 7(3): Research Article

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

Implementation of a Face Recognition System for Interactive TV Control System

Histograms of Oriented Gradients for Human Detection p. 1/1

Road-Sign Detection and Recognition Based on Support Vector Machines. Maldonado-Bascon et al. et al. Presented by Dara Nyknahad ECG 789

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Discriminative classifiers for image recognition

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Object Tracking using HOG and SVM

Component-based Face Recognition with 3D Morphable Models

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image

Research of Image Registration Algorithm By corner s LTS Hausdorff Distance

A ROBUST DISCRIMINANT CLASSIFIER TO MAKE MATERIAL CLASSIFICATION MORE EFFICIENT

Pupil Localization Algorithm based on Hough Transform and Harris Corner Detection

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Face Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Automatic Shadow Removal by Illuminance in HSV Color Space

An Angle Estimation to Landmarks for Autonomous Satellite Navigation

Real Time Motion Detection Using Background Subtraction Method and Frame Difference

CS4670: Computer Vision

Color Local Texture Features Based Face Recognition

Non-rigid body Object Tracking using Fuzzy Neural System based on Multiple ROIs and Adaptive Motion Frame Method

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Verification: is that a lamp? What do we mean by recognition? Recognition. Recognition

What do we mean by recognition?

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Human detection solution for a retail store environment

A Study of Medical Image Analysis System

Metric for the Fusion of Synthetic and Real Imagery from Multimodal Sensors

A Novel Extreme Point Selection Algorithm in SIFT

Object Detection Design challenges

Categorization by Learning and Combining Object Parts

Classification of objects from Video Data (Group 30)

A Novel Image Semantic Understanding and Feature Extraction Algorithm. and Wenzhun Huang

Study on Image Position Algorithm of the PCB Detection

VEHICLE RECOGNITION USING VIBE AND SVM

HOG-based Pedestriant Detector Training

PEOPLE IN SEATS COUNTING VIA SEAT DETECTION FOR MEETING SURVEILLANCE

Category vs. instance recognition

Histogram of Oriented Gradients for Human Detection

Face and Nose Detection in Digital Images using Local Binary Patterns

Auto-Digitizer for Fast Graph-to-Data Conversion

An adaptive container code character segmentation algorithm Yajie Zhu1, a, Chenglong Liang2, b

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

International Journal of Modern Engineering and Research Technology

Research of Traffic Flow Based on SVM Method. Deng-hong YIN, Jian WANG and Bo LI *

Robotics Programming Laboratory

Lane Markers Detection based on Consecutive Threshold Segmentation

Finger Vein Biometric Approach for Personal Identification Using IRT Feature and Gabor Filter Implementation

Research on Robust Local Feature Extraction Method for Human Detection

Announcements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14

A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods

Local features: detection and description. Local invariant features

Local Image Features

International Journal of Advance Research in Engineering, Science & Technology

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

Progress Report of Final Year Project

Part-based Face Recognition Using Near Infrared Images

Human Motion Detection and Tracking for Video Surveillance

Co-occurrence Histograms of Oriented Gradients for Pedestrian Detection

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion

Real-Time Human Detection using Relational Depth Similarity Features

A New Strategy of Pedestrian Detection Based on Pseudo- Wavelet Transform and SVM

Fast Image Matching Using Multi-level Texture Descriptor

Tracking. Hao Guan( 管皓 ) School of Computer Science Fudan University


Object Category Detection: Sliding Windows

A Two-Stage Template Approach to Person Detection in Thermal Imagery

Temperature Calculation of Pellet Rotary Kiln Based on Texture

Speeding up the Detection of Line Drawings Using a Hash Table

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Liver Image Mosaicing System Based on Scale Invariant Feature Transform and Point Set Matching Method

Local Patch Descriptors

Feature Descriptors. CS 510 Lecture #21 April 29 th, 2013

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

FACIAL RECOGNITION BASED ON THE LOCAL BINARY PATTERNS MECHANISM

Face Detection using Hierarchical SVM

AN EXAMINING FACE RECOGNITION BY LOCAL DIRECTIONAL NUMBER PATTERN (Image Processing)

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

EFFICIENT REPRESENTATION OF LIGHTING PATTERNS FOR IMAGE-BASED RELIGHTING

Measurement of Pedestrian Groups Using Subtraction Stereo

Research on QR Code Image Pre-processing Algorithm under Complex Background

A New Algorithm for Shape Detection

Multi-Scale Kernel Operators for Reflection and Rotation Symmetry: Further Achievements

Pedestrian Detection using Infrared images and Histograms of Oriented Gradients

Multi-feature face liveness detection method combining motion information

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Transcription:

Open Access Library Journal 2018, Volume 5, e4573 ISSN Online: 2333-9721 ISSN Print: 2333-9705 Pedestrian Detection with Improved LBP and Hog Algorithm Wei Zhou, Suyun Luo Automotive Engineering College, Shanghai University of Engineering Science, Shanghai, China How to cite this paper: Zhou, W. and Luo, S.Y. (2018) Pedestrian Detection with Improved LBP and Hog Algorithm. Open Access Library Journal, 5: e4573. https://doi.org/10.4236/oalib.1104573 Received: April 4, 2018 Accepted: April 25, 2018 Published: April 28, 2018 Copyright 2018 by authors and Open Access Library Inc. This work is licensed under the Creative Commons Attribution International License (CC BY 4.0). http://creativecommons.org/licenses/by/4.0/ Open Access Abstract This article aims to improve the HOG + SVM pedestrian detection method proposed by previous researchers. The speed of HOG + SVM to detect pedestrians is relatively slow, and the detection accuracy is not very good. This paper proposes a PCA (principal component analysis) dimension reduction for HOG and also interpolates it. The article combines the dimensions of individual HOG features and improves their accuracy, and fuses them with improved LBP features. The features of the fusion of HOG features and LBP features can both express pedestrian profile information and obtain pedestrian texture information. This can improve the speed of pedestrian detection and improve the accuracy of detection, which is beneficial to reduce false detection and missed detection. Although some researchers have combined the two features of HOG and LBP, after simple fusion of these two features, the experimental results show that the detection effect is not much improved. This article is aimed at different formats of video detection material, an application program written on the MFC platform, making pedestrian detection of the material quickly verified, which is conducive to pedestrian detection results data analysis and recording. Subject Areas Computer Engineering Keywords HOG Feature, Improved LBP Feature, MFC 1. Pedestrian Detection of Specific Programs As a future trend of smart driving, its complex structure and what can be explored is very much. The article mainly focuses on the pedestrian detection part. DOI: 10.4236/oalib.1104573 Apr. 28, 2018 1 Open Access Library Journal

The whole smart car pedestrian detection can be divided into the following several parts: information acquisition part, comprehensive feature extraction, classification training, implementation detection as shown in Figure 1. 2. Establish an Improved LBP Model Figure 2 is a basic LBP feature schematic, which is based on a pixel-by-pixel review of the image. With this pixel as the center, and with this pixel s size value set as a threshold, then its surrounding 3*3 range of pixels is compared (binarized), and its binarization result is specified Arrange regularly to get a set of binary values and use this binary value as the output point for this point. Its definition is shown in (2-1): g In the formula, S( x) p ( p = 0,1, p 1) ( ) LBP S g g (2-1) p 1, 2 p PR = p= 0 p c 1, if x T =, g c is the gray value of the center pixel, 0, otherwise is the grayscale value of p surrounding pixels, and T is the threshold value. For example, for the center in the range of 3*3 in Figure 2, its gray value is 68, with 68 as the threshold. He binarizes his eight fields and sets the value of the binarization into a new value, i.e., 10001011, in a clockwise direction from the top left (the order of the specific directions can be self-determined, as long as the laws follow a certain rule). That is, decimal 139 and 139 as output points. After the overall scan is over, there will be an LBP output image. The histogram of this output image is the LBP histogram, which is often used as the recognition feature of the later work and is therefore also called the LBP feature. Its code implementation in open cv and test results are shown in Figure 3. LBP Rotation Invariant Mode The basic LBP has good robustness to illumination (i.e., grayscale invariance), but it does not have rotation invariance. Therefore, researchers have extended the above basis and proposed LBP features with rotation invariance. The idea is to make the LBP feature in the circular neighborhood continue to rotate, and then get different LBP eigenvalues, find the smallest LBP value from the rotated LBP eigenvalues, and use this value as the characteristic value of the last center pixel. The specific process is shown in Figure 4. Information Collection ImprovedLBP feature extraction HOGfeatureextrac tion Combine two features to get a reliable feature model and classify them by SVM Realize detection Figure 1. Pedestrian inspection system. DOI: 10.4236/oalib.1104573 2 Open Access Library Journal

Figure 2. Schematic diagram of hog feature extraction. Figure 3. Schematic diagram of the extraction of hog blocks. Figure 4. Rotate invariant lbp describe subflow diagram. Its definition is as follows (2-2). { ( PR ) LBP = min ROR LBP, i i = 0,1,, Q 1 (2-2) ri QR,, DOI: 10.4236/oalib.1104573 3 Open Access Library Journal

Among them, ROR(x,i) means to cycle x to the right by moving the i bit. Here, there is no provision for which point to start from. It can be seen from Figure 5 that it is clockwise rotated. The rotation-invariant LBP descriptor not only has the robustness of the illumination of the basic LBP descriptor, but also has the advantages of rotation invariance and fewer model types, making the LBP texture more simplified. 3. SVM Classifier The support vector machine (svm) was proposed by Vapnik and Core [1] based on statistical VC dimension theory and structural risk minimization in 1995. Its advantage lies in its ability to solve small sample, nonlinear and high-dimensional pattern recognition and get good results. The SVM can find a good balance between the learning accuracy of a given training sample (complexity of the model) and the ability to identify its sample (learning ability) as accurately as possible based on a limited sample content. Get the best practicality. 1) Linear separable SVM The initial development of SVM begins with a linearly separable optimal classification surface. Divide the positive and negative samples in the sample into two parts accurately, and also maximize the separation interval. The SVM strives to obtain a hyperplane that keeps the points in the sample as far away from the face as possible, that is, the area where the largest margin formed by the faces where the positive and negative samples are far apart from each other. Points H1, H2 on the separation plane parallel to the hyperplane and passing the positive and negative samples, such a point (training sample). We call him the support vector. Figure 5 shows the classification line in the case of linear separability: 2) Linear Inseparable SVM For linearly inseparable problems we analyze and deal with the following examples. As shown in Figure 6 below: Define the points in the blue part Figure 5. Classification of linear separable cases. DOI: 10.4236/oalib.1104573 4 Open Access Library Journal

Figure 6. Positive and negative sample set. between points A and B on the number axis as positive samples, and the points in the yellow parts of both sides as negative samples. A linear function (straight line) in two-dimensional space cannot find a straight line to separate positive and negative samples. 2 But we can find a curve g( x) = a0 + ax 1 + ax 2 to separate positive and negative samples, as shown in Figure 7 below. Obviously this curve can separate positive and negative samples, but he is not a linear function and is a general quadratic function. In order to make it a linear function, it is rebuilt to define a variable y and b equivalence as (2-3): y1 1 c1 a0 y = y 2 x b c 2 a = = = 1 2 y 3 x c 3 a 2 (2-3) Then g(x) can be equivalent to f(y) = <b,y>a, i.e., g(x) = f(y) = c_1 y_1 + c_2 y_2 + c_3 y_3, it can be seen that g(x) becomes The linear function, its difference with the quadratic function is that the dimension becomes higher, and here we get a method that encounters a linearly inseparable sample, trying to increase the dimension of the function, so that it becomes linearly separable. The above is the principle knowledge used in this article. 4. Pedestrian Detection with Fusion IHOGP-LBP Feature Multiple Training The previous section mainly studied the improved method of HOG algorithm. Through the simplified three-line interpolation and PCA dimension reduction [2] of HOG, the calculation speed of HOG is improved and the accuracy of its detection is also improved. Its effect can be reflected in the following experiments. HOG can describe the edges and gradients of objects very well during feature extraction but lacks description of texture information for some pedestrians. Here we will fuse LBP descriptors, combine pedestrian texture information, and better express pedestrian information through the integration of multi-feature integration graphs, making the detection effect more perfect [3]. Firstly prepare the positive and negative samples, then extract the IHOG features of the positive and negative samples and then reduce the dimension. After training, the IHOGP detector is obtained. Then the negative samples are detected by the detector and then the features of the hard example are extracted, and the IHOGP characteristics before the fusion are obtained. Continue training and eventually get the appropriate detector. The right frame shows the process of extracting LBP. Its process is the same as the model training process on the left [4]. The final result is the DOI: 10.4236/oalib.1104573 5 Open Access Library Journal

Figure 7. Positive and negative sample set classification. LBP detection operator. Pedestrian detection of the main line of thought is the middle of the framework of the order, the left and right sides of the middle of the process. In the specific algorithm, the detection scheme for the fusion IHOGP-LBP feature multiple training is shown in Figure 8. Pedestrian detection of the main line of thought is the middle of the framework of the order, the left and right sides of the middle of the process. It can be seen from the above figure that after the picture is input, the IHOGP feature is extracted first, and then input into the SVM classifier to train to obtain a suspicious pedestrian area, but it is not sure whether it is a pedestrian. The LBP descriptors are then extracted and classified to obtain suspicious positive samples. Finally, the two characteristics are combined to train, and a more reliable pedestrian detector is obtained. Through the last pedestrian detector, the pedestrian in the image is detected, which can accurately detect the location of a person. Figure 9 is a schematic diagram of IHOGP-LBP feature fusion. The above figure is the process of feature fusion and can be expressed by Equation (3-1). F _ IHOGP LBP I = F _ IHOGP I + F _ LBP I (3-1) ( )( ) ( ) ( ) Where I is represented as a sample, F _ ( ) IHOGP feature of the sample, and F _ ( ) IHOGP I is represented as the LBP I is the lbp feature of the sample. The samples are first extracted from the IHOGP features, then the LBP features are extracted, and finally they are combined in parallel to form a fusion feature. From the above figure, we can see that the feature histogram of fusion has become more prominent, which shows that the features of the pedestrian after fusion are more obvious, making the probability of detecting pedestrians even higher. 5. Realization of Video Pedestrian Detection System in Driving Environment In the above method for pedestrian detection in video, the source code and video file format need to be modified for each scene detection. In practice, it seems to be tedious. This article will develop a simple application program that will make video in various formats quickly available and detect pedestrians. The application development environment for this article is windows 7, 64-bit operating system, memory 4G, and the processor is Intel(R) core(tm) i5. The DOI: 10.4236/oalib.1104573 6 Open Access Library Journal

W. Zhou, S. Y. Luo Figure 8. Improved algorithm for pedestrian detection. developed software is MFC in visual studio 2010. MFC is a packaged windows API library provided by Microsoft Corporation [5]. Its biggest advantage is to provide the framework of the application program. This makes the program developers write their own programs in the existing framework, lost the tedious programming of the underlying program, but also makes programmers quickly familiar with the framework. MFC provides a large number of classes to facilitate different project development. Everything is double-sided, MFC because of the package of a large number of C + + classes, the existence of the package, making a lot of things disappeared. This makes it easier for people who are getting started to understand basic knowledge. This article completes a pedestrian detection application program under the MFC framework. Its main function is to open the video material in the file, and then detect the pedestrians in the material video and pass the window box to the pedestrian [6]. The interface designed in this article is easy to operate. The final rendering is shown in Figure 10. The main class used in this article is CIVSDlg, which contains the video playback dialog [7]. The functions that define many video operations and the corresponding variables are shown in Figure 11. In the open control, in addition to opening the video in the file, a message processing program needs to be inserted. The message processing is to enable it to run an image processing program, namely the above-mentioned detector in DOI: 10.4236/oalib.1104573 7 Open Access Library Journal

Figure 9. Fusion of IHOGP-lBP features. Figure 10. Pedestrian detection APP interface. the text, so that it can detect edestrians in the video. This article inserts a pedestrian detection handler and runs the test. Get the results shown in Figure 12. 6. Pedestrian Detection Results The experiment in this paper compares the detection effectiveness of the two detection methods, hog + svm and ihogp-pca + svm, in different scenarios. This article selects various scenarios, and detects the effects of two detection methods in different scenarios. As shown in Figure 13. DOI: 10.4236/oalib.1104573 8 Open Access Library Journal

Figure 11. CIVSDlg class diagram. Figure 12. Application use map. Figure 13. Hog and method detection time comparison. From the figure above, we can see that in the feature extraction time, with the increase of resolution, the hog extraction time becomes longer and longer, and the improved feature extraction performs better in this aspect without much time extension. DOI: 10.4236/oalib.1104573 9 Open Access Library Journal

References [1] Oren, M., Papageorgiou, C., Sinha, P., et al. (1997) Pedestrian Detection Using Wavelet Template. CVPR. [2] Xu, D., Li, X. and Liu, Z. (2005) Recognition Letteretal. Cast Shadow Detection in Video Segmentation. Pattern, 26, 91-99. [3] Shashua, A., Gdalyahu, Y. and Hayun, G. (2004) Pedestrian Detection for Driving Assistance Systems: Single-Frame Classification and System Level Performance. Proceedings of IEEE Intelligent Vehicles Symposium, 1-6. [4] Sun, H., Hua, C.-Y. and Luo, Y.-P. (2004) A Multi-Stage Classifier Based Algorithm of Pedestrian Detection in Night with a Near Infrared Camera in a Moving Car. Proceedings of 3rd IEEE International Conference on Image and Graphics, USA, 120-123. [5] Mikolajczyk, K., Schmid, C., Zisserman, A., et al. (2004) Human Detection Based on a Probablilistic Assembly of Robust Part Detectors. ECCV, 69-82. [6] Lipton, A., Kanade, T., Fujiyoshi, H., et al. (2000) A System for Video Surveillance and Monitoring. Carnegie Mellon University, the Robotics Institute, Pittsburg. [7] Tons, M., Doerfler, R., Meinecke, M., et al. (2004) Radar Sensors and Sensor Platform Used for Pedestrian Protection in the EC-Funded Project SAVE-U. Proceedings of IEEE Intelligent Vehicles Symposium, USA, 813-818. DOI: 10.4236/oalib.1104573 10 Open Access Library Journal