A Novel Image Semantic Understanding and Feature Extraction Algorithm. and Wenzhun Huang

Similar documents
Research on Multi-sensor Image Matching Algorithm Based on Improved Line Segments Feature

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Pedestrian Detection with Improved LBP and Hog Algorithm

An Angle Estimation to Landmarks for Autonomous Satellite Navigation

Application of Geometry Rectification to Deformed Characters Recognition Liqun Wang1, a * and Honghui Fan2

Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models

Object Tracking using HOG and SVM

The Population Density of Early Warning System Based On Video Image

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

Face Recognition Technology Based On Image Processing Chen Xin, Yajuan Li, Zhimin Tian

An efficient face recognition algorithm based on multi-kernel regularization learning

A Rapid Automatic Image Registration Method Based on Improved SIFT

Video annotation based on adaptive annular spatial partition scheme

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

Generic Face Alignment Using an Improved Active Shape Model

Algorithm research of 3D point cloud registration based on iterative closest point 1

Liver Image Mosaicing System Based on Scale Invariant Feature Transform and Point Set Matching Method

A Novel Extreme Point Selection Algorithm in SIFT

Chapter 7. Conclusions and Future Work

Sketchable Histograms of Oriented Gradients for Object Detection

An Introduction to Content Based Image Retrieval

Local Feature Detectors

Open Access Moving Target Tracking Algorithm Based on Improved Optical Flow Technology

Discriminative classifiers for image recognition

Implementation and Comparative Analysis of Rotation Invariance Techniques in Fingerprint Recognition

TEXTURE CLASSIFICATION METHODS: A REVIEW

Rotation Invariant Finger Vein Recognition *

Beyond Bags of Features

Combining Appearance and Topology for Wide

An adaptive container code character segmentation algorithm Yajie Zhu1, a, Chenglong Liang2, b

A paralleled algorithm based on multimedia retrieval

Improving Image Segmentation Quality Via Graph Theory

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

MPEG-7 Visual shape descriptors

Frequent Inner-Class Approach: A Semi-supervised Learning Technique for One-shot Learning

Feature Detectors and Descriptors: Corners, Lines, etc.

EE795: Computer Vision and Intelligent Systems

Announcements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Multiple-Choice Questionnaire Group C

Scale Invariant Feature Transform

Outline 7/2/201011/6/

Content Based Image Retrieval (CBIR) Using Segmentation Process

Pixel-Pair Features Selection for Vehicle Tracking

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

Combining Selective Search Segmentation and Random Forest for Image Classification

Multimodal Information Spaces for Content-based Image Retrieval

III. VERVIEW OF THE METHODS

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

A Comparison of SIFT, PCA-SIFT and SURF

Scale Invariant Feature Transform

Algorithm Optimization for the Edge Extraction of Thangka Images

Video Inter-frame Forgery Identification Based on Optical Flow Consistency

Application of partial differential equations in image processing. Xiaoke Cui 1, a *

Human Detection and Action Recognition. in Video Sequences

Improvement of SURF Feature Image Registration Algorithm Based on Cluster Analysis

Research of Image Registration Algorithm By corner s LTS Hausdorff Distance

2 Proposed Methodology

CAP 5415 Computer Vision Fall 2012

An algorithm of lips secondary positioning and feature extraction based on YCbCr color space SHEN Xian-geng 1, WU Wei 2

An Adaptive Threshold LBP Algorithm for Face Recognition

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Why study Computer Vision?

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

High Resolution Remote Sensing Image Classification based on SVM and FCM Qin LI a, Wenxing BAO b, Xing LI c, Bin LI d

AN ENHANCED ATTRIBUTE RERANKING DESIGN FOR WEB IMAGE SEARCH

Graph Matching Iris Image Blocks with Local Binary Pattern

The SIFT (Scale Invariant Feature

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Object Recognition Algorithms for Computer Vision System: A Survey

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

An Efficient Approach for Color Pattern Matching Using Image Mining

The design of medical image transfer function using multi-feature fusion and improved k-means clustering

A New Method in Shape Classification Using Stationary Transformed Wavelet Features and Invariant Moments

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

Contents I IMAGE FORMATION 1

ImgSeek: Capturing User s Intent For Internet Image Search

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Advanced Digital Image Forgery Detection by Using SIFT

Motion illusion, rotating snakes

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

Research on adaptive network theft Trojan detection model Ting Wu

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

EVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION

Sparse Models in Image Understanding And Computer Vision

Latest development in image feature representation and extraction

Preliminary Local Feature Selection by Support Vector Machine for Bag of Features

Image Retrieval System Based on Sketch

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features

Research on Evaluation Method of Video Stabilization

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

Su et al. Shape Descriptors - III

Image processing and features

Multi-focus Image Fusion Using Stationary Wavelet Transform (SWT) with Principal Component Analysis (PCA)

Naïve Bayes for text classification

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

Transcription:

A Novel Image Semantic Understanding and Feature Extraction Algorithm Xinxin Xie 1, a 1, b* and Wenzhun Huang 1 School of Information Engineering, Xijing University, Xi an 710123, China a 346148500@qq.com, b huangwenzhun@xijing.edu.cn *The corresponding author Keywords: Image semantic; Algorithm; Feature extraction; Semantic understanding Abstract. In this paper, we propose a novel image semantic understanding and the feature extraction algorithm. Image stitching algorithm based on feature is extracted from the image of basic certain characteristics as matching primitives, usually in the form of general dot, line, area, or some special structures, etc., rather than directly using the image gray level itself, and then to these features as a model for registration. Thus, because the image matching algorithm based on feature extraction and image just some of the aspects of characteristics, namely extraction pixel is less, and makes the image registration time is greatly reduced. Under this theoretical support, we proposed the image semantic understanding and the feature extraction algorithm with the experimental verification that proves the effectiveness of the method that is meaningful. Introduction In the theory of computer vision, image segmentation, feature extraction and basic target recognition constituted by the lower to the top three major tasks. Target recognition and feature extraction in the image segmentation as the foundation, the image segmentation result is good or bad will directly affect the subsequent feature extraction and target recognition [1-3]. In recent years, with the new challenge scene classification is put forward, the variety of the scene classification techniques emerged. In order to more clearly reflect the core development of the scene classification techniques that shows as much as possible the existing classification technology of the whole scene. The technical evolution process in the scene classification can be roughly divided into the three stages: from the original method, based on the underlying feature gradually transition to the method based on high-level features, the development to the present based on the characteristics of the learning. At each stage of different period also appeared a variety of different scene classification method, which based on high-level feature classification method and can be divided into the method based on scenario model and the method based on semantic scene. To be sure, this is just a scene classification technology evolution trend, does not represent a row behind the method of one stage must than the previous phase of the method is effective. For example, now new scene of underlying characteristics still scene classification is an important breakthrough, and based on the deep learning method is not necessarily in some simple are very effective on small data sets. Automatic image annotation is a very challenging work in the area of the image retrieval research, is the key to realize image semantic retrieval. If can realize automatic image annotation, the image retrieval problem can be converted into the mature text retrieval problem, the automatic image annotation in computer vision, machine learning, information retrieval and other aspects of content which has a strong research value [4]. According to the literature review, the state-of-the-art image semantic understanding algorithms can be summarized as the listed aspects. (1) The automatic image annotation algorithm based on the probability correlation model. Intuitively, if two images with high visual similarity, they mark the higher the probability of the similar keywords sequence. The characteristics of this approach is that without learning mechanism for each semantic keyword set up a corresponding said low-level visual features. (2) Based on the learning algorithm of basic automatic image annotation. Based on mapping algorithm is a kind of a semi-supervised learning algorithm, a known class mark training data and the 2016. The authors Published by Atlantis Press 0113

unknown class mark test data will be involved in the learning process of the algorithm. And compared to the traditional supervised learning and unsupervised learning, a semi-supervised learning can use more information in the learning phase, such as the distribution of data characteristics, it is suitable for according to the total amount is larger, a relatively small quantity of the labeled training data. (3) Based on the classification algorithm of the automatic image annotation. Existing labeling algorithm based on classification are simply from the angle of the sample learning more or learn more marks to describe and solve the problem of annotation. Although these methods have different characteristics in the specific expression, but their core idea is consistent, namely using the known annotation data to establish a model to describe the potential connection between text and image characteristics of the vocabulary or mapping relationship, and predict the unknown image annotation. Figure 1. The Image Semantic Understanding Database In this article, we propose a novel image semantic understanding and feature extraction algorithm. This paper presents a new image ontology annotations framework. We will image semantic concept and high-level abstractions can be defined as attributes, quadratic mark method was adopted to realize automatic tagging for image semantics. In the mark for the first time, then give the building area based on Bayesian network object classes and properties of the ontology concept, the concept of relationship and combining with the image of the ontology concepts, logical semantic relationship between for the secondary labeling method. In the later sections, we will discuss in detail. The Proposed Algorithm The Principles of the Image Semantic. Visual language model is a model of analysis of the image block space information. An image can be divided into many areas, each area may, in accordance with its visual characteristics expressed in a visual words. Not only is there the spatial dependence between these visual words, there are a total of frequency by using these features, between visual words can easily dig image semantics [5-6]. Image semantic mining research focus in the field of the image classification is based on image semantics, in which the image semantic mining of human cognitive behavior, its core is that from an existing in the training data to high-level semantic concepts and the relationship between low-level visual features of image and the automatic modeling, using the learned model for automatic mining unclassified image semantics. By estimating the conditional probability distribution of visual words can obtain image block interdependence between semantic relations as the follows. 2016. The authors Published by Atlantis Press 0114

p w w, C i i1 t F w i1 i t F w, w C i1 C t In image semantic mining, said image semantic dimension of feature space is generally high, at the same time there are a lot of synonymous visual words and visual word polysemy, the existence of the redundant information, not only lead to the distribution of the model based on the visual language of basic visual words calculation result is invalid, and cannot be used for image semantic mining well for automatic classification of image semantics. As the satisfactory tool for analyzing the corresponding issues, the SVM can be introduced for optimization. The SVM classification performance is affected by many of factors, of which the following two factors is the key. (1) Form of kernel function. Different kernel functions lead to the generalization performance of SVM is different, so how to choose according to the specific data of kernel function is directly related to the performance of the SVM classification. (2) The error penalty parameter C and kernel functions parameters. In determining the feature subspace of C value of small said the penalties for experience error is small, the complexity of the learning machine small and empirical risk value; C in infinity, all constraint conditions must be met, this means that training sample must be classified. Technology of application of the kernel function, the nonlinear problem of input space, through function mapped to high-dimensional feature space, linear discriminant function in high-dimensional space construction, commonly used kernel function has the following several categories. k x x xx linear (2), i i 2, i exp i k x x x x RBF (3) The Image Feature Extraction. Image matching technology is to realize the image fusion, image correction, image Mosaic, and one of the key steps, target recognition and tracking has been widely used in the field of image recognition and image reconstruction, etc. Image matching for different source image matching, is difficult in the infrared image and visible light image are the most common and often used in military field, and visible light and infrared image imaging mechanism and different shooting conditions, there are large difference between the edge of the image contour, even very fuzzy, noise is bigger, therefore, the research of the high precision, fast correct matching rate is high, anti-jamming strong matching algorithm has the very vital significance. (1) Figure 2. The Image Feature Extraction Procedure Demonstration In order to overcome the disadvantage of image matching method based on gray level, people put forward the method based on feature, based on the characteristics of the method first extracted from to match images contain important features of image feature points, with the similarity measure between image feature point matching that is commonly used in basic image matching features such as edges, outline, linear and angular point. Matching based on feature for image distortion, such as block has certain robustness, but its matching performance depends largely on the quality of feature extraction, and this kind of method of feature extraction is easily affected by noise [7]. 2016. The authors Published by Atlantis Press 0115

6th International Conference on Management, Education, Information and Control (MEICI 2016) x2 y 2 (4) G x, y, exp 2 2 2 2 Scale invariant feature transform is a method of feature detection based on the feature description. The SIFT algorithm is zoom, rotate and affine invariability, can resist some illumination changes and perspective transformation. At the same time, this characteristic also has the high distinguish ability, is advantageous to the subsequent matching. The algorithm firstly using the feature point detection method to extract the two corners of the image, and secondly by the Taylor series to eliminate pseudo angular point, in turn, determine the feature points, to determine the core feature points on each curve calculating the gradient vector and direction angle correction, then the feature points neighborhood part adopts nearby projection principle, and the gradient direction histogram to determine the main direction of the feature points, finally in polar constraint bidirectional matching method was designed on the basis of feature point matching as the formula 5. 1 2 D 1 D (5) x 2 x At this time make the other side of the pixel candidate feature points, and then repeat the above calculation, get the new subpixel location precision. Finally, by using the subpixel accuracy position replace all scale candidate feature points location [8]. The Image Semantic Understanding. Parallel regional technology is image processing and analysis theory of the technique of image analysis. In the study of image and application, people often only interested in some of the images. These parts are often referred to as the target or prospects, the rest of the known as the background as they are usually in a specific area, the unique quality of image. In order to identify and analyze the target, need to separate out the relevant area, on the basis of this is possible to increase the use of the basic target, such as feature extraction and measurement, image segmentation is refers to divide the image into the distinctive regional and extract the interested target technology and process, this feature can be gray, color, texture, etc., the target can correspond to a single area that also can correspond to multiple regions. x Figure 3. The Verification of the Image Scene Understanding The complex scene of the video monitoring system, the tracking technology is not perfect enough, such as multiple target tracking of the occlusion may make track interrupt each other, and complex behavior variety, path length difference is big, therefore, is not suitable for the complex behavior of trajectory modeling directly. The complex behavior by landmark tracks down into simple behavior we can use the same simple behavior training as the follows. O flow Ov1, Ov2,, OvT 2016. The authors Published by Atlantis Press (6) 0116

GIST features low computational complexity and easy to use, good at depicting the scene of the overall layout, but in the background of complex or the dense poor performance in the scene. SIFT features to translation, scaling, scale transform, brightness change, perspective transformation has certain stability, but not very well describes some of the overall layout of core complex scene. HOG features can be described on the edge of the shape and the target contour, but in the shape of the scene structure under the condition of that unstable failure. Characteristic to have with local features by describing the scene geometry information of scene, said the global but the characteristics of scaling and rotation invariance, and ignore the color information of the scene, so the complex scenario in some poor performances. Images of the underlying features in simple scene classification task is a kind of effective, says but as a result of this representation are usually with a few semantic information, making them under the condition of the scene is more complex expression is very limited, so how to understand high-level semantic became a scene of scene classification new breakthrough. In view of scene and changeable situation, it removes the limitation of the whole space structure, in depicting the image area into the local spatial structure characteristics, the second stage is the classifier and the categorizing scenario testing completely separated, the image area for classifier to further optimize the weight parameters make the classification more effective. But it also has its own limitations, so that the trained model to different data sets is versatile of course this is also the most classification method has a defect. Summary and Conclusion In this paper, we propose a novel image semantic understanding and the feature extraction algorithm. When local content under the conditions of different imaging features of obvious differences that will lead to feature descriptor is difficult to match, for this, put forward a kind of image feature descriptor based on stable area. Maximum stable extremal region detection algorithm to extract image stability region, combined with the location and scale of image feature points, according to the characters of the stable area calculation principal direction, in stable region adapted to the feature points within the larger neighborhood scale, generated based on the characteristics of the logarithmic polar coordinates description operator. Under this background, we then apply the corresponding issues on the image semantic understanding tasks that assist to achieve the better performance. References [1] X. M. Liu, H. X. Yao, R. R. Ji, P. F. Xu and X. S. Sun: Multimedia tools and applications, Vol. 64 (2013) No.1, p.53. [2] S. Gupta, P. ArbeláezArbelaez, R. Girshick and J. Malik: International Journal of Computer Vision, Vol.112 (2015) No.2, p.133. [3] S. Aditya, Y. Aloimonos, C. Baral, C. Fermuller and Y. Z. Yang: 2015 AAAI Spring Symposium Series (Stanford University, California, March 23, 2015). [4] X. F. Wang and J. H. Li: Proceedings of the 2nd International Conference on Computer Science and Electronics Engineering (Hangzhou, P. R. China, March 22-23, 2013). p.1669. [5] T. D. Kelley and S. McGhee: Proc. SPIE 8756, Multisensor, Multisource Information Fusion: Architectures, Algorithms, and Applications 2013 (Baltimore, Maryland, USA, April 29, 2013). [6] P. Molino and L. M. Aiello: SMIR@ SIGIR 2014 (Gold Coast, QLD, Australia). p.38. [7] A. Amnal, A. Dewabharata, S. Y. Chou and M. Erawan: Proceedings of the Institute of Industrial Engineers Asian Conference 2013 (Springer Singapore, July 12, 2013). p.1081. [8] J. L. Martin: To boldly proceed: Papers from the 39th International Systemic Functional Congress (Edited by John S. Knox, Sydney, Australia, 2012). p.183. 2016. The authors Published by Atlantis Press 0117