Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm

Size: px
Start display at page:

Download "Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm"

Transcription

1 Journal of Applied Science and Engineering, Vol. 21, No. 2, pp (2018) DOI: /jase _21(2).0011 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm Chi-Yi Tsai* and Chao-Chun Yu Department of Electrical and Computer Engineering, Tamkang University, Tamsui, Taiwan 251, R.O.C. Abstract Textureless object recognition is a difficult task in computer vision because the object-ofinterest (OOI) may not have enough texture information for extracting object features. To address this problem, this paper presents a textureless object recognition method based on the existing Line2D algorithm. The proposed method employs an edge-based hierarchical template matching method to detect and identify a wide variety of textureless objects. Given a reference template image of an OOI, a hierarchical edge-template database containing different 2D poses of the OOI was firstly created by applying affine transformation with different rotating and scaling settings to the reference template image. Next, an edge-based template matching process is performed to detect and recognize the OOI by searching matches between the hierarchical edge-template database and the input image. Finally, the position and angle posture of the OOI can be determined by the best match having the highest similarity measure. Experimental results show that the proposed method not only can efficiently recognize the type, quantity, position, and angle information of various textureless objects in the image, but also can achieve real-time performance about 24 frames per second (fps) in processing 640x480 images. Therefore, the proposed algorithm has the potential to be used in many computer vision applications. Key Words: Textureless Object Detection, Textureless Object Recognition, Template Matching, Pose Identification, Line2D Algorithm 1. Introduction Object recognition is one of the important tasks in the field of computer vision, and most of the conventional methods use texture information of objects to produce feature descriptors for object recognition process. However, in many practical applications, the object to be recognized may not have enough texture information for extracting keypoint descriptors [1 3], greatly increasing the difficulty of the object recognition task. This problem generally is referred to as textureless object recognition. Therefore, how to solve the problem of textureless *Corresponding author. chiyi_tsai@mail.tku.edu.tw object recognition remains an important issue for many computer vision applications. Several approaches have been proposed to address the textureless object recognition problem from a single image, and we examine them into three categories: feature-based approaches, view-based approaches, and shape-based approaches. The feature-based approaches [4 6] use edge features or line segments to recognize the textureless OOI by matching 2D image features to the corresponding 3D model points of the OOI. In [4], Costa and Shapiro proposed an object recognition system named relational indexing of objects (RIO), which performs a feature-based alignment algorithm to retrieve matches between an edge image of the OOI and 2D view-class

2 230 Chi-Yi Tsai and Chao-Chun Yu object models from a large database of object models. In [5], the authors defined a 3D invariant space, in which object recognition and pose estimation can be achieved by matching a set of 3D invariant model points of a given 3D model and a set of invariant light rays in an image of the OOI. In [6], David and DeMenthon proposed a model-based object recognition algorithm that uses correspondences between model and image line features to recognize the OOI in a clutter environment. One advantage of the feature-based approaches is that the 3D pose of the OOI can be calculated from the match result directly. However, finding the correspondences between model and image features is still a challenge in these methods. The view-based approaches [7 10] compare a given image of the OOI with pre-computed 2D views of the OOI represented by using aspect graphs [7]. These methods can roughly determine the object pose; however, they have to deal with a search problem in a large full geometric search space. To reduce the complexity of the aspect graph, Cyr and Kimia proposed a similarity-based aspect graph method [8] that groups similar 2D views of the OOI into a few aspects, each of them representing a single prototype of the OOI. In the recognition process, a shape similarity metric was employed to measure the similarity between a given OOI image and these few prototypes. Eggert et al. proposed a scale-space aspect graph approach [9] to reduce the number of aspects in the graph by increasing the scale of the object model. In [10], Ulrich et al. integrated the scale-space and similaritybased aspect graph approaches to develop a hierarchical view-based 3D object recognition algorithm. To improve robustness of object recognition, the authors used a robust similarity metric proposed in [11] instead of the original one. Moreover, the scale-space concept was realized by generating a hierarchical model from computer-aided design (CAD) model of the object within some predefined pose ranges. However, the complexity of the view-based approaches is still very high for many applications. The shape-based approaches [12 16] are based on template matching of edge segments. In [12], Hinterstoisser et al. proposed a dominant orientation template matching method (also called Line2D method) for realtime textureless object detection without a time consuming training stage. They also combined the Line2D method with a matching approach in range data to form a multimodal detection algorithm, called LineMod method [13,14]. The Line2D and LineMod methods can detect multiple objects in real-time; however, both methods require users to pre-build multimodal templates of each object manually. Cai et al. proposed a two-stage cascaded object detection method [15], which is based on an edge-based sparse image descriptor and can provide a rough estimation of object pose through a voting procedure. However, this method requires a large database recording the training templates of each OOI with different poses. In [16], Tombari et al. proposed a Bunch of Lines Descriptor (BOLD), which encodes a group of neighboring line segments obtained from a fast line segment detector [17] to produce an edge-based scale- and rotation-invariant descriptor. The BOLD keypoint allows improving robustness of edge-based object detection algorithms; however, it is difficult to deal with detection of curvilinear or simple objects in a heavy clutter scene. To detect textureless objects in heavily cluttered-occluded scenes, Chan et al. proposed a Bounding Oriented-Rectangle Descriptor for Enclosed Regions (BORDER), which is a discriminative rotation-invariant descriptor based on a Linelets technique to stabilize keypoint repeatability in occlusion [18]. However, the BORDER method is computationally expensive especially when large number of keypoints needs to be processed. Moreover, the precision of BORDER keypoint is restricted to its fairly-low block resolution [4 4]. Recently, the same authors also proposed a Binary Integrated Net Descriptor (BIND), which uses binary coding to encode multi-layered regional information into binary strings for high precision edge-based description [19]. Because BIND is a binarybased textureless object detector, its recognition speed is faster than BORDER about 2.2 times. In this paper, a textureless object recognition method is proposed based on the existing Line2D algorithm. An edge-based hierarchical template matching method is proposed to efficiently detect and identify a wide variety of textureless objects. Given a reference template image of an OOI, a hierarchical edge-template database containing various 2D poses of the OOI was firstly created by applying affine transformation with different rotating and scaling settings to the reference template image.

3 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm 231 Next, an edge-based template matching process is performed to detect and recognize the OOI by searching matches between the hierarchical edge-template database and the input image. Finally, the 2D pose of the OOI can be determined by the best match having the highest similarity measure. Experimental results show that the proposed method not only can efficiently recognize the type, quantity, position, and angle information of various textureless objects in the image, but also can identify up-down relationship between the recognized objects. In addition, the proposed method achieves real-time performance about 24 frames per second (fps) in processing images. These advantages make the proposed algorithm having the potential to be used in many computer vision applications. The rest of this paper is organized as follows. Section 2 introduces the proposed edge-based hierarchical template matching algorithm. Section 3 presents the proposed database generation method to automatically produce a hierarchical edge-template database of a given OOI. Experimental results are reported in Section 4 to evaluate the performance and robustness of the proposed textureless object detection and recognition method. Section 5 concludes the contributions of this paper. 2. The Proposed Edge-based Hierarchical Template Matching Algorithm Figure 1 shows the block diagram of the proposed textureless object detection and recognition system, which is constructed by an efficient image enhancement algorithm [20], an edge-based template matching algorithm and a database consisting of hierarchical edge templates of the OOI and response map of each edge orientation for similarity computation. This section focuses on the design of the edge-based template matching algorithm, which is divided into five stages: color edge detection, edge orientation quantization, one-hot coding of edge orientation, orientation spreading by dilation and similarity measurement. The following subsections briefly explain the function of each stage in the proposed edgebased template matching process. 2.1 Color Edge Detection To detect textureless objects, image edges are often used as image cues because they are the most stable features on textureless objects. To improve detection performance, we use edges on each color channel to detect the OOI in the input image. Given an RGB color image I rgb, two directional edge maps of each color channel can be computed by using Sobel operator such that (1) where the symbol denotes 2D convolution operator, c {R, G, B} denotes one of the color channels of the color image, and S h and S v are the horizontal and vertical convolution masks defined as For each image location, the edge with largest magni- Figure 1. Block diagram of the proposed textureless object detection and recognition system.

4 232 Chi-Yi Tsai and Chao-Chun Yu tude is chosen as the color edge of the image, denoted by g,suchthat (2) c c c T where g (, x y) [ Ih (, x y) Iv (, x y)] is the gradient vectorofcolorchannelc at image location (x, y) oftheinput color image. 2.2 Edge Orientation Quantization After extracting color edges of the input image, an edge orientation map can be computed by evaluating color gradient vector such that (3) where d min is a threshold to remove noises, and atan2(y, x) is a four quadrant arctangent function returning tan 1 (y/x) with a proper quadrant [21]. Next, the orientation of color edges is quantized into eight directions {g 0, g 1,, g 7 } with equal spaces as shown in Figure 2(a). To speed up the processing time, we divide the edge orientation map into N blocks, each of them containing eight bins to count edge orientations in the block such that (4) where i {0,1,,7},j {0, 1,..., N 1},and q = 22.5 is a constant angle interval to quantify the edge orientation. Then, the quantized edge orientation of the j-th block can be determined such that (5) where g i =(i +0.5) q is the i-th quantized edge orientation. 2.3 One-hot Coding of Edge Orientation To improve computational efficiency, the quantized edge orientation map obtained from Eq. (5) is encoded into a binarized image by one-hot coding. That is, each bit of the binary code corresponds to one quantized orientation, and the bit corresponding to the dominant edge orientation of a block is set to 1 as shown in Figure 2(b). Figure 2. (a) Quantifying the gradient orientation into eight directions {g 0, g 1,,g 7 }. (b) Binarized image of the quantized edge orientation map using one-hot coding.

5 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm 233 These binary codes help to accelerate the computation of similarity measurement through a simple lookup table indexing operation, which is presented in section Orientation Spreading by Dilation As suggested by [13], one can diffuse the quantized edge orientation of each block to its adjacent locations to obtain a binarized spreading image, which is more robust to noise. To achieve this, the authors in [13] first computed a bit map for each quantized orientation and then combined these bit maps using shift and OR operations. Here, we propose another method to achieve the same purpose. Figure 3 shows concept of the proposed orientation spreading method, which applies morphological dilation operation on the one-hot coded binarized image to produce the binarized spreading image directly. The proposed method is very efficient because the dilation processing is bitwise, allowing to accomplish the orientation spreading operation for all eight bit-planes simultaneously. 2.5 Similarity Measurement Similarity measure is the last step of the template matching process. To measure the similarity between one block of the binarized edge orientation map and one of the quantized edge orientations g i, a similarity criterion is proposed such that (6) where B =(b 7 b 6 b 0 ) 2 is the binary code of the block to be measured, and B g ={b 7 g 7, b 6 g 6,..., b 0 g 0 } is a quantized edge orientation set associated with the binary code B.Let ref g denote the quantized edge orientation map of a given reference template, and B q is the binarized edge orientation map of a query template. Suppose that each of them contains N blocks. Based on the criterion (6), the similarity between g ref and B q can be determined by (7) where g j is the quantized edge orientation of the j-th block of g ref,andb j is the binary code of j-th block of B q. To speedup the computation of the similarity measurement (7), an efficient acceleration method was presented in [13] based on lookup tables. That is, for each quantized edge orientation, a lookup table can be precomputed with respect to all combinations of the binary code. Then, the computation of the similarity criterion (6) can be simplified as a lookup-table indexing operation, as shown in Figure 4. In this case, the quantized edge orientation is assumed as g 0 and the quantized edge orientation set is given by B g ={g 7, g 2, g 1 }. By doing so, the computations of equation (7) can be significantly reduced once the lookup table of each quantized edge orientation is pre-computed a priori. 3. Hierarchical Edge-template Database Generation As shown in Figure 1, the proposed system requires Figure 3. Binarized spreading image obtained by applying dilation operation on the one-hot coded binarized image of the quantized edge orientation map.

6 234 Chi-Yi Tsai and Chao-Chun Yu not only can deal with scale and rotation variations of the OOI, but also can efficiently determine the scale factor and rotation angle of the OOI according to the pose information of the best matched edge template in the database. The robustness and pose estimation performance of the proposed textureless object detection and recognition system is validated in the experiment section. 4. Experimental Results Figure 4. Efficiently accelerating the computation of similarity criterion (6) with respect to the quantized edge orientation g 0 using a lookup-table indexing operation. a hierarchical edge-template database of the OOI to perform the edge template matching process described in the previous section. This section thus presents an offline process to produce the hierarchical edge-template database to be used in the proposed system. Since the matching process is very sensitive to scale and rotation changes of the OOI, increasing the number of edge templates of the OOI under different poses helps to improve robustness of the proposed system against scale and rotation variations of the OOI. Given a reference image of an OOI, an efficient way to produce a variant of reference images of the OOI is setting a group of scaling s [s 1, s n ] and rotating [ 1, m ] parameters a priori. Then, applying these settings on the reference image generates a 2D hierarchical model constructed by hierarchical image templates of the OOI with different scales and rotations via affine transformation [22] such that The following experiments comprise three parts. The first part presents detection performance of the proposed method, including multi-objects detection, object classification and orientation estimation. The second part presents comparisons between the proposed method and other existing methods. The third part presents the computational efficiency of the proposed method running on a popular desktop platform. 4.1 Performance Evaluation Table 1 lists four reference images of the textureless objects (a screw nut, a screw, a key, and a scissor) used in the experiments, each of them is assigned a Class ID from 0 to 3, respectively. The database totally contains 1,944 reference images (504 screw nut images, 720 screw images, 504 key images, and 216 scissor images) (8) where p ij and p ref are pixel positions of the transformed template and the reference template, respectively. Figure 5 shows an example to produce the hierarchical image templates with different poses of the OOI. By doing so, the proposed edge-based template matching method Figure 5. An offline process to generate a 2D hierarchical model constructed by hierarchical image templates of a textureless OOI with different scale and rotation settings, s [s 1, s n ] and [ 1, m ].

7 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm 235 Table 1. Objects-of-interest used in the experiments Reference image of textureless objects Image resolution (width height) Range of orientation angles (Degree) Number of templates Class ID to to to to generated by the proposed method described in the previous section. The orientation angle of the OOI is measured in degree, where a counterclockwise rotation is positive. The threshold d min used in Eq. (3) is set as 7.5 in the experiments. Table 2 shows the experimental result of the proposed method to detect and recognize multiple Class 0 and Class 1 textureless objects. In this experiment, several Class 0 and Class 1 objects were randomly placed on the table. From Table 2, it is clear that the proposed method successfully detects and recognizes these objects, each of them having a high similarity score greater than 87%. Moreover, the image position and orientation angle of each object also can be measured simultaneously. Note that in this paper the image position of the object is referred to the top-left point of the detected region-of-interest (ROI). Similar results can be observed in Table 3, in which the proposed method was used to detect and recognize multiple Class 2 and Class 3 textureless objects randomly placed on the table. Each detected object has a high similarity score over 86%, and its ROI position and orientation angle can be measured as well. The proposed method is able to be applied on a complex scene to detect the OOI recorded in the database. Table 4 shows the experimental results in this case, in which the Class 2 and Class 3 objects were randomly placed on the table together with other unknown objects. It is clear from Table 4 that the proposed method correctly extracts these two textureless objects from the complex scene. The similarity score of both detected objects is larger than 90%, and the measured ROI position

8 236 Chi-Yi Tsai and Chao-Chun Yu deals with the occlusion problem during textureless object detection and recognition tasks. One can see from Table 5 that the proposed method correctly distinguishes the multiple occluded Class 0 objects placed on the table. Moreover, the top-most object also has a higher similarity score than that of the occluded objects. This implies that the value of similarity score not only can be used to detect and recognize the OOI recorded in the database, but also can determine the top-most object in the case of detecting multiple occluded objects. This property may extend the proposed method to robotic manipulation applications such as random bin picking [23], pick-andplace [24] tasks. Therefore, the proposed method is suitable to be integrated in a robotic vision system to recognize textureless objects in a clutter scene. 4.2 Comparison to Other Existing Methods In the experiment, the performance of the proposed method was evaluated by comparing with the Speeded Up Robust Features (SURF) detection method [2], the Oriented FAST and Rotated BRIFE (ORB) detection method [3], and the BOLD method [16]. Figure 6 shows the experimental results of detecting and recognizing multiple textureless objects. It is clear from Figure 6 that the SURF and ORB detection methods fail to detect these and orientation angle are consistent with the pose of each OOI appeared in the captured image. The proposed method has robustness against occlusion in object detection and recognition. Table 5 presents the experimental results of the proposed method which

9 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm 237 (a) (c) (b) (d) textureless model 0, model 5, and model 7, respectively. It can be seen from Figure 8 that the proposed method successfully detects and recognizes the target textureless object in a clutter scene. Therefore, the above experimental results validate the performance of the proposed method on the application of multiple textureless object detection and recognition, even in a complex and clutter scene. Figure 6. Comparisons between the proposed and the other existing methods: experimental result obtained by (a) the SURF detection method, (b) the ORB detection method, (c) the BOLD method, and (d) the proposed method. objects because both methods are based on the keypoint matching result, which usually contains many incorrect matches for textureless objects. Although the BOLD method is able to recognize the type of these objects, it detects only one object for each type. On the contrary, the proposed method not only succeeds to recognize the type of different textureless objects, but also correctly detects the quantity of multiple textureless objects. To further evaluate the performance of the proposed method, we also compared the proposed method with two recently published methods: the BORDER method [18] and the BIND method [19]. Figures 7(a), 7(b), and 7(c) illustrate the experimental results obtained from the BORDER method, BIND method, and the proposed method, respectively. From Figure 7, one can see that the BORDER method fails to detect both types of the textureless objects. The BIND method successfully recognizes the screw nut object, but it only finds one of the multiple screw nut objects. By contrast, the proposed method not only can recognize both types of the textureless objects, but also returns the correct number of the textureless objects in the scene. Figure 8 shows the experimental result of the proposed method tested on an online public textureless dataset [25], which includes nine textureless models and fifty-four test scenes with clutter and occlusions. Figures 8(a), 8(b), and 8(c) show the recognition result of the 4.3 Computational Efficiency The real-time performance of the proposed edge-based template matching algorithm has been tested on a Windows 8.1 machine equipped with a 2.5 GHz Intel Core (TM) i7-4710mq CPU and 8 GB of memory. Table 6 tabulates the average processing time and average frame rate of the proposed method with respect to the number of objects in the image. From Table 6, it is clear that the processing speed of the proposed method is dependent on the number of objects. That is, the more objects in the image, the longer the processing time is required. When the number of objects is increased from one to seven in the image, the average processing time of the proposed method is increased from 39 ms to 42 ms. The overall av- Figure 7. Comparisons between the proposed and two recently published methods: experimental results obtained by (a) the BORDER method, (b) the BIND method, and (c) the proposed method.

10 238 Chi-Yi Tsai and Chao-Chun Yu Figure 8. Experimental results of the proposed method tested on the D-textureless dataset [25]: recognition result of (a) the textureless model 0, (b) the textureless model 5, and (c) the textureless model 7. Table 6. Average processing time (in milliseconds) and average frame rate (in fps) of the proposed method Number of detected objects Average processing time (ms) Average frame rate (fps) Overall average erage processing time of the proposed method is about 41ms. Therefore, the average frame rate of the proposed textureless object detection system achieves real-time performance about 24 fps in processing 640-by-480 video streams for detecting up to seven textureless objects in the image. This advantage makes the proposed method applicable to realistic products. 5. Conclusions and Future Work In this paper, a novel textureless object recognition algorithm is proposed based on an edge-based template matching method to detect and recognize a wide variety of textureless objects. To improve the detection accuracy of the proposed system, a template image database of an OOI is automatically generated by applying a group of presetting affine transformations on the template image of the OOI. To detect and recognize the OOI, an efficient edge-based template matching process is performed to find the best match having the highest similarity measure between the template image database and the input image. Experimental results show that the proposed method can efficiently detect and recognize multiple textureless objects in the image. Moreover, the up-down relationship between the recognized objects also can be identified according to the similarity measures. The proposed system achieves real-time performance in processing VGA video streams, which is suitable for many applications of vision-based robotic manipulation. In future work, the proposed method will be integrated into a robot manipulator system to accomplish random binpicking function for picking textureless objects. Acknowledgements The authors would like to thank Jacob Chan of Nanyang Technological University, Singapore, for providing the BORDER and BIND programs and Han-Po Lin of Tamkang University, Taiwan, in collecting the experimental data for this research. This work was supported by the Ministry of Science and Technology (MOST) of Taiwan, ROC under grant MOST E References [1] Lowe, D. G., Distinctive Image Features from Scaleinvariant Keypoints, International Journal of Computer Vision, Vol. 60, No. 2, pp (2004). doi: /B:VISI [2] Bay, H., Tuytelaars, T. and Gool, L. V., SURF: Speeded Up Robust Features, Computer Vision and Image Understanding, Vol. 110, No. 3, pp (2008). doi: /j.cviu [3] Rublee, E., Rabaud, V., Konolige, K. and Bradski, G., ORB: an Efficient Alternative to SIFT or SURF, IEEE International Conference on Computer Vision, Colorado Springs, USA, Jun , pp (2011). doi: /ICCV [4] Costa, M. S. and Shapiro, L. G., 3D Object Recognition and Pose with Relational Indexing, Computer Vi-

11 Real-time Textureless Object Detection and Recognition Based on an Edge-based Hierarchical Template Matching Algorithm 239 sion and Image Understanding, Vol. 79, No. 3, pp (2000). doi: /cviu [5] Weiss, I. and Ray, M., Model-based Recognition of 3D Objects from Single Images, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 23, No. 2, pp (2001). doi: / [6] David, P. and DeMenthon, D., Object Recognition in High Clutter Images Using Line Features, IEEE International Conference on Computer Vision, Beijing, China, Oct , pp (2005). doi: /ICCV [7] Schiffenbauer, R. D., A Survey of Aspect Graphs, Department of Computer and Information Science, Polytechnic University, Technical Report TR-CIS , Brooklyn, New York (2001). [8] Cyr, C. M. and Kimia, B. B., A Similarity-based Aspect-graph Approach to 3D Object Recognition, International Journal of Computer Vision, Vol. 57, No. 1, pp (2004). [9] Eggert, D. W., Bowyer, K. W., Dyer, C. R., Christensen, H. I. and Goldgof, D. B., The Scale Space Aspect Graph, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 15, No. 11, pp (1993). doi: / [10] Ulrich, M., Wiedemann, C. and Steger, C., Combining Scale-space and Similarity-based Aspect Graphs for Fast 3D Object Recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 34, No. 10, pp (2012). doi: /TPAMI [11] Steger, C., Occlusion, Clutter, and Illumination Invariant Object Recognition, International Archives of the Photogrammetry, Remote Sensing, Vol. 34, Part 3A, pp (2002). [12] Hinterstoisser, S., Lepetit, V., Ilic, S. and Fua, P., Dominant Orientation Templates for Real-time Detection of Texture-less Objects, IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, USA, Jun , pp (2010). doi: /CVPR [13] Hinterstoisser, S., Cagniart, C., Ilic, S., Sturm, P., Navab, N., Fua, P. and Lepetit, V., Gradient Response Maps for Real-time Detection of Textureless Objects, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 34, No. 5, pp (2012). doi: /TPAMI [14] Hinterstoisser, S., Holzer, S. and Cagniart, C., Multimodal Templates for Real-time Detection of Textureless Objects in Heavily Cluttered Scenes, IEEE International Conference on Computer Vision, Barcelona, Spain, Nov. 6 13, pp (2011). doi: / ICCV [15] Cai, H., Werner, T. and Matas, J., Fast Detection of Multiple Textureless 3-D Objects, International Conference on Computer Vision Systems, St. Petersburg, Russia, Jul , pp (2013). doi: / _11 [16] Tombari, F., Franchi, A. and Di, L., BOLD Features to Detect Texture-less Objects, IEEE International Conference on Computer Vision, Sydney, Australia, Dec. 1 8, pp (2013). doi: /ICCV [17] von Gioi, R. G., Jakubowicz, J., Morel, J. M. and Randall, G., LSD: a Fast Line Segment Detector with a False Detection Control, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 32, No. 4, pp (2010). doi: /TPAMI [18] Chan, J., Lee, J. A. and Qian, K., BORDER: an Oriented Rectangles Approach to Texture-less Object Recognition, IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, Jun , pp (2016). doi: /CVPR [19] Chan, J., Lee, J. A. and Qian, K., BIND: Binary Integrated Net Descriptors for Texture-less Object Recognition, IEEE Conference on Computer Vision and Pattern Recognition, Hawaii, USA, Jul , pp (2017). doi: /CVPR [20] Tsai, C.-Y. and Huang, C.-H., Real-time Implementation of an Adaptive Simultaneous Dynamic Range Compression and Local Contrast Enhancement Algorithm on the GPU, Journal of Real-Time Image Processing, Online Published (2015). doi: / s [21] Ali, M. A., Park, H. A. and Lee, C. S. G., Closed-form Inverse Kinematic Joint Solution for Humanoid Robots, IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, Oct , pp (2010). doi: /IROS [22] Rogers, D. F. and Adams, J. A., Mathematical Elements for Computer Graphics, 2 nd Ed., McGraw-Hill,

12 240 Chi-Yi Tsai and Chao-Chun Yu New York, pp (1990). [23] Tsai, C.-Y., Yu, C.-J., Yu, C.-C. and Wong, C.-C., Real-time Implementation of a Keypoint-based Automatic Target Selection and Tracking System for Applications of Random Bin Picking, ICIC Express Letters, Part B: Applications, Vol. 7, No. 5, pp (2016). [24] Tsai, C.-Y., Wong, C.-C., Yu, C.-J., Liu, C.-C. and Liu, T.-Y., A Hybrid Switched Reactive-based Visual Servo Control of 5-dof Robot Manipulators for Pick-andplace Tasks, IEEE Systems Journal, Vol. 9, No. 1, pp (2015). doi: /JSYST [25] D-textureless dataset on BOLD/dataset/. Manuscript Received: Aug. 29, 2017 Accepted: Jan. 5, 2018

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES Pin-Syuan Huang, Jing-Yi Tsai, Yu-Fang Wang, and Chun-Yi Tsai Department of Computer Science and Information Engineering, National Taitung University,

More information

Part-Based Skew Estimation for Mathematical Expressions

Part-Based Skew Estimation for Mathematical Expressions Soma Shiraishi, Yaokai Feng, and Seiichi Uchida shiraishi@human.ait.kyushu-u.ac.jp {fengyk,uchida}@ait.kyushu-u.ac.jp Abstract We propose a novel method for the skew estimation on text images containing

More information

Ensemble of Bayesian Filters for Loop Closure Detection

Ensemble of Bayesian Filters for Loop Closure Detection Ensemble of Bayesian Filters for Loop Closure Detection Mohammad Omar Salameh, Azizi Abdullah, Shahnorbanun Sahran Pattern Recognition Research Group Center for Artificial Intelligence Faculty of Information

More information

Key properties of local features

Key properties of local features Key properties of local features Locality, robust against occlusions Must be highly distinctive, a good feature should allow for correct object identification with low probability of mismatch Easy to etract

More information

Local Patch Descriptors

Local Patch Descriptors Local Patch Descriptors Slides courtesy of Steve Seitz and Larry Zitnick CSE 803 1 How do we describe an image patch? How do we describe an image patch? Patches with similar content should have similar

More information

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract

More information

VEHICLE MAKE AND MODEL RECOGNITION BY KEYPOINT MATCHING OF PSEUDO FRONTAL VIEW

VEHICLE MAKE AND MODEL RECOGNITION BY KEYPOINT MATCHING OF PSEUDO FRONTAL VIEW VEHICLE MAKE AND MODEL RECOGNITION BY KEYPOINT MATCHING OF PSEUDO FRONTAL VIEW Yukiko Shinozuka, Ruiko Miyano, Takuya Minagawa and Hideo Saito Department of Information and Computer Science, Keio University

More information

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction Chieh-Chih Wang and Ko-Chih Wang Department of Computer Science and Information Engineering Graduate Institute of Networking

More information

Fast Image Matching Using Multi-level Texture Descriptor

Fast Image Matching Using Multi-level Texture Descriptor Fast Image Matching Using Multi-level Texture Descriptor Hui-Fuang Ng *, Chih-Yang Lin #, and Tatenda Muindisi * Department of Computer Science, Universiti Tunku Abdul Rahman, Malaysia. E-mail: nghf@utar.edu.my

More information

Specular 3D Object Tracking by View Generative Learning

Specular 3D Object Tracking by View Generative Learning Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp

More information

A Hybrid Feature Extractor using Fast Hessian Detector and SIFT

A Hybrid Feature Extractor using Fast Hessian Detector and SIFT Technologies 2015, 3, 103-110; doi:10.3390/technologies3020103 OPEN ACCESS technologies ISSN 2227-7080 www.mdpi.com/journal/technologies Article A Hybrid Feature Extractor using Fast Hessian Detector and

More information

The SIFT (Scale Invariant Feature

The SIFT (Scale Invariant Feature The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical

More information

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm Group 1: Mina A. Makar Stanford University mamakar@stanford.edu Abstract In this report, we investigate the application of the Scale-Invariant

More information

Outline 7/2/201011/6/

Outline 7/2/201011/6/ Outline Pattern recognition in computer vision Background on the development of SIFT SIFT algorithm and some of its variations Computational considerations (SURF) Potential improvement Summary 01 2 Pattern

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Discriminatively Trained Templates for 3D Object Detection: A Real Time Scalable Approach

Discriminatively Trained Templates for 3D Object Detection: A Real Time Scalable Approach 2013 IEEE International Conference on Computer Vision Discriminatively Trained Templates for 3D Object Detection: A Real Time Scalable Approach Reyes Rios-Cabrera a,b a CINVESTAV, Robotics and Advanced

More information

Invariant Feature Extraction using 3D Silhouette Modeling

Invariant Feature Extraction using 3D Silhouette Modeling Invariant Feature Extraction using 3D Silhouette Modeling Jaehwan Lee 1, Sook Yoon 2, and Dong Sun Park 3 1 Department of Electronic Engineering, Chonbuk National University, Korea 2 Department of Multimedia

More information

Local features and image matching. Prof. Xin Yang HUST

Local features and image matching. Prof. Xin Yang HUST Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

CS 223B Computer Vision Problem Set 3

CS 223B Computer Vision Problem Set 3 CS 223B Computer Vision Problem Set 3 Due: Feb. 22 nd, 2011 1 Probabilistic Recursion for Tracking In this problem you will derive a method for tracking a point of interest through a sequence of images.

More information

Lecture 10 Detectors and descriptors

Lecture 10 Detectors and descriptors Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of detectors SIFT Shape context Silvio Savarese Lecture 10-26-Feb-14 From the 3D to 2D & vice versa P =

More information

CS 231A Computer Vision (Fall 2012) Problem Set 3

CS 231A Computer Vision (Fall 2012) Problem Set 3 CS 231A Computer Vision (Fall 2012) Problem Set 3 Due: Nov. 13 th, 2012 (2:15pm) 1 Probabilistic Recursion for Tracking (20 points) In this problem you will derive a method for tracking a point of interest

More information

Determinant of homography-matrix-based multiple-object recognition

Determinant of homography-matrix-based multiple-object recognition Determinant of homography-matrix-based multiple-object recognition 1 Nagachetan Bangalore, Madhu Kiran, Anil Suryaprakash Visio Ingenii Limited F2-F3 Maxet House Liverpool Road Luton, LU1 1RS United Kingdom

More information

Computer Vision for HCI. Topics of This Lecture

Computer Vision for HCI. Topics of This Lecture Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi

More information

Detection and Fine 3D Pose Estimation of Texture-less Objects in RGB-D Images

Detection and Fine 3D Pose Estimation of Texture-less Objects in RGB-D Images Detection and Pose Estimation of Texture-less Objects in RGB-D Images Tomáš Hodaň1, Xenophon Zabulis2, Manolis Lourakis2, Šťěpán Obdržálek1, Jiří Matas1 1 Center for Machine Perception, CTU in Prague,

More information

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

SCALE INVARIANT FEATURE TRANSFORM (SIFT) 1 SCALE INVARIANT FEATURE TRANSFORM (SIFT) OUTLINE SIFT Background SIFT Extraction Application in Content Based Image Search Conclusion 2 SIFT BACKGROUND Scale-invariant feature transform SIFT: to detect

More information

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify

More information

Image Features: Detection, Description, and Matching and their Applications

Image Features: Detection, Description, and Matching and their Applications Image Features: Detection, Description, and Matching and their Applications Image Representation: Global Versus Local Features Features/ keypoints/ interset points are interesting locations in the image.

More information

Image Matching Using SIFT, SURF, BRIEF and ORB: Performance Comparison for Distorted Images

Image Matching Using SIFT, SURF, BRIEF and ORB: Performance Comparison for Distorted Images Image Matching Using SIFT, SURF, BRIEF and ORB: Performance Comparison for Distorted Images Ebrahim Karami, Siva Prasad, and Mohamed Shehata Faculty of Engineering and Applied Sciences, Memorial University,

More information

Object Tracking Algorithm based on Combination of Edge and Color Information

Object Tracking Algorithm based on Combination of Edge and Color Information Object Tracking Algorithm based on Combination of Edge and Color Information 1 Hsiao-Chi Ho ( 賀孝淇 ), 2 Chiou-Shann Fuh ( 傅楸善 ), 3 Feng-Li Lian ( 連豊力 ) 1 Dept. of Electronic Engineering National Taiwan

More information

BOLD features to detect texture-less objects

BOLD features to detect texture-less objects TPR features to detect texture-less objects Federico Tombari DISI, University of Bologna federico.tombari@unibo.it Alessandro Franchi Datalogic Automation alessandro.franchi@datalogic.com Luigi Di Stefano

More information

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images Karthik Ram K.V & Mahantesh K Department of Electronics and Communication Engineering, SJB Institute of Technology, Bangalore,

More information

A Fuzzy Brute Force Matching Method for Binary Image Features

A Fuzzy Brute Force Matching Method for Binary Image Features A Fuzzy Brute Force Matching Method for Binary Image Features Erkan Bostanci 1, Nadia Kanwal 2 Betul Bostanci 3 and Mehmet Serdar Guzel 1 1 (Computer Engineering Department, Ankara University, Turkey {ebostanci,

More information

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment

More information

Motion Estimation and Optical Flow Tracking

Motion Estimation and Optical Flow Tracking Image Matching Image Retrieval Object Recognition Motion Estimation and Optical Flow Tracking Example: Mosiacing (Panorama) M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 Example 3D Reconstruction

More information

Local Features Tutorial: Nov. 8, 04

Local Features Tutorial: Nov. 8, 04 Local Features Tutorial: Nov. 8, 04 Local Features Tutorial References: Matlab SIFT tutorial (from course webpage) Lowe, David G. Distinctive Image Features from Scale Invariant Features, International

More information

Fast Detection of Multiple Textureless 3-D Objects

Fast Detection of Multiple Textureless 3-D Objects Fast Detection of Multiple Textureless 3-D Objects Hongping Cai, Tomáš Werner, and Jiří Matas Center for Machine Perception, Czech Technical University, Prague Abstract. We propose a fast edge-based approach

More information

Learning 6D Object Pose Estimation and Tracking

Learning 6D Object Pose Estimation and Tracking Learning 6D Object Pose Estimation and Tracking Carsten Rother presented by: Alexander Krull 21/12/2015 6D Pose Estimation Input: RGBD-image Known 3D model Output: 6D rigid body transform of object 21/12/2015

More information

Texture Segmentation by Windowed Projection

Texture Segmentation by Windowed Projection Texture Segmentation by Windowed Projection 1, 2 Fan-Chen Tseng, 2 Ching-Chi Hsu, 2 Chiou-Shann Fuh 1 Department of Electronic Engineering National I-Lan Institute of Technology e-mail : fctseng@ccmail.ilantech.edu.tw

More information

III. VERVIEW OF THE METHODS

III. VERVIEW OF THE METHODS An Analytical Study of SIFT and SURF in Image Registration Vivek Kumar Gupta, Kanchan Cecil Department of Electronics & Telecommunication, Jabalpur engineering college, Jabalpur, India comparing the distance

More information

Object Recognition Algorithms for Computer Vision System: A Survey

Object Recognition Algorithms for Computer Vision System: A Survey Volume 117 No. 21 2017, 69-74 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu Object Recognition Algorithms for Computer Vision System: A Survey Anu

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

A Novel Extreme Point Selection Algorithm in SIFT

A Novel Extreme Point Selection Algorithm in SIFT A Novel Extreme Point Selection Algorithm in SIFT Ding Zuchun School of Electronic and Communication, South China University of Technolog Guangzhou, China zucding@gmail.com Abstract. This paper proposes

More information

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT Oct. 15, 2013 Prof. Ronald Fearing Electrical Engineering and Computer Sciences University of California, Berkeley (slides courtesy of Prof. John Wawrzynek)

More information

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS 8th International DAAAM Baltic Conference "INDUSTRIAL ENGINEERING - 19-21 April 2012, Tallinn, Estonia LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS Shvarts, D. & Tamre, M. Abstract: The

More information

Video Processing for Judicial Applications

Video Processing for Judicial Applications Video Processing for Judicial Applications Konstantinos Avgerinakis, Alexia Briassouli, Ioannis Kompatsiaris Informatics and Telematics Institute, Centre for Research and Technology, Hellas Thessaloniki,

More information

A Comparison of SIFT, PCA-SIFT and SURF

A Comparison of SIFT, PCA-SIFT and SURF A Comparison of SIFT, PCA-SIFT and SURF Luo Juan Computer Graphics Lab, Chonbuk National University, Jeonju 561-756, South Korea qiuhehappy@hotmail.com Oubong Gwun Computer Graphics Lab, Chonbuk National

More information

Color Local Texture Features Based Face Recognition

Color Local Texture Features Based Face Recognition Color Local Texture Features Based Face Recognition Priyanka V. Bankar Department of Electronics and Communication Engineering SKN Sinhgad College of Engineering, Korti, Pandharpur, Maharashtra, India

More information

DeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material

DeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material DeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material Yi Li 1, Gu Wang 1, Xiangyang Ji 1, Yu Xiang 2, and Dieter Fox 2 1 Tsinghua University, BNRist 2 University of Washington

More information

State-of-the-Art: Transformation Invariant Descriptors. Asha S, Sreeraj M

State-of-the-Art: Transformation Invariant Descriptors. Asha S, Sreeraj M International Journal of Scientific & Engineering Research, Volume 4, Issue ş, 2013 1994 State-of-the-Art: Transformation Invariant Descriptors Asha S, Sreeraj M Abstract As the popularity of digital videos

More information

Shape Descriptor using Polar Plot for Shape Recognition.

Shape Descriptor using Polar Plot for Shape Recognition. Shape Descriptor using Polar Plot for Shape Recognition. Brijesh Pillai ECE Graduate Student, Clemson University bpillai@clemson.edu Abstract : This paper presents my work on computing shape models that

More information

Scale Invariant Feature Transform

Scale Invariant Feature Transform Scale Invariant Feature Transform Why do we care about matching features? Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Image

More information

Local Image Features

Local Image Features Local Image Features Ali Borji UWM Many slides from James Hayes, Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial Overview of Keypoint Matching 1. Find a set of distinctive key- points A 1 A 2 A 3 B 3

More information

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing CS 4495 Computer Vision Features 2 SIFT descriptor Aaron Bobick School of Interactive Computing Administrivia PS 3: Out due Oct 6 th. Features recap: Goal is to find corresponding locations in two images.

More information

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK 1 Po-Jen Lai ( 賴柏任 ), 2 Chiou-Shann Fuh ( 傅楸善 ) 1 Dept. of Electrical Engineering, National Taiwan University, Taiwan 2 Dept.

More information

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features 1 Kum Sharanamma, 2 Krishnapriya Sharma 1,2 SIR MVIT Abstract- To describe the image features the Local binary pattern (LBP)

More information

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Jung H. Oh, Gyuho Eoh, and Beom H. Lee Electrical and Computer Engineering, Seoul National University,

More information

A Modified Mean Shift Algorithm for Visual Object Tracking

A Modified Mean Shift Algorithm for Visual Object Tracking A Modified Mean Shift Algorithm for Visual Object Tracking Shu-Wei Chou 1, Chaur-Heh Hsieh 2, Bor-Jiunn Hwang 3, Hown-Wen Chen 4 Department of Computer and Communication Engineering, Ming-Chuan University,

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE) Features Points Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Finding Corners Edge detectors perform poorly at corners. Corners provide repeatable points for matching, so

More information

Fast Natural Feature Tracking for Mobile Augmented Reality Applications

Fast Natural Feature Tracking for Mobile Augmented Reality Applications Fast Natural Feature Tracking for Mobile Augmented Reality Applications Jong-Seung Park 1, Byeong-Jo Bae 2, and Ramesh Jain 3 1 Dept. of Computer Science & Eng., University of Incheon, Korea 2 Hyundai

More information

Object Detection by Point Feature Matching using Matlab

Object Detection by Point Feature Matching using Matlab Object Detection by Point Feature Matching using Matlab 1 Faishal Badsha, 2 Rafiqul Islam, 3,* Mohammad Farhad Bulbul 1 Department of Mathematics and Statistics, Bangladesh University of Business and Technology,

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

Object Recognition with Invariant Features

Object Recognition with Invariant Features Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user

More information

Improving Visual SLAM Algorithms for use in Realtime Robotic Applications

Improving Visual SLAM Algorithms for use in Realtime Robotic Applications Improving Visual SLAM Algorithms for use in Realtime Robotic Applications Patrick Benavidez, Mohan Kumar Muppidi, and Mo Jamshidi, Ph.D.Lutcher Brown Endowed Chair Professor Autonomous Control Engineering

More information

[2006] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image

[2006] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image [6] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image Matching Methods, Video and Signal Based Surveillance, 6. AVSS

More information

An Approach for Real Time Moving Object Extraction based on Edge Region Determination

An Approach for Real Time Moving Object Extraction based on Edge Region Determination An Approach for Real Time Moving Object Extraction based on Edge Region Determination Sabrina Hoque Tuli Department of Computer Science and Engineering, Chittagong University of Engineering and Technology,

More information

The Crucial Components to Solve the Picking Problem

The Crucial Components to Solve the Picking Problem B. Scholz Common Approaches to the Picking Problem 1 / 31 MIN Faculty Department of Informatics The Crucial Components to Solve the Picking Problem Benjamin Scholz University of Hamburg Faculty of Mathematics,

More information

Object detection using non-redundant local Binary Patterns

Object detection using non-redundant local Binary Patterns University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2010 Object detection using non-redundant local Binary Patterns Duc Thanh

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image SURF CSED441:Introduction to Computer Vision (2015S) Lecture6: SURF and HOG Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Speed Up Robust Features (SURF) Simplified version of SIFT Faster computation but

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

Scale Invariant Feature Transform

Scale Invariant Feature Transform Why do we care about matching features? Scale Invariant Feature Transform Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Automatic

More information

Local Feature Detectors

Local Feature Detectors Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,

More information

Fuzzy based Multiple Dictionary Bag of Words for Image Classification

Fuzzy based Multiple Dictionary Bag of Words for Image Classification Available online at www.sciencedirect.com Procedia Engineering 38 (2012 ) 2196 2206 International Conference on Modeling Optimisation and Computing Fuzzy based Multiple Dictionary Bag of Words for Image

More information

A Fast Caption Detection Method for Low Quality Video Images

A Fast Caption Detection Method for Low Quality Video Images 2012 10th IAPR International Workshop on Document Analysis Systems A Fast Caption Detection Method for Low Quality Video Images Tianyi Gui, Jun Sun, Satoshi Naoi Fujitsu Research & Development Center CO.,

More information

Robotic Grasping Based on Efficient Tracking and Visual Servoing using Local Feature Descriptors

Robotic Grasping Based on Efficient Tracking and Visual Servoing using Local Feature Descriptors INTERNATIONAL JOURNAL OF PRECISION ENGINEERING AND MANUFACTURING Vol. 13, No. 3, pp. 387-393 MARCH 2012 / 387 DOI: 10.1007/s12541-012-0049-8 Robotic Grasping Based on Efficient Tracking and Visual Servoing

More information

PEOPLE IN SEATS COUNTING VIA SEAT DETECTION FOR MEETING SURVEILLANCE

PEOPLE IN SEATS COUNTING VIA SEAT DETECTION FOR MEETING SURVEILLANCE PEOPLE IN SEATS COUNTING VIA SEAT DETECTION FOR MEETING SURVEILLANCE Hongyu Liang, Jinchen Wu, and Kaiqi Huang National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Science

More information

Point Pair Feature based Object Detection for Random Bin Picking

Point Pair Feature based Object Detection for Random Bin Picking Point Pair Feature based Object Detection for Random Bin Picking Wim Abbeloos and Toon Goedemé KU Leuven, Department of Electrical Engineering, EAVISE Leuven, Belgium Email: wim.abbeloos@kuleuven.be, toon.goedeme@kuleuven.be

More information

Evaluation and comparison of interest points/regions

Evaluation and comparison of interest points/regions Introduction Evaluation and comparison of interest points/regions Quantitative evaluation of interest point/region detectors points / regions at the same relative location and area Repeatability rate :

More information

CAP 5415 Computer Vision Fall 2012

CAP 5415 Computer Vision Fall 2012 CAP 5415 Computer Vision Fall 01 Dr. Mubarak Shah Univ. of Central Florida Office 47-F HEC Lecture-5 SIFT: David Lowe, UBC SIFT - Key Point Extraction Stands for scale invariant feature transform Patented

More information

Generic Face Alignment Using an Improved Active Shape Model

Generic Face Alignment Using an Improved Active Shape Model Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn

More information

N.Priya. Keywords Compass mask, Threshold, Morphological Operators, Statistical Measures, Text extraction

N.Priya. Keywords Compass mask, Threshold, Morphological Operators, Statistical Measures, Text extraction Volume, Issue 8, August ISSN: 77 8X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Combined Edge-Based Text

More information

An indirect tire identification method based on a two-layered fuzzy scheme

An indirect tire identification method based on a two-layered fuzzy scheme Journal of Intelligent & Fuzzy Systems 29 (2015) 2795 2800 DOI:10.3233/IFS-151984 IOS Press 2795 An indirect tire identification method based on a two-layered fuzzy scheme Dailin Zhang, Dengming Zhang,

More information

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

Implementation and Comparison of Feature Detection Methods in Image Mosaicing IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p-ISSN: 2278-8735 PP 07-11 www.iosrjournals.org Implementation and Comparison of Feature Detection Methods in Image

More information

SIFT: Scale Invariant Feature Transform

SIFT: Scale Invariant Feature Transform 1 / 25 SIFT: Scale Invariant Feature Transform Ahmed Othman Systems Design Department University of Waterloo, Canada October, 23, 2012 2 / 25 1 SIFT Introduction Scale-space extrema detection Keypoint

More information

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking Feature descriptors Alain Pagani Prof. Didier Stricker Computer Vision: Object and People Tracking 1 Overview Previous lectures: Feature extraction Today: Gradiant/edge Points (Kanade-Tomasi + Harris)

More information

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving

More information

Rotation Invariant Finger Vein Recognition *

Rotation Invariant Finger Vein Recognition * Rotation Invariant Finger Vein Recognition * Shaohua Pang, Yilong Yin **, Gongping Yang, and Yanan Li School of Computer Science and Technology, Shandong University, Jinan, China pangshaohua11271987@126.com,

More information

Image Matching Using Run-Length Feature

Image Matching Using Run-Length Feature Image Matching Using Run-Length Feature Yung-Kuan Chan and Chin-Chen Chang Department of Computer Science and Information Engineering National Chung Cheng University, Chiayi, Taiwan, 621, R.O.C. E-mail:{chan,

More information

Human Motion Detection and Tracking for Video Surveillance

Human Motion Detection and Tracking for Video Surveillance Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,

More information

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Human Detection Systems based on Sliding Windows Approach-A Review Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg

More information

Extracting Road Signs using the Color Information

Extracting Road Signs using the Color Information Extracting Road Signs using the Color Information Wen-Yen Wu, Tsung-Cheng Hsieh, and Ching-Sung Lai Abstract In this paper, we propose a method to extract the road signs. Firstly, the grabbed image is

More information

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK Xiangyun HU, Zuxun ZHANG, Jianqing ZHANG Wuhan Technique University of Surveying and Mapping,

More information

Improving Recognition through Object Sub-categorization

Improving Recognition through Object Sub-categorization Improving Recognition through Object Sub-categorization Al Mansur and Yoshinori Kuno Graduate School of Science and Engineering, Saitama University, 255 Shimo-Okubo, Sakura-ku, Saitama-shi, Saitama 338-8570,

More information

Online Tracking Parameter Adaptation based on Evaluation

Online Tracking Parameter Adaptation based on Evaluation 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Online Tracking Parameter Adaptation based on Evaluation Duc Phu Chau Julien Badie François Brémond Monique Thonnat

More information

Pixels. Orientation π. θ π/2 φ. x (i) A (i, j) height. (x, y) y(j)

Pixels. Orientation π. θ π/2 φ. x (i) A (i, j) height. (x, y) y(j) 4th International Conf. on Document Analysis and Recognition, pp.142-146, Ulm, Germany, August 18-20, 1997 Skew and Slant Correction for Document Images Using Gradient Direction Changming Sun Λ CSIRO Math.

More information

International Journal of Advanced Research in Computer Science and Software Engineering

International Journal of Advanced Research in Computer Science and Software Engineering ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: Fingerprint Recognition using Robust Local Features Madhuri and

More information

Computers and Mathematics with Applications. An embedded system for real-time facial expression recognition based on the extension theory

Computers and Mathematics with Applications. An embedded system for real-time facial expression recognition based on the extension theory Computers and Mathematics with Applications 61 (2011) 2101 2106 Contents lists available at ScienceDirect Computers and Mathematics with Applications journal homepage: www.elsevier.com/locate/camwa An

More information

Designing Applications that See Lecture 7: Object Recognition

Designing Applications that See Lecture 7: Object Recognition stanford hci group / cs377s Designing Applications that See Lecture 7: Object Recognition Dan Maynes-Aminzade 29 January 2008 Designing Applications that See http://cs377s.stanford.edu Reminders Pick up

More information