3D Model Matching Based on Silhouette Image Matching

Similar documents
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Absolute Scale Structure from Motion Using a Refractive Plate

Measurement of Pedestrian Groups Using Subtraction Stereo

CS 354R: Computer Game Technology

DISTRIBUTED AUDIO-VIDEO SHARING BY COPY-AND-TRANSFER OPERATION FOR NETWORK 3D GAMES

Module 6: Pinhole camera model Lecture 32: Coordinate system conversion, Changing the image/world coordinate system

Section 4.2 selected answers Math 131 Multivariate Calculus D Joyce, Spring 2014

Choose the best answer. For 1 3, use the figure. For 4 5, use the figure. Which face is parallel to face DEF? BCEF

3.3 Optimizing Functions of Several Variables 3.4 Lagrange Multipliers

Software correction of geometric errors for multiaxis

Ray tracing based fast refraction method for an object seen through a cylindrical glass

Rectangular Coordinates in Space

In-plane principal stress output in DIANA

Lecture 5. If, as shown in figure, we form a right triangle With P1 and P2 as vertices, then length of the horizontal

Human Motion Detection and Tracking for Video Surveillance

Model-based segmentation and recognition from range data

Evaluation of Hardware Oriented MRCoHOG using Logic Simulation

Task analysis based on observing hands and objects by vision

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Three-Dimensional Viewing Hearn & Baker Chapter 7

Computer Graphics (CS 543) Lecture 13b Ray Tracing (Part 1) Prof Emmanuel Agu. Computer Science Dept. Worcester Polytechnic Institute (WPI)

Artistic Rendering of Function-based Shape Models

Similarity Image Retrieval System Using Hierarchical Classification

Expanding gait identification methods from straight to curved trajectories

Contents. MATH 32B-2 (18W) (L) G. Liu / (TA) A. Zhou Calculus of Several Variables. 1 Homework 1 - Solutions 3. 2 Homework 2 - Solutions 13

Fatigue Crack Growth Simulation using S-version FEM

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

RECONSTRUCTION OF REGISTERED RANGE DATA USING GEODESIC DOME TYPE DATA STRUCTURE

Camera-Based Document Image Retrieval as Voting for Partial Signatures of Projective Invariants

Time-to-Contact from Image Intensity

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

CSE328 Fundamentals of Computer Graphics

Vector Calculus: Understanding the Cross Product

Multi prototype fuzzy pattern matching for handwritten character recognition

A Novel Video Retrieval Method to Support a User's Recollection of Past Events Aiming for Wearable Information Playing

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Robotics Programming Laboratory

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

4.5 VISIBLE SURFACE DETECTION METHODES

The Five Rooms Project

T O B C A T C A S E E U R O S E N S E D E T E C T I N G O B J E C T S I N A E R I A L I M A G E R Y

Math 233. Lagrange Multipliers Basics

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Ray Tracer I: Ray Casting Due date: 12:00pm December 3, 2001

A simple method for interactive 3D reconstruction and camera calibration from a single view

Simulating a 3D Environment on a 2D Display via Face Tracking

374 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 2, APRIL 2012

Introduction to Functions of Several Variables

CS 4620 Midterm 1. Tuesday 22 October minutes

3D Object Model Acquisition from Silhouettes

A 100Hz Real-time Sensing System of Textured Range Images

If the center of the sphere is the origin the the equation is. x y z 2ux 2vy 2wz d 0 -(2)

OBJECT detection in general has many applications

CS6716 Pattern Recognition

Introduction to Supervised Learning

You may know these...

Using surface markings to enhance accuracy and stability of object perception in graphic displays

Ray Tracer Due date: April 27, 2011

Interactive Collision Detection for Engineering Plants based on Large-Scale Point-Clouds

Algebra 1 Notes Quarter

Hand-Eye Calibration from Image Derivatives

Chapter 5. Projections and Rendering

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992

Binary Histogram in Image Classification for Retrieval Purposes

CS602 Midterm Subjective Solved with Reference By WELL WISHER (Aqua Leo)

The Collision-free Workspace of the Tripteron Parallel Robot Based on a Geometrical Approach

DOWN PLUNGE CROSS SECTIONS

Lecture 3 Sections 2.2, 4.4. Mon, Aug 31, 2009

lecture 10 - depth from blur, binocular stereo

3.6: First Person Computer Games

A Representation of Human Visual Acuity on an Immersive Projection Display

SUPPLEMENTARY FILE S1: 3D AIRWAY TUBE RECONSTRUCTION AND CELL-BASED MECHANICAL MODEL. RELATED TO FIGURE 1, FIGURE 7, AND STAR METHODS.

Capturing, Modeling, Rendering 3D Structures

y (s,e) y=x+b (i, j) y=x+a y=x

AP* Optics Free Response Questions

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Chapter 6 Practice Test

Pattern Recognition Methods for Object Boundary Detection

Conspicuous Character Patterns

Clustering Billions of Images with Large Scale Nearest Neighbor Search

3. Draw the orthographic projection (front, right, and top) for the following solid. Also, state how many cubic units the volume is.

Overview of Computer Graphics

Chapter 24. Creating Surfaces for Displaying and Reporting Data

Efficient Stereo Image Rectification Method Using Horizontal Baseline

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

Grasping Known Objects with Aldebaran Nao

(1) Given the following system of linear equations, which depends on a parameter a R, 3x y + 5z = 2 4x + y + (a 2 14)z = a + 2

ALGEBRA 1 NOTES. Quarter 3. Name: Block

Coarse-to-fine image registration

Visualising Solid Shapes

Ray Tracing through Viewing Portals

Introduction to Computer Graphics. Modeling (3) April 27, 2017 Kenshi Takayama

Overview. A real-time shadow approach for an Augmented Reality application using shadow volumes. Augmented Reality.

Mathematics Stage 4 Diagnostic Tasks Answers with Common Errors

INTERACTIVE ENVIRONMENT FOR INTUITIVE UNDERSTANDING OF 4D DATA. M. Murata and S. Hashimoto Humanoid Robotics Institute, Waseda University, Japan

The mathematics behind projections

STEREO BY TWO-LEVEL DYNAMIC PROGRAMMING

Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera

Cluster Analysis. Angela Montanari and Laura Anderlucci

Interactive Video Retrieval System Integrating Visual Search with Textual Search

Transcription:

D Model Matching Based on Silhouette Image Matching YOSHIHIRO OKADA Graduate School of Information Science and Electrical Engineering, Kyushu University 6-1 Kasuga Koen, Kasuga, Fukuoka 816-8580 JAPAN Abstract: - Recently a lot of D s have been created and stored because of the great demand in CG animation productions and D game productions. In this situation, we need the D database system that allows us to retrieve required s easily and accurately. So the authors have been studying D database systems. The main questions are that what is the best feature as the similarity measure among D s and what is the best way to specify the query. In this paper, the authors propose the silhouette image matching as the answer of the first question and the hand sketch image as the answer of the second question. The authors have developed a D database system using the silhouette image matching as the similarity measure, and it accepts the hand sketch image as the query. This paper proposes such a database system and describes its availability showing its evaluation results. Key-Words: - D database, Hand sketch images, Silhouette image matching, D matching. 1 Introduction Recent advances in computer hardware technology made it possible to render D images in real time even if using a personal computer. As a result, many computer animation creation products were made and computer animations were created in many game productions and movie productions. In this situation, we need the D database system that allows us to retrieve required s easily and accurately. So we have been studying D database systems. The main questions are that what is the best feature as the similarity measure among D s and what is the best way to specify the query. This paper proposes the silhouette image matching as the answer of the first question and the hand sketch image as the answer of the second question. We have developed a D database system using the silhouette image matching as the similarity measure and it accepts the hand sketch image as the query. In this paper, we introduce such a database system and describe its availability showing its evaluation results. Already some researches on the D matching have been done. Paquet and Rioux propose a D database system. This system employs many popular D matching algorithms [1]. Osada et al proposed D matching algorithm using distribution of distances between any two random points on the surface of a D []. This method is called D and it has good evaluation results reported in their paper. Hiraga et al proposed new topology matching technique using Multi-resolutional Reeb Graphs (MRGs) []. Bandlow et al proposed recognition method of objects using their silhouette [4]. In this case, objects are not D s but D images. We have not met researches on the D matching using their silhouette images. Our proposed method employs the silhouette image matching for the D matching and we obtained good evaluation results. The remains of this paper are organized as follows: Section briefly describes overview of the D database system and explains detail of the silhouette image matching. Section shows experimental results to clarify the availability of the system. Finally Section 4 concludes the paper. D Model Matching Basic concept is very simple as shown in Figure 1. First of all, the system automatically generates three orthogonal silhouette images of each D for all D s and stores them as their feature information. In this case, silhouette images for each D are generated along with its three principal axis directions. Next, the user enters hand sketch images as the query to search his/her required s. After that, the system retrieves the best match s according to the silhouette image matching result.

Fig. 1 System overview..1 Principal axes To generate silhouette images from a automatically, it is necessary to determine view directions from the eye-position. Very common way is to use the principal axes [1][5]. Principal axis directions are determined as eigen vectors of the moment of a D calculated by the following expression (1). I xx I xy I xz I = I yx I yy I yz (1) I zx I zy I zz Here, each element q, r x, y, z )of the matrix I ( { } qr I is calculated by the following equation (). n 1 I qr = [ Si ( qi qcm ) ( ri rcm ] () n i= 1 n is the number of faces, Si is the area of i -th face. q i, r i are q, r elements of i -th face, q CM, rcm are q, r elements of the center of mass. Finally eigen vectors are calculated by the following equation. I a i λ i ai i 1,, () The eigen vector having the biggest eigen value is the first principal and the eigen vector having the second biggest eigen value is the second principal. The remainder is the third principal. Positive or negative directions are determined by distribution of amount of surface area. =, { }. Silhouette image generation As shown in Figure, the system generates three black and white silhouette images each of those is corresponding to each of three principal axes by actually rendering such D scenes on a computer screen. A background color is set white and a D color is set black. Each of three scenes is rendered with varying its view direction Fig. Two sets of normalized images, (A) and (B). corresponding to each of three principal axes. Its calculation cost is very small because the most parts of its rendering process are supported by the hardware. After rendering one scene, its image is normalized into a square shape in a specific resolution, e.g., eight by eight, 16 16,, and then stored in a storage device. The file size of the image of pixels resolution is 18 bytes. The total file size of three images is 84 bytes. This value is enough small in comparison with the file size of a D. As shown in Figure, to generate a square image, there are two kinds of normalizations, (A) and (B). The (A) uses the maximum length of a D s bounding box as a size of the square. In this case, information of ratio among a width, height and depth size of D s bounding box is kept. Therefore, this information influences similarity measure strongly. On the other hand, the case (B), a D s bounding box is normalized into a cube shape and the D is also normalized. Then three silhouette images are generated and stored. In this case, information of ratio among a width, height and depth size of D s bounding box is ignored. To compensate this, scaling information is used as another part of similarity measure as explained later.. Silhouette image matching The similarity measure is treated as that if two D s have a smaller error between their silhouette images, they have a larger similarity...1 Error between two s silhouette images Error between two s silhouette images is determined as the ratio of number of the different pixels to the total pixels as follows.

Fig. Original and its three alternative silhouette images. Error = two images different pixels total pixels (4) When calculating error between two 's silhouette images, actually minimum value is used among four error values of four kinds of images because four image patterns exist as shown in Figure. They are an original, a vertical flip, a horizontal flip, and a 180 degrees rotation image. As previously mentioned, positive or negative directions of first and second principal axis are determined by distribution of amount of the surface area of a. So if these directions are different between two s to be compared, their three image directions become different as shown in Figure. Case 1 is that the first principal axis direction is opposite, Case is that second principal axis direction is opposite, and Case is that both first and second principal axis are opposite. Then these three images, i.e., a vertical flip, a horizontal flip, and a 180 degrees rotation image, and the original image are necessary to compare. Each D has its three silhouette images. Total error is determined as mean value of three image matching errors as follows: =1 = i Error i Error (5) However, even if one D is very similar to the other to be compared, their three principal axes are sometimes different. Therefore, we have to calculate equation (5) for all combination patterns as shown in Table 1. There are six combination patterns. Each combination pattern's errer PErrork is calculated as follows. =1 = j Error j PError (6) k For example, in the case of PError1 between a A and a B, Error1 is an error value between A's front image and front image, Error is an error value between A's top image and top image, and Error is an error value between A's side image and side image. In the other case of PError, Error1 is between A's front image and front image, Error is between A's top image and side image, and Error is between A's side image and top image. Then we obtain final error value between two D s by the next equation. Error image = min( PError1, PError, L, PError6 ) (7) Table 1 Six combination patterns between A's three images and three images. j= A's k= 1 4 5 6 1 Front Front Front Top Top Side Side Top Top Side Front Side Top Front Side Side Top Side Front Front Top.. Error for scaling factor between two s As previously mentioned, there are two sets of normalized silhouette images (A) and (B) as shown in Figure. As a result of similarity evaluation, the set (B) has a good result. However, in this case, the scaling factor of a D disappears. To compensate this defect, information of ratio among a width, height and depth size of a D should be used. Then we employ two scaling values, i.e., the ratio of the second principal axis length to the first principal axis length and the ratio of the third principal axis length to the first principal axis length. Next equation (8) means an error between two D s concerning their scaling factor. Mean square root value is used since it is popular measure.

Error scale = ( l m (8) m ) + ( l ) Here l,l are the ratio of the second principal axis length to the first principal axis length and the ratio of the third principal axis length to the first principal axis length respectively for one of the two D s to be compared. Similarly m, m are the same elements for the other of the two D s... Total error between two s Finally we calculate a total error value using the next equation (9). w 1 Error image + w w 1 + w Error scale Error = (9) Here w1, w are weight values those the user specifies arbitrarily..4 Prototype system Figure 4 shows a screen image of the prototype system. The system generates three silhouette images for each D automatically by actually rendering them on a computer screen. Then the user can enter the query as his/her hand sketch images on a computer screen and retrieve required D s interactively. The three right upper small windows are for entering the query of three hand sketch images. The main window on the center shows a beathoven and its bounding box rendered according to its three principal axes. Fig. 4 A screen image of the prototype system.. Evaluation results This section presents evaluation results of silhouette image matching as the similarity measure for D matching to clarify its effectiveness. An error value is calculated using the equation (9). Here we use these values as w =, w 1. 1 =.1 Preparation We made a D database by collecting its D s from Internet. It consists of 150 s as shown in Figure 5. We downloaded most of them from DCAFE [6]. Then we classified 150 s into 0 classes so that each class consists of five same kinds of s. Our prototype system generated three silhouette images for each and totally 450 silhouette images for all s. The silhouette image resolution is pixels. Its generation time for one set of three silhouette images is about two seconds with using a standard PC ( Pentium II 450MHz, 56MB memory ).. Evaluation measures We use the same three evaluation measures described in []. They are "First Tier", "Second Tier" and "Nearest Neighbor" as follows. First Tier: This criteria means the percentage of top ( k 1) matches (excluding the query) from the query's class, where k is the size of the class, i.e., k is five in our D DB. This criteria is stringent, since each

Fig. 5 D database (150s = 0 classes 5 s/class) in the class has only one chance to be in the first tier. Top ( k 1) matches FirstTier = (10) k 1 Second Tier: This criteria is the same type of result, but for the top ( k 1) matches. Top ( k 1)matches SecondTier = (11) k 1 Nearest Neighbor: This criteria means the percentage of test in which the top match was from the query's class.. Results and discussion Table shows experimental results. Mean values, 0.45, 0.58 and 0.707, are corresponding to First Tire, Second Tire, and Nearest Neighbor respectively. Each of these three values is each mean of all the D s values. These values are better than values in the paper [] since the corresponding values of D actually applied on the same D database are 0.17, 0.418 and 0.447. Osada et al present their own evaluation results in their paper. These values are 0.49, 0.66 and 0.66 for their own D database. These values are depending on a D database. Any way, our results are very close to them. Our total search time for 150 queries is 87 seconds using PC ( Pentium II 450 MHz CPU, 56MB memory ) so that one search time is 0.58 second. This includes 150 times compares and 150 times file accesses. This value is not so bad since D also takes about 0.15 second for 150 times compares since one compare consumes 0.1 msec actually reported in the paper. Furthermore, the size of a secondary information file concerning D seems more than 4 Kbytes since it generates 104 int values for one D. This value is more than the size of our silhouette images file. Therefore, it can be said that our silhouette image matching is efficient as similarity measure for D s. Table Evaluation result 1 (silhouette image resolution : pixels) First Second Nearest Classes Tier Tier Neighbor Jet Plane 0.5 0.95 0.6 Helicopter 0.4 0.55 0.6 Space Ship 0. 0.5 0.4 Human 0.7 0.8 0.8

Head 0.5 0.6 0.4 Fish 0.4 0.45 0.8 Bird 0. 0. 0.6 Animal(4 legs) 0.45 0.75 1 Computer 0.15 0. 0.4 Chair 0.5 0.55 0.8 Sofa 0.5 0.7 0.6 Bed 0. 0. 0.8 Cup 0.5 0.5 0.6 Glass 0.5 0. 0.6 Bottle 0.95 1 0.8 Can 0.65 0.85 0.8 Air Force 0.45 0.65 0.8 Missile 0.75 0.85 0.8 Guitar 0.75 0.9 0.8 Tree 0. 0. 0.6 Flower 0.45 0.6 0.8 Ball 0.7 1 1 Car 1 1 1 Motor Bike 0.1 0.5 0.4 Submarine 0.55 0.7 1 Hand Gun 0.55 0.7 0.8 Rifle 0.5 0.4 0.4 Knife 0.5 0.5 0.8 Sword 0.6 0.7 1 Door 0.1 0.1 0.4 Mean value 0.45 0.58 0.707 Table Evaluation result (silhouette image resolution varying) Search Resolution time First Second Nearest ( pixels ) (sec) Tier Tier Neighbor 8 8 0.08 0.50 0.50 0.560 16 16 0.18 0.456 0.60 0.670 0.58 0.45 0.58 0.707 64 64 1.5 0.46 0.585 0.667 Deep color columns are classes that have a greater value than 0.5, i.e., a better value rather than others. These classes apt to have three principal axes those are clearly distinguished each other. So our method is especially efficient for s belong to such classes. Table shows evaluation results according to silhouette image resolution varying. As you see this table, at most 16 16 resolution is enough. Of course, search time is also better than the cases of high resolutions. All cases the same PC ( Pentium II 450MHz CPU, 56MB memory ) is used. Consequently 16 16 resolution is the best for our method. 4. Concluding remarks This paper proposed a D database system based on the silhouette image matching and presented experimental results to clarify the effectiveness of proposed method. Especially this paper described that it is useful to enter hand sketch images as the query for retrieval of required D s. To create hand sketch images is very intuitive manner so that using this prototype system, the user can enter the query easily and rapidly. This is one feature of our proposed system. As future works, our D matching is based on the D image matching. Already many researches on the D image matching have been done [7]. Using such more efficient algorithms, we will make our D database system able to retrieve D s more accurately. Acknowledgement We wish to thank Mr. Fumiki Nakano for his help to make D database. References: [1] Eric Paquet, Marc Rioux: Nefertiti: a query by content system for three-dimensional and image databases management, Image and Vision Computing 17, 157-166, 1999. [] Osada, R, et. al.: Matching D Models with Shape Distributions, Shape Modeling International, May 001. [] M. Hiraga, Y. Shinagawa, T. Kohmura and T. L. Kunii : Topology Matching for Fully Automatic Similarity Estimation of D Shapes, Proc. SIGGRAPH001, pp.0-1, 001. [4] Thorsten Bandlow, Alexa Hauck, Tobias Einsele, and Georg Färber: Recognising Objects by their Silhouette. In IMACS Conf. on Comp. Eng. in Systems Appl. (CESA'98), pages 744-749, April 1998. [5] Gottschalk, S., Lin, M. C. and Manocha, D.: OBBTree: A Hierarchical Structure for Rapid Interference Detection, Computer Graphics (SIGGRAPH '96 Proceedings), pp. 171-180, 1996. [6] http://www.dcafe.com/asp/freestuff.asp [7] R. C. Veltkamp and M. Hagedoorn: State-of-the-art in Shape Matching, Technical Report UU-CS-1999-7, Utrecht University, the Netherlands, 1999.