Mouse Ray Picking Explained

Similar documents
Overview. By end of the week:

The Viewing Pipeline adaptation of Paul Bunn & Kerryn Hugo s notes

COMS 4160: Problems on Transformations and OpenGL

E.Order of Operations

Describe the Orthographic and Perspective projections. How do we combine together transform matrices?

Transforms 3: Projection Christian Miller CS Fall 2011

3D Viewing Episode 2

Projection Matrix Tricks. Eric Lengyel

Game Architecture. 2/19/16: Rasterization

Evening s Goals. Mathematical Transformations. Discuss the mathematical transformations that are utilized for computer graphics

Rasterization Overview

Viewing transformations. 2004, Denis Zorin

Notes on Assignment. Notes on Assignment. Notes on Assignment. Notes on Assignment

Transforms 1 Christian Miller CS Fall 2011

CS 381 Computer Graphics, Fall 2012 Midterm Exam Solutions. The Midterm Exam was given in class on Tuesday, October 16, 2012.

Lecture 4. Viewing, Projection and Viewport Transformations

Announcements. Submitting Programs Upload source and executable(s) (Windows or Mac) to digital dropbox on Blackboard

Transformation Pipeline

Rendering Pipeline and Coordinates Transforms

CPSC 314, Midterm Exam. 8 March 2010

Perspective transformations

3D Viewing. CS 4620 Lecture 8

Today. Rendering pipeline. Rendering pipeline. Object vs. Image order. Rendering engine Rendering engine (jtrt) Computergrafik. Rendering pipeline

CSE 167: Lecture #4: Vertex Transformation. Jürgen P. Schulze, Ph.D. University of California, San Diego Fall Quarter 2012

CPSC 314, Midterm Exam 1. 9 Feb 2007

CS230 : Computer Graphics Lecture 6: Viewing Transformations. Tamar Shinar Computer Science & Engineering UC Riverside

3D Viewing Episode 2

INTRODUCTION TO COMPUTER GRAPHICS. It looks like a matrix Sort of. Viewing III. Projection in Practice. Bin Sheng 10/11/ / 52

Surface Graphics. 200 polys 1,000 polys 15,000 polys. an empty foot. - a mesh of spline patches:

CS 130 Final. Fall 2015

3D Graphics Pipeline II Clipping. Instructor Stephen J. Guy

Viewing. Reading: Angel Ch.5

Three-Dimensional Viewing Hearn & Baker Chapter 7

Drawing in 3D (viewing, projection, and the rest of the pipeline)

Projection and viewing. Computer Graphics CSE 167 Lecture 4

Pipeline Operations. CS 4620 Lecture Steve Marschner. Cornell CS4620 Spring 2018 Lecture 11

CS 498 VR. Lecture 18-4/4/18. go.illinois.edu/vrlect18

INTRODUCTION TO COMPUTER GRAPHICS. cs123. It looks like a matrix Sort of. Viewing III. Projection in Practice 1 / 52

S U N G - E U I YO O N, K A I S T R E N D E R I N G F R E E LY A VA I L A B L E O N T H E I N T E R N E T

THE VIEWING TRANSFORMATION

COMP Computer Graphics and Image Processing. 5: Viewing 1: The camera. In part 1 of our study of Viewing, we ll look at ˆʹ U ˆ ʹ F ˆʹ S

CSE 690: GPGPU. Lecture 2: Understanding the Fabric - Intro to Graphics. Klaus Mueller Stony Brook University Computer Science Department

CS 381 Computer Graphics, Fall 2008 Midterm Exam Solutions. The Midterm Exam was given in class on Thursday, October 23, 2008.

Prof. Feng Liu. Fall /19/2016

Computer Graphics. Chapter 10 Three-Dimensional Viewing

Pipeline Operations. CS 4620 Lecture 14

OpenGL Transformations

CSE 167: Introduction to Computer Graphics Lecture #5: Projection. Jürgen P. Schulze, Ph.D. University of California, San Diego Fall Quarter 2017

Perspective matrix, OpenGL style

CS4620/5620: Lecture 14 Pipeline

1 Transformations. Chapter 1. Transformations. Department of Computer Science and Engineering 1-1

SUMMARY. CS380: Introduction to Computer Graphics Projection Chapter 10. Min H. Kim KAIST School of Computing 18/04/12. Smooth Interpolation

The Graphics Pipeline. Interactive Computer Graphics. The Graphics Pipeline. The Graphics Pipeline. The Graphics Pipeline: Clipping

Lecture 5: Viewing. CSE Computer Graphics (Fall 2010)

Motivation. Culling Don t draw what you can t see! What can t we see? Low-level Culling

GRAFIKA KOMPUTER. ~ M. Ali Fauzi

Spring 2013, CS 112 Programming Assignment 2 Submission Due: April 26, 2013

3D Viewing. CS 4620 Lecture Steve Marschner. Cornell CS4620 Spring 2018 Lecture 9

CPSC 314, Midterm Exam. 8 March 2013

Reading on the Accumulation Buffer: Motion Blur, Anti-Aliasing, and Depth of Field

Computer Graphics. Chapter 7 2D Geometric Transformations

Robust Stencil Shadow Volumes. CEDEC 2001 Tokyo, Japan

Overview of Projections: From a 3D world to a 2D screen.

Shadows in the graphics pipeline

CSE 167: Introduction to Computer Graphics Lecture #4: Vertex Transformation

CS4202: Test. 1. Write the letter corresponding to the library name next to the statement or statements that describe library.

The Graphics Pipeline and OpenGL I: Transformations!

COMP3421. Introduction to 3D Graphics

C P S C 314 S H A D E R S, O P E N G L, & J S RENDERING PIPELINE. Mikhail Bessmeltsev

Orthogonal Projection Matrices. Angel and Shreiner: Interactive Computer Graphics 7E Addison-Wesley 2015

Geometry: Outline. Projections. Orthographic Perspective

Drawing in 3D (viewing, projection, and the rest of the pipeline)

CSE528 Computer Graphics: Theory, Algorithms, and Applications

COMP3421. Introduction to 3D Graphics

Fachhochschule Regensburg, Germany, February 15, 2017

CS Computer Graphics: Transformations & The Synthetic Camera

Models and The Viewing Pipeline. Jian Huang CS456

蔡侑庭 (Yu-Ting Tsai) National Chiao Tung University, Taiwan. Prof. Wen-Chieh Lin s CG Slides OpenGL 2.1 Specification

CSE452 Computer Graphics

COMP 175 COMPUTER GRAPHICS. Ray Casting. COMP 175: Computer Graphics April 26, Erik Anderson 09 Ray Casting

Projection: Mapping 3-D to 2-D. Orthographic Projection. The Canonical Camera Configuration. Perspective Projection

Computer Graphics Shadow Algorithms

For each question, indicate whether the statement is true or false by circling T or F, respectively.

Visibility: Z Buffering

CS 4204 Computer Graphics

Computer Viewing. CS 537 Interactive Computer Graphics Prof. David E. Breen Department of Computer Science

TSBK03 Screen-Space Ambient Occlusion

Viewing COMPSCI 464. Image Credits: Encarta and

COMP3421. Introduction to 3D Graphics

Advanced Computer Graphics (CS & SE )

Graphics pipeline and transformations. Composition of transformations

Motivation. What we ve seen so far. Demo (Projection Tutorial) Outline. Projections. Foundations of Computer Graphics

CSE328 Fundamentals of Computer Graphics

To Do. Demo (Projection Tutorial) Motivation. What we ve seen so far. Outline. Foundations of Computer Graphics (Fall 2012) CS 184, Lecture 5: Viewing

Institutionen för systemteknik

3D Graphics for Game Programming (J. Han) Chapter II Vertex Processing

Perspective Mappings. Contents

To Do. Motivation. Demo (Projection Tutorial) What we ve seen so far. Computer Graphics. Summary: The Whole Viewing Pipeline

Viewing/Projections IV. Week 4, Fri Feb 1

3D Viewing. CMPT 361 Introduction to Computer Graphics Torsten Möller. Machiraju/Zhang/Möller

Transcription:

Mouse Ray Picking Explained Brian Hook http://www.bookofhook.com April 5, 2005 1 Introduction There comes a time in every 3D game where the user needs to click on something in the scene. Maybe he needs to select a unit in an RTS, or open a door in an RPG, or delete some geometry in a level editing tool. This conceptually simple task is easy to screw up since there are so many little steps that can go wrong. The problem, simply stated, is this: given the mouse s position in window coordinates, how can I determine what object in the scene the user has selected with a mouse click? One method is to generate a ray using the mouse s location and then intersect it with the world geometry, finding the object nearest to the viewer. Alternatively we can determine the actual 3-D location that the user has clicked on by sampling the depth buffer (giving us (x, y, z) in viewport space) and performing an inverse transformation. Technically there is a third approach, using a selection or object ID buffer, but this has numerous limitations that makes it impractical for widespread use. This article describes using the inverse transformation to derive world space coordinates from the mouse s position on screen. Before we worry about the inverse transformation, we need to establish how the standard forward transformation works in a typical graphics pipeline. 1

2 The View Transformation The standard view transformation pipeline takes a point in model space and transforms it all the way to viewport space 1. It does this by transforming the original point through a series of coordinate systems: Model W orld V iew Clip N ormalizeddevice V iewport Not each step is discrete. OpenGL has the GL MODELVIEW matrix, (M), that transforms a point p from model space to a point v in view space. Both model and view space use a right-handed coordinate system with +Y up, +X to the right, and -Z into the screen. Mp = v (1) Another matrix, GL PROJECTION (P ), then transforms the view space point to the homogeneous clip space point c. Clip space is a right-handed coordinate system (+Z into the screen) contained within a canonical clipping volume extending from ( 1, 1, 1) to (+1, +1, +1). P v = c (2) 1 viewport coordinates are also sometimes known as window coordinates or, for systems without a window system, screen coordinates 2

After clipping is performed the perspective divide transforms the homogeneous coordinate c back into the Cartesian point n in normalized device space. Normalized device coordinates are left-handed, where n w = 1, and are contained within the canonical view frustum from ( 1, 1, 1) to (+1, +1, +1). n = c c w (3) Finally there is the viewport scale and translation, V, which transforms n into the final viewport window coordinates w. Another axis inversion occurs here; this time +Y goes down instead of up. 2. Viewport depth values are calculated by rescaling NDC Z coordinates from the range ( 1, 1) to (0, 1), with 0 at the near clip plane and 1 at the far clip plane. It s important to note that any user specified depth bias may impact our calculations later. V n = w (4) Using this pipeline we can take a model space point, apply a series of transformations, and get a viewport coordinate. Our goal is to transform the mouse position in viewport coordinates all the way back to world space. Since we re not rendering a model, model space and world space are the same. 3 The Inverse View Transformation To go from mouse coordinates to world coordinates we have to do the exact opposite of the view transformation: V iewport NDC Clip V iew W orld/model That s a lot of steps, and it s easy to screw up, and if you screw up just a little that s enough to blow everything apart. 2 Some less common window systems may place the origin at another location, such as the bottom left of the window, so this isn t always true 3

3.1 V iewport NDC Clip The first step is to transform the viewport coordinates into clip coordinates. The viewport transformation V takes a normalized device coordinate n and transforms it into a viewport coordinate. Given viewport width w and height h then our viewport coordinate v is: V n = v = n x+1w 2 1 n y h 2 n z+1 2 (5) So we need to do the inverse of this process by rearranging Equation 5 to solve for n: 2v x 1 w 2v y n = h 2v z 1 (6) 1 Okay, not so bad. The only real question is what is v z? We can either calculate v z by reading it back from the framebuffer, or ignore it altogether and substitute 0, in which case we ll be computing a ray passing through v that we ll then have to intersect with world geometry to find the corresponding point in 3-space. From here we need to go to clip coordinates, which, if you recall, are the homogeneous versions of the NDC coordinates (i. e. w 1.0). Since w is already 1.0 and the transformation back to clip coordinates is a scale by w, this step can be skipped and we can assume that our NDC coordinates are the same as our clip coordinates. 3.2 Clip V iew A vector v in view space is transformed to clip coordinates c by multiplying it against the GL PROJECTION matrix P : P v = c (7) Given this, we can do the opposite by multiplying the clip coordinate by the inverse of the GL PROJECTION matrix. This isn t as scary as it 4

sounds we can avoid computing a true 4x4 matrix inverse if we just construct the inverse projection matrix at the same time we build the projection matrix. A typical OpenGL perspective projection matrix P takes the form: a 0 0 0 P = 0 b 0 0 0 0 c d (8) 0 0 e 0 The specific coefficient values depend on the nature of the perspective projection matrix (for more information I recommend you look at the man pages for gluperspective). These co-efficients should scale and bias v x, v y, and v z into clip space while assigning v z to c w. To transform from view coordinate v to clip coordinates c: So solving for v we get: P v = c = v = c z d av x bv y cv z + dv w ev z c xa cy b c we ccw de (9) (10) Encoding Equation 10 in matrix form gives us the inverse projection matrix: 1 0 0 0 a P 1 = 0 1 0 0 b 1 0 0 0 (11) e 0 0 1 c d de Computing the view coordinate from a clip coordinate is now simply: P 1 c = v (12) 5

There s no guarantee that v w will be 1, so we ll want to rescale v appropriately: v = v v w (13) 3.3 V iew Model Finally we just need to go from view coordinates to world coordinates by multiplying the view coordinates against the inverse of the modelview matrix. Again we can avoid doing a true inverse if we just logically break down what the modelview transform accomplishes when working with the camera: it is a translation (centering the universe around the camera) and then a rotation (to reflect the camera s orientation). The inverse of this is reversed rotation (accomplished with a transpose) followed by a translation with the negation of the modelview matrix s translation component after it has been rotated by the inverse rotation. If given our initial modelview matrix M consisting of a 3x3 rotation submatrix R and a 3-element translation vector t: R 11 R 12 R 13 t x M = R 21 R 22 R 23 t y R 31 R 32 R 33 t z (14) 0 0 0 1 Then we can construct the inverse modelview M 1 using the transpose of the rotation submatrix R T and the camera s translation vector t. R T t = t (15) R T 11 R T 12 R T 13 t x M 1 = R T 21 R T 22 R T 23 t y R T 31 R T 32 R T 33 t (16) z 0 0 0 1 If you re specifying the modelview matrix directly, for example by using glloadmatrix, then you already have it lying around and you can build the inverse as described in Equation 15. If, on the other hand, the modelview matrix is built dynamically using something like glulookat or a sequence of gltranslate, glrotate, and glscale calls, you can use glgetfloatv to retrieve the current modelview matrix. 6

Now that we have the inverse modelview matrix we can use it to transform our view coordinate v into world space, giving us w. M 1 v = w (17) If depth value under the mouse was used to construct the original viewport coordinate, then w should correspond to the point in 3-space where the user clicked. If the depth value was not read then we have an arbitrary point in space with which we can construct a ray from the viewer s position, a: r = a + t(w a) (18) However, there s a trick we can use to skip Equation 18 altogether. Setting v w in Equation 17 to 0 right before the inverse modelview transformation any translation components are removed. This means we ll be taking a ray in view coordinates and getting a ray back in world coordinates. Of course this is only relevant if we re trying to compute a pick ray instead of back projecting an actual point in space. 4 Picking We should now have one of two things: either an actual point in world space corresponding to the location of the mouse click, or a world space ray representing the direction of the mouse click. If we have an actual point we can search against all geometry in the world and see which piece it s closest to. If we have the ray, then we ll need to perform an intersection test between the ray and the world geometry and find the geometry closest to the near Z clip plane. Either method should be reasonably simple to implement. 5 Conclusion Picking objects in a 3D scene using a mouse is a common task, but there are very few papers that describe pragmatic approaches to accomplishing this. Hopefully this paper helps someone trying to muddle through this on their own, God knows I could have used it a few weeks ago. 7

6 Greets A shout out and greets to my boys Casey Muratori and Nichola Vining for helping me sort through this shit and hopefully not sounding like a dumbass. Yo. 8