Practice Exam Sample Solutions

Similar documents
Final Exam Study Guide

Types of Edges. Why Edge Detection? Types of Edges. Edge Detection. Gradient. Edge Detection

ENG 7854 / 9804 Industrial Machine Vision. Midterm Exam March 1, 2010.

lecture 10 - depth from blur, binocular stereo

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

(Sample) Final Exam with brief answers

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Physics 345 Pre-Lab 5 Lenses Part II

COMPUTER AND ROBOT VISION

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

Lecture 9: Hough Transform and Thresholding base Segmentation

CS 787: Assignment 4, Stereo Vision: Block Matching and Dynamic Programming Due: 12:00noon, Fri. Mar. 30, 2007.

Segmentation I: Edges and Lines

Lecture #11: The Perceptron

A function: A mathematical relationship between two variables (x and y), where every input value (usually x) has one output value (usually y)

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Chapter 1. Linear Equations and Straight Lines. 2 of 71. Copyright 2014, 2010, 2007 Pearson Education, Inc.

CPSC 425: Computer Vision

Perspective and vanishing points

CS201 Computer Vision Lect 4 - Image Formation

CIRCULAR MOIRÉ PATTERNS IN 3D COMPUTER VISION APPLICATIONS

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Final Review CMSC 733 Fall 2014

Lecture 15: Segmentation (Edge Based, Hough Transform)

Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy

Introduction to 3D Imaging: Perceiving 3D from 2D Images

Supervised Learning in Neural Networks (Part 2)

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

Robert Collins CSE486, Penn State Lecture 08: Introduction to Stereo

Other Linear Filters CS 211A

Range Sensors (time of flight) (1)

Stereo Vision. MAN-522 Computer Vision

CS4495/6495 Introduction to Computer Vision

Edge and local feature detection - 2. Importance of edge detection in computer vision

Image Processing: Final Exam November 10, :30 10:30

Geometry of Multiple views

Unit 1, Lesson 1: Moving in the Plane

Working with Transformations on the Coordinate Plane

Edge and corner detection

Draw Guide. Chapter 7 Working with 3D Objects

Miniature faking. In close-up photo, the depth of field is limited.

COS Lecture 10 Autonomous Robot Navigation

Back propagation Algorithm:

IN5520 Digital Image Analysis. Two old exams. Practical information for any written exam Exam 4300/9305, Fritz Albregtsen

Example 1: Regions. Image Segmentation. Example 3: Lines and Circular Arcs. Example 2: Straight Lines. Region Segmentation: Segmentation Criteria

Lecture 14: Computer Vision

Lecture 6: Edge Detection

Binocular cues to depth PSY 310 Greg Francis. Lecture 21. Depth perception

Robot vision review. Martin Jagersand

Grade 6 Math Circles. Spatial and Visual Thinking

Translations SLIDE. Example: If you want to move a figure 5 units to the left and 3 units up we would say (x, y) (x-5, y+3).

Recitation Handout 10: Experiments in Calculus-Based Kinetics

Digital Image Processing COSC 6380/4393

Computer and Machine Vision

Edge detection. Convert a 2D image into a set of curves. Extracts salient features of the scene More compact than pixels

Local Feature Detectors

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Objectives. Materials

CS4758: Rovio Augmented Vision Mapping Project

Lecture 7: Most Common Edge Detectors

Assignment 2: Stereo and 3D Reconstruction from Disparity

CoE4TN4 Image Processing

Chapter - 2 : IMAGE ENHANCEMENT

1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)

Announcements. Hough Transform [ Patented 1962 ] Generalized Hough Transform, line fitting. Assignment 2: Due today Midterm: Thursday, May 5 in class

3.5 Equations of Lines and Planes

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

Homework #1. Displays, Image Processing, Affine Transformations, Hierarchical Modeling

Weka ( )

Chapter 3 Image Registration. Chapter 3 Image Registration

Example 2: Straight Lines. Image Segmentation. Example 3: Lines and Circular Arcs. Example 1: Regions

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Basic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Why study Computer Vision?

EE795: Computer Vision and Intelligent Systems

CS 4620 Midterm, March 21, 2017

Multimedia Computing: Algorithms, Systems, and Applications: Edge Detection

Multilayer Feed-forward networks

CS4442/9542b Artificial Intelligence II prof. Olga Veksler

2D/3D Geometric Transformations and Scene Graphs

Computer Vision. Image Segmentation. 10. Segmentation. Computer Engineering, Sejong University. Dongil Han

Think-Pair-Share. What visual or physiological cues help us to perceive 3D shape and depth?

Feature Tracking and Optical Flow

Dietrich Paulus Joachim Hornegger. Pattern Recognition of Images and Speech in C++

Midterm Exam! CS 184: Foundations of Computer Graphics! page 1 of 13!

3D-OBJECT DETECTION METHOD BASED ON THE STEREO IMAGE TRANSFORMATION TO THE COMMON OBSERVATION POINT

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

Homework 3: Shading, Raytracing, and Image Processing

Handout 1: Viewing an Animation

Midterm Exam CS 184: Foundations of Computer Graphics page 1 of 11

Light, Photons, and MRI

Transcription:

CS 675 Computer Vision Instructor: Marc Pomplun Practice Exam Sample Solutions Note that in the actual exam, no calculators, no books, and no notes allowed. Question 1: out of points Question 2: out of points Question 3: out of points Question 4: out of points Question 5: out of points Question 6: out of points Question 7: out of points Question 8: out of points Total Score: Grade:

Question 1: True or False? Tell whether each of the following statements is true or false by checking the appropriate box. Do not check any box if you do not know the right answer, because you will lose points for incorrect answers. Statement True False a) For mean shift segmentation, we need to know the number of [ ] [X] regions in advance.. b) The Hough transform can only be used to detect straight lines. [ ] [X] c) The median filter is especially well-suited for removing salt-andpepper noise. d) The difference code of a curve does not change when we rotate the curve by 90 degrees. e) The Laplacian convolution filter computes the second derivative of the input image. f) When we move backward, the direction of optical flow throughout our visual field is perpendicular to the gradient of its magnitude. g) Textures with the same gray co-occurrence matrices are always identical. h) When we expand a binary image A and then shrink the result, we always obtain image A again. [X] [ ] [X] [ ] [X] [ ] [ ] [X] [ ] [X] [ ] [X] i) Local binary patterns are represented by 256-bin histograms. [ ] [X] j) There is no 2D shape with greater compactness (i.e. a smaller value for P 2 /A) than a circle. [X] [ ]

Question 2: Efficient Range Imaging Imagine that you want to do range imaging through triangulation. In order to be efficient, you decide to use a pattern of eight vertical lines that only needs to be shifted by the distance between neighboring lines in order to measure the depth of all visible points in the scene. In order to avoid associating points with an incorrect vertical line, you take multiple images of the same scene with the pattern in the same position. In each picture, a different subset of lines is projected. a) What is the minimum number of pictures that you need to take for each projection angle in order to be sure to determine the number of the line that any point belongs to? Why? You need 4 pictures. Although 3 pictures would suffice to represent 2 3 = 8 unique on/off patterns, it would mean that one of the lines would be off in all 3 pictures and thus could not be localized. b) Derive a general formula that tells you for a given number of lines the minimum number of pictures necessary. Explain how you derived the formula. Given the answer to (a), it follows that p pictures can uniquely identify up to l = (2 p 1) lines. This gives us: l = 2 p 1 => l + 1 = 2 p => log2(l + 1) = p Since we want to know the minimum number pmin, we use the ceiling function: => log2(l + 1) = p

Question 3: Polyline Splitting Describe the algorithm for polyline splitting to approximate the shape of a contour by a sequence of connected straight lines. Use diagrams to illustrate the concept. Here you should put something similar to slides 2 to 5 from March 6.

Question 4: Slope Density Functions (a) Draw the slope density function of the following contour (let us say that we are going in counterclockwise direction and an angle of 0 means downward): 0 90 180 270 (b) Describe how this function changes if you mirror the contour along a horizontal axis, i.e., turn it upside-down. You can describe it in words or draw another diagram. It is like a 90 counterclockwise rotation, so everything is shifted by 90 to the right (and wrapping around): 0 90 180 270

Question 5: Errors in Depth Perception Having two eyes, we are able to perceive the depth (z-distance) of an object through binocular disparity information. As you know, we can also give our computer vision system two cameras and let it do the same thing. The question is: How accurate is its estimation of depth, and on what factors does this accuracy depend? The main problem here is the limited accuracy and resolution of the cameras. Let us say that the actual position of an object in the camera image may be up to one millimeter to the left or to the right of its actual position. For example, if the left camera measures xl = 5.3 mm, it means that the object s actual position could be anywhere between xl = 5.2 mm and xl = 5.4 mm. (a) Given this camera accuracy, what is the z-range that an object could have (i.e., the minimum and maximum z-distance possible), if the cameras with baseline b = 10 cm and focal length f = 20 cm measure positions xl = 6.1 and xr = 5.1? If you do not remember the equation, try to derive it; it is not very difficult. (b) What do you think will happen if the object is much further away from the system? Will the error in z-distance measurement (i.e., the z-range) increase or decrease? Why? (c) What do you think will happen if we keep the object in the same place as in (a), but increase the distance between the cameras, i.e., the baseline b? Will the error in z- distance measurement (i.e., the z-range) increase or decrease? Why? Now the z-distance error will decrease, because the same variation in the camera image corresponds to smaller variation in the z-distance measurement. It s harder to see in the equations this time, because the baseline, the focal length, and the difference between xl and xr will change. But do the math, it s a great exercise.

(d) If we increase the baseline, another problem becomes more and more difficult. What problem is that? The stereo matching, i.e., determining the correspondence between points in the left and the right image, becomes more difficult. This is because we are now looking at objects from more distinct angles, and so the two images of the same object look less similar. Question 6: Teaching the Artificial Brain Describe the basic principles underlying supervised learning in artificial neural networks. For example, how is the network presented with the information that it is supposed to learn? How does the learning happen, i.e., what changes in the network, and based on what criteria? Use at least two paragraphs to outline the most important ideas. Do not use any equations. To teach the network to compute a function, we need to provide information in the form of exemplars, i.e., desired input-output pairs. If, for example, the network is supposed to output 1 for the input (1, 0, 1), then the corresponding exemplar is ((1, 0, 1), 1). Then we randomly pick one of the exemplars and feed its input part into the input layer neurons of our network. The network then performs all its computations, i.e., each neuron multiplies each of its inputs with the corresponding synaptic weight and applies the output function, usually a sigmoid function, to determine its output. Then we compare the network s output with the desired output specified in the exemplar. If there is a discrepancy, then we use a rule (e.g., the backpropagation learning rule) to change the weights in such a way that next time the same input is presented, the output will be closer to the desired one. Such rules typically use gradient-descent, i.e., they move against the error gradient in weight space. After each exemplar was shown to the network and its weights changed accordingly, one epoch is finished. If the overall network error, i.e., the deviation between desired and actual output values, is still larger than a certain threshold, we perform another epoch, and so on, until the error falls below the threshold. Then the network should have learned the function and be able to predict appropriate output values even for inputs that it has never seen before. (This text is more detailed than what you would need to write in order to get the full score.)

Question 7: Signature, Please! Use the diagram on the right to sketch the signature of the contour on the left. It does not have to be perfectly scaled. Choose any starting point you like and indicate it on the contour on the left. Question 8 (Bonus Question): Tricking the Difference Image Technique In difference images, pixel values greater than zero (or greater than a certain threshold) are thought to reflect moving objects in the scene. However, this does not always have to be the case. There are other factors besides object motion that can cause non-zero values in a difference image. List as many such factors as you can think of. Change in lighting Noise in the images varying over time Motion of the camera