ERAD Optical flow in radar images. Proceedings of ERAD (2004): c Copernicus GmbH 2004
|
|
- Tobias Dawson
- 5 years ago
- Views:
Transcription
1 Proceedings of ERAD (2004): c Copernicus GmbH 2004 ERAD 2004 Optical flow in radar images M. Peura and H. Hohti Finnish Meteorological Institute, P.O. Box 503, FIN Helsinki, Finland Abstract. Optical flow is one of the standard computer vision methods for computing motion vectors. The key idea is to derive motion vectors from intensity gradients and differences of subsequent images. This model is analogous to advection equations in meteorology which motivates testing the method for weather radar data. Also the computational simplicity of optical flow makes it an interesting alternative to autocorrelation based methods. Our final target is to apply motion vectors in nowcasting precipitation for 0 3 h. (One should keep in mind that motion of a precipitation area generally differs from Doppler measurable motion of droplets.) Related products typically exceed pixels in size and the maximal allowed processing time is below a minute for deriving the motion vectors. In this paper, we briefly review the optical flow model and suggest techniques for fast, quality weighted computing in extraction of motion vectors. We emphasize weather radar data specific issues such as the discontinuity and multi-scale nature of precipitation. Nevertheless, our first experiments suggest that the proposed techniques are applicable also in a wider context of motion extraction problems. 1 Introduction Nowcasting of precipitation is the most central application in operational radar meteorology. As to automated nowcasting of the next couple of hours, direct radar data extrapolation techniques are typically superior to physical models applied in numerical weather prediction (NWP). Principally, generalpurpose image extrapolation schemes can be considered as far as attention is paid on application specific constraints. In processing weather radar data, we would like to stress the following three issues. Correspondence to: M. Peura (Markus.Peura@fmi.fi) First, accepted processing time for radar image products is typically from one to five minutes. Hence, the applied algorithms should be computationally light. Second, the objects appearing in radar data are distinct, separated by large clear areas. This is due to discontinuous nature of precipitation and also due to sensitivity limits of radar. However, motion fields are continuous; motion vectors should be assigned also to echo free areas. Third, precipitation objects have large differences in size, ranging from convective cells in half-kilometre scale to precipitation areas hundreds of kilometres wide. A motion detection algorithm should however work consistently regardless of target size. This far, for extracting motion in image sequences, a standard technique has been to apply autocorrelation due to its conceptual clarity and robustness (Holmlund, 1998). However, an alternative approach, the optical flow technique (Sonka et al., 1993) seems to be simpler and hence faster than autocorrelation. While autocorrelation is essentially a matching technique, optical flow applies differential computing that requires fewer computation loops. The problem remains to find out whether this speedup compromises the quality of results. An excellent discussion on different optical flow methods is presented by J. L. Barron and Fleet (1994). We apply a general optimization problem outlined in that paper and adopt respective notations. In this paper, we focus on computational speed and utilization of quality information. We start by reviewing the optical flow model (Sect. 2) and outline our modified scheme for weather radar data (Sect. 3). In Sect. 3.1, we show how radar data can be treated as a continuous flow by smoothing input images. Both pre-smoothing and actual optical flow computation can be accelerated by sliding window techniques discussed in Sec The proposed computation scheme supports generating and using quality information in both extraction and application of motion vectors; two quality descriptors are suggested in Sect. 3.3.
2 M. Peura and H. Hohti: Optical flow in radar images 455 Fig. 1. The proposed overall scheme. 2 Optical flow f 1 (5th July 2004, 14:30) f (5th July 2004, 14:45) Fig. 2. Sample images from FMI Anjalankoski radar. In optical flow, the basic idea is to derive a continuous motion field a flow from temporal and spatial derivatives of two subsequent images Sonka et al. (1993). Formally, flow of quantity f = f (x, y, t) can be modelled as df dt = f t + f x u + f y v = f t + f v. (1) where df/dt is the change observed in the coordinate system flowing with data and the partial derivatives indicate the changes in the rigid image frame coordinates. Motion vector v = [u v] T is the quantity to be derived. This model is analogous to meteorological advection equations, which suggests using this approach in related problems as well. Practically, one assumes no changes in the data, df/dt = 0, and hence f v + f t = 0. This single equation contains two unknowns, u and v, which suggest solving a larger set of respective equations within some neighborhood of each (x, y). This overdetermined set leads to minimizing a squared sum of type w ( f v + f t ) 2 (2) where w is a weighting function defined in image coordinates and/or neighborhood coordinates. The role and definition of w is discussed further in Sect After some matrix manipulation, we obtain [ ] 1 Gyy G v = xt G xy G yt G 2 xy G (3) xxg yy G xy G xt + G xx G yt where G xx = wf x f x, G xy = wf y f x, G yy = wfy f y, G xt = wf x f t, and G yt = wf y f t (within ).
3 456 M. Peura and H. Hohti: Optical flow in radar images Fig. 3. Image f obtained through multi-scale-smoothing of f (Fig. 2, bottom). Fig. 4. Gradient quality image w(x, y) of f. Light areas contain pixels of high quality. 3 Proposed modifications The proposed scheme is shown in Fig. 1, illustrating the separate roles of motion vector extraction and application. The details discussed in this section refer to this image. Likewise, we shall focus on two sample images shown in Fig Pre-smoothing Direct application of optical flow solution (3) is possible for data that are flow-like: smoothly continuous thus differentiable. Optical flow approach needs gradient information to work. However, weather radar data contains objects with distinct edges, and the objects are separated from each other by echo-free areas. This suggests that information of the edges should be somehow spread into echo-free areas. This can be achieved Fig. 5. Motion vector quality q (top), motion vector component u (center) and its q-weighted-averaged version u (bottom). for example by smoothing input images with an averaging convolution mask. As a result, not only the effective scope of the objects will increase but also discrete edges will become more differentiable. Specifically, we suggest multiscale smoothing, which means blurring the original image f
4 M. Peura and H. Hohti: Optical flow in radar images 457 f +3 f +12 f +6 f +15 Fig. 6. Frames from an image series extrapolated from f 1 and f. f +9 f +18
5 458 M. Peura and H. Hohti: Optical flow in radar images with an averaging window operator A and mixing the result in the original image: f = cf + (1 c)a{f } with some c ]0, 1[. (One can apply this recursively.) As shown in Fig. 3, the resulting image contains both original details and smoothed regions. As a rule of thumb, the dimensions of the averaging window should be comparable to those of the object displacements between subsequent image frames. 3.2 Accelerating computation by algorithm design Image processing operations engaging computation of cumulants can be often accelerated. For example, performing the above-mentioned smoothing for an N N image with an n n window requires a computational effort proportional to N 2 n 2 using a brute force algorithm (with four nesting loops). However, the same result can be obtained by two subsequent averaging operations of window dimensions n 1 and 1 n, with effort proportional to N 2 n. Moreover, using a sliding window technique, that is, by incrementally updating the intensity sum in a window, the computational complexity is of order N 2 +2n N 2 which is remarkable in the case of large n. Also the actual optical flow computation can be accelerated. The crucial point here is that if w = w(x, y, t) in image coordinates, also the cumulants in (3) can be updated using a sliding-window technique, yielding complexity of order N 2 n instead of N 2 n 2. (If w = w(i, j) in coordinates, sliding is impossible.) 3.3 Weighting computation with quality descriptors If available, quality information should be taken into account in computations, especially in compositing multi-source data but also in spatial or temporal interpolations of single-source data. Sometimes quality information is available for input data, sometimes a computation scheme provides it as a byproduct, with negligible extra effort, and sometimes it is a central component of a system (Holmlund, 1998; Peura, 2002). In this context, we consider quality information that shares the same format with the data, i.e. an image array. Then, one can apply weighted averaging or weighted median filtering for the original image. As a result, data with high quality smoothly overrides (spatially or temporally adjacent) data of lower quality. In applying optical flow, there seems to be at least two stages where to use data quality. First, weighting function w mentioned in (2) and (3) should describe the quality that is, usability or confidence of each image location (x, y) for the optical flow algorithm. We suggest that this function is a gradient stability measure of the form w(x, y) = w o ( f (x, y) f 1 (x, y) ) where w o is an increasing function. A related image is shown in Fig. 4. Zero or small gradient change tells that that we are probably still on the same slope while changed gradient indicates that the slope has gone past (x, y). Second, there is a simple way to associate quality information to motion vectors. If neighborhood window is small compared to the size of details in image data, the gradient values inside are nearly same, and the calculation of a motion vector is ill-determined. This is the so called aperture problem. The nominator in (3) is actually an inverted and negated determinant (Peura and Hohti, 2004); the value of the determinant is zero if the motion vector is ambiguous. Hence, we propose a monotonous mapping of type q(x, y) = q o (G xx G yy G 2 xy ) where q o is an increasing function. A related image is shown in Fig Extrapolation As we mentioned in the introduction, we aim at nowcasting products based on direct extrapolation of weather radar data. Samples from an image series extrapolated from the two images of Fig. 2 are shown in Fig Conclusions We presented how optical flow approach can be applied in extracting motion vectors in radar images. Pre-smoothing is recommended, because radar images contain discontinuities at the edges of precipitation areas. The basic optical flow method is simple as such, but with minor efforts in algorithm design it can made even faster. Hence, in time critical applications, optical flow should be considered as an alternative to autocorrelation based methods. As to quality computation, optical flow algorithm both applies and produces quality information fluently. The results obtained this far encourage continuing this research. In the future, we will study parallel processing for convective and widespread rain as well as detection of divergence and converge. References Holmlund, K.: The Utilization of Statistical Properties of Satellite- Derived Atmospheric Motion Vectors to Derive Quality Indicators, Weather and Forecasting, pp , Barron, S. S. B. and Fleet, D. J.: On Optical Flow, in 6th Int. Conf. on Artificial Intelligence and Information-Control Systems of Robots (AIICSR), edited by I. Plander, pp. 3 14, World Scientific, Bratislava, Slovakia, citeseer.nj.nec.com/ article/barron94optical.html, sept , 1994, Smolenice Castle, Slovakia, Peura, M.: Computer vision methods for anomaly removal, in Second European Conference on Radar Meteorology (ERAD02), pp , Copernicus Gesellschaft, Peura, M. and Hohti, H.: Motion vectors in Weather Radar Images, in Seventh International Winds Workshop, Eumetsat, to appear, Sonka, M., Hlavac, V., and Boyle, R.: Image Processing, Analysis and Computer Vision, Chapman & Hall Computing, 1993.
Rack a program for anomaly detection, product generation, and compositing
Rack a program for anomaly detection, product generation, and compositing Markus Peura Finnish Meteorological Institute, Helsinki, Finland (first.last@fmi.fi) (Dated: May 30, 2012 ) 1 Introduction Rack
More informationMatching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.
Matching Compare region of image to region of image. We talked about this for stereo. Important for motion. Epipolar constraint unknown. But motion small. Recognition Find object in image. Recognize object.
More informationHand-Eye Calibration from Image Derivatives
Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed
More informationSpatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University
Spatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University 1 Outline of This Week Last topic, we learned: Spatial autocorrelation of areal data Spatial regression
More informationVisual Tracking (1) Tracking of Feature Points and Planar Rigid Objects
Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationDisplacement estimation
Displacement estimation Displacement estimation by block matching" l Search strategies" l Subpixel estimation" Gradient-based displacement estimation ( optical flow )" l Lukas-Kanade" l Multi-scale coarse-to-fine"
More informationEE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm
EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm Group 1: Mina A. Makar Stanford University mamakar@stanford.edu Abstract In this report, we investigate the application of the Scale-Invariant
More informationMotion Estimation. There are three main types (or applications) of motion estimation:
Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion
More informationMarcel Worring Intelligent Sensory Information Systems
Marcel Worring worring@science.uva.nl Intelligent Sensory Information Systems University of Amsterdam Information and Communication Technology archives of documentaries, film, or training material, video
More informationMotion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi
Motion and Optical Flow Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationComparison Between The Optical Flow Computational Techniques
Comparison Between The Optical Flow Computational Techniques Sri Devi Thota #1, Kanaka Sunanda Vemulapalli* 2, Kartheek Chintalapati* 3, Phanindra Sai Srinivas Gudipudi* 4 # Associate Professor, Dept.
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More informationOutlines. Medical Image Processing Using Transforms. 4. Transform in image space
Medical Image Processing Using Transforms Hongmei Zhu, Ph.D Department of Mathematics & Statistics York University hmzhu@yorku.ca Outlines Image Quality Gray value transforms Histogram processing Transforms
More informationMoving Object Tracking in Video Using MATLAB
Moving Object Tracking in Video Using MATLAB Bhavana C. Bendale, Prof. Anil R. Karwankar Abstract In this paper a method is described for tracking moving objects from a sequence of video frame. This method
More informationThe Lucas & Kanade Algorithm
The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest
More informationCOMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE
COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationScientific Visualization Example exam questions with commented answers
Scientific Visualization Example exam questions with commented answers The theoretical part of this course is evaluated by means of a multiple- choice exam. The questions cover the material mentioned during
More informationIntroduction to Digital Image Processing
Introduction to Digital Image Processing Ranga Rodrigo June 9, 29 Outline Contents Introduction 2 Point Operations 2 Histogram Processing 5 Introduction We can process images either in spatial domain or
More informationOutline 7/2/201011/6/
Outline Pattern recognition in computer vision Background on the development of SIFT SIFT algorithm and some of its variations Computational considerations (SURF) Potential improvement Summary 01 2 Pattern
More informationVIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural
VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD Ertem Tuncel and Levent Onural Electrical and Electronics Engineering Department, Bilkent University, TR-06533, Ankara, Turkey
More informationC18 Computer vision. C18 Computer Vision. This time... Introduction. Outline.
C18 Computer Vision. This time... 1. Introduction; imaging geometry; camera calibration. 2. Salient feature detection edges, line and corners. 3. Recovering 3D from two images I: epipolar geometry. C18
More informationSURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD
SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD M.E-II, Department of Computer Engineering, PICT, Pune ABSTRACT: Optical flow as an image processing technique finds its applications
More informationVisual Tracking (1) Feature Point Tracking and Block Matching
Intelligent Control Systems Visual Tracking (1) Feature Point Tracking and Block Matching Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationExperiments with Edge Detection using One-dimensional Surface Fitting
Experiments with Edge Detection using One-dimensional Surface Fitting Gabor Terei, Jorge Luis Nunes e Silva Brito The Ohio State University, Department of Geodetic Science and Surveying 1958 Neil Avenue,
More informationINVESTIGATIONS OF CROSS-CORRELATION AND EUCLIDEAN DISTANCE TARGET MATCHING TECHNIQUES IN THE MPEF ENVIRONMENT. Greg And Ken Holmlund # ABSTRACT
INVESTIGATIONS OF CROSS-CORRELATION AND EUCLIDEAN DISTANCE TARGET MATCHING TECHNIQUES IN THE MPEF ENVIRONME Greg Dew @ And Ken Holmlund # @ Logica # EUMETSAT ABSTRACT Cross-Correlation and Euclidean Distance
More informationSIFT - scale-invariant feature transform Konrad Schindler
SIFT - scale-invariant feature transform Konrad Schindler Institute of Geodesy and Photogrammetry Invariant interest points Goal match points between images with very different scale, orientation, projective
More informationImage Processing. Filtering. Slide 1
Image Processing Filtering Slide 1 Preliminary Image generation Original Noise Image restoration Result Slide 2 Preliminary Classic application: denoising However: Denoising is much more than a simple
More informationComputer Vision for HCI. Topics of This Lecture
Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014
SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image
More informationBabu Madhav Institute of Information Technology Years Integrated M.Sc.(IT)(Semester - 7)
5 Years Integrated M.Sc.(IT)(Semester - 7) 060010707 Digital Image Processing UNIT 1 Introduction to Image Processing Q: 1 Answer in short. 1. What is digital image? 1. Define pixel or picture element?
More informationCHAPTER 5 MOTION DETECTION AND ANALYSIS
CHAPTER 5 MOTION DETECTION AND ANALYSIS 5.1. Introduction: Motion processing is gaining an intense attention from the researchers with the progress in motion studies and processing competence. A series
More informationLOCAL-GLOBAL OPTICAL FLOW FOR IMAGE REGISTRATION
LOCAL-GLOBAL OPTICAL FLOW FOR IMAGE REGISTRATION Ammar Zayouna Richard Comley Daming Shi Middlesex University School of Engineering and Information Sciences Middlesex University, London NW4 4BT, UK A.Zayouna@mdx.ac.uk
More informationVC 11/12 T11 Optical Flow
VC 11/12 T11 Optical Flow Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos Miguel Tavares Coimbra Outline Optical Flow Constraint Equation Aperture
More informationEECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline
EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT Oct. 15, 2013 Prof. Ronald Fearing Electrical Engineering and Computer Sciences University of California, Berkeley (slides courtesy of Prof. John Wawrzynek)
More informationDense Image-based Motion Estimation Algorithms & Optical Flow
Dense mage-based Motion Estimation Algorithms & Optical Flow Video A video is a sequence of frames captured at different times The video data is a function of v time (t) v space (x,y) ntroduction to motion
More informationMotion Estimation for Video Coding Standards
Motion Estimation for Video Coding Standards Prof. Ja-Ling Wu Department of Computer Science and Information Engineering National Taiwan University Introduction of Motion Estimation The goal of video compression
More informationLocal Image Registration: An Adaptive Filtering Framework
Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,
More informationCorrelation and Relaxation Labelling---An Experimental Investigation on Fast Algorithms
Correlation and Relaxation Labelling---An Experimental Investigation on Fast Algorithms Q. X. Wu*, S. J. McNeill, & D. Pairman Landcare Research New Zealand PO Box 38491, Wellington Mail Centre New Zealand
More informationThe SIFT (Scale Invariant Feature
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical
More informationOptical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.
Optical Flow-Based Motion Estimation Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. 1 Why estimate motion? We live in a 4-D world Wide applications Object
More informationCHAPTER 3 IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN
CHAPTER 3 IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN CHAPTER 3: IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN Principal objective: to process an image so that the result is more suitable than the original image
More informationELEC Dr Reji Mathew Electrical Engineering UNSW
ELEC 4622 Dr Reji Mathew Electrical Engineering UNSW Review of Motion Modelling and Estimation Introduction to Motion Modelling & Estimation Forward Motion Backward Motion Block Motion Estimation Motion
More informationECE 176 Digital Image Processing Handout #14 Pamela Cosman 4/29/05 TEXTURE ANALYSIS
ECE 176 Digital Image Processing Handout #14 Pamela Cosman 4/29/ TEXTURE ANALYSIS Texture analysis is covered very briefly in Gonzalez and Woods, pages 66 671. This handout is intended to supplement that
More informationChapter 9 Object Tracking an Overview
Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging
More informationPeripheral drift illusion
Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationOn-line and Off-line 3D Reconstruction for Crisis Management Applications
On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be
More informationCoarse-to-fine image registration
Today we will look at a few important topics in scale space in computer vision, in particular, coarseto-fine approaches, and the SIFT feature descriptor. I will present only the main ideas here to give
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationNon-Rigid Image Registration III
Non-Rigid Image Registration III CS6240 Multimedia Analysis Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (CS6240) Non-Rigid Image Registration
More informationLocal Features: Detection, Description & Matching
Local Features: Detection, Description & Matching Lecture 08 Computer Vision Material Citations Dr George Stockman Professor Emeritus, Michigan State University Dr David Lowe Professor, University of British
More informationImage Enhancement in Spatial Domain (Chapter 3)
Image Enhancement in Spatial Domain (Chapter 3) Yun Q. Shi shi@njit.edu Fall 11 Mask/Neighborhood Processing ECE643 2 1 Point Processing ECE643 3 Image Negatives S = (L 1) - r (3.2-1) Point processing
More informationCorner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology
Corner Detection Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology rhody@cis.rit.edu April 11, 2006 Abstract Corners and edges are two of the most important geometrical
More informationVivekananda. Collegee of Engineering & Technology. Question and Answers on 10CS762 /10IS762 UNIT- 5 : IMAGE ENHANCEMENT.
Vivekananda Collegee of Engineering & Technology Question and Answers on 10CS762 /10IS762 UNIT- 5 : IMAGE ENHANCEMENT Dept. Prepared by Harivinod N Assistant Professor, of Computer Science and Engineering,
More informationSchool of Computing University of Utah
School of Computing University of Utah Presentation Outline 1 2 3 4 Main paper to be discussed David G. Lowe, Distinctive Image Features from Scale-Invariant Keypoints, IJCV, 2004. How to find useful keypoints?
More informationComputer Vision I - Basics of Image Processing Part 2
Computer Vision I - Basics of Image Processing Part 2 Carsten Rother 07/11/2014 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationAnnouncements. Edge Detection. An Isotropic Gaussian. Filters are templates. Assignment 2 on tracking due this Friday Midterm: Tuesday, May 3.
Announcements Edge Detection Introduction to Computer Vision CSE 152 Lecture 9 Assignment 2 on tracking due this Friday Midterm: Tuesday, May 3. Reading from textbook An Isotropic Gaussian The picture
More informationAn Approach for Reduction of Rain Streaks from a Single Image
An Approach for Reduction of Rain Streaks from a Single Image Vijayakumar Majjagi 1, Netravati U M 2 1 4 th Semester, M. Tech, Digital Electronics, Department of Electronics and Communication G M Institute
More informationLecture 20: Tracking. Tuesday, Nov 27
Lecture 20: Tracking Tuesday, Nov 27 Paper reviews Thorough summary in your own words Main contribution Strengths? Weaknesses? How convincing are the experiments? Suggestions to improve them? Extensions?
More informationVisual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.
Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania 1 What is visual tracking? estimation of the target location over time 2 applications Six main areas:
More informationArbib: Slides for TMB2 Section 7.2 1
Arbib: Slides for TMB2 Section 7.2 1 Lecture 20: Optic Flow Reading assignment: TMB2 7.2 If, as we walk forward, we recognize that a tree appears to be getting bigger, we can infer that the tree is in
More informationALGORITHM OF PRELIMINARY SESSILE DROP PROCESSING FOR NOISE SUPPRESSION X
ALGORITHM OF PRELIMINARY SESSILE DROP PROCESSING FOR NOISE SUPPRESSION X Y.P. KULYNYCH, L.I. MURAVSKY, R.S. BATCHEVSKY, O.G. KUTS Physic-Mechanical Institute of the National Academy of Sciences of Ukraine,
More informationAN ADAPTIVE MESH METHOD FOR OBJECT TRACKING
AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING Mahdi Koohi 1 and Abbas Shakery 2 1 Department of Computer Engineering, Islamic Azad University, Shahr-e-Qods Branch,Tehran,Iran m.kohy@yahoo.com 2 Department
More informationA Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation
, pp.162-167 http://dx.doi.org/10.14257/astl.2016.138.33 A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation Liqiang Hu, Chaofeng He Shijiazhuang Tiedao University,
More informationCS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow
CS 565 Computer Vision Nazar Khan PUCIT Lectures 15 and 16: Optic Flow Introduction Basic Problem given: image sequence f(x, y, z), where (x, y) specifies the location and z denotes time wanted: displacement
More informationEdge and corner detection
Edge and corner detection Prof. Stricker Doz. G. Bleser Computer Vision: Object and People Tracking Goals Where is the information in an image? How is an object characterized? How can I find measurements
More informationRegularization and Markov Random Fields (MRF) CS 664 Spring 2008
Regularization and Markov Random Fields (MRF) CS 664 Spring 2008 Regularization in Low Level Vision Low level vision problems concerned with estimating some quantity at each pixel Visual motion (u(x,y),v(x,y))
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationMotion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation
Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion
More informationMotion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures
Now we will talk about Motion Analysis Motion analysis Motion analysis is dealing with three main groups of motionrelated problems: Motion detection Moving object detection and location. Derivation of
More informationVideo Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin
Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods
More informationBlur Space Iterative De-blurring
Blur Space Iterative De-blurring RADU CIPRIAN BILCU 1, MEJDI TRIMECHE 2, SAKARI ALENIUS 3, MARKKU VEHVILAINEN 4 1,2,3,4 Multimedia Technologies Laboratory, Nokia Research Center Visiokatu 1, FIN-33720,
More informationDigital Image Processing
Digital Image Processing Lecture # 6 Image Enhancement in Spatial Domain- II ALI JAVED Lecturer SOFTWARE ENGINEERING DEPARTMENT U.E.T TAXILA Email:: ali.javed@uettaxila.edu.pk Office Room #:: 7 Local/
More informationProceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives
Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns CAIP'95, pp. 874-879, Prague, Czech Republic, Sep 1995 Direct Obstacle Detection and Motion from Spatio-Temporal Derivatives
More informationVisual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys
Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes Motion and perceptual organization Sometimes, motion is the onl cue Motion and perceptual organization Sometimes, motion is the
More informationVisual Tracking (1) Pixel-intensity-based methods
Intelligent Control Systems Visual Tracking (1) Pixel-intensity-based methods Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationTracking Weather Storms using 3D Doppler Radial Velocity Information
Tracking Weather Storms using 3D Doppler Radial Velocity Information X. Tang 1 and J.L. Barron 1,R.E.Mercer 1,andP.Joe 2 1 Department of Computer Science University of Western Ontario London, Ontario,
More informationFace Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm
Face Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm Dirk W. Wagener, Ben Herbst Department of Applied Mathematics, University of Stellenbosch, Private Bag X1, Matieland 762,
More informationDigital Image Processing (CS/ECE 545) Lecture 5: Edge Detection (Part 2) & Corner Detection
Digital Image Processing (CS/ECE 545) Lecture 5: Edge Detection (Part 2) & Corner Detection Prof Emmanuel Agu Computer Science Dept. Worcester Polytechnic Institute (WPI) Recall: Edge Detection Image processing
More informationDynamic Texture Recognition Using Normal Flow and Texture Regularity
223 Dynamic Texture Recognition Using Normal Flow and Texture Regularity Renaud Péteri 1 and Dmitry Chetverikov 2 1 Centre for Mathematics and Computer Science (CWI), Kruislaan 413, 1098SJ Amsterdam, The
More informationMotion Estimation (II) Ce Liu Microsoft Research New England
Motion Estimation (II) Ce Liu celiu@microsoft.com Microsoft Research New England Last time Motion perception Motion representation Parametric motion: Lucas-Kanade T I x du dv = I x I T x I y I x T I y
More informationCS4442/9542b Artificial Intelligence II prof. Olga Veksler
CS4442/9542b Artificial Intelligence II prof. Olga Veksler Lecture 2 Computer Vision Introduction, Filtering Some slides from: D. Jacobs, D. Lowe, S. Seitz, A.Efros, X. Li, R. Fergus, J. Hayes, S. Lazebnik,
More informationSpatio-Temporal Stereo Disparity Integration
Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 WRI C225 Lecture 04 130131 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Histogram Equalization Image Filtering Linear
More informationImage features. Image Features
Image features Image features, such as edges and interest points, provide rich information on the image content. They correspond to local regions in the image and are fundamental in many applications in
More informationEdge Detection Lecture 03 Computer Vision
Edge Detection Lecture 3 Computer Vision Suggested readings Chapter 5 Linda G. Shapiro and George Stockman, Computer Vision, Upper Saddle River, NJ, Prentice Hall,. Chapter David A. Forsyth and Jean Ponce,
More informationpacket-switched networks. For example, multimedia applications which process
Chapter 1 Introduction There are applications which require distributed clock synchronization over packet-switched networks. For example, multimedia applications which process time-sensitive information
More informationComputer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier
Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 1. IMAGE PROCESSING Computer Vision 2 Dr. Benjamin Guthier Content of this Chapter Non-linear
More informationHomogenization and numerical Upscaling. Unsaturated flow and two-phase flow
Homogenization and numerical Upscaling Unsaturated flow and two-phase flow Insa Neuweiler Institute of Hydromechanics, University of Stuttgart Outline Block 1: Introduction and Repetition Homogenization
More informationComputer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.
Announcements Edge and Corner Detection HW3 assigned CSE252A Lecture 13 Efficient Implementation Both, the Box filter and the Gaussian filter are separable: First convolve each row of input image I with
More informationLecture 2.2 Cubic Splines
Lecture. Cubic Splines Cubic Spline The equation for a single parametric cubic spline segment is given by 4 i t Bit t t t i (..) where t and t are the parameter values at the beginning and end of the segment.
More informationAN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe
AN EFFICIENT BINARY CORNER DETECTOR P. Saeedi, P. Lawrence and D. Lowe Department of Electrical and Computer Engineering, Department of Computer Science University of British Columbia Vancouver, BC, V6T
More informationOptic Flow and Basics Towards Horn-Schunck 1
Optic Flow and Basics Towards Horn-Schunck 1 Lecture 7 See Section 4.1 and Beginning of 4.2 in Reinhard Klette: Concise Computer Vision Springer-Verlag, London, 2014 1 See last slide for copyright information.
More informationKeeping features in the camera s field of view: a visual servoing strategy
Keeping features in the camera s field of view: a visual servoing strategy G. Chesi, K. Hashimoto,D.Prattichizzo,A.Vicino Department of Information Engineering, University of Siena Via Roma 6, 3 Siena,
More informationImage Differentiation
Image Differentiation Carlo Tomasi September 4, 207 Many image operations, including edge detection and motion analysis in video, require computing the derivatives of image intensity with respect to the
More informationContent-Based Image Retrieval of Web Surface Defects with PicSOM
Content-Based Image Retrieval of Web Surface Defects with PicSOM Rami Rautkorpi and Jukka Iivarinen Helsinki University of Technology Laboratory of Computer and Information Science P.O. Box 54, FIN-25
More informationLeow Wee Kheng CS4243 Computer Vision and Pattern Recognition. Motion Tracking. CS4243 Motion Tracking 1
Leow Wee Kheng CS4243 Computer Vision and Pattern Recognition Motion Tracking CS4243 Motion Tracking 1 Changes are everywhere! CS4243 Motion Tracking 2 Illumination change CS4243 Motion Tracking 3 Shape
More information