Efficient Block Matching Algorithm for Motion Estimation

Similar documents
Joint Adaptive Block Matching Search (JABMS) Algorithm

Video Compression System for Online Usage Using DCT 1 S.B. Midhun Kumar, 2 Mr.A.Jayakumar M.E 1 UG Student, 2 Associate Professor

Local Image Registration: An Adaptive Filtering Framework

Enhanced Hexagon with Early Termination Algorithm for Motion estimation

A New Fast Motion Estimation Algorithm. - Literature Survey. Instructor: Brian L. Evans. Authors: Yue Chen, Yu Wang, Ying Lu.

Module 7 VIDEO CODING AND MOTION ESTIMATION

Prediction-based Directional Search for Fast Block-Matching Motion Estimation

An Adaptive Cross Search Algorithm for Block Matching Motion Estimation

Video Compression An Introduction

Fast Optical Flow Using Cross Correlation and Shortest-Path Techniques

Adaptive Multi-Stage 2D Image Motion Field Estimation

AN ADJUSTABLE BLOCK MOTION ESTIMATION ALGORITHM BY MULTIPATH SEARCH

EE 5359 Low Complexity H.264 encoder for mobile applications. Thejaswini Purushotham Student I.D.: Date: February 18,2010

Motion Vector Estimation Search using Hexagon-Diamond Pattern for Video Sequences, Grid Point and Block-Based

Fobe Algorithm for Video Processing Applications

Multimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

Motion Estimation for Video Coding Standards

Module 7 VIDEO CODING AND MOTION ESTIMATION

Optical flow and tracking

Redundancy and Correlation: Temporal

Compression of Stereo Images using a Huffman-Zip Scheme

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN

EECS 556 Image Processing W 09

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV

Lecture 7, Video Coding, Motion Compensation Accuracy

Express Letters. A Simple and Efficient Search Algorithm for Block-Matching Motion Estimation. Jianhua Lu and Ming L. Liou

A Novel Hexagonal Search Algorithm for Fast Block Matching Motion Estimation

Chapter 10. Basic Video Compression Techniques Introduction to Video Compression 10.2 Video Compression with Motion Compensation

Motion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi

Context based optimal shape coding

AN EFFICIENT VIDEO WATERMARKING USING COLOR HISTOGRAM ANALYSIS AND BITPLANE IMAGE ARRAYS

Rate Distortion Optimization in Video Compression

Image Compression Algorithm and JPEG Standard

Leow Wee Kheng CS4243 Computer Vision and Pattern Recognition. Motion Tracking. CS4243 Motion Tracking 1

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Compression of Light Field Images using Projective 2-D Warping method and Block matching

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami

Visual Tracking (1) Feature Point Tracking and Block Matching

FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS

EE 264: Image Processing and Reconstruction. Image Motion Estimation I. EE 264: Image Processing and Reconstruction. Outline

arxiv: v1 [cs.cv] 28 Sep 2018

Feature Tracking and Optical Flow

MOTION COMPENSATION IN BLOCK DCT CODING BASED ON PERSPECTIVE WARPING

Semi-Hierarchical Based Motion Estimation Algorithm for the Dirac Video Encoder

Feature Tracking and Optical Flow

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Visual Tracking (1) Pixel-intensity-based methods

Three Dimensional Motion Vectorless Compression

An Efficient Mode Selection Algorithm for H.264

Video Quality Analysis for H.264 Based on Human Visual System

AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING

Frequency Band Coding Mode Selection for Key Frames of Wyner-Ziv Video Coding

Module 05 Motion Analysis / Overview. Advanced Video Analysis and Imaging (5LSH0), Module 05. Motion Applications / Video Coding (1)

IMAGE COMPRESSION USING HYBRID QUANTIZATION METHOD IN JPEG

EE Low Complexity H.264 encoder for mobile applications

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

DIGITAL video compression is essential for the reduction. Two-Bit Transform for Binary Block Motion Estimation

Image Error Concealment Based on Watermarking

Implementation and analysis of Directional DCT in H.264

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

Complexity Reduction Tools for MPEG-2 to H.264 Video Transcoding

Moving Object Tracking in Video Using MATLAB

Digital Image Stabilization and Its Integration with Video Encoder

Comparison Between The Optical Flow Computational Techniques

A High Quality/Low Computational Cost Technique for Block Matching Motion Estimation

Transactions Briefs. An Adaptive Search Length Algorithm for Block Matching Motion Estimation

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

A Robust Two Feature Points Based Depth Estimation Method 1)

Geometric transform motion compensation for low bit. rate video coding. Sergio M. M. de Faria

Fast Natural Feature Tracking for Mobile Augmented Reality Applications

Multiview Image Compression using Algebraic Constraints

Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:

By Charvi Dhoot*, Vincent J. Mooney &,

Fast Motion Estimation for Shape Coding in MPEG-4

EE 5359 MULTIMEDIA PROCESSING SPRING Final Report IMPLEMENTATION AND ANALYSIS OF DIRECTIONAL DISCRETE COSINE TRANSFORM IN H.

MOTION estimation is one of the major techniques for

Video Compression MPEG-4. Market s requirements for Video compression standard

1-2 Feature-Based Image Mosaicing

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

An Improvement of the Occlusion Detection Performance in Sequential Images Using Optical Flow

VIDEO AND IMAGE PROCESSING USING DSP AND PFGA. Chapter 3: Video Processing

Integrating 3D Vision Measurements into Industrial Robot Applications

Image Based Rendering. D.A. Forsyth, with slides from John Hart

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Digital Video Processing

JPEG 2000 vs. JPEG in MPEG Encoding

Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives

A MOTION MODEL BASED VIDEO STABILISATION ALGORITHM

Video Image Sequence Super Resolution using Optical Flow Motion Estimation

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

An Empirical Study of Block Matching Techniques for the Detection of Moving Objects. Abstract

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

MOTION ESTIMATION IN MPEG-2 VIDEO ENCODING USING A PARALLEL BLOCK MATCHING ALGORITHM. Daniel Grosu, Honorius G^almeanu

International Journal of Advanced Research in Computer Science and Software Engineering

Variable Temporal-Length 3-D Discrete Cosine Transform Coding

Transcription:

Efficient Block Matching Algorithm for Motion Estimation Zong Chen International Science Inde Computer and Information Engineering waset.org/publication/1581 Abstract Motion estimation is a key problem in video processing and computer vision. Optical flow motion estimation can achieve high estimation accuracy when motion vector is small. hree-step search algorithm can handle large motion vector but not very accurate. A joint algorithm was proposed in this paper to achieve high estimation accuracy disregarding whether the motion vector is small or large, and keep the computation cost much lower than full search. Keywords Motion estimation, Block Matching, Optical flow, hree step search. I. INRODUCION OION is a prominent source of temporal variations in Mimage sequences. Motion in image sequences acquired by a video camera is induced by movements of objects in a 3D scene and by camera motion. hus, camera parameters, such as its 3D motion (rotation, translation) or focal length, play an important role in image motion modeling. If these parameters are known precisel only object motion needs to be recovered. However, this scenario is rather rare and both object and camera motions usually need to be computed. he 3D motion of objects and camera induces 2D motion on the image plane via a suitable projection system. It is this 2D motion, also called optical flow, which needs to be recovered from intensity and color information of a video sequence. 2D motion finds diverse applications in video processing and compression as well as in computer vision. In video processing, motion information is used for standards conversion (motion-compensated 3D sampling structure conversion), noise suppression (motion-compensated filtering) or deblurring (motion-compensated restoration). In computer vision, 2D motion usually serves as an intermediary in the recovery of camera motion or scene structure. In video compression, the knowledge of motion helps remove temporal data redundancy and therefore attain high compression ratios. In video compression, consecutive video frames are similar except for changes induced by objects moving within the frames. In the trivial case of zero motion between frames (and no other differences caused by noise, etc.), it is easy for the encoder to efficiently predict the current frame as a duplicate of the prediction frame. When this is done, the only information necessary to transmit to the decoder becomes the syntactic overhead necessary to reconstruct the picture from Zong Chen is with School of Computer Sciences and Engineering, Fairleigh Dickinson Universit eaneck, NJ 07666, USA (email: zchen@fdu.edu). the original reference frame. When there is motion in the images, the displacement of moving objects between successive frames will be estimated (motion estimation) first. he resulting motion information is then exploited in efficient inter-frame predictive coding (motion compensation). Consequentl the prediction error is transmitted. he motion information also has to be transmitted, unless the decoder is able to estimate the motion field. An efficient representation of the motion is thus critical tin order to reach high performance in video compression. In the field of motion estimation, many techniques have been applied [1]-[5]. Basicall these techniques can be divided into four main groups: gradient techniques, pixelrecursive techniques, block matching techniques, and frequency-domain techniques. Among those four groups, block matching is particularly suitable in video compression schemes based on discrete cosine transform (DC) such as those adopted by the recent standards H.261, H.263 and MPEG family [6]. Optical flow estimation has been extensively researched. Although most optical flow estimation algorithms are gradient (for each pixel) based algorithms, they also can be applied to blocks. Optical flow methods such as Lucas-Kanade s [7] achieve high accuracy for scenes with small displacements but fail when the displacements are large. In order to achieve high accurate motion estimation without fail when displacements are large, a joint algorithm is proposed in this paper. he rest of the paper is organized as follows. Optical flow algorithm and block matching algorithms will be described in section II and III. he joint algorithm will be presented and the performance will be evaluated in section IV. he simulation results demonstrate the significant achievement of the proposed joint algorithm. II. OPICAL FLOW he hypothesis of optical flow estimation is that the image luminance is invariant between frames. E( X, E( X X, t (1) where X=[ y]. he aylor series expansion of the right hand side of (1) gives E( X X, t E( X, E X Et t (2) where E = [(E/x), (E/y)] is the gradient operator. 2655

International Science Inde Computer and Information Engineering waset.org/publication/1581 Assuming t0, and defining the motion vector v = (v x, v y ) =X/t, we obtain E v E / t 0 (3) Equation (3) is known as the spatio-temporal constraint equation or the optical flow constraint equation. As the image intensity change at a point due to motion gives only one constraint, while the motion vector at the same point has two components, the motion field cannot be computed without an additional constraint. Lucas and Kanade [7] implemented a least square fit of local first-order constraints (3) to a constant model for v by minimizing 2 L( v) [ E( pi ) v E( pi ) / t] (4) he solution of (4) is given by ( A A) v A b (5) where, for a block with n n pixels at time t, A )] [ E( p1 ), E( p2),..., E( p n 2 (6) b ] [ E( p1) / t, E( p2 ) / t,..., E( p n 2 ) / t (7) If the position of p i is ( y) in frame t, then E( x 1, E( E( p i ) (8) E( y 1, E( E( p i ) / t E( t 1) E( (9) For a block with n n pixels, it requires 2n 2 addition to calculate the matrix A (Ex and Ey), and n 2 addition to calculate the vector b (-E t ). he solution of (5) is v = (A A) -1 A b, which is solved in closed form when (A A) is nonsingular. Otherwise, v = (A A) + A b, where (A A) + is the generalized inverse matrix of (A A). In general, for a matrix multiplication A mn B nk = C mk requires mkn multiplication and mk(n-1) addition. herefore, it needs 22n 2 multiplication and 22(n 2 1) addition to calculate (A 2n 2 A 2 2n ), 21n 2 multiplication and 21(n 2 1) addition to calculate (A 2n 2 2 b n 1 ). Also, it needs one 22 matrix inverse for (A A) -1 22, plus 212 multiplication and 211 addition to calculate (A A) -1 22 (A b) 21. herefore, it requires total 6n 2 multiplication and about 9n 2 addition, plus one 22 matrix inverse. III. BLOCK MACHING In block matching algorithms, each macroblock (8 8 pixels) in the new frame is compared with shifted regions of the same size from the previous frame, and the shift which results in the minimum error is selected as the best motion vector for that macro-block. he motion compensated prediction frame is then formed from all the shifted regions from the previous decoded frame. In most systems, the best match is found using the sum of absolute error (SAE) criterion. he SAE of two blocks, X and Y, with n n pixels is defined as n n SAE( X, Y ) X ( i, j) Y ( i, j) (10) i1 j1 he computation cost for each SAE calculation is 2n 2 additions. here are several well-known algorithms that perform the block matching motion estimation. he full search algorithm (FSA) [3]-[5] that determines the motion vector of a macroblock by computing the SAE at each location in the search area (basically 7 pixels). his is the simplest method, it provides the best performance, but at a very high computational cost: 15 15=225 SAE calculation, i.e. 225 2n 2 =450n 2 additions. Significant computation savings can be made with hierarchical search algorithm, also called three-step search algorithm [8]. In three-step search algorithm, eight positions around a center are tested and the position of minimum distortion becomes the center of the next stage. For a center [c cy] and step size s, the positions [cx-s, cy-s], [cx-s, cy], [cx-s, cy+s], [c cy-s], [c cy], [c cy+s], [cx+s, cy-s], [cx+s, cy], [cx+s, cy+s] are examined. After each stage the step size is reduced. Fig 1. An example of three-step search algorithm. he maximum displacement is ±6 pixels in both directions (as in Figure 1) with step size 3, 2, 1. his allows the algorithm to halt in three steps and hence its name. here are also variants allow the maximum displacement to be ±7 pixels with step size 4, 2, 1. hree-step search algorithm needs to calculate total 9 3=27 SAE, i.e., its computation cost is 27 2n 2 =54n 2 additions. 2656

IV. SIMULAION AND HE JOIN ALGORIHM In order to evaluate the performance of optical flow, full block search and three-step search algorithm, two video sequences were used. One is the Claire video sequence, which is composed of one woman talking with small motions before a flat background. Another is the able ennis video sequence, which is composed of a high speed moving objects. Figure 2 and 3 show two scenes from these two video sequences. (b) Prediction errors of three step search International Science Inde Computer and Information Engineering waset.org/publication/1581 Fig 2. Claire video sequence Fig 3. able tennis video sequence he SAE of optical flow, three-step search, and full search algorithms for Claire video sequence are 80.28, 87.94 and 72.09. he prediction errors of optical flow, three-step search, and full search algorithms for Claire video sequence are shown in figure 4. (a) Prediction errors of optical flow (c) Prediction errors of full search Fig 4. Prediction errors for Claire he SAE of optical flow, three-step search, and full search algorithms for able ennis video sequence are 685.33, 293.58, and 220.67. he prediction errors of optical flow, three-step search, and full search algorithms for able ennis video sequence are shown in figure 5. (a) Prediction errors of optical flow (b) Prediction errors of three step search 2657

ABLE IPREDICION ERRORS OF OPICAL FLOW, HREE-SEP SEARCH, FULL SEARCH, AND HE JOIN ALGORIHMS FOR VIDEO CLAIRE AND ABLE ENNIS Optical Flow hree-step Search Full Search Joint Algorithm Claire 80.28 87.94 72.09 73.32 able ennis 685.33 293.58 220.67 238.40 (c) Prediction errors of full search Fig 5. Prediction errors for able tennis International Science Inde Computer and Information Engineering waset.org/publication/1581 Optical flow method achieves high accuracy for scenes with small motion vector and small displacements (< 1-2 pixels/frame) but fail when the motion vector and displacements are large [9]. It is understandable; because the invariant luminance hypothesis of optical flow is invalid when there is a large motion vector and displacement like the example in figure 6. Fig 6. Example of luminance change between frames. he full search algorithm has the most accurate estimation result. However, the computation cost is much higher than the others. hree-step search algorithm can handle large displacements efficiently but are not very accurate [10] [11], because it only search sub-optimal result in each step, instead of searching optimal result from global area. he best scenario would have high estimation accuracy disregarding whether the motion vector is small or large, and the computation cost is not too much. A straightforward idea is using three-step search algorithm and optical flow estimation at the same time. After compute the prediction error, the motion vector which produces smaller error would be used as the estimation result. he algorithm can be written as follows: 1) Use optical flow algorithm to estimate the motion vector V opt, and calculate the prediction error Err opt. 2) Use three-step search algorithm to estimate the motion vector V 3step, and calculate the prediction error Err 3step. 3) If Err opt < Err 3step, V= V opt, otherwise, V= V 3step. 4) Return V as the final estimation result. Claire and able ennis video sequences also were used to evaluate the performance of this joint algorithm. able I shows the simulation results of optical flow, three-step search, full search, and the joint algorithms. he prediction error images are shown in Figure 7. (a) Prediction errors of "Claire" (b) Prediction errors of "able tennis" Fig 7. Prediction errors of the Joint Algorithm From able1, it shows clearly that the prediction error of the proposed joint algorithm is nearly as small as the full search algorithm. At the same time, the computation cost of the joint algorithm is only the sum of optical flow and threestep search, which is (54+9) n 2 =63n 2 addition, 6n 2 multiplication, plus one 22 matrix inverse, for a block with n n pixels. It is still much lower than the computation of full search, which requires 450n 2 addition for a block with n n pixels. V. CONCLUSION An efficient algorithm with joint of optical flow and threestep search is presented in this paper. he joint algorithm can achieve high accuracy disregarding whether the motion vector is small or large, and the computation cost is not too much. wo video sequences, Claire which only contains slow moving objects, and able ennis which contains high speed moving objects, were used to test the performance of optical flow, three-step search, full search, and the joint algorithm. he result shows that the joint algorithm works nearly as well as the full search algorithm, while cost much lower computation. 2658

International Science Inde Computer and Information Engineering waset.org/publication/1581 REFERENCES [1] J. R. Jain and A. K. Jain, Displacement measurement and its application in interframe image coding, IEEE rans. Commun., Vol. COM-29, pp. 1799-1808, Dec. 1981. [2] H. G. Musmann, P. Pirsch and H.-J. Grallert, Advances in picture coding, Proc. IEEE, Vol. 73, No. 4, pp. 523-548, 1985. [3] V. Bashkaran and K. Konstantinides, Image and video compression standards: algorithms and architectures, Kluwer Academic Publishers, 1995. [4] A. Murat ekalp, Digital video processing, Prentice-Hall, 1995. [5] K. R. Rao and J. J. Hwang, echniques and standards for image, video and audio Coding, Prentice-Hall, 1996. [6] F. Dufaux and F. Moscheni, Motion estimation techniques for digital V: A review and a new contribution, Proc. IEEE, Vol. 83, No. 6, June 1995. [7] B. D. Lucas and. Kanade, An iterative image registration technique with an application to stereo vision, in Proceedings of DARPA Image Understanding, pp. 121-130, 1981. [8]. Koga, K. Iinuma, A. Hirano, Y. Iijima,. Ishiguro, Motioncompensated interframe coding for video conferencing, Proceedings NC'81 (IEEE), pp.5.3.1 - G.5.3.4 [9] A. Singh, Optical flow: a unified perspective, IEEE Computer Society Press, Loas Alamitos, CA, 1991. [10] J.L. Barron, D.J. Fleet, and S.S. Beauchemin, Performance of Optical Flow echniques, in International Journal of Computer Vision, February 1994, vol. 12(1), pp. 43 77. [11] P. Anandan, A computational framework and an algorithm for the measurement of visual motion, in International Journal of Computer Vision, Vol. 2, pp. 283-310, January 1989. 2659