Measuring Horizontal Ground Deformation Using Optical Satellite Images Part 1 of 2

Similar documents
Measuring Ground Deformation using Optical Imagery

Interferometric processing. Rüdiger Gens

The 2017 InSAR package also provides support for the generation of interferograms for: PALSAR-2, TanDEM-X

Sentinel-1 Toolbox. Interferometry Tutorial Issued March 2015 Updated August Luis Veci

Geometric Accuracy Evaluation, DEM Generation and Validation for SPOT-5 Level 1B Stereo Scene

Terrain correction. Backward geocoding. Terrain correction and ortho-rectification. Why geometric terrain correction? Rüdiger Gens

Geometric Rectification of Remote Sensing Images

Geometric Correction

Filtering, unwrapping, and geocoding R. Mellors

Sentinel-1 Toolbox. TOPS Interferometry Tutorial Issued May 2014

GEOG 4110/5100 Advanced Remote Sensing Lecture 4

fraction of Nyquist

Correction and Calibration 2. Preprocessing

Interferometry Tutorial with RADARSAT-2 Issued March 2014 Last Update November 2017

POSITIONING A PIXEL IN A COORDINATE SYSTEM

In addition, the image registration and geocoding functionality is also available as a separate GEO package.

DIGITAL SURFACE MODELS OF CITY AREAS BY VERY HIGH RESOLUTION SPACE IMAGERY

Motion compensation and the orbit restitution

Individual Interferograms to Stacks!

SAR Interferometry. Dr. Rudi Gens. Alaska SAR Facility

Sentinel-1 Toolbox. Offset Tracking Tutorial Issued August Jun Lu Luis Veci

PSI Precision, accuracy and validation aspects

A Correlation Test: What were the interferometric observation conditions?

Individual Interferograms to Stacks

Memorandum. Clint Slatton Prof. Brian Evans Term project idea for Multidimensional Signal Processing (EE381k)

Scene Matching on Imagery

InSAR Operational and Processing Steps for DEM Generation

Course Outline (1) #6 Data Acquisition for Built Environment. Fumio YAMAZAKI

DINSAR: Differential SAR Interferometry

Do It Yourself 8. Polarization Coherence Tomography (P.C.T) Training Course

Capturing, Modeling, Rendering 3D Structures

COORDINATE TRANSFORMATION. Lecture 6

Repeat-pass SAR Interferometry Experiments with Gaofen-3: A Case Study of Ningbo Area

GAMMA Interferometric Point Target Analysis Software (IPTA): Users Guide

Video Georegistration: Key Challenges. Steve Blask Harris Corporation GCSD Melbourne, FL 32934

Interferometric Synthetic-Aperture Radar (InSAR) Basics

PolSARpro v4.03 Forest Applications

WIDE AREA DEFORMATION MAP GENERATION WITH TERRASAR-X DATA: THE TOHOKU-OKI EARTHQUAKE 2011 CASE

GEOG 4110/5100 Advanced Remote Sensing Lecture 4

Co-registration and complex interpolation

Coherence Based Polarimetric SAR Tomography

Lateral Ground Movement Estimation from Space borne Radar by Differential Interferometry.

EE795: Computer Vision and Intelligent Systems

Synthetic Aperture Radar Interferometry (InSAR)

Computational Aspects of MRI

Calibration of IRS-1C PAN-camera

PROBLEMS AND LIMITATIONS OF SATELLITE IMAGE ORIENTATION FOR DETERMINATION OF HEIGHT MODELS

InSAR DEM; why it is better?

Image Processing and Analysis

IMAGINE OrthoRadar. Accuracy Evaluation. age 1 of 9

MULTI-TEMPORAL INTERFEROMETRIC POINT TARGET ANALYSIS

Computer Graphics. Sampling Theory & Anti-Aliasing. Philipp Slusallek

DETECTION AND QUANTIFICATION OF ROCK GLACIER. DEFORMATION USING ERS D-InSAR DATA

DEVELOPMENT OF CAMERA MODEL AND GEOMETRIC CALIBRATION/VALIDATION OF XSAT IRIS IMAGERY

Exterior Orientation Parameters

Ice surface velocities using SAR

MONO-IMAGE INTERSECTION FOR ORTHOIMAGE REVISION

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning

INTERFEROMETRIC MULTI-CHROMATIC ANALYSIS OF HIGH RESOLUTION X-BAND DATA

University of Technology Building & Construction Department / Remote Sensing & GIS lecture

Automatic DEM Extraction

fmri pre-processing Juergen Dukart

WIDE BASELINE INTERFEROMETRY WITH VERY LOW RESOLUTION SAR SYSTEMS

Generate Glacier Velocity Maps with the Sentinel-1 Toolbox

SPOT-1 stereo images taken from different orbits with one month difference

Automatic DEM Extraction

ENVI Automated Image Registration Solutions

ELEC Dr Reji Mathew Electrical Engineering UNSW

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

CPSC 425: Computer Vision

Automatic DEM Extraction

Development and Applications of an Interferometric Ground-Based SAR System

Digital Processing of Synthetic Aperture Radar Data

DIGITAL HEIGHT MODELS BY CARTOSAT-1

Reading. 2. Fourier analysis and sampling theory. Required: Watt, Section 14.1 Recommended:

Basic Algorithms for Digital Image Analysis: a course

Interferometric SAR Processing

Using Hierarchical Warp Stereo for Topography. Introduction

Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16

The main goal of Computer Graphics is to generate 2D images 2D images are continuous 2D functions (or signals)

2D Image Processing Feature Descriptors

CORRECTING RS SYSTEM DETECTOR ERROR GEOMETRIC CORRECTION

Anno accademico 2006/2007. Davide Migliore

Remote Sensing Introduction to the course

9/14/2011. Contents. Lecture 3: Spatial Data Acquisition in GIS. Dr. Bo Wu Learning Outcomes. Data Input Stream in GIS

What have we leaned so far?

Lecture 20: Tracking. Tuesday, Nov 27

Radar Coherent Backscatter!

Training i Course Remote Sensing Basic Theory & Image Processing Methods September 2011

Improving wide-area DEMs through data fusion - chances and limits

Concept and methodology of SAR Interferometry technique

A Study on Ortho-rectification of SPOT6 Image Guo-dong YANG, Xiu-wen XIN and Qiong WU*

Fourier analysis and sampling theory

Chapter N/A: The Earthquake Cycle and Earthquake Size

VALIDATION OF A NEW 30 METER GROUND SAMPLED GLOBAL DEM USING ICESAT LIDARA ELEVATION REFERENCE DATA

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

What is Frequency Domain Analysis?

WORCESTER POLYTECHNIC INSTITUTE

IMAGE CORRECTIONS 11.1 INTRODUCTION. Objectives. Structure 11.1 Introduction Concept of Image Distortion and Correction

Transcription:

Measuring Horizontal Ground Deformation Using Optical Satellite Images Part 1 of 2 Sébastien Leprince California Institute of Technology Ge 167 Lecture Notes May 12, 2009

Overview Objective: Monitoring natural phenomena involving Earth s surface dynamics, e.g., earthquakes, volcanoes, glacier flow, landslides, sand dunes migration, etc... Motivation: To validate/calibrate/refine physical models. To improve early evaluation of damage for large disasters Approach: Measuring horizontal ground deformations from optical satellite images: SPOT 1-2-3-4 (10 m), SPOT 5 (5 m and 2.5 m), ASTER (15 m), Quickbird (0.7 m), Aerial photographs (0.25-1 m)

Measuring deformation: a toy example Unit length NS correlation image dy EW correlation image -0.3 +0.25 dx

Measuring Horizontal Ground Displacement, Methodology Flow P r e - e a r t h q u a k e i m a g e Inputs: Raw images Orbits, platform attitudes, camera model Digital Elevation Model Orthorectification: Images must superimpose accurately Correlation: Sub-pixel Correlation P o s t - e a r t h q u a k e i m a g e Displacement in rows and columns provide the E/W and N/S components of the ground deformation Outputs: N/S offset field E/W offset field SNR The Signal to Noise Ratio assesses the measure quality.

The 1999 Mw 7.1 Hector Mine Earthquake 34 10'N 34 20'N 34 30'N 34 40'N ³ A 0 5 10 20 km Offset in meters -3.0 m +3.0-3 A A' -4 0 2 4 6 8 10 12 Location in km 116 30'W 116 20'W 116 10'W 116 0'W 3 2 1 0-1 -2 A' 34 10'N 34 20'N 34 30'N 34 40'N ³ 0 5 10 20 km -3.0 m +3.0 116 30'W 116 20'W 116 10'W 116 0'W The Hector Mine horizontal coseismic field (NS and EW) derived from 10m SPOT4 1998 and 10m SPOT2 2000 images.

In this 2-session lecture + 1 lab, you will (should...) learn: The basic theories, methodologies, and trade-offs involved to produce horizontal deformation maps from optical data, How to interpret/estimate the quality of the deformation maps produced by understanding potential biases, How to use the COSI-Corr (Co-registration of Optically Sensed Images and Correlation) software, which will assist you in all the processing tasks involved.

The viewing geometry of optical sensors Questions: Identify the move between these two shots. What did your brain had to do to come up with the solution?

The viewing geometry of optical sensors Image acquired with almost similar viewing geometries

The viewing geometry of optical sensors Image acquired with almost similar viewing geometries

Images must be compared in the same geometry Images must be re-projected in a common geometry to be compared. Infinitely many possible projections: any arbitrary common viewing geometry is theoretically valid: in which geometry images should be analyzed? Try to find a viewing geometry that minimizes geometrical stretch on both images during re-projection: application dependent, but global trends exist: Radar interferometry: images acquired with short baseline. Almost similar viewing geometry, common viewing geometry defined as the viewing geometry of one of the images (master/slave) Same idea also encountered for optical sensors. However, optical sensors have no baseline constraints and images can have very different viewing geometries. Can define an intermediate viewing geometry, independent of the images (more flexible to compare images from several different sensors). Images analyzed in orthorectified geometry.

SAR vs Optical viewing geometries Image 1 Virtual image Image 2 SAR images are usually analyzed (e.g., InSAR applications) by taking one of the image as the reference geometry. Due to the largest disparities in viewing geometries, optical images are usually analyzed using intermediate projections. They are commonly orthorectified, meaning that they are reprojected on the ground, accounting for the topography. In Orthoimages, each pixel is as if it was seen exactly from above. Proper projection of images depends on the proper modeling of the acquisition sensor.

Common types of optical sensor geometries Acquisition geometry of a frame camera. The image is acquired in a single shot. Archive images are film-based and need to be scanned to be processed; more modern cameras use digital frame sensors. Frame camera: very high resolution imaging (1-100 cm GSD), limited footprint (1-10 km) - aerial surveys, old spy satellites, etc... Pushbroom systems (main focus in this lecture): high resolution imaging (0.5-3 0m GSD), medium swath width (15-60km) - SPOT, ASTER, Quickibird, Ikonos, etc... Whiskbroom systems: medium-low resolution (30-1000 m GSD), large swath (100-1000 km), multispectral - Landsat, Modis, AVHRR, etc...

Orthorectification Model Pushbroom acquisition geometry satellite velocity look direction of pixel (c, r) absolute pointing error from ancillary data M PSat = O CCD acquiring column c CCD array O, optical center in space M, ground point seen by pixel p u 1 pixel pointing model R(p) 3D rotation matrix, roll, pitch, yaw at p T(p) Terrestrial coordinates conversion δ correction on the look directions to insure coregistration λ > 0 M(p) = O(p) + λ [ ] T(p)R(p) u 1 (p) + δ(p) }{{} u 3 (p)

Orthorectification: Look Direction Correction u 3Th (x 0,y 0 ) P(y 0 ) u 3Sat (x 0,y 0 ) M 0 du 3 (x 0,y 0 ) Look direction discrepancy Topography M δ(p) = du3 (x 0, y 0 ) = u 3Th (x 0, y 0 ) u 3Sat (x 0, y 0 ) Reference ellipsoid Ground Control Point GCP: {p(x 0, y 0 ), M 0 }

Orthorectification: An Irregular Mapping Trajectory image plane Topography Reference ellipsoid d i d j Image pixels are assumed to regularly sample the image plane, but they sample the ground irregularly

Orthorectification: An Irregular Mapping Trajectory d i d j image plane Topography Reference ellipsoid d i d j We introduce an inverse mapping to facilitate resampling

Orthorectification: An Irregular Mapping Trajectory d i d j image plane Topography Reference ellipsoid d i d j To avoid aliasing in the resampled signal, ideal resampling kernel: h d (x) = πx sin d πx, with d = max(1, {d i }) d

Orthorectification: Inverse Model Principle Projection plane P (M) P Solution: find pixel (x, y) that minimizes Reference ellipsoid h Φ M M' u Φ(x, y) = OM OM (x, y) 2 with Earth center O Topography OM (x, y) = OP(y) + λ u(x, y) and λ such that M belongs to P(M) Given M, which pixel (x, y) saw M?

Resampling: general approach

Resampling: practical implementation

Resampling: 1D ideal case 1 0.5 0 s 0 (t) h T0 (t) T 0 t s 0 (t) bandlimited at Nyquist frequency π T 0, to be resampled at a new sampling period T 1 S(ω) Ideal resampling kernel: Nyquist frequencies π T 0 π T 1 0 π T 1 π T 0 ω h d (t) = sin πt d πt d T 1 > T 0 d = T 1 S(ω) To avoid aliasing in the resampled signal: d = max(t 0, T 1 ) π π π π 0 ω T 1 T 0 T 0 T 1 Nyquist T 1 < T 0 d = T frequencies 0 d effective signal resolution

Resampling: popular convolution-based interpolation kernels Nearest-Neighbor Linear Piecewise-cubic

Ortho-Resampling: a simple particular case using NN kernel Inverse orthorectification mapping Raw satellite image Ortho-image to be built N output pixel = area weigthed average of input pixels covered by inverse mapping of output pixel Kernel locally warped according to local mapping warp (warping usually linearized locally = Jacobian of ortho-mapping) In practice, use higher order (sinc-like) kernels for better resampling quality

Image Correlation: problem formulation Given two images i 1 and i 2 such that i 2 (x, y) = i 1 (x x, y y ) how to retrieve the translation ( x, y )? Generally, problem formulated as: Find ( x, y ) such that S(i 1 (x, y), i 2 (x + x, y + y )) is maximum, for some similarity measure S. Which similarity measure S should we use? How to achieve sub-pixel accuracy?

Image Correlation: popular similarity measures Sum of Squared Differences (SSD) SSD i1,i 2 ( x, y ) = N x N y [i 1 (x, y) i 2 (x + x, y + y )] 2 for N x and N y centered about x and y Normalized Cross-Correlation (NCC), or Correlation Coefficient Nx Ny (i 1 (x, y) i 1 )(i 2 (x + x, y + y ) i 2 ) NCC i1,i 2 ( x, y ) = Nx Ny (i 1 (x, y) i 1 ) 2 Nx Ny (i 2 (x + x, y + y ) i 2 ) 2 Phase Correlation based on Fourier Shift Theorem i 2 (x, y) = i 1 (x x, y y ) I 2 (ω x, ω y ) = I 1 (ω x, ω y )e j(ω x x +ω y y ) F 1{ I 1 (ω x, ω y )I2 (ω x, ω y ) } I 1 (ω x, ω y )I2 (ω = F 1{ e j(ω x x +ω y y ) } = δ(x + x, y + y ) x, ω y )

Image Correlation: popular similarity measures

Image Correlation: subpixel accuracy Correlation surface computed for integer values of ( x, y ) comprised in a search area. Then interpolation of the correlation surface near the maximum, or fit to a smooth surface, e.g., Gaussian. Easy implementation but not very accurate because interpolation model for correlation surface not always well known. By introducing a subpixel shift between images to be correlated via interpolation/resampling. The image is shifted by resampling, the goal being to find the best resampling shift that maximizes the similarity function. Implementation more complex but accuracy can be up to an order of magnitude better than previous method. If we solve the registration problem in the Fourier domain directly, image resampling becomes implicit and can lead to faster algorithms using FFT (best of both worlds?).

Image Correlation: what we do in COSI-Corr Fourier Shift Theorem i 2 (x, y) = i 1 (x x, y y ) I 2 (ω x, ω y ) = I 1 (ω x, ω y )e j(ω x x +ω y y ) Normalized Cross-spectrum C i1 i 2 (ω x, ω y ) = I 1(ω x, ω y )I 2 (ω x, ω y ) I 1 (ω x, ω y )I 2 (ω x, ω y ) = ej(ω x x +ω y y ) Finding the relative displacement φ( x, y ) = π π W(ω x, ω y ) C i1 i 2 (ω x, ω y ) e j(ω x x +ω y y ) 2 ω x = π ω y = π W weighting matrix. ( x, y ) such that φ minimum.

Image Correlation: so which similarity measure is best finally? The NCC method usually performs better than the SSD because it is insensitive to linear transformations in image intensities. Indeed, photometric differences due to change in illumination, which are always encountered in practice with images acquired at different time, can be locally modeled as linear contrast changes. NCC and SSD are based on second order statistics and can be viewed, from a Bayesian approach, as the best solution when images are corrupted with additive Gaussian white noise. In other words, these methods consider that the displacement to be found is exact, and that images amplitude only, contains noise. Normalized phase correlation methods are also insensitive to linear contrast change. However, the noise formulation is quite different. From a Bayesian point of view, the noise can be considered Gaussian-like on the displacement, while images are considered noise free! Indeed, we can write: Φ( ) = ω e jω x e jω 2 = 2 [1 cos(ω( x ))] ω Then minimizing Φ is equivalent to maximizing, over, ω e W (ω) cos(ω( x)). C(ω) This is Von-Mises distribution on the phase shift, then Gaussian-like distribution on.

Image Correlation: so which similarity measure is best finally? Conclusion: The phase correlation method will provide more accurate results when the images have a low noise level, and when some deviation from the rigid translation model is expected. This is often the case in practice and the measure provided can be seen as the average displacement over the correlation window. At high noise level, NCC methods usually perform better than phase correlation methods, but using iterative re-weighted methods on W mitigates the strong noise-free requirement of the images in the phase correlation formulation and adds some robustness to the solution.

Processing Chain Select Image Registration Points from raw image Process with manual input Automatic process Orthorectify patches centered at RP Resample image patches 1st ref: Shaded DEM Deduce viewing correction δ for co-registration Orthorectify / resample full raw image 1 Correlate patches, find relative displacement with reference image

Processing Chain Select Image Registration Points from raw image Process with manual input Automatic process Orthorectify patches centered at RP Resample image patches 2nd ref: 1st ortho image Correlate patches, find relative displacement with reference image Deduce viewing correction δ for co-registration Orthorectify / resample full raw image 1 Orthorectify / resample full raw image 2 Correlation on sliding windows Horizontal deformation map

1999 Mw 7.1 Hector Mine Earthquake, CA

1999 Mw 7.1 Hector Mine Earthquake, CA

1999 Mw 7.1 Hector Mine Earthquake, CA

The 1999 Mw 7.1 Hector Mine Earthquake 34 10'N 34 20'N 34 30'N 34 40'N ³ A 0 5 10 20 km Offset in meters -3.0 m +3.0-3 A A' -4 0 2 4 6 8 10 12 Location in km 116 30'W 116 20'W 116 10'W 116 0'W 3 2 1 0-1 -2 A' 34 10'N 34 20'N 34 30'N 34 40'N ³ 0 5 10 20 km -3.0 m +3.0 116 30'W 116 20'W 116 10'W 116 0'W The Hector Mine horizontal coseismic field (NS and EW) derived from 10m SPOT4 1998 and 10m SPOT2 2000 images.

The 1999 Mw 7.1 Hector Mine Earthquake 700 600 Epicenter Field measurements SPOT measurements Right lateral slip in cm 500 400 300 200 100 0 0 5 10 15 20 25 30 35 N153 E Distance (km) Horizontal slip vectors measured from linear least square adjustment on each side of the fault. Perpendicular profiles are stacked over a width of 880 m and a length of 8 km. Field measurements from J.A. Treiman et al., BSSA, 2002

The 1999 Mw 7.1 Hector Mine Earthquake M. Simons et al., BSSA, 2002

Optical Correlation and SAR Complementarity Optical Image Correlation: The larger the displacement, the higher the technique SNR. Technique mostly usefully to measure large deformation gradients, Mostly sensitive to horizontal component of ground deformation, Can be used to measure change over long time periods, Images from sensors with different geometry, orbit, and resolution can be correlated, potentially huge archive Can correlate on sand, ice even with some ice melt (passive sensor) Images contain aliasing, which limits correlation accuracy Very sensitive to weather condition (clouds), no night imaging for visible bands SAR: Amplitude correlation on speckle pattern Azimuth (and range) offsets Fairly insensitive to weather conditions + night and day imaging Analysis only from images acquired with identical geometry. Shadow casting depends on radar orientation and surface reflection properties Images with little aliasing Little backscattering on sand, or different backscattering on ice melt InSAR: Sensitive to ground deformation in range direction Very accurate but loss of coherence when displacement gradient too large (depends on radar wavelength and pixel size), Temporal decorrelation due to change in scatterer geometry Path delays due to tropospheric/ionospheric structure