Introduction to Objective Analysis

Similar documents
Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

10.4 Linear interpolation method Newton s method

Watershed Sciences 4930 & 6920 GEOGRAPHIC INFORMATION SYSTEMS

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

LOESS curve fitted to a population sampled from a sine wave with uniform noise added. The LOESS curve approximates the original sine wave.

Introduction to Design Optimization: Search Methods

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Spatial Interpolation & Geostatistics

Introduction to ANSYS DesignXplorer

Spatial Interpolation - Geostatistics 4/3/2018

Dijkstra's Algorithm

Approximation Methods in Optimization

EE795: Computer Vision and Intelligent Systems

Geol 588. GIS for Geoscientists II. Zonal functions. Feb 22, Zonal statistics. Interpolation. Zonal statistics Sp. Analyst Tools - Zonal.

f xx + f yy = F (x, y)

Machine Learning / Jan 27, 2010

5 Classifications of Accuracy and Standards

Estimating Parking Spot Occupancy

Clustering Lecture 5: Mixture Model

Clustering Color/Intensity. Group together pixels of similar color/intensity.

An Adaptive Stencil Linear Deviation Method for Wave Equations

1. Practice the use of the C ++ repetition constructs of for, while, and do-while. 2. Use computer-generated random numbers.

1.2 Numerical Solutions of Flow Problems

READ ME FIRST. Investigations 2012 for the Common Core State Standards A focused, comprehensive, and cohesive program for grades K-5

Gridded Data Speedwell Derived Gridded Products

Figure (5) Kohonen Self-Organized Map

3. Data Structures for Image Analysis L AK S H M O U. E D U

A nodal based evolutionary structural optimisation algorithm

Scattered Data Problems on (Sub)Manifolds

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing

An introduction to interpolation and splines

Research on time optimal trajectory planning of 7-DOF manipulator based on genetic algorithm

Spatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University

Applying Supervised Learning

Smoothers. < interactive example > Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems

Unsupervised Learning

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017

Geostatistics 2D GMS 7.0 TUTORIALS. 1 Introduction. 1.1 Contents

A METHOD TO PREDICT ACCURACY OF LEAST SQUARES SURFACE MATCHING FOR AIRBORNE LASER SCANNING DATA SETS

ELEC Dr Reji Mathew Electrical Engineering UNSW

Until now we have worked with flat entities such as lines and flat polygons. Fit well with graphics hardware Mathematically simple

GEO-SLOPE International Ltd , 700-6th Ave SW, Calgary, AB, Canada T2P 0T8 Main: Fax:

Gas Distribution Modeling Using Sparse Gaussian Process Mixture Models

AN INTRODUCTION TO GRID ANALYSIS: HOTSPOT DETECTION

Modeling Plant Succession with Markov Matrices

Motion Capture & Simulation

k-means, k-means++ Barna Saha March 8, 2016

Lesson 5 overview. Concepts. Interpolators. Assessing accuracy Exercise 5

Tuning an Algorithm for Identifying and Tracking Cells

Chapter 4. Clustering Core Atoms by Location

DATA MINING AND WAREHOUSING

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.

This work is about a new method for generating diffusion curve style images. Although this topic is dealing with non-photorealistic rendering, as you

Computer Vision I - Filtering and Feature detection

= f (a, b) + (hf x + kf y ) (a,b) +

Spatial and multi-scale data assimilation in EO-LDAS. Technical Note for EO-LDAS project/nceo. P. Lewis, UCL NERC NCEO

Intensity Augmented ICP for Registration of Laser Scanner Point Clouds

3 Nonlinear Regression

Summary of Publicly Released CIPS Data Versions.

3 Nonlinear Regression

Chapter 6. Semi-Lagrangian Methods

Surface Reconstruction. Gianpaolo Palma

Interactive Graphics. Lecture 9: Introduction to Spline Curves. Interactive Graphics Lecture 9: Slide 1

VARIANCE REDUCTION TECHNIQUES IN MONTE CARLO SIMULATIONS K. Ming Leung

CPSC 340: Machine Learning and Data Mining. More Linear Classifiers Fall 2017

Data Acquisition. Chapter 2

Analysis of different interpolation methods for uphole data using Surfer software

What is Multigrid? They have been extended to solve a wide variety of other problems, linear and nonlinear.

Improvements in Continuous Variable Simulation with Multiple Point Statistics

Summer 2016 Internship: Mapping with MetPy

THE preceding chapters were all devoted to the analysis of images and signals which

Using Subspace Constraints to Improve Feature Tracking Presented by Bryan Poling. Based on work by Bryan Poling, Gilad Lerman, and Arthur Szlam

CPSC 340: Machine Learning and Data Mining. Regularization Fall 2016

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Serial Data, Smoothing, & Mathematical Modeling. Kin 304W Week 9: July 2, 2013

AMS527: Numerical Analysis II

Spatial Interpolation & Geostatistics

Petrel TIPS&TRICKS from SCM

Surface Shape Regions as Manifestations of a Socio-economic Phenomenon

University of Wisconsin-Madison Spring 2018 BMI/CS 776: Advanced Bioinformatics Homework #2

CS 109 C/C ++ Programming for Engineers w. MatLab Summer 2012 Homework Assignment 4 Functions Involving Barycentric Coordinates and Files

ADAPTIVE FINITE ELEMENT

Opening the Black Box Data Driven Visualizaion of Neural N

13. Geospatio-temporal Data Analytics. Jacobs University Visualization and Computer Graphics Lab

ECMWF contribution to the SMOS mission

Carnegie Learning Math Series Course 2, A Florida Standards Program

Computer Graphics. Curves and Surfaces. Hermite/Bezier Curves, (B-)Splines, and NURBS. By Ulf Assarsson

Introduction to optimization methods and line search

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation

Geological modelling. Gridded Models

Mid-Year Report. Discontinuous Galerkin Euler Equation Solver. Friday, December 14, Andrey Andreyev. Advisor: Dr.

MODELING FOR RESIDUAL STRESS, SURFACE ROUGHNESS AND TOOL WEAR USING AN ADAPTIVE NEURO FUZZY INFERENCE SYSTEM

MA 323 Geometric Modelling Course Notes: Day 21 Three Dimensional Bezier Curves, Projections and Rational Bezier Curves

A brief history of time for Data Vault

Implicit Surfaces. Misha Kazhdan CS598b

CPSC 340: Machine Learning and Data Mining. Density-Based Clustering Fall 2016

Scan Matching. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

1. Assumptions. 1. Introduction. 2. Terminology

Geometric Rectification of Remote Sensing Images

Transcription:

Chapter 4 Introduction to Objective Analysis Atmospheric data are routinely collected around the world but observation sites are located rather randomly from a spatial perspective. On the other hand, most computer forecast models use some type of uniform grid for their calculations. If you are going to routinely use observed data for any type of finite difference calculations, there is a need to transform or interpolate these irregularly spaced observations to the uniform grid without human intervention. This process is called objective analysis (OA). This process is related to the subjective analysis process (line drawing and human interpretation) described in Chapters 1 and 3, but is totally free of human intervention. In the late 1940s, as numerical weather prediction was getting started, the initial values for the model grids were determined by first subjectively drawing charts and then interpolating between isolines to determine grid values. From an operational perspective, this was a slow, labor-intensive process. OA was the answer. It eliminated the human from the process, was considerably faster, and likely more accurate. The objective analysis processes described in this chapter focus on spatial interpolation. Temporal interpolation can also be used if you want to blend data over time onto a grid. Temporal interpolation is not addressed here. Types of Objective Analysis Objective analysis has evolved over the last 50 years and can be divided into three main categories. Function Fitting Model: In this approach a set of observations is mathematically fit to a set of relatively smooth equations. Usually some type of least square error technique is used. An example of this approach is the polynomial expansion of Panosky (1949). Successive Corrections Method: This approach is an iterative process. It starts with a first guess or background field. It then uses a linear weighting scheme to adjust the initial guess 1

and subsequent estimates to fit the grid point values to the observations to within a specified tolerance. Statistical Interpolation Method (also called optimal or optimum interpolation): In this approach the ensemble average of the squared difference between the gridded and observed data is minimized using a least-square methodology. The evolution of OA since the late 1950s has led to the current method for initializing forecast models: data assimilation. Data assimilation uses statistical interpolation methods that combine observations with short range forecasts to produce, as accurately as possible, a dataset that represents the current state of the atmosphere. The assimilation model produces a short forecast, usually 3 or 6 hours into the future, and then incorporates recent data into the model to adjust the model to the current observations. The successive corrections approach will be discussed in detail in this chapter. Empirical Linear Interpolation Introduction: In this approach grid point values are determined as a linear combination of scalar variable values from nearby observing sites. The basic form of the linear equation is: X grid = j=1 to m (h j X oj ) X grid = interpolated grid point value h j = scalar weight at point j X oj = observation at point j m = number of observations The scalar weight has the general form: h j = w j / [ j=1 to m (w j )] w j = weighting function applied to point j Note that the sum of all h j equals one [ (h j )=1]. The weight function (w j ) can be determined in any number of ways. Statistical functions are currently in vogue. This chapter will focus on historically and empirically derived functions. In general, empirically derived weight functions use distance 2

weighting where observations that are closer to the grid point are given more weight than observations that are farther away. Scale-selection properties can also be incorporated into these functions. Practical Perspective: Ideally, all available observations can be used to determine the value at a grid point. If a weighting function falls off with distance, there is some distance away from the grid point where the weight becomes so small that the observations have a very little influence on the grid point value. To save computer time and eliminate these low-influence observations, only observations within a specified distance from the grid point are used in the calculation. This distance is called the radius of influence. The size of the radius of influence also determines the degree of smoothing that is incorporated into the interpolation. As a general rule, smaller radii have less smoothing; larger radii create more smoothing. Weighting Function Properties: Weighting functions should have the following properties: The influence of an observation should be higher for closer observations and lower for observations farther from the grip point. If an observation is coincident with a grid point, it should have a unit weight. The resulting grid values should be differentiable (a desirable property). The more observations included in the interpolation, the higher is the smoothing level. Applying these properties result in a wide variety of weighting functions. The exact form of the weighting function is often determined by trial and error, hence the name, empirical. An Example: Let s consider the example shown in Figure 4-1. There are 12 observations in the dataset. Only four observations are within the radius of influence: D, E, F and G. These will be used to calculate the value at the grid point (gp). Thus, for grid point, X gp : X gp = (h D *X D ) + (h E *X E ) + (h F *X F ) + (h G *X G ) Note that in this particular example, no consideration is given to the distribution of the observation about the grip point. Weighting functions can be designed with this in mind. 3

Let s define the weighting function, w j = exp(-r) = e -r, for r less than a 3.5 units, where r is the distance from the grid point to the observation (in arbitrary units). Figure 4-1: Example of Radius of Influence The following table summarizes the calculations: Point X oj =Ob r w j =e -r h j h j *X j D 74 3.35 0.0351 0.0536 3.97 E 73 1.25 0.2865 0.4375 31.94 F 69 2.80 0.0608 0.0928 6.40 G 77 1.30 0.2725 0.4161 32.04 Sum 0.6549 1.0000 74.35 The four observation values and corresponding distances from the observation point to the grid point are listed in the first three columns on the left. First, calculate the value of the weighting function. These values are shown in the w j column. Then sum up the weighting functions for all observations. This value is at the bottom of the w j column, 0.6549. Next, calculate the weight for each observation, h j, by dividing individual weighting function value by the sum of the weight functions. Note that the sum of the weights adds up to one. Finally, multiply the observation values by the corresponding weight to get the 4

contribution of that observation to the final interpolated values. These values are shown in the last column on the right. The sum of the individual contributions gives interpolated values of 74.35. You have now done one grid point. Are you ready to do the rest by hand or should we let a computer do it? Successive Corrections The empirical linear interpolation uses the actual observation values when calculating grid point values. The successive corrections approach begins with a first guess or background grid and then uses the difference between these background values and the observations to determine the final grid point values via an iterative process. This method is useful for grid areas where observations are sparse or non-existent. The successive corrections procedure includes a series of steps: Estimate the background value at each observation point [X bj ]. Calculate the prediction error, that is, the difference between the background value at an observation point and the corresponding observation value [X oj X bj ]. Use empirical linear interpolation to distribute the prediction errors to the grid points. Correct the grid point value based on the distributed errors. [X grid(k+1) = X grid(k) + j=1 to m (h j *{X oj X bj })], where k is the iteration step index. Repeat the above steps until the value of all predicted errors is below a specified amount. This iterative process allows the grid point values to converge to the observation values. At each iteration step the value of the weight can be modified or the radius of influence can be changed. Cressman Scheme The Cressman (1959) scheme is described here because it is the first operational successive corrections OA scheme used by the Weather Bureau (now the National Weather Service) in the early days of operational numerical weather prediction. It is similar to the successive corrections scheme described in the previous section. It uses a non-zero weighting function 5

within a prescribed radius of influence. This radius of influence decreases with each successive scan. After a large number of scans the grid values converge to the observations. In areas of high observation density, the grid point values reflect the observations, while in low-density areas the grid values are closer to the first guess or background field. The Cressman weighting function takes the form: w j = ( d 2 r 2 ) / ( d 2 + r 2 ) r < d w j = 0 r > d r = distance from the grid point to the observation d = radius of influence This original weighting function was modified slightly and is referred to as the Cressman Model Extension: w j = [( d 2 r 2 ) / ( d 2 + r 2 )] α r < d w j = 0 r > d where α > 1. Selecting the proper values for d and α is somewhat empirical and depends upon the data spacing and the desired level of smoothing. In general, a smaller radius of influence has less smoothing; a larger radius of influence produces a higher degree of smoothing. For the original Cressman scheme a value for d that is approximately twice the average spacing of the observations tends to be a reasonable compromise between undersmoothing and over-smoothing. See Thiébaux and Pedder (1987), page 94-99, for examples. Barnes Scheme The Barnes (1964) proposed a scheme which could be used to interpolate randomly spaced data onto a grid with a desired level of detail. The Barnes weighting function is given by: w j = exp [ -r 2 / 4k ] 6

r = distance from the grid point to the observation k = parameter used to define the response of the weighting function The value of k is chosen so that when r=d (d is the radius of influence), the weighting function is e -4 times maximum value of data at r=0. The radius of influence is chosen so that it is greater than the average spacing of the observations. The iterative process recommended by Barnes for determining grid point values of some variable is as follows: On the first scan, estimate the value of the variable at each grid point using the following scheme: x i,j = [ ( n=1 to N w(r,d)*x on ) / w(r,d) ] o x i,j = grid point value of the variable o X on = observed value of the variable o N = number of observations o w(r,d) = Barnes weighting function Next, estimate the variable value for each observation location by averaging the four closest grid values. Calculate the difference between the variable estimate and the observation value (residual). Distribute the differences using the linear interpolation scheme described above with the Barnes weighting function. Repeat the last three steps until the residual is less than a prescribed precision factor. Smoothing Smoothing is a process that averages data in space and/or time so that fluctuations of a scale smaller than a specified amount are reduced or removed. In some ways, smoothing is a type of OA. Smoothing can be easily applied to a spatial field or time series. For example, a running mean is often used with time series to smooth out short-term fluctuations so that longer term cycles can be identified. For example, Figure 4-2 shows a time series of high temperature data for Kansas City International Airport (KMCI). The daily high temperatures are shown in blue and the 7-day running mean is in red. Even though day-to-day 7

fluctuations are large, the general warmer and cooler periods can be seen in the blue time series. The 7-day running mean damps out the small-scale changes and makes identification of the warmer and cooler period much easier. In fact, one could infer a periodicity of approximately two weeks. Figure 4-2: KMCI High Temperature Data Smoothing schemes can take on a wide variety of forms. Details are beyond the scope of this chapter. Draft: 5-04-2010 Final: 5-11-2011 8