Nonlinear Noise Reduction

Similar documents
arxiv:chao-dyn/ v1 13 Dec 1998

Edge and local feature detection - 2. Importance of edge detection in computer vision

Artificial Intelligence for Robotics: A Brief Summary

Parameter Estimation in Differential Equations: A Numerical Study of Shooting Methods

Linear Methods for Regression and Shrinkage Methods

The Pre-Image Problem in Kernel Methods

Effects of multi-scale velocity heterogeneities on wave-equation migration Yong Ma and Paul Sava, Center for Wave Phenomena, Colorado School of Mines

Automatic Singular Spectrum Analysis for Time-Series Decomposition

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis

Spectral Compression: Weighted Principal Component Analysis versus Weighted Least Squares

ECG782: Multidimensional Digital Signal Processing

The Curse of Dimensionality

AN ALGORITHM FOR BLIND RESTORATION OF BLURRED AND NOISY IMAGES

CSE 547: Machine Learning for Big Data Spring Problem Set 2. Please read the homework submission policies.

Dimension Reduction CS534

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei

Chapter 4. Clustering Core Atoms by Location

Basis Functions. Volker Tresp Summer 2017

Multicomponent f-x seismic random noise attenuation via vector autoregressive operators

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

ELEG Compressive Sensing and Sparse Signal Representations

Function approximation using RBF network. 10 basis functions and 25 data points.

General Instructions. Questions

Extensions of One-Dimensional Gray-level Nonlinear Image Processing Filters to Three-Dimensional Color Space

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.

MTTTS17 Dimensionality Reduction and Visualization. Spring 2018 Jaakko Peltonen. Lecture 11: Neighbor Embedding Methods continued

x' = c 1 x + c 2 y + c 3 xy + c 4 y' = c 5 x + c 6 y + c 7 xy + c 8

LOESS curve fitted to a population sampled from a sine wave with uniform noise added. The LOESS curve approximates the original sine wave.

Unsupervised Learning

Diffusion Wavelets for Natural Image Analysis

IMAGE DE-NOISING IN WAVELET DOMAIN

Lecture 8 Object Descriptors

Ultrasonic Multi-Skip Tomography for Pipe Inspection

Spatial Interpolation & Geostatistics

Clustering and Visualisation of Data

The Singular Value Decomposition: Let A be any m n matrix. orthogonal matrices U, V and a diagonal matrix Σ such that A = UΣV T.

GLOSSARY OF TERMS. Commutative property. Numbers can be added or multiplied in either order. For example, = ; 3 x 8 = 8 x 3.

Comparative Analysis in Medical Imaging

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

Spatial Interpolation - Geostatistics 4/3/2018

Supplementary Figure 1. Decoding results broken down for different ROIs

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Non-linear dimension reduction

Spatial Outlier Detection

MAPI Computer Vision. Multiple View Geometry

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

Segmentation and Grouping

3 Nonlinear Regression

Classroom Tips and Techniques: Stepwise Solutions in Maple - Part 2 - Linear Algebra

CS 664 Slides #11 Image Segmentation. Prof. Dan Huttenlocher Fall 2003

Lecture 6: Edge Detection

5.6 Self-organizing maps (SOM) [Book, Sect. 10.3]

arxiv: v1 [physics.data-an] 27 Sep 2007

Dense Image-based Motion Estimation Algorithms & Optical Flow

Lecture 26: Missing data

Non-Cooperative Target Identification Based on Singular Value Decomposition

3 Nonlinear Regression

Image Restoration and Reconstruction

CoE4TN4 Image Processing. Chapter 5 Image Restoration and Reconstruction

Overview. Efficient Simplification of Point-sampled Surfaces. Introduction. Introduction. Neighborhood. Local Surface Analysis

Compressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI

Numerical Analysis and Statistics on Tensor Parameter Spaces

Inital Starting Point Analysis for K-Means Clustering: A Case Study

Obtaining Feature Correspondences

SUMMARY. denoise the original data at each iteration. This can be

Model parametrization strategies for Newton-based acoustic full waveform

XRDUG Seminar III Edward Laitila 3/1/2009

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Filtering Images. Contents

Principal Component Analysis

Figure (5) Kohonen Self-Organized Map

Lab # 2 - ACS I Part I - DATA COMPRESSION in IMAGE PROCESSING using SVD

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2

Chapter 2 Basic Structure of High-Dimensional Spaces

Computer vision: models, learning and inference. Chapter 13 Image preprocessing and feature extraction

Machine Learning / Jan 27, 2010

Robust Ring Detection In Phase Correlation Surfaces

White Paper. Abstract

Updated Sections 3.5 and 3.6

Anadarko Public Schools MATH Power Standards

Uncovering the Lagrangian of a system from discrete observations

GEOMETRIC MANIFOLD APPROXIMATION USING LOCALLY LINEAR APPROXIMATIONS

CoE4TN3 Medical Image Processing

Image Processing Lecture 10

Subdivision. Outline. Key Questions. Subdivision Surfaces. Advanced Computer Graphics (Spring 2013) Video: Geri s Game (outside link)

Experiments with Edge Detection using One-dimensional Surface Fitting

Automatic basis selection for RBF networks using Stein s unbiased risk estimator

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Edge detection. Convert a 2D image into a set of curves. Extracts salient features of the scene More compact than pixels

Dense Matrix Algorithms

Cluster Analysis (b) Lijun Zhang

3D Surface Recovery via Deterministic Annealing based Piecewise Linear Surface Fitting Algorithm

Probability Models.S4 Simulating Random Variables

Image Restoration and Reconstruction

SUPPLEMENTARY FILE S1: 3D AIRWAY TUBE RECONSTRUCTION AND CELL-BASED MECHANICAL MODEL. RELATED TO FIGURE 1, FIGURE 7, AND STAR METHODS.

CS 195-5: Machine Learning Problem Set 5

A Topography-Preserving Latent Variable Model with Learning Metrics

Transcription:

Chapter 3 Nonlinear Noise Reduction Many natural and engineered systems generate nonlinear deterministic time series that are contaminated by random measurement or dynamic noise. Even clean time series generated from simulations have inherent digitization noise, which magnitude is at least half of the digital resolution. Generally, one can express the measured time series as: x n = h(x n + η n ) + γ n, (3.) where h is a scalar smooth measurement function, η n random variable is the source of systematic noise and reflects the uncertainty in the underlying dynamic state variable, and γ n random variable is additive measurement noise. Generally, dealing with measurement noise is less problematic since, in reality, there is actual deterministic trajectory underlying the measurements. In contrast, the dynamic noise can reflect some random variations in system parameters, and, thus, we cannot claim that there is a deterministic attractor for the system. In some cases, the practical effect of both noise sources can be treated as just an additive noise. But, in general, the type and source of noise source will determine the extent of predictability and properties of the system. In the remainder of this chapter we will assume that our measurements are contaminated by the measurement noise and if there is any dynamical noise it can be treated as additive measurement noise. Chaotic time series have inherently broad spectra and are not amenable to conventional spectral noise filtering. Two main methods for filtering noisy chaotic time series [36, 37, 45] are: () model based filtering [3], and (2) projective noise reduction [36]. While model based reduction has some merit for systems with known analytical models [3], projective noise reduction provides a superior alternative [36] for experimental data. There are other methods like adaptive filtering or the wavelet shrinkage [7] method that can work in certain cases but require an opportune selection of parameters or some a priori knowledge of the system or the noise mechanisms [22]. In this chapter, we describe projective methods for nonlinear noise reduction that are based on a local subspace identification [3, 9, 36] in the reconstructed phase space [48, 6]. In particular, 45

46 CHAPTER 3. NONLINEAR NOISE REDUCTION Figure 3.: Schematic illustration of projective noise reduction idea. The blue curve shows the true phase space trajectory, and the tangent space to it at point ỹ n is shown by the red straight line. The noisy measured trajectory is shown by the black dots and line. The noise reduction idea is to project noisy y n onto the tangent space to obtain better approximation ȳ n ỹ n. we will describe POD and SOD based projective noise reductions. 3. Projective Noise Reduction Both POD-based local projective noise reduction [36], and SOD-based smooth projective noise reduction [4] work in the reconstructed phase space of the system generating the noisy time series. The basic idea is that, at any point on the attractor, noise leaks out to higher dimensions than the actual local dimension of the attractor s tangent space at that point. Thus, by embedding data into the higher d-dimensional space and then projecting it down to the tangent subspace of q-dimensions reduces the noise in the data as illustrated in Fig. 3.. The differences between the methods are in the way this tangent space is identified. For both noise reduction methods, the noisy time series {x i } n i= is embedded into a d-dimensional global phase space using delay coordinate embedding generating the vector-valued trajectory: y i = [ x i, x i+τ,... x i+(d )τ ] T, (3.2) where τ is the delay time, and we get a set of {y i } n (d )τ i= reconstructed phase space points. The selection of the global embedding dimension d, and the dimension of the tangent subspace q are important steps in both noise reduction schemes. For a noise-free time series, a method of false nearest neighbors [, 58] is usually used to estimate the minimum necessary embedding dimension D as described in Chapter 9. However, for the experimental time series we want proceed conservatively by embedding in large d and observing results for various q-dimensional projections. While POD based projective noise reduction works for both map or flow generated time series,

3.. PROJECTIVE NOISE REDUCTION 47 SOD based noise reduction is only applicable to continuous flow generated data. This is due to the fact that, for the SOD, we need to estimate data-based time derivative of the time series, which would be problematic for the map generated data. In addition, the described SOD method can also be modified to use POD instead of SOD when dealing with the continuous time series ob obtain the energy based subspace instead of smooth subspace for noise reduction. 3.. POD Based Local Projective Noise Reduction Standard methodology to approximate nonlinear map governing the evolution of a deterministic signal is to use local linear maps. Assuming the original map F(y) is a smooth nonlinear function of the reconstructed coordinates, we can linearize it about a point of interest: F(y) F(y i ) + DF(y i )(y y i ) = y i+ + DF(y i )(y y i ), (3.3) or y i+ = A i y i+ + b i, (3.4) where A i is the Jacobian DF(y i ). Usually A i is determined by a least squares fit to experimental data. The basic idea is to adjust y i by projecting it down to a subspace span by the dominant eigenvectors of A i which should approximate the tangent space to the actual dynamics at time i as Figure 3.2: Schematic of the local projective noise reduction. Consider a cloud of points on a k-dimensional hypersurface which are corrupted by noise. The noise spreads the points above and below the hypersurface so that in d-dimensional space they form a d-dimensional object of small but finite thickness. We can approximate the distribution of points by an ellipsoid and thus identify the direction of smallest extent. Then we project the points onto the subspace orthogonal to it.

48 CHAPTER 3. NONLINEAR NOISE REDUCTION illustrated in Fig. 3.2. In the local projective noise algorithm [34], for each point y i in the reconstructed d-dimensional phase space r temporarily uncorrelated nearest neighbor points y (j) i { } r j= are determined. This is usually done utilizing fast nearest neighbor search algorithms such a kd-tree based searches as was the case for the false nearest neighbor and fixed mass based algorithms. The POD is applied to this cloud of nearest neighbor points, and the optimal q-dimensional projection of the cloud is obtained providing the needed filtered ȳ i (which is also adjusted to account for the shift in the projection due to the trajectory curvature at y i [37]). The filtered time series { x i } n i= is obtained by averaging the appropriate coordinates in the adjusted vector-valued time series {ȳ i }. POD Subspace Identification and Data Projection POD is solved using singular value decomposition of a matrix Y i = containing the r nearest neighbors of point y i : [ y () i, y (2) i ],..., y (r) i R d r Y T i = P i Σ i Ξ T i, (3.5) where POMs are given by the columns of Ξ i R d d, POCs are given by the columns of P i R r d and SOVs are diagonal terms of Σ i squared. We will assume that the POVs are arranged in descending order (σ 2 σ2 2 σd 2). The most energy in q-dimensional approximation to Y (q < d) can be obtained by retaining only q columns of interest in the matrices P and Ξ i, and reducing Σ i to the corresponding q q minor. In effect, we are looking for a projection of Y i onto a q-dimensional most energetic subspace. The embedding of this q-dimensional projection into d-dimensional space (Ȳi) is constructed using the corresponding reduced matrices: P i R n q, Σ i R q q, and Ξ i R d q, as follows Ȳ T i = P i Σ i Ξ T i. (3.6) Adjusting for Bias due to Curvature The above approximation would be acceptable if our true clean trajectory passed through the center of mass of the cloud of nearest neighbors. However, due to the inherent trajectory curvature, the clean trajectory would pass through a point shifted outwards from the center of mass. Therefore, the identified POD subspaces will not be tangent to the actual trajectory s manifold, they would actually intersect it. Thus, we need to adjust our linear projection subspace towards the point of our interest by eliminating the shift caused by the curvature. As suggested by Kantz and Schreiber [38] one can do so by moving the projection as 2ȳ n ȳ n, where ȳ n = (ȳ n + ȳ n+ )/2 is the average of the neighborhood mass centers as illustrated in Fig. 3.3.

3.. PROJECTIVE NOISE REDUCTION 49 Figure 3.3: Consider a cloud of noisy points y k around a one dimensional curve going through our noise free point of interest y n. The local linear approximations are not tangents to the actual trajectory and the centre of mass (ȳ) of the neighborhood and is shifted inward with respect to the curvature. We can adjust for it by shifting the centre of mass ȳ n outward with respect to the curvature by 2ȳ n ȳ n. 3..2 SOD Based Noise Reduction In contrast, smooth projective noise reduction works with short strands of the reconstructed phase space trajectory. Therefore, this method can only be applicable for data coming from a continuous flow and not a map. These strands are composed of (2l + ) time-consecutive reconstructed phase space points, where l > d is a small natural number. Namely, each reconstructed point y k has an associated strand S k = [y k l,..., y k+l ], and an associated bundle of r nearest neighbor strands {S j k }r j=, including the original. This bundle is formed by finding (r ) nearest neighbor points {y j k }r j= for y k and the corresponding strands. SOD is applied to each of these strands in the bundle and the corresponding q-dimensional smoothest approximations to the strands { S j k }r j= are obtained. The filtered ȳ k point is determined by the weighted average of points {ỹ j k }r j= in the smoothed strands of the bundle. Finally, the filtered time series { x i } n i= are obtained just as in the local projective noise reduction. The procedure can be applied repeatedly for further smoothing or filtering. 3..3 Smooth Subspace Identification and Data Projection SOD is solved using a generalized singular value decomposition of the matrix pair S i and Ṡ i : S T i = U i C i Φ T i, Ṡ T i = V i G i Φ T i, C T i C i + G T i G i = I, (3.7) where SOMs are given by the columns of Φ i R d d, SOCs are given by the columns of Q i = U i C i R n d and SOVs are given by the term-by-term division of diag(c T i C i) and diag(g T i G i). The resulting SPMs are the columns of the inverse of the transpose of SOMs: Ψ i = Φ T i R d d. In this paper, we will assume that the SOVs are arranged in descending order (λ λ 2 λ d ).

5 CHAPTER 3. NONLINEAR NOISE REDUCTION The magnitude of the SOVs quadratically correlates with the smoothness of the SOCs [3]. The smoothest q-dimensional approximation to S i (q < d) can be obtained by retaining only q columns of interest in the matrices U i and Φ i, and reducing C i to the corresponding q q minor. In effect, we are looking for a projection of S i onto a q-dimensional smoothest subspace. The embedding of this q-dimensional projection into d-dimensional space ( S i ) is constructed using the corresponding reduced matrices: Ū i R n k, C i R k k, and Φ i R d k, as follows S T i = Ū i C i Φ T i. (3.8) 3..4 Data Padding to Mitigate the Edge Effects SOD-based noise reduction is working on (2l + )-long trajectory strands. Therefore, for the first and last l points in the matrix Y, we will not have full-length strands for calculations. In addition, the delay coordinate embedding procedure reconstructs the original n-long time series into Y which has only [n (d )τ]-long rows. Therefore, the first and last (d )τ points in {x i } n i= will not have all d components represented in Y, while other points will have their counterparts in each column of Y. To deal with these truncations, which cause unwanted edge effects in noise reduction, we pad both ends of Y by [l + (d )τ]-long trajectory segments. These trajectory segments are identified by finding the nearest neighbor points to the starting and the end points inside Y itself (e.g., y s and y e, respectively). Then, the corresponding trajectory segments are extracted from Y: Y s = [y s l (d )τ,..., y s ] and Y e = [y e+,..., y e+l+(d )τ ], and are used to form a padded matrix Ŷ = [Y s, Y, Y e ]. Now, the procedure can be applied to all points in Ŷ starting at l + and ending at n + (d )τ + l points. 3.2 Smooth Noise Reduction Algorithm Smooth noise reduction algorithm is schematically illustrated in Fig. 3.4. While we usually work with data in column form, in this figure for illustrative purpose we arranged arrays in row form. The particular details are explained as follows:. Delay Coordinate Embedding (a) Estimate the appropriate delay time τ and the minimum necessary embedding dimension D for the time series {x i } n i=. (b) Determine the local (i.e., projection), q D, and the global (i.e., embedding), d D, dimensions for trajectory strands. (c) Embed the time series {x i } n i= using global embedding parameters (τ, d) into the reconstructed phase space trajectory Y R d (n (d )τ).

3.2. SMOOTH NOISE REDUCTION ALGORITHM 5 Figure 3.4: Schematic illustration of smooth projective noise reduction algorithm (d) Partition the embedded points in Y into a kd-tree for fast searching. 2. Padding and Filtering: (a) Pad the end points of Y to have appropriate strands for the edge points, which results in Ŷ R d [n+2l+(d )τ]. In addition, initialize a zero matrix Ȳ R d [n+2l+(d )τ] to hold the smoothed trajectory. (b) For each point {ŷ i } n+(d )τ+l i=l+ construct a (2l+)-long trajectory strand S i = [ŷ i l,..., ŷ i+l ] R d (2l+). The next step is optional intended for dense but noisy trajectories only. (c) If r >, for the same point ŷ i, look up (r ) nearest neighbor points that are temporarily uncorrelated, and the corresponding nearest neighbor strands {S j i }r j=2. (d) Apply SOD to each of the r strands in this bundle, and obtain the corresponding q- dimensional smooth approximations to all d-dimensional strands in the bundle { S j } r j=. (e) Approximate the base strand by taking the weighted average of all these smoothed strands S i = S j i j, using weighting that diminishes contribution with the increase in the distance from the base strand. (f) Update the section corresponding to ŷ i point in the filtered trajectory matrix as Ȳ i = Ȳi + S i W, (3.9)

52 CHAPTER 3. NONLINEAR NOISE REDUCTION tent exponential - comb -2-3 -4-5 - -8-6 -4-2 2 4 6 8 Figure 3.5: Sample weighting functions for l = size strand where Ȳi is composed of (i l)-th through (i + l)-th columns of Ȳ: Ȳ i = [ȳ i l,..., ȳ i,..., ȳ i+l ], (3.) and W = diag(w i ) (i = l,..., l) is the appropriate weighting along the strand with the peak value at the center as illustrated in Fig. 3.5. 3. Shifting and Averaging: (a) Replace the points {ŷ i } n+(d )τ+l i=+l at the center of each base strand by its filtered approximation ȳ i determined above to form Ȳ R d [n+2l+(d )τ]. (b) Average each d smooth adjustment to each point in the time series to estimate the filtered point: x i = d d Ȳ (k, i + l + (k )τ). k= 4. Repeat the first three steps until data is smoothed out, or some preset criterion is met. 3.3 Application of the Algorithms In this section we evaluate the performance of the algorithm by testing it on a time series generated by Lorenz model [68] and a double-well Duffing oscillator [46]. The Lorenz model used to generate chaotic time series is: ẋ = 8 x + yz, ẏ = (y z), and ż = xy + 28y z. (3.) 3 The chaotic signal used in the evaluation is obtained using the following initial conditions (x, y, z ) = (, 5, 5), and the total of 6, points were recorded using. sampling time period. The

3.4. RESULTS AND DISCUSSION 53 double-well Duffing equation used is: ẍ +.25 ẋ x + x 3 =.3 cos t. (3.2) The steady state chaotic response of this system was sampled thirty times per forcing period and a total of 6, points were recorded to test the noise reduction algorithms. In addition, a 6, point, normally distributed, random signal was generated, and was mixed with the chaotic signals in /, /, /5, 2/5, and 4/5 standard deviation ratios, which respectively corresponds to 5,,, 4, and 8 % noise in the signal or SNRs of 26.2,, 3.98, 7.96, and.94 db. This results in total of five noisy time series for each of the models. The POD-based projective and SOD-based smooth noise reduction procedures were applied to all noise signals using total of ten iterations. To evaluate the performance of the algorithms, we used several metrics that include improvements in SNRs and power spectrum, estimates of the correlation sum and short-time trajectory divergence rates (used for estimating the correlation dimension and short-time largest Lyapunov exponent, respectively [33]). In addition, phase portraits were examined to gain qualitative appreciation of noise reduction effects. 3.4 Results and Discussion 3.4. Lorenz Model Based Time Series Using average mutual information estimated from the clean Lorenz time series, a delay time of 2 sampling time periods is determined. The false nearest neighbors algorithm (see Fig. 3.6) indicates that the attractor is embedded in D = 3 dimensions for the noise-free time series. Even for the noisy time series the trends show clear qualitative change around four or five dimensions. Therefore, six is 8 % noise 5% noise % noise % noise 4% noise 8%noise FNNs (%) 6 4 2 3 4 5 6 7 8 9 Embedding Dimension (size) Figure 3.6: False nearest neighbor algorithm results for Lorenz time series

54 CHAPTER 3. NONLINEAR NOISE REDUCTION x i + 2 n i + 2 Power/frequency (db/hz) Original x i n i Frequency (Hz) Original x i + 2 x i + 2 Power/frequency (db/hz) x i x i Frequency (Hz) Figure 3.7: Phase portrait from Lorenz signal (a); random noise phase portrait (b); power spectrum of clean and noisy signals (c); POD-filtered phase portrait (d) and SOD-filtered phase portrait (e) of 5% noise added data; and the corresponding power spectrums (f). used for global embedding dimension (d = 6), and k = 3 is used as local projection dimension. The reconstructed phase portraits for the clean Lorenz signal and the added random noise are shown in Fig. 3.7 (a) and (b), respectively. The corresponding power spectral densities for the clean and the noise-added signals are shown in Fig. 3.7 (c). The phase portraits of POD- and SOD-filtered signals with 5% noise are shown in Fig. 3.7 (d) and (e), respectively. Fig. 3.7 (f) shows the corresponding power spectral densities. In all these and the following figures, 64 nearest neighbor points are used for POD-based algorithm, and bundles of 9 eleven-point-long trajectory strands are used for SOD-based algorithm. The filtered data shown is for successive applications of the noise reduction algorithms. The decrease in the noise floor after filtering is considerably more dramatic for the SOD algorithm when compared to the POD. The successive improvements in SNRs after each iteration of the algorithms are shown in Fig. 3.8, and the corresponding numerical values are listed in Table 3. for the 5% and % noise added signals. While the SNRs are monotonically increasing for the SOD algorithm after each successive application, they peak and then gradually decrease for the POD algorithm. In addition, the rate of increase is considerably larger for the SOD algorithm compared to the POD, especially during the initial applications. As seen from the Table 3., the SOD-based algorithm provides about 3 db improvement in SNRs, while POD-based algorithm manages only 7 8 db improvements.

3.4. RESULTS AND DISCUSSION 55 SNR db 4 3 5% noise % noise % noise 4% noise 8% noise 2 4 6 8 Iteration Number SNR db 4 3 5% noise % noise % noise 4% noise 8% noise 2 4 6 8 Iteration Number Figure 3.8: Noisy Lorenz time series SNRs versus number of application of POD-based (left) and SOD-based (right) nonlinear noise reduction procedures The estimates of the correlation sum and short-time trajectory divergence for the noise-reduced data and the original clean data are shown in Fig. 3.9. For the correlation sum, both POD- and SOD-based algorithms show similar scaling regions and improve considerably on the noisy data. For the short-time divergence both methods manage to recover some of the scaling regions, but SOD does a better job at small scales, where POD algorithm yields large local fluctuations in the estimates. log (C ( )) 3 Clean Data POD filtered SOD filtered log (C ( )) 3 Clean Data POD filtered SOD filtered.5.5.5 log ( ).5.5.5 log ( ) 3 3 2 2 ln δ n Clean Data POD filtered SOD filtered 3 4 n (timesteps) ln δ n Clean Data POD filtered SOD filtered 3 4 n (timesteps) Figure 3.9: Correlation sum (top plots) and short-time divergence (bottom plots) for 5% (left plots) and % (right plots) noise contaminated Lorenz time series The qualitative comparison of the phase portraits for both noise reduction methods are shown in Fig. 3.. While POD does a decent job al low noise levels, it fails at higher noise levels where

56 CHAPTER 3. NONLINEAR NOISE REDUCTION Table 3.: SNRs for the noisy time series from Lorenz model, and for noisy data filtered by PODand SOD-based algorithms Signal Lorenz + 5% Noise Lorenz + % Noise Algorithm None POD SOD None POD SOD SNR (db) 26.6 33.835 39.658 3.9794 22.3897 27.438 no deterministic structure is recovered at 8% noise level. In contrast, SOD provides smoother and cleaner phase portraits. Even at 8% noise level, the SOD algorithm is able to recover large deterministic structures in the data. While SOD still misses small scale features at 8% noise level, topological features are still similar to the original phase portrait in Fig. 3.7(a). 5% Noise % Noise % Noise 4% Noise 8% Noise x i + 2 POD Filtered x i + 2 5 5 SOD Filtered x i + 2 x i x i x i x i x i Figure 3.: Phase space reconstructions for noisy and noise-reduced Lorenz time series after ten iterations of the POD- and SOD-based algorithms 3.4.2 Duffing Equation Based Time Series Using the noise-free Duffing time series and the average mutual information, a delay time of seven sampling time periods is determined. In addition, false nearest neighbors algorithm indicates that the attractor is embedded in D = 4 dimensions for % noise (see Fig. 3.). Six is used for global embedding dimension (d = 6), and k = 3 is used as local projection dimension. The reconstructed

3.4. RESULTS AND DISCUSSION 57 FNNs (%) 8 6 4 % noise 5% noise % noise % noise 4% noise 8%noise 2 3 4 5 6 7 8 9 Embedding Dimension (size) Figure 3.: False nearest neighbor algorithm results for Duffing time series phase portraits for the clean Duffing signal and the added random noise are shown in Fig. 3.2 (a) and (b), respectively. The corresponding power spectral densities for the clean and noise-added signals are shown in Fig. 3.2 (c). The phase portraits of POD- and SOD-filtered signals with 5% noise are shown in Fig. 3.2 (d) and (e), respectively. Fig. 3.2 (f) shows the corresponding power spectral densities. In all these and the following figures, 64 nearest neighbor points were used for POD-based algorithm, and bundles of nine -point-long trajectory strands were for SODbased algorithm. The filtered data shown is after successive applications of the noise reduction algorithms. As before, the decrease in the noise floor after filtering is considerably more dramatic for the SOD algorithm when compared to the POD. The successive improvements in SNRs after each iteration of the algorithms are shown in Fig. 3.3, x i + 7 n i + 7 2 Power/frequency (db/hz) Original (a) (b) (c) 2 x i n i Frequency (Hz) 2 Original x i + 7 x i + 7 Power/frequency (db/hz) (d) (e) (f) x i x i Frequency (Hz) 2 Figure 3.2: Reconstructed phase portrait from Duffing signal (a); random noise phase portrait (b); power spectrum of clean and noisy signals (c); POD-filtered phase portrait of 5% noise added (d); SOD-filtered phase portrait of 5% noise added (e); and the corresponding power spectrums (f).

58 CHAPTER 3. NONLINEAR NOISE REDUCTION SNR db 4 3 5% noise % noise % noise 4% noise 8% noise 2 4 6 8 Iteration Number SNR db 4 3 5% noise % noise % noise 4% noise 8% noise 2 4 6 8 Iteration Number Figure 3.3: Noisy Duffing time series SNRs versus number of application of POD-based (left) and SOD-based (right) nonlinear noise reduction procedures and the corresponding numerical values are listed in Table 3.2 for 5% and % noise added signals. Here, SNRs for both methods are peaking at some point and then gradually decrease. In addition, the rate of increase is considerably larger for the SOD compared to the POD algorithm, especially during the initial iterations. SOD is usually able to increase SNR by about db, while POD provides only about 4 db increase on average. Table 3.2: SNRs for the noisy time series from Duffing model, and for noisy data filtered by PODand SOD-based algorithms Signal Duffing + 5% Noise Duffing + % Noise Filtering None POD SOD None POD SOD SNR (db) 26.6 3.7583 36.687 3.9794 9.5963 24.5692 The estimates of the correlation sum and short-time trajectory divergence for the noisy, noise reduced data, and the original clean data from Duffing oscillator are shown in Fig. 3.4. For the correlation sum, both POD- and SOD-based algorithms show similar scaling regions and improve substantially on the noisy data, with SOD following the noise trend closer than POD. For the shorttime divergence rates both methods manage to recover some of the scaling regions, but SOD does a better job at small scales, where POD algorithm causes large local fluctuations in the estimates. The qualitative comparison of the Duffing phase portraits for both noise reduction methods are shown in Fig. 3.5. For consistency all noise-reduced phase portraits are obtained after consecutive applications of the algorithms. While POD does a decent job at low noise levels, it fails at high noise levels where no deterministic structure is recovered at 4% or 8% noise levels. In contrast, SOD provides smoother and cleaner phase portraits at every noise level. Even at 8% noise level, the SOD algorithm is able to recover some large deterministic structures in the data,

3.4. RESULTS AND DISCUSSION 59 log (C ( )) 3 Clean Data POD filtered SOD filtered log (C ( )) 3 Clean Data POD filtered SOD filtered.5.5 log ( ).5.5 log ( ) ln δ n Clean Data POD filtered SOD filtered 3 3 4 n (timesteps) ln δ n Clean Data POD filtered SOD filtered 3 3 4 n (timesteps) Figure 3.4: Correlation sum (top plots) and short-time divergence (bottom plots) for 5% (left plots) and % (right plots) noise contaminated Duffing time series providing topologically similar phase portrait. In summary, the decrease in the noise floor observed in the power spectral density for the SODbased method was at least db larger then for the POD-based method, which itself improved on the original signals noise floor at best by db. SNRs show similar trends, where SOD-based method manages to improve it by about db, while POD-based manages only 4 5 db increase. In addition, SOD accomplishes this with considerably fewer applications of the noise reduction scheme, while POD works more gradually and requires many more iterations at higher noise levels. Estimating nonlinear metrics used in calculations of correlation dimension and short-time Lypunov exponent showed that both methods are adequate at low noise levels. However, SOD performed better at small scales, where POD showed large local fluctuations. At larger noise levels both methods perform similarly for the correlation sum, but SOD provides larger scaling regions for short-time trajectory divergence rates. POD still suffers from large variations at small length/time scales for larger noise levels. The above quantitative metrics do not tell a whole story, however; we also need to look at trajectories themselves. Visual inspection of the reconstructed phase portraits showed superior results for SOD. POD trajectories had small local variations even at low noise levels, and completely failed to recover any large dynamical structures at large noise levels. In contrast, SOD method provides very good phase portraits up to % noise levels, and for larger noise levels the reconstructions still have clear topological consistency with the original phase portrait. However, SOD also fails to capture

6 CHAPTER 3. NONLINEAR NOISE REDUCTION 5% Noise % Noise % Noise 4% Noise 8% Noise x i + 7.5.5.5 2 2 2 2 POD Filtered x i + 7 SOD Filtered x i + 7 x i x i x i x i x i Figure 3.5: Phase space reconstructions for noisy and noise-reduced Duffing time series after ten iterations of the POD- and SOD-based algorithms finer phase space structures for 8% noise level. Problems Problem 3. Create an artificial time series by integrating the Chua s circuit by Matsumoto et al. (985): ẋ = α (y x + bx + 2 ) (a b) ( x + x ) (3.3) ẏ = x y + z (3.4) ż = βy (3.5) using these parameters: α = 9, β = /7, a = 8/7, and b = 5/7. Start the simulation using the following initial conditions: x() =, y() =, and z() =.6. In particular, using the z time series containing at least 6, points:. Make sure your time series is appropriately sampled, make it zero mean and unit standard deviation. 2. Determine the needed embedding parameters.

3.4. RESULTS AND DISCUSSION 6 3. Make noisy time series with %, 5 %, and % additive Gaussian noise (percents are in terms of relative standard deviations of the signal and the noise). 4. Apply both projective and smooth noise reduction schemes to the noisy time series. 5. Compare the results using phase portraits, power spectrum, correlation sum, and Lypunov exponents. 6. Interpret your results.