ENHANCED MONITORING USING MULTISCALE EXPONENTIALLY WEIGHTED MOVING AVERAGE CONTROL CHARTS

Size: px
Start display at page:

Download "ENHANCED MONITORING USING MULTISCALE EXPONENTIALLY WEIGHTED MOVING AVERAGE CONTROL CHARTS"

Transcription

1 ENHANCED MONITORING USING MULTISCALE EXPONENTIALLY WEIGHTED MOVING AVERAGE CONTROL CHARTS A Thesis by MD. ALAMGIR MOJIBUL HAQUE Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE Chair of Committee, Co-chair of Committee, Committee Members, Head of Department, Mohamed N. Nounou Hazem N. Nounou Luc Vechot M. Nazmul Karim August 2016 Major Subject: Chemical Engineering Copyright 2016 Md. Alamgir Mojibul Haque

2 ABSTRACT The exponentially weighted moving average (EWMA) method is a widely used univariate process monitoring technique. This conventional EWMA technique is normally designed to optimize the out of control average run length (ARL1) specific to a fixed in control average run length (ARL0). This design procedure of EWMA technique is based on some assumptions the evaluated process residuals are Gaussian, independent and contain moderate level of noise. Violation of these assumptions may adversely affect its fault detection abilities. Wavelet based multiscale representation of data is a powerful data analysis tool and has inherent properties that can help deal with these violations of assumptions, which thus improve the performance of EWMA through satisfying its assumptions. The main purpose of this work is to develop a multiscale EWMA technique with improved performance over the conventional technique and establish a design procedure for this method to optimize its parameters by minimizing the out of control average run length for different fault sizes and using a specified in control average run length assuming that the residuals are contaminated with zero mean Gaussian noise. Through several comparative studies using Monte Carlo simulations, it has been shown that the multiscale EWMA technique provides a better performance over the conventional method. Multiscale EWMA is shown to provide smaller ARL1 and missed detection rate with a slightly higher false alarm rate compared to the conventional EWMA technique not only when both the techniques are designed to perform optimally but also ii

3 when data violate the assumptions of the EWMA chart. The advantages of the multiscale EWMA method over the conventional method are also illustrated through their application to monitor a simulated distillation column. iii

4 DEDICATION I dedicate this thesis to my parents, Md. Samsul Haque and Nurjahan Beguam, for their inspiration and unconditional love. I would also like to dedicate this work to all my teachers who taught to me to understand the true meaning of being educated. iv

5 ACKNOWLEDGEMENTS I would like to express my gratitude to my committee chair, Dr. Mohamed Nounou, for his insightful ideas and valuable guidance which had helped me overcoming the difficult phases of my Master s thesis at Texas A&M University. Without his persistent help and continuous inspiration, this dissertation work would not have been possible. I would also like to thank my committee co-chair Dr. Hazem Nounou and committee member Dr. Luc Vechot for their important advice and feedback through the course of my research. I am also grateful to Dr. Majdi Mansouri and Md. Ziyan Sheriff for helping me in completing the thesis by providing various resources. In addition, I am deeply obliged to Texas A&M University at Qatar for providing me the opportunity to do research. And last but not the least, I acknowledge the motivation and inspiration given to me by my parents. v

6 NOMENCLATURE SPM CUSUM MA EWMA PLS PCA CVA SPC UCL LCL ARL1 ARL0 AR SW ACF Statistical Process Monitoring Cumulative Sum Moving Average Exponentially Weighted Moving Average Partial Least Squares Principal Component Analysis Canonical Variate State Space Statistical Process Control Upper Control Limit Lower Control Limit Out of Control Average Run Length In Control Average Run Length Autoregressive Shapiro-Wilk Autocorrelation Function vi

7 TABLE OF CONTENTS vii Page ABSTRACT... ii DEDICATION... iv ACKNOWLEDGEMENTS... v NOMENCLATURE... vi TABLE OF CONTENTS...vii LIST OF FIGURES... ix LIST OF TABLES... xiii 1. INTRODUCTION Literature review Conventional univariate monitoring techniques Indicators for monitoring process performance Research objectives MONITORING USING EWMA CHARTS Design procedure of the conventional EWMA technique Assessing the performance of the EWMA chart under violation of assumptions Assessing the impact of high noise levels in the data on the performance of the EWMA chart Assessing the impact of autocorrelation in the data on the performance of EWMA chart Assessing the impact of deviation from normality in the data on the performance of the EWMA chart WAVELET BASED MULTISCALE REPRESENTATION Introduction to wavelet based multiscale representation Advantages of multiscale representation of data Noise feature separation Decorrelation of autocorrelated data Data are closer to normal at multiple scales... 38

8 Page 4. WAVELET BASED MULTISCALE EWMA CHART Process monitoring using Multiscale EWMA Chart Design procedure of optimizing the parameters of the multiscale EWMA fault detection technique Optimizing the multiscale EWMA parameters Design steps for the multiscale EWMA technique PERFORMANCE COMPARISON BETWEEN THE CONVENTIONAL AND MULTISCALE EWMA CHARTS Comparison between the performance of the conventional and multiscale EWMA techniques under no violations of the EWMA assumptions Comparison between the performances of the conventional and multiscale EWMA techniques under violations of the EWMA assumptions Comparison of performance using data with different level of noise Comparison of performance using autocorrelated data Comparison of performance using non-gaussian (chi-square) data APPLICATION OF THE MULTISCALE EWMA CHART CONCLUSIONS AND FUTURE DIRECTIONS Concluding remarks Future directions REFERENCES viii

9 LIST OF FIGURES Page Figure 1: A schematic representation of the weightings used to compute the detection statistics used in various univariate SPC charts....7 Figure 2: Schematic representation of false alarm and missed detection Figure 3: Schematic representation of average run length Figure 4: Trade-off between false alarm rate and missed detection rate Figure 5: Combination of λ and L for ARL Figure 6: Optimal λ for different fault size for ARL Figure 7: Impact of noise level on ARL1 values for the conventional EWMA chart Figure 8: Impact of noise level on the false alarm and missed detection rates of the conventional EWMA chart Figure 9: EWMA statistics for different noise levels Figure 10: Impact of autocorrelation on ARL1 for the conventional EWMA chart Figure 11: Impact of autocorrelation on the false alarm and missed detection rates for the conventional EWMA chart Figure 12: Impact of autocorrelation on the performance of EWMA chart for an autoregressive coefficient value of Figure 13: Impact of autocorrelation on the performance of EWMA chart for an autoregressive coefficient value of Figure 14: Impact of deviation from normality on ARL1 for the conventional EWMA chart Figure 15: Impact of deviation from normality on the false alarm and missed detection rates for the conventional EWMA chart Figure 16: Impact of deviation from normality on the performance of EWMA chart, SW = Figure 17: Impact of deviation from normality on the performance of EWMA chart, SW stat = ix

10 x Page Figure 18: Schematic diagram for multiscale representation of data Figure 19: Decorrelation of autocorrelated data at multiple scales Figure 20: Distribution of chi-squared data at multiple scales Figure 21: A schematic diagram of the multiscale EWMA fault detection algorithm Figure 22: Combination of λ and L for in control average run length Figure 23: 3-D plot of the optimization of parameters for fault size 2.5σ for multiscale EWMA chart for ARL0 value of Figure 24: Optimal λ's for the multiscale EWMA chart for different fault sizes for ARL0 = Figure 25: Effect of decomposition depth on the ARL1 of the multiscale EWMA chart Figure 26: Effect of decomposition depth on false alarm and missed detection rates of the multiscale EWMA chart Figure 27: Performance comparison between the conventional and multiscale EWMA charts in terms of ARL1 for different fault sizes Figure 28: Performance comparison between the conventional and multiscale EWMA charts in terms of false alarm rate for different fault sizes Figure 29: Performance comparison between the conventional and multiscale EWMA charts in terms of missed detection rate for different fault sizes Figure 30: Trade-off between false alarm and missed detection rate for the multiscale EWMA chart Figure 31: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL1 for different level of noise in the data Figure 32: Performance comparison between the multiscale and conventional EWMA techniques in terms of false alarm and missed detection rate for different level of noise in the data Figure 33: Detection statistics for the conventional and multiscale methods in the case where the noise standard deviation equals

11 Page Figure 34: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL1 for different level of autocorrelation in the data Figure 35: Performance comparison between the multiscale and conventional EWMA technique in terms of false alarm and missed detection rates for different level of autocorrelation in the data Figure 36: Detection statistic for the conventional and multiscale methods in the case where the autoregressive coefficient equal to Figure 37: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL1 for non-gaussian data (Chi-square) Figure 38: Performance comparison between the multiscale and conventional EWMA technique in terms of false alarm and missed detection rates for non- Gaussian data (Chi-square) Figure 39: Detection statistics for the conventional and multiscale methods in the case where the Shapiro Wilk equals Figure 40: Comparing the performances of the conventional and multiscale EWMA chart for a step fault of magnitude ±2σ in the residuals of simulated distillation column data Figure 41: Comparing the performances of the conventional and multiscale EWMA chart for a step fault of magnitude ±σ in the residuals of simulated distillation column data xi

12 LIST OF TABLES Page Table 1: Comparison of optimum values of L and λ with those obtained from literature xii

13 1. INTRODUCTION Process monitoring is essential for proper operation of various engineering systems. The goal of statistical process monitoring (SPM) is to detect the occurrence of faults (fault detection) and the nature of operational changes that cause a process to deviate from its desired target (fault diagnosis). Thus, process monitoring involves two main tasks: fault detection and fault diagnosis. There are various types of fault detection techniques, which will be discussed later. However, fault detection needs to be followed by fault diagnosis that aims at locating the root cause of the process change and enables the process operators to take necessary actions to correct the situation (process recovery), thereby returning the process back to its desired operation. In this way, the product quality can be maintained and safe operation can be assured. In this work, improving the task of fault detection is addressed. The term fault here is generally defined as a shift from the target value of a variable or a calculated parameter associated with a process [1]. Fault detection methods can be classified into three categories [2]: I. Model-based methods II. III. Data based methods Knowledge based methods In model based approaches, measured data of a process variable are compared with a model. This model is obtained from basic understanding about the process and is expressed in terms of mathematical relationships between the process inputs and outputs 1

14 [3][4]. Diagnostic observers, parity relations, Kalman filters and parameter estimation techniques are some of the frequently used model based approaches. The effectiveness of the model based methods depends on the accuracy of the process model. But in practice, it can be very difficult task to derive an accurate model, especially if the process involves a large number of inputs and outputs. Knowledge based methods provide an alternative approach to the model based approach in the case of complex processes with incomplete knowledge or when analytical models are not available. Examples of knowledge based methods include causal analysis and expert systems [2][5][6]. Data based methods, on the other hand, rely on the availability of process data, from which process information can be extracted and used for fault detection and diagnosis [7][8][9]. In this category of methods, historical process data are collected by measuring key process variables under fault free conditions which are then used to construct an empirical model. This empirical model is later used to find out the residuals, which quantify the difference between the observed value of a variable and the expected value of that variable predicted by the process model. These residuals are then evaluated to monitor the process. Several data based techniques can be found in the literature. These techniques are divided into two different classes depending on the number of variables that are being monitored: univariate and multivariate techniques[10]. Univariate process monitoring techniques are used to monitor a single variable, while multivariate techniques are used to monitor multiple process variables. The univariate techniques include the Shewhart, 2

15 cumulative sum (CUSUM) chart, moving average (MA) and the exponentially weighted moving average (EWMA) chart. Multivariate monitoring techniques, on the other hand, include partial least square (PLS) [11], principal component analysis (PCA) [12], canonical variate state space (CVA) [13], and others. The scope of this work is limited to the univariate SPM techniques. Among the univariate techniques, Shewhart chart is one of the simplest charts. It evaluates the raw residuals, i.e., it doesn t use any filter to process the residuals. The Shewhart chart has been shown capable of detecting large faults or large shifts in the mean of a process variable. This is because the Shewhart chart only considers the most recent data sample in fault detection, which makes it not very sensitive to small changes in the data. Other univariate techniques apply linear filters on the residuals to improve their sensitivities to small shifts, such as the CUSUM and EWMA techniques. Another advantage of these filters is that they reduce the noise content in the data. However, it is known that linear filters are not very effective in removing noise from real data because they operate at a single scale, i.e., they work at a fixed scale or frequency and discard all features in the data that are above a certain frequency level [14][15]. In practice, however, process data are multiscale in nature due to the changes that occur in the process at different times and different frequencies. So the mismatch between the nature of the measured data and the nature of the linear filters makes the traditional statistical process control (SPC) methods inappropriate to deal with practical data. Non-linear filters like wavelet based multiscale filtering have shown much promise in dealing with real data [15][16] [17][18]. 3

16 Another limitation associated with the existing univariate fault detection methods is that these methods assume the measured data are independent and normally distributed (Gaussian). Real process data, however, don t usually follow these assumptions, which deteriorates the performance of these conventional monitoring methods. Multiscale wavelet based method has showed inherent ability to deal with those assumptions [15], which helps improve the effectiveness of process monitoring, especially for correlated data [19][20]. Therefore, the objective of this work is to utilize the advantages of multiscale wavelet based representation to improve the monitoring performance of the EWMA control chart, especially for the detection of faults with small magnitudes. A design procedure for optimizing the EWMA parameters (which is based on minimizing the out of control average run length) will be developed. Also, the performance of the developed multiscale EWMA technique will be assessed and compared with its conventional time domain counterpart in the cases where the data are autocorrelated and non-gaussian. The different techniques will be compared using three performance indices, which include the missed detection rate, false alarm rate, and the average run length. In the next sections, a review on some of the popular univariate control charts will be presented, followed by a description of the main research objectives of this work Literature review In this section, descriptions of the commonly used univariate control charts are presented along with, their advantages and limitations. 4

17 Conventional univariate monitoring techniques Shewhart chart Walter Shewhart first developed Shewhart chart in the 1920s [21]. It was intended to monitor quality of a manufacturing process at different stages. It is widely used in the field of statistical quality control because of its computational simplicity which makes it easy to implement. Only three features are needed to design a Shewhart chart a center line (C) or a mean value, the upper control limit (UCL) and the lower control limit (LCL) [22]. The shewhart chart can be of different types depending on the parameters that are being monitored. To monitor the average level of a process variable, the mean chart (x ) is used. On the other hand, the Range chart or standard deviation (S) chart is used to monitor the sample process variation or spread. Shewhart chart is usually used to monitor the sample mean. In some occasions, the mean chart is coupled with the range chart or S chart when robustness against the variability in the observations is required. The simultaneous use of both the mean and range chart ensures the capture of almost all important features hidden in the data as only one chart may not be able to do that. The sample mean of a particular process variable, x can be computed by the following equations [22]: x i = x ij n 5 n j=1 k x = x i k i=1

18 where, n and k represent the subgroup size and the number of subgroups respectively. The number of subgroups usually represents the number of sensors that measure a particular variable. So, the subgroup size k is equal to 1 when only one sensor measures the variable and then x i equals to x. When a single variable is being monitored by multiple sensors, then sub-grouping is required. The upper and lower control limits are defined as follows: UCL, LCL = x + L n L n = cσ n where, σ represents the standard deviation of the priori data when there are no faults in the data and c is a constant which is computed using a nomogram. When a process sample observation falls outside the control limits, then this indicates that the process mean has shifted from the target value, x. So, it is evident that the control limits have to be chosen carefully, which depend on the Ln value. Generally a value of 3σ is used for the parameter Ln [23] because finding accurate values of c from nomogram for different processes is very difficult. Similar equations have been used to compute the center line, upper and lower control limits for the R and S chart[24]. As indicated earlier, the Shewhart chart can t detect relatively small faults, so it is not advisable to use it in fault detection when small deviations from the process mean are expected. In fact, the Shewhart chart is only able to detect faults larger than three times the standard deviation of the original signal [25], which is a major drawback of the Shewhart chart. This inability to detect small mean shifts is due to the short memory of 6

19 the Shewhart chart as it only considers the current process measurement. Other control charts have longer memory as they average past samples to compute their detection statistics. The weighing used to compute the detection statistic for various charts are shown in Figure 1. Figure 1: A schematic representation of the weightings used to compute the detection statistics used in various univariate SPC charts CUSUM chart An effective alternative to the Shewhart chart is the cumulative sum (CUSUM) chart. It was first introduced by Page in 1954 [26]. The design of the conventional CUSUM chart involves computing the CUSUM statistic which is defined by the following equation [27]: 7

20 i S i = (x j μ 0 ) j=1 This quantity Si is plotted against the sample number i where μ 0 is the mean value of the process variable under fault free conditions. The CUSUM statistic can also be computed recursively as follows: S i = (x i μ 0 ) + S i 1 When several observations are available at each time sample, then the observation xj is replaced by average of all the observations at a particular sample time, x. j To detect a particular fault in the process, the one-sided CUSUM charts are used by plotting the following statistic [27]: S i = max[0,x i (μ 0 + K)] where, K is the reference value to detect a shift in the mean of size Δ. K is defined as follows: K = 2 When Si exceeds a decision interval H, then it is assumed that mean of the process variable has shifted from the targeted value by a margin of Δ. The value of H can be computed as: H = d 2 where d = ( 2 β δ2) ln(1 ) and δ = α σ x where, α and β are the type I and type II error probabilities respectively and σ x is the standard deviation of the process variable. The most popular form of the CUSUM chart, however, is the two-sided CUSUM chart. The positive and negative CUSUM statistics are calculated as follows: 8

21 S H(i) = max[0,x i (μ 0 + K) + S H(i 1) ] S L(i) = max[0,(μ 0 K) x i + S L(i 1) ] When SH(i) and SL(i) exceed the decision interval or the control limits, then it is assumed that the process is not in control. Computation of the control limits require knowledge about probability density function of the distribution of process variable, which is usually hard to obtain. So, a practical experience is important in designing a CUSUM chart. Control limits of 4σ or 5σ are suggested to provide a reasonable detection for a mean shift of around 1σ in the process data [23]. The CUSUM chart can perform better than the Shewhart chart in detecting relatively small mean shifts although it can result in more false alarms or type I errors EWMA chart The exponentially weighted moving average chart was first introduced in the literature by Roberts in 1959 [28]. Since then it has been widely used as forecasting tool [29] and also as a tool for process monitoring and diagnosis [30]. The EWMA control scheme is easy to implement. The design of EWMA control scheme includes computation of the EWMA statistic and the upper and lower control limits. The EWMA statistic can be computed as follows Z i = λx i + (1 λ)z i 1, 0 < λ 1 where, λ is called a smoothing parameter, which changes the memory of the detection statistic. 9

22 EWMA is also known as a geometric moving average method, because the EWMA statistic can also be written as a moving average of the current and past observations as follows: i 1 Z i = λ (1 λ) j x i j + (1 λ) i Z 0 j=0 The above equation shows that the weights assigned to past observations decrease exponentially, giving the name of the EWMA technique. The upper and lower control limits are defined in terms of the standard deviation of the EWMA statistic and are computed as follows: UCL, LCL = µ 0 ± Lσ z = µ 0 ± Lσ λ (2 λ) [1 (1 λ)2i ] where, µ is the standard deviation of the observations. The last term in the bracket quickly converges to zero as the number of observation increases and thus the control limits can be computed as follows: UCL,LCL = µ 0 ± Lσ λ (2 λ) Whenever, the control statistic falls outside the range of the control limits, the process is considered to be out of control. To use the EWMA method, the choice of the smoothing parameter needs to be made carefully. Generally, a value in the range of 0.2 to 0.3 is found reasonable [27]. However, in practice, the optimum choice of the smoothing parameter depends on the size of the mean shift to be detected. For large mean shifts, large values of λ are needed, while smaller values of λ are needed to detect smaller mean shifts 10

23 more quickly [31][32]. This can be attributed to the fact that, when λ equals 1, the value of the EWMA statistic only uses the most recent observation. This makes the EWMA chart equivalent to the Shewhart chart, which is only capable of detecting large faults. On the other hand, for very small values of λ, the EWMA method becomes similar to the CUSUM, which is more capable of detecting smaller shifts. Furthermore, all the fault detection charts described earlier rely on some assumptions, which include: There is a moderate level of noise in the data. The process data are independent, i.e., uncorrelated The process data follow a normal or Gaussian distribution. Practical data, however, don t usually satisfy these assumptions. As a result, the performance of these control chart deteriorates. Several indicators are used to analyze the performance of a control chart. Those indicators are described in the next section Indicators for monitoring process performance The most commonly used indicators for assessing the performance of process monitoring performance include the out of control average run length (ARL1) and false alarm rate. In this work, these two indicators will be used along with the missed detection rate, which quantifies the effectiveness of detection achieved by the fault detection method. False alarm, which is also known as a type I error, represents the case where the SPC declares the presence of a fault when in reality there isn t any fault in the process. In 11

24 other words, control chart shows an out of control signal when the process is actually in control. Missed detection, on the other hand, means the SPC chart fails to detect a fault when actually a fault exists in the process, which is also a type II error. Missed detection and false alarms are illustrated in the Figure 2, in which the highlighted area represents a fault region, while the all other areas are fault free. Therefore, points in the fault free region that fall outside the control limits are false alarms, while any point inside the fault region, but within the control limits, is a missed detection. Figure 2: Schematic representation of false alarm and missed detection. The average run length, on the other hand, is the average number of samples a fault detection method takes before it declares the presence of a fault. Average run length can be used to characterize both types of error, I and II. The in control average run length (ARL0) is the average number of observations a control chart takes to show an out of 12

25 control signal when the process is in control. ARL0 corresponds to a type I error. On the other hand, out of control average run length (ARL1) represents the average number of observations that a control chart takes to declare a fault after a fault occurs, which corresponds to a type II error. These average run lengths are illustrated in figure 3. Figure 3: Schematic representation of average run length For a good control chart performance, all three indicators (ARL1, missed detection rate and false alarm rate) need to be as small as possible. But in practice, however, a control chart that is designed to respond quickly to certain changes in the process mean value will become sensitive to high frequency effects or noise. As a result, false alarm rate during normal operation will increase [33]. On the other hand, if one wants to reduce the false alarm rate by expanding the control width this can eventually increase the missed detection rate and out of control average run length. So this means that there is a trade-off between the false alarm rate and missed detection rate. This is illustrated in the following 13

26 figure where control width and smoothing parameter of EWMA chart has been varied for a fixed fault size of 1σ. Figure 4: Trade-off between false alarm rate and missed detection rate. Figure 4 shows that, it s not possible to decrease all the indicators at the same time, so it is very important to prioritize the indicators before designing the EWMA chart. By doing so, one can make sure that the control chart gives optimum performance in terms of a selected indicator. The selection of the indicator normally depends on the process requirement Research objectives As discussed earlier, violating the assumptions of the conventional univariate monitoring techniques (such as EWMA) degrades their performances. Multiscale representation has inherent abilities to deal with those assumptions and thus can help 14

27 improve the effectiveness of these techniques. Therefore, the main objective of this work is to utilize the advantages of multiscale representation of data to improve the fault detection abilities of the EWMA control chart, especially under violation of its assumptions, i.e., when the data have large noise content, autocorrelation, or non-gaussian distribution. Specifically, the following objectives will be sought in this work: Assess the performance of EWMA under the violation of its assumptions Develop a multiscale EWMA fault detection method that combines the advantages of multiscale representation and those of the EWMA technique Develop a design procedure for optimizing the parameters of the multiscale EWMA technique based on ARL1 Compare the performances of the multiscale EWMA and the connventional EWMA techniques using their optimum parameters Comparing the performances of the multiscale and conventional EWMA techniques under the violation of the main assumptions Provide possible directions for future research work 15

28 2. MONITORING USING EWMA CHARTS While the Shewhart chart takes into account only the current data sample to evaluate the process performance, the CUSUM and EWMA charts consider a weighted sum of past observations. The CUSUM chart gives equal weight to all past observations, while the EWMA chart gives more importance to the more recent observations [27]. Both of the CUSUM and EWMA charts perform almost equally in detecting small mean shift but the EWMA chart is somewhat easier to set up and operate. Moreover, since the EWMA statistic is weighted average of all past and current observations, it is less sensitive to the normality assumption [23]. For these reasons, the EWMA chart has been chosen to study for this work as a model of conventional univariate technique Design procedure of the conventional EWMA technique EWMA technique has been studied extensively by many researchers, and its properties and design procedures are well established [34][31][35]. The following design procedure of the conventional EWMA technique based on ARL1 has been developed [35]: Choose an acceptable value of ARL0, in control average run length Specify the minimum fault size that needs to be detected as quickly as possible, and determine the value of λ which produces the lowest ARL1 for that specific fault size. Find the value of the control width L, which along with the value of λ (found from previous step) provides the required ARL0 value. 16

29 Plots that can be used to find these optimum EWMA L and λ combinations for different ARL0 values are available in [35]. In this work, an attempt is made to reproduce those curves by simulation so that similar plots can be constructed for the multiscale EWMA technique in later stages. To illustrate how to reproduce these plots, an ARL0 value of 500 has been selected as an example. Then, training fault free data consisting of 8192 zero mean Gaussian observations having a unit standard deviation are used to find out different combination of λ and L values such that each combination gives an in control average run length (ARL0) of 500. A Monte Carlo simulation of 5000 realizations is used for each combination to be sure that these combinations in fact give the specified ARL0. The following figure is constructed using all of these combinations. Width, L λ Figure 5: Combination of λ and L for ARL

30 λ To select the optimum combination that gives the lowest ARL1, faulty testing data having the same length as the training data and a fault of size equal to the standard deviation of the data (i.e., 1σ) are generated. Then the EWMA chart is applied to the testing data using all the combinations of λ and L values shown in Figure 5 to see which combination gives the lowest ARL1 value. This combination of L and λ values is the optimum for a fault size of 1σ. A Monte Carlo simulation is used for each combination to get statistically meaningful results. The same procedure is repeated for different fault sizes, which provides the results shown in Figure Fault size Figure 6: Optimal λ for different fault size for ARL Optimum values of λ and L for different fault sizes found by simulation are compared with the values obtained from Crowder[35] in Table 1 below. The values are sufficiently close which validate the simulations in this work. 18

31 Table 1: Comparison of optimum values of L and λ with those obtained from literature. Fault size ARL0 Optimum width, L Optimum λ Simulated value Reported in Crowder [35] Simulated value Reported in Crowder [35] Having verified the design procedure for the conventional EWMA technique based on lowest ARL1 values for different fault sizes, EWMA method can be used to assess how it performs under violation of its assumptions. The assessment will be performed in terms of all three indicators, ARL1, false alarm rate and missed detection rate Assessing the performance of the EWMA chart under violation of assumptions Assessing the impact of high noise levels in the data on the performance of the EWMA chart In this section, the impact of different measurement noise levels on the performance of EWMA technique proposed by the previous section will be assessed. For that purpose, training data (without fault) consisting of 8192 samples are generated. These data have zero mean and, unit variance and follow a Gaussian distribution. These training data are used to compute the control limits using a EWMA chart. Then, testing data having 19

32 the same length are generated and faults of magnitudes of ±1 are introduced between samples and The control limits obtained from the training data are applied on the EWMA statistics computed using the testing data to detect the fault and to compute the three indicators- ARL1, false alarm rate and missed detection rate. A Monte Carlo simulation of 5000 realizations is performed to get statistically meaningful results. This whole simulation is then repeated for different values of noise ranging from 0.03 to 2 times standard deviation (σ) of the data. The results of this Monte Carlo simulation, which are shown in Figures 7 and 8 show that- the ARL1 and missed detection rate increase, while the false alarm rate remains relatively constant with a slight decrease at higher noise levels. Figure 7: Impact of noise level on ARL 1 values for the conventional EWMA chart. 20

33 Figure 8: Impact of noise level on the false alarm and missed detection rates of the conventional EWMA chart. To get proper understanding of what actually happens when the noise level increases, one can take a look at the EWMA control chart of an individual realization. The EWMA statistic of the testing data along with its control limits are shown for different noise levels (σ = 0.5,1, and 1.5) in Figure 9. These figures show that at a very high noise level (e.g., σ = 1.5), the EWMA chart fails to detect the fault most of the time. This is due to the fact the high noise level masks the fault and makes it harder to properly detect the fault, which results in higher missed detection rates and ARL1. 21

34 Figure 9: EWMA statistics for different noise levels. The false alarm rate, on the other hand, remains relatively constant for different noise levels because the control chart itself adjusts the control width by widening and shrinking depending on the level of noise in the data. The initial relative high rate of false alarm for low noise level is due to the combination of two reasons narrow control width and inertia effect. The inertia effect normally occurs for low values of λ. In this simulation low value of λ (obtained from the plot) is used because fault size is relatively low and it is known that low values of λ can detect small faults quickly. Low value of λ means it gives 22

35 low weight to the new data. As a result, when a shift occurs, the EWMA statistic takes some time to detect this sudden shift. In Figure 9, for a noise level of 0.5σ, when the process returns to the normal value from a negative shift, the EWMA chart fails to track this change instantaneously, which causes few false alarms. For high noise levels, this inertia effect gets nullified by the widened control width Assessing the impact of autocorrelation in the data on the performance of EWMA chart Autocorrelation is the presence of correlation between a data samples and previous samples. The run length properties of traditional EWMA technique can be affected by the presence of autocorrelation in the data. For example, the ARL0 value can be much smaller than what it is designed for when the data have positive correlation [36][37]. Since the design procedure for conventional EWMA chart assumes independent samples, its performance can degrade in the case of autocorrelated data. In this section, performance of EWMA chart is assessed in the presence of autocorrelation. Autocorrelation in the data can be quantified using various models. A commonly used model is the autoregressive (AR) model, in which the data are represented by a linear sum of previous measurements and random noise. An AR model of order p is defined as follows [38]: z t = 1 z t p z t p + ξ t where, i are the autoregressive coefficients for different lagged measurements, z t is the deviation from the targeted process variable µ and ξ t is random noise which is usually assumed to be a random Gaussian variable with zero mean and unit variance. For this 23

36 work, a simple AR (1) model is used to simulate autocorrelated data, and is assumed to have the following form: z t = 1 z t 1 + ξ t To assess the effect of autocorrelation, training and testing data, both of 8192 observations, are generated using an AR (1) model, and a fault of ±1σ size is introduced on the testing data between samples and samples The control limits calculated using the training data, which are then applied on the testing data statistics to find out the ARL1, false alarm rate and missed detection rate. A Monte Carlo simulat ion of 5000 realizations is performed to have meaningful results. This same simulation is repeated for different autoregressive coefficients ranging from 0.1 to 1. Smaller values of coefficient mean lower autocorrelation and bigger values represent higher autocorrelation. The results of these simulations are illustrated in Figures 10 and 11. Figure 10 shows that, out of control average run length remains almost constant at a value of 10 for a wide range of autoregressive coefficient, 1. This value of ARL1 is actually equal to the optimum ARL1 value designed to detect a fault of size of 1 in presence of no autocorrelation in the data. Figure 10 also shows that ARL1 increases at relatively very high values of autoregressive coefficients. Such high level of autocorrelation is not very common in practice, so it can be concluded that the effect of autocorrelation on ARL1 is not significant in practice. 24

37 Figure 10: Impact of autocorrelation on ARL1 for the conventional EWMA chart. Figure 11: Impact of autocorrelation on the false alarm and missed detection rates for the conventional EWMA chart. On the other hand, the false alarm rate and missed detection rate increase gradually for larger autoregressive coefficients as shown in Figure 11. To further illustrate the 25

38 deterioration in performance of the EWMA chart in the presence of autocorrelation in the data, the EWMA control chart of individual realizations for two different values of 1 are shown in Figures 12 and 13. These Figures clearly show that the performance of EWMA deteriorates at larger autocorrelation levels. Figure 12: Impact of autocorrelation on the performance of EWMA chart for an autoregressive coefficient value of 0.3. Figure 13: Impact of autocorrelation on the performance of EWMA chart for an autoregressive coefficient value of

39 Assessing the impact of deviation from normality in the data on the performance of the EWMA chart The conventional EWMA chart assumes that residuals follow a normal or Gaussian distribution. However, this assumptions may not always be true. In this section, the impact of deviation from normality in the data on the performance of EWMA chart is assessed. To simulate non-gaussian data, several distributions can be used, which include the chi-square distribution, the lognormal distribution, the gamma distribution, the Weibull distribution etc. In this work, a chi-square distribution with varying degrees of non-normality is used to simulate non-gaussian data. The chi-square distribution is a special case of the gamma distribution [39]. The probability density function of a gamma distributed random variable, x, is defined as follows: f(x) = λr x r 1 e λx г(r), for x > 0 where, λ > 0 and r > 0. This distribution becomes chi-square distribution when λ = 1 2 and r = k. So, the probability density function of the chi-square distribution has the following 2 form: f(x) = x k 2 1 e x 2 k 22 г(r) 27

40 where, k represents the degrees of freedom value. Different degree of non-normality can be produced by varying this degree of freedom value. The degree of non-normality in data can be measured in various ways, which can be categorized as follows [40]: Graphical methods Moment type tests methods Other tests designed specifically to test for normality Graphical methods include plots of either the raw data or the plot of probability density functions of the data. Example of raw data plots are histogram, stem and leaf plots, box plots (skeletal). On the other hand, probability plots include normal quantile plots (Q- Q), percentile plots (P-P) etc. These graphical methods visualize differences between the empirical distribution and the theoretical distribution like a normal distribution of the data and are not convenient to investigate the deviation from normality of the process data. Moment type tests include skewness and kurtosis tests and are frequently used to quantify the degree of normality of a particular distribution of data. Skewness roughly check normality by measuring the degree of symmetry of a distribution. A normally distributed data is symmetrical. So, skewness of a normally distributed data is zero. A deviation from this zero value, either positive or negative, represents a skewed distribution with long tail to right or left side of the distribution respectively [41]. The skewness coefficients bigger that 1 or less than -1 indicate fair amount of skewness and thus deviation from normality. Kurtosis measures the peakedness of the distribution or heaviness of the tail. If a variable is normally distributed then its kurtosis value is 3. A kurtosis value of less than 3 indicates a thicker tailed and lower peaked distribution 28

41 compared to a normal distribution [42]. Though this sample moment type tests are commonly used, they are not adopted in this work to measure the deviation from normality. Because both the kurtosis and skewness are sensitive to outliers. These outliers can cause inaccurate values of kurtosis or skewness, which consist of a very little portion of the data sample and may result from measurement errors. Besides the graphical method and moment type tests method, various other methods are available to test normality and include sample entropy, Kullback-Leiber, relative entropy and similar metrics [43][44][45]. In this work, the Shapiro-Wilk test is used to measure the degree of non-normality which also falls in this category and is a powerful univariate normality test. The Shapiro-Wilk metric value has a range between 0 and 1. Values closer to 1 means the data follow a distribution closer to normal, while values closer to 0 represents distributions that are further away from normality. In this simulation, which is intended to assess the effect of deviation from normality on the performance of EWMA, training and testing data sets consisting of 8192 observations each, are generated using a chi-square distribution, and faults having magnitudes of ±1σ are introduced in the testing data between samples and samples The control limits are computed using the training fault free data, and are then used to evaluate the faulty testing data using the three indicators - ARL1, false alarm rate and missed detection rate. This simulation is performed for different degrees of freedom values which correspond to different Shapiro-Wilk statistics, i.e., different degrees of non-normality. A Monte Carlo simulation of 5000 realizations is performed to obtain statistically meaningful results, which are shown in Figures 14 and

42 Figure 14: Impact of deviation from normality on ARL 1 for the conventional EWMA chart. Figure 15: Impact of deviation from normality on the false alarm and missed detection rates for the conventional EWMA chart. It is seen from Figures 14 and 15 that all three indicators, ARL1, false alarm rate and missed detection rate, do not change significantly with the Shapiro-Wilk statistic. These indicators remain almost constant even though the data deviate from normality. 30

43 These simulation results validate the fact that the EWMA technique is insensitive to nonnormality in the data [27]. To explain these results, individual EWMA realizations for two different values of Shapiro-Wilk statistics are investigated. Figures 16 and 17 show the EWMA control charts along with their histogram for Shapiro-Wilk statistics and respectively. Figure 16: Impact of deviation from normality on the performance of EWMA chart, SW =

44 Figure 17: Impact of deviation from normality on the performance of EWMA chart, SW stat = From Figure 16 and 17, it is clear that there is not much difference in the plots of EWMA statistics for the two cases even though the histograms show that the two data sets have different degrees of non-normality. In summary, it can be seen from the simulations performed to assess the impact of violating the normality, independence, and noise content assumptions made in the development of the EWMA fault detection method that these violations in the independence and noise level assumptions can seriously degrade its performance. In this work, wavelet based multiscale representation of data will be used to deal with these assumptions and thus to help improve the performance of EWMA. More information about multiscale representation and its advantages for process monitoring will be presented in the next chapter. 32

45 3. WAVELET BASED MULTISCALE REPRESENTATION In this chapter, wavelet based multiscale representation of data will be introduced, followed by a discussion on its advantages for process monitoring that can help improve the performance of EWMA Introduction to wavelet based multiscale representation Real process data or signals are normally a combination of various features such as measurement noise, process disturbances, process dynamics, and faults etc. These features usually contain varying contributions over time and frequency. For instance, a step fault in a signal is localized in time domain, but spans a wide range in the frequency domain, while correlated noise spans a wide range in the time domain but a small range in the frequency domain. So, effective feature extraction from such data requires representing the data at both time and frequency, which can be achieved by decomposing the data at multiple scales using wavelets. Multiscale decomposition algorithm was first developed by Mallat in 1989 [46], in which a signal is represented at multiple resolutions by expressing the data as a weighted sum of the orthonormal basis functions, called wavelets and scaling functions, that have the following form [47][48]: θ(t) = 1 s θ (t u s ) where, s and u represent the dilation and translation parameters respectively and θ is the mother wavelet. A number of basis functions are available which can be used as wavelet 33

46 functions in multiscale decomposition, such as, the Daubechies and Haar basis functions [49][50]. In wavelet based multiscale data decomposition, low pass and high pass filters are applied on the data. For example, applying a low pass filter on the original data provides a coarser approximation of the data, which is called the first scaled signal (see Figure 18). The low pass filter is derived from a scaling function of the form: φ ij (t) = 2 j φ(2 j t k) where, j and k represent the discretized dilation and translation parameters respectively. The difference between the first scaled signal and the original data (called first detail signal, see Figure 18) can be computed by applying a high pass filter that is derived from a wavelet function of the form: ψ ij (t) = 2 j ψ(2 j t k) Repeating the application of the low pass and high pass filter provides scaled and detailed signal at various levels, which correspond to different frequencies. After applying the low pass and high pass filters, the original signal or data can be expressed as the sum of the last scaled signal and all detail signals from all scales, which can be mathematically expressed as follows [51]: n2 J x(t) = a JK φ JK k=1 J n2 J + d jk ψ jk (t) j=1 k=1 where, J and n represent the maximum possible decomposition depth and the length of the signal. This multiscale data representation procedure has been illustrated in Figure 18 [52]. 34

47 Figure 18: Schematic diagram for multiscale representation of data. Wavelet based multiscale decomposition is an effective data analysis tool that has been widely used in various applications including physical, medical, engineering and social sciences. It has also been found useful in improving the effectiveness of various fault detection methods such as PCA by developing a multiscale PCA monitoring method [48][53], which has been used in practice to improve monitoring wastewater treatment processes [54]. Some of the advantages of multiscale representation in process monitoring are discussed next Advantages of multiscale representation of data Noise feature separation Multiscale representation has the ability to separate noise from important features in the data. When data is decomposed at multiple scales by passing through low pass and 35

48 high pass filters, noise is effectively separated from the important features. Random noise in a signal normally are present over all the coefficients, while deterministic features in the data are captured in a few, but relatively large coefficients. The important features in the data are usually captured by the last scaled signals as well as any large wavelet coefficient (in the detail signals), while other small wavelet coefficients usually correspond to noise [14][55]. Thus, multiscale provides an effective method for noisefeature separation as shown in 18. This advantage of noise-feature separation of multiscale representation of data has been used effectively for various application such as filtering time series genomic data [56] Decorrelation of autocorrelated data Another advantage of multiscale wavelet based representation is that the wavelet coefficients of detail signals at different scales become approximately decorrelated even though the original data are autocorrelated [51]. To demonstrate the effect of multiscale decomposition on the level of autocorrelation in the detail signals at multiple scales, the autocorrelation function (ACF) is used. The ACF quantifies the magnitude of the correlation between data samples as a function of their separation [38]. Thus, it measures the memory of stochastic processes [51]. For uncorrelated data, the ACF shows zero values for all lags (time difference between two samples) except for lag zero, where it shows a value of unity. On the other hand, for correlated processes, ACF shows non-zero values for lags other that zero. 36

49 Figure 19 shows the detail signals obtained from a correlated signal (that represents an AR(1) model with an autoregressive model parameter of 0.7) and the corresponding ACF s at different scales. Figure 19 clearly shows that even though the time-domain data are autocorrelated (where its ACF has non-zero values at lags other than zero), the detail signals are approximately decorrelated. Figure 19: Decorrelation of autocorrelated data at multiple scales. 37

50 Data are closer to normal at multiple scales Multiscale wavelet decomposition also makes the distribution of data closer to normal or Gaussian at multiple scales even if the original data follow non-normal distribution. Even though the effect of distribution on the performance of EWMA is not significant, transforming the data to be closer to normal helps satisfy its assumption better. To show the advantage of multiscale representation in providing detail signals that are closer to normal at multiple scales, histograms of a chi-square distributed signal as well as its detail signals at multiple scales are shown in Figure 20. The original timedomain data have a Shapiro-Wilk statistic of , which means it has a high degree of non-normality. Figure 20 shows that, as the decomposition depth increases, the detailed signals become more and more Gaussian. 38

51 Figure 20: Distribution of chi-squared data at multiple scales. So, wavelet based multiscale decomposition helps transform the data to be closer to normal at multiples scales despite their distribution in the time-domain doesn t follow normal distribution. The above described advantages of multiscale representation clearly show that they can help satisfy the independence, normality, and noise level assumptions made by 39

52 various univariate fault detection methods, such as EWMA which is the method of interest in this work. Thus, utilizing multiscale representation should provide improvements to the performance of EWMA. However, performing multiscale fault detection in two separate steps (multiple filtering and then fault detection) may not provide the sought improvements since filtering may remove features that are important for fault detection. Thus, an algorithm that integrates fault detection using EWMA and multiscale representation is needed [52]. A multiscale EWMA fault detection method is presented in the next chapter. 40

53 4. WAVELET BASED MULTISCALE EWMA CHART In this chapter, an algorithm for wavelet based multiscale EWMA fault detection will be described. Then, a design procedure for optimizing the parameters of the multiscale EWMA technique based on the lowest ARL1 value will be presented Process monitoring using Multiscale EWMA Chart The proposed multiscale EWMA monitoring technique consists of two phases [52] as shown in the Figure 21. In the first phase, fault free training data are normalized so that they have zero mean and unit variance, and are then decomposed at multiple scale using wavelet based multiscale decomposition. Then, the EWMA chart is applied to the detail signals at different scales as well as to the last scaled signal, and the control limits are computed at all scales. These control limits are used then to threshold the wavelet coefficients of detail signals. If any wavelet coefficient violates the control limits at a certain scale, all the wavelet coefficients at that scale are retained. If no violation of the limits occur at a certain scale, then all wavelet coefficients at that scale are ignored. The retained detail signals and the last scaled signals are then reconstructed to get the final reconstructed signal. Finally, EWMA is applied on the reconstructed signal to obtain the final multiscale EWMA detection statistic and the control limits. In the second phase, the testing data are decomposed at multiple scales using the same wavelet filters used in the training phase after normalizing the data using the same mean standard deviation obtained in training. The control limits obtained from the training phase are then applied to the detailed signals of the testing data at the respective scales 41

54 and also to the last scaled signal. At any scale, the wavelet coefficients that violate the control limits are retained while other that don t violate the limits are ignored. Then, a reconstructed signal from all the retained coefficients is obtained. Finally, the previously obtained control limits from reconstructed training data are then applied on the EWMA statistic of the reconstructed testing data to detect possible faults. This algorithm is illustrated schematically in Figure 21. Figure 21: A schematic diagram of the multiscale EWMA fault detection algorithm. 42

55 The multiscale EWMA fault detection algorithm presented in Figure 21 provides a general framework for the implementation of multiscale EWMA, but it does not provide a strategy for selecting its parameters, namely the smoothing parameter, λ, and the control width, L. Applying the multiscale EWMA algorithm without optimizing these parameters may not provide better performance compared to the well-designed conventional EWMA technique for detecting specific fault sizes. In the next section, a procedure for selecting the optimum parameters used in the multiscale EWMA fault detection technique will be presented 4.2. Design procedure of optimizing the parameters of the multiscale EWMA fault detection technique As indicated earlier, it is not possible to minimize all the monitoring performance indicators ARL1, false alarm and missed detection rate at the same time for the conventional EWMA fault detection method, which is also true for the multiscale EWMA technique. So, to establish a design procedure for the multiscale EWMA technique, first of all, a method for optimizing its parameters based on an indicator needs to be established. The parameters optimized to provide lower ARL1 values, will not give lower false alarm rates and vice versa. In this work, ARL1 is used as a design basis to optimize the multiscale EWMA parameters and to establish a design procedure for its implementation. Selection of the ARL1 as a design basis allows faster detection of faults, which is desirable in online process monitoring. 43

56 In this section, a procedure of optimizing the parameters of the multiscale EWMA technique for a fixed ARL0 value will be discussed and then a design procedure of the implementation of the multiscale EWMA technique will be established Optimizing the multiscale EWMA parameters The conventional EWMA fault detection method is associated with two parameters control width, L and smoothing parameter, λ. The multiscale EWMA methods, however, brings in another parameter, which is the multiscale decomposition depth. For establishing a design procedure for implementing the multiscale EWMA method, all three parameters needed to be optimized to give a minimum ARL1 value, which is the used design criteria. To optimize the parameters L and λ, a fault free training data consisting of 8192 Gaussian samples having a zero mean and unit standard deviation are generated. The decomposition depth is selected to be 6 (around half the maximum decomposition depth, which is 13 since 2 13 = 8192). Then, the multiscale EWMA algorithm is used to find out different combinations of λ and L values such that each combination gives an in control average run length (ARL0) of 500. A Monte Carlo simulation of 5000 realizations is used for each combination to make sure that these combinations in fact give the specified ARL0. To find out these combinations, first of all, a fixed value of λ is assumed, and then different values of L are used to compute the ARL0 value. Among all of these combinations, the pair of L and λ that results in an in control average run length value of 500 is stored. This process is repeated for different values of λ, and in each case the value of L that results in an in-control average run length of 500 is stored. These obtained combinations of L and λ are used to generate the plots shown in Figure

57 Figure 22: Combination of λ and L for in control average run length 500. Now, to select the optimum combination that gives the lowest ARL1, faulty testing data having the same length as the training data and a fault of size equal to the standard deviation of the data (i.e., σ) between samples 2000 to 3000 and a fault of size equal to - 1σ between samples 4500 to 5000 are generated. The size of the fault (±2.5σ) is selected here randomly, the same process will be repeated later for other fault sizes. Then, the multiscale EWMA algorithm is applied to the testing data using all the combinations of λ and L values shown in Figure 22 to see which combination gives the lowest ARL1 value, which is the optimum combination for a fault size of σ. A Monte Carlo simulation is used for each combination to get statistically meaningful results. The results of this optimization process are shown in Figure 23 that shows that optimum L and λ combination that minimize the ARL1 criterion for a fault size of 2.5σ. 45

58 Figure 23: 3-D plot of the optimization of parameters for fault size 2.5σ for multiscale EWMA chart for ARL 0 value of 500. The same procedure is repeated for different fault sizes, which provides the results shown in Figure 24, which is used to find out the optimum value of λ for different fault sizes. Figure 24: Optimal λ's for the multiscale EWMA chart for different fault sizes for ARL 0 =

59 As the optimum values of L and λ are available now for different fault sizes, the next step is to optimize the decomposition depth. Normally, noise can be reduced more by increasing the decomposition level. However, a larger decomposition depth will cause delay in detecting a fault in the time domain [20]. So selecting the optimum decomposition level is important to get full advantage of the multiscale algorithm. To examine the effect of decomposition depth on the performance of the multiscale EWMA technique, the following simulation study is performed. In this simulation, a training data set consisting of 8192 observations that follow zero mean Gaussian distribution is generated. Then the multiscale EWMA algorithm is applied on the training data using decomposition depth of one and the control limits for all detailed as well as for the reconstructed signal are computed. Then, multiscale EWMA is applied on a testing data set that is generated the same way as the training data set but with two faults having magnitudes of ±1σ between samples and samples from The control limits obtained from the training data are used to detect faults in the testing data, and the ARL1, false alarm rate and missed detection rate are computed. The same simulation is repeated using different decomposition depths ranging from A Monte Carlo simulation using 5000 realization is performed for this study and the results are shown in Figures 25 and 26. Figure 25 shows that, the ARL1 value decreases sharply from a decomposition depth of 0 (which corresponds to the time domain or conventional EWMA) to a depth of 1. This is an indication that multiscale EWMA method outperforms the conventional EWMA method with respect to ARL1. Then, the ARL1 reaches its lowest value at 47

60 decomposition depth of 6 and then it increases again for larger decomposition depths. These result show that increasing the decomposition depth improves the ARL1 (i.e., improves the speed of detection, but up to a certain depth beyond which the speed of detection deteriorates). Figure 25: Effect of decomposition depth on the ARL1 of the multiscale EWMA chart. The effect of decomposition depth on the false alarm and missed detection rates, on the other hand, is illustrated in Figure 26, which shows that multiscale EWMA provides improvement in terms of the missed detection rate and a slight increase in the false alarm rate with respect to the conventional method. Figure 26 also shows that both missed detection and false alarm rates are not significantly affected by the decomposition depth. Thus, Figures 25 and 26 show that the multiscale EWMA algorithm provides its best performance at a decomposition depth around half the maximum decomposition depth 48

61 (which is 6 in this simulated study). It is important to note that, applying multiscale EWMA, as in any multiscale technique, requires a dyadic data set. Figure 26: Effect of decomposition depth on false alarm and missed detection rates of the multiscale EWMA chart. With the help of these simulated results, a design procedure for the multiscale EWMA can be established, which will be presented in the next section Design steps for the multiscale EWMA technique steps: The design procedure of the multiscale EWMA technique consists of the following Choose an acceptable value of in control average run length (ARL0), which is taken as 500, in this work Select the decomposition depth based on the length of the data (half the maximum decomposition depth is recommended) -data has to be dyadic 49

62 Specify the minimum fault size that needs to be detected as quickly as possible, and determine the value of λ which provides the lowest ARL1 for that specific fault size. (using Figure 24) Find the value of the control width L, which along which the value of λ (found from previous step) provides the required ARL0 value. (using Figure 22) In the next chapter, the performance of the multiscale EWMA described earlier will be compared with that of the conventional method from different perspectives to see which one gives better performance in fault detection. 50

63 5. PERFORMANCE COMPARISON BETWEEN THE CONVENTIONAL AND MULTISCALE EWMA CHARTS Both the conventional and multiscale EWMA techniques are designed to detect faults in the shortest possible time by minimizing the out-of-control average run length (ARL1). In this chapter, the performances of the two techniques will be compared using their optimally designed parameters. The two techniques will be compared using all three indices, ARL1, missed detection rate and false alarm rate. First, a comparison will be performed at different sizes of mean shifts when the assumptions of the conventional EWMA method are satisfied, which include the normality and independence of the residuals. Then, the performance will be compared when these assumptions are violated Comparison between the performance of the conventional and multiscale EWMA techniques under no violations of the EWMA assumptions In this section, the performances of the conventional and multiscale EWMA techniques are compared at different sizes of mean shifts and using their optimal parameters to make sure that each technique provides its best performance. To perform this comparison, training data consisting of 8192 independent zero mean Gaussian observations are generated. The testing data are also generated in a similar manner but with additive faults of magnitude 1 between samples and of magnitude -1 between samples Then, the conventional EWMA technique is applied using L and λ values of and 0.12, respectively, which are the optimum values to detect 51

64 fault of size 1. Similarly, the multiscale EWMA technique is applied on the same data with its own optimum parameter values, which are 2.41, 0.11 and 6 for L, λ and the depth, respectively. All three indices, ARL1, false alarm rate and missed detection rate are computed for both techniques and a Monte Carlo simulation of 5000 realization is performed to get meaningful results. Then, the same procedure is repeated for different fault sizes ranging from 0.5 to 4, and the simulation results are shown in Figures 27, 28 and 29. Figure 27 shows that, for smaller fault sizes, the multiscale EWMA detects the faults significantly quicker than the conventional EWMA technique. For example, for a fault of size 0.5, the multiscale EWMA has an ARL1 value of which is much smaller than the ARL1 value of the conventional EWMA, which is However, for larger faults, both the techniques perform almost equally. Figure 27: Performance comparison between the conventional and multiscale EWMA charts in terms of ARL 1 for different fault sizes. 52

65 Figure 28, on the other hand, which shows the comparison results in terms of false alarm rate, shows that the multiscale EWMA technique results in a slight increase in the false alarm rate compared to the conventional method, especially for small shift sizes. However, the difference between the false alarm rates for the two techniques are not too high and the maximum false alarm rate that the multiscale technique produces (for fault size of 0.5) is equal to 4.3%. These relatively higher false alarm rates are expected though as the technique is designed to give a minimum ARL1 which can increase the false alarm rate (see section 1.1.2). Figure 28: Performance comparison between the conventional and multiscale EWMA charts in terms of false alarm rate for different fault sizes. The comparison results between the conventional and multiscale EWMA techniques in terms of missed detection rate are shown in Figure 29, which shows that the multiscale EWMA technique provides improved performance over the conventional one 53

66 for most of the fault sizes except for very high fault sizes (for fault sizes of 3 4), where both the techniques perform almost equally. Figure 29: Performance comparison between the conventional and multiscale EWMA charts in terms of missed detection rate for different fault sizes. To explain the nonlinear behavior of the plots shown in Figure 29, it is important to understand how the false alarm and missed detection rates change for different values of L and λ for the multiscale EWMA technique. The effect of changing the control width (L) and smoothing parameter (λ) on the missed detection and false alarm rates are illustrated in Figure 30 for L values ranging from 1-4 and λ values ranging from In this analysis, the noise level and fault sizes are kept fixed at 1 and ±1 respectively. Figure 30 clearly shows a nonlinear behavior for the missed detection and false alarm rates at different L and λ values, which explains the nonlinear behavior observed in Figure

67 Figure 30: Trade-off between false alarm and missed detection rate for the multiscale EWMA chart. In summary, Figures 28 and 29 show that although both the multiscale and conventional EWMA technique are designed to optimize the ARL1 criterion, the multiscale EWMA provides improvement in the missed detection rate, besides ARL1. So, even though the false alarm rate increases slightly, it is evident that the multiscale algorithm outperforms its conventional time domain EWMA Comparison between the performances of the conventional and multiscale EWMA techniques under violations of the EWMA assumptions In this section, the performance of the multiscale EWMA chart will be assessed and compared with that of the conventional EWMA method under violation of EWMA chart assumptions. To compare the performances of the two EWMA techniques, the same data that were used in section 2.2 to assess the effect of noise level, autocorrelation and deviation from normality will be used here. The optimum parameters are used for the 55

68 respective techniques, which correspond to the fault size introduced in the data, which is 1σ for all the cases Comparison of performance using data with different level of noise In this study, the effect of the noise level on the performances of the conventional and multiscale EWMA methods is assessed. Here, the fault size is fixed at 1 and the noise standard deviation is varied from , and then in each case the ARL1, missed detection rate and false alarm rate are computed, and the results are shown in Figures 31 and 32. Figure 31 shows that- the multiscale EWMA chart gives better performance in terms of ARL1 value than the conventional method, especially at larger noise levels. The missed detection and false alarm rates, on other hand, are illustrated in Figure 32, which shows that the multiscale EWMA technique outperforms the conventional method with respect to the missed detection rate, while the false alarm rates are comparable for both techniques with a slight advantages in favor of the conventional method. 56

69 Figure 31: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL 1 for different level of noise in the data. Figure 32: Performance comparison between the multiscale and conventional EWMA techniques in terms of false alarm and missed detection rate for different level of noise in the data. To better explain the relative performances of the conventional and multiscale EWMA techniques, the detection statistics used in both methods are shown in Figure 33. It shows that the multiscale EWMA technique helps remove noise from the data through the 57

70 application of EWMA at multiple scales, which narrows the control limits of the final reconstructed signal. This helps provide a quicker detection of faults and a smaller missed detection rate, but at the expense of a slight increase in the false alarm rate compared to the conventional technique. Therefore, the multiscale EWMA method provides a better performance than the conventional method, especially at higher noise levels. Figure 33: Detection statistics for the conventional and multiscale methods in the case where the noise standard deviation equals Comparison of performance using autocorrelated data In this section, the conventional and multiscale techniques are compared through their application on autocorrelated data, which are generated using an autoregressive model, AR(1). The value of the autoregressive parameter, is varied between 0.1 and 0.95 and at each value the ARL1, false alarm rate, and missed detection rate are computed which 58

71 are shown in Figures 34 and 35. Figure 34 shows that- the multiscale EWMA method consistently provides smaller ARL1 values compared to the conventional EWMA almost for all values of the autoregressive coefficients. Figure 34 also shows that for the multiscale EWMA method, the ARL1 value remains almost constant except very high level of autocorrelation, where it still provides a quicker detection than the conventional EWMA technique. Figure 34: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL 1 for different level of autocorrelation in the data. Figure 35, on the hand, shows that the multiscale EWMA technique provides smaller missed detection than the conventional method for all levels of autocorrelation. The false alarm rate for the conventional method is smaller than that of the multiscale 59

72 technique, especially at high level of autocorrelation, i.e., large autoregressive model parameter. Figure 35: Performance comparison between the multiscale and conventional EWMA technique in terms of false alarm and missed detection rates for different level of autocorrelation in the data. The advantages of the multiscale EWMA chart can be better illustrated by comparing the detection statistics for both techniques, which are illustrated in Figure 36. It shows that, again narrower control limits are obtained using the multiscale EWMA chart, which helps achieve a smaller ARL1 value and missed detection rate, but with a relative increase in the false alarm rate, especially at high levels of autocorrelation. In summary, the multiscale EWMA method outperforms its conventional counterpart using autocorrelated data. 60

73 Figure 36: Detection statistic for the conventional and multiscale methods in the case where the autoregressive coefficient equal to Comparison of performance using non-gaussian (chi-square) data In this section, the effect of deviation from normality is examined for the multiscale and conventional EWMA methods through their application on chi-square data at different degrees of non-normality, i.e., different values of Shapiro Wilk ranging from 0.69 to 1. The results of the assessment are shown in Figures 37 and 38. Figure 37 shows that, for both techniques the ARL1 remains almost constant at different Shapiro Wilk values, and that the multiscale technique consistently provides smaller ARL1 values (i.,e, quicker detection) irrespective of the degree of non-normality. 61

74 Figure 37: Performance comparison between the multiscale and conventional EWMA technique in terms of ARL1 for non-gaussian data (Chi-square). Similarly, the false alarm and missed detection rates also remain almost constant for both the techniques at different degrees of non-normality (see Figure 38). However, the multiscale EWMA method provides smaller missed detection rates but higher false alarm rates. 62

75 Figure 38: Performance comparison between the multiscale and conventional EWMA technique in terms of false alarm and missed detection rates for non- Gaussian data (Chi-square). The performances of the conventional and multiscale EWMA techniques can also be compared by comparing their detection statistics, which are shown in Figure 39, which shows that, the multiscale EWMA technique provides faster detection by shrinking the control limits which result in higher false alarm rates. 63

76 Figure 39: Detection statistics for the conventional and multiscale methods in the case where the Shapiro Wilk equals In summary, the multiscale EWMA technique outperforms the conventional EWMA technique not only when both the techniques are designed to perform optimally but also when data violate the assumptions of the EWMA chart. In the next chapter, the performances of both techniques will be evaluated using real application data. 64

77 6. APPLICATION OF THE MULTISCALE EWMA CHART In this chapter, real application data will be used to assess the effectiveness of the multiscale EWMA method as a monitoring technique. To do so, simulated distillation column data will be used. To simulate real process data, a distillation column consisting of 32 theoretical stages with a total condenser and a reboiler was simulated using Aspen Tech 7.2. The feed stream, which enters the column in a saturated liquid form at stage 16 with a mass flow rate of 1kg/mole.s and a temperature of 322 K, has compositions of 40 mole % propane and 60 mole % isobutene. The nominal operating conditions of the distillation column can be found in [57]. Dynamic data of the distillation column are generated by changing the feed and reflux flow rates from their nominal operating values, which is done by introducing step changes of magnitudes ±2% in the feed and reflux flow rates. After each step change, the process is given sufficient time to settle to a new steady state after which another step change is introduced. After introducing all step changes and reaching to a final steady state, 1024 observations that represent the process behavior are generated. All simulated data are assumed to be noise free, and therefore, the simulated data are contaminated with zero mean Gaussian noise to represent measurement errors. The objective of this example is to compare the performances of the conventional and multiscale EWMA methods through monitoring the composition of propane in the distillate stream using faults of different magnitudes. Since both techniques require evaluating residuals, a partial least squares (PLS) model is developed to predict the 65

78 residuals of the propane composition using temperature data at various trays of the column as well as the feed and reflux rates. The residuals obtained by this process are then divided into two set training and testing, each consisting of 512 samples. Two step faults of magnitude ±2σ are then introduced to the residuals between samples and samples of the testing data respectively, where σ is the standard deviation of the residuals. Then, the control limits are computed from the training data by using the corresponding optimum parameters for a fault size of 2σ of both techniques. Those control limits are then used to compute ARL1, false alarm rate and missed detection rates for both the techniques using the testing data. Figure 40, which shows the detection statistics for both the conventional and multiscale EWMA techniques, clearly shows that the multiscale EWMA technique provides a smaller missed detection rate (5%) compared to one obtained using the conventional technique (33%). However, the false alarm rate obtained using the multiscale technique (3.9%) is a little larger than what is obtained using the conventional technique (0.6%). On the other hand, the multiscale EWMA provides a little smaller ARL1 value (4) compared to the one obtained using the conventional technique (5). 66

79 Figure 40: Comparing the performances of the conventional and multiscale EWMA chart for a step fault of magnitude ±2σ in the residuals of simulated distillation column data. A similar comparison is performed using smaller fault sizes (±σ), and the results are shown in Figure 41. The detection statistics shown in Figure 41 show that the missed detection rate for the multiscale EWMA is smaller (19%) than the conventional EWMA (39%). The multiscale EWMA method provides relatively faster detection (ARL1=16) than the conventional method (ARL1=19).The false alarm rate is higher for the multiscale technique (6%) than for the conventional technique (1%). 67

80 Figure 41: Comparing the performances of the conventional and multiscale EWMA chart for a step fault of magnitude ±σ in the residuals of simulated distillation column data. 68

IMPROVED SHEWHART CHART USING MULTISCALE REPRESENTATION. A Thesis MOHAMMED ZIYAN SHERIFF

IMPROVED SHEWHART CHART USING MULTISCALE REPRESENTATION. A Thesis MOHAMMED ZIYAN SHERIFF IMPROVED SHEWHART CHART USING MULTISCALE REPRESENTATION A Thesis by MOHAMMED ZIYAN SHERIFF Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of

More information

Control Charts. An Introduction to Statistical Process Control

Control Charts. An Introduction to Statistical Process Control An Introduction to Statistical Process Control Course Content Prerequisites Course Objectives What is SPC? Control Chart Basics Out of Control Conditions SPC vs. SQC Individuals and Moving Range Chart

More information

UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication

UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication Citation for published version (APA): Abbas, N. (2012). Memory-type control

More information

ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART

ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART Vadhana Jayathavaj Rangsit University, Thailand vadhana.j@rsu.ac.th Adisak

More information

Cpk: What is its Capability? By: Rick Haynes, Master Black Belt Smarter Solutions, Inc.

Cpk: What is its Capability? By: Rick Haynes, Master Black Belt Smarter Solutions, Inc. C: What is its Capability? By: Rick Haynes, Master Black Belt Smarter Solutions, Inc. C is one of many capability metrics that are available. When capability metrics are used, organizations typically provide

More information

MINI-PAPER A Gentle Introduction to the Analysis of Sequential Data

MINI-PAPER A Gentle Introduction to the Analysis of Sequential Data MINI-PAPER by Rong Pan, Ph.D., Assistant Professor of Industrial Engineering, Arizona State University We, applied statisticians and manufacturing engineers, often need to deal with sequential data, which

More information

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Exploratory data analysis tasks Examine the data, in search of structures

More information

CREATING THE DISTRIBUTION ANALYSIS

CREATING THE DISTRIBUTION ANALYSIS Chapter 12 Examining Distributions Chapter Table of Contents CREATING THE DISTRIBUTION ANALYSIS...176 BoxPlot...178 Histogram...180 Moments and Quantiles Tables...... 183 ADDING DENSITY ESTIMATES...184

More information

Box-Cox Transformation for Simple Linear Regression

Box-Cox Transformation for Simple Linear Regression Chapter 192 Box-Cox Transformation for Simple Linear Regression Introduction This procedure finds the appropriate Box-Cox power transformation (1964) for a dataset containing a pair of variables that are

More information

Workload Characterization Techniques

Workload Characterization Techniques Workload Characterization Techniques Raj Jain Washington University in Saint Louis Saint Louis, MO 63130 Jain@cse.wustl.edu These slides are available on-line at: http://www.cse.wustl.edu/~jain/cse567-08/

More information

Multivariate Capability Analysis

Multivariate Capability Analysis Multivariate Capability Analysis Summary... 1 Data Input... 3 Analysis Summary... 4 Capability Plot... 5 Capability Indices... 6 Capability Ellipse... 7 Correlation Matrix... 8 Tests for Normality... 8

More information

Table of Contents (As covered from textbook)

Table of Contents (As covered from textbook) Table of Contents (As covered from textbook) Ch 1 Data and Decisions Ch 2 Displaying and Describing Categorical Data Ch 3 Displaying and Describing Quantitative Data Ch 4 Correlation and Linear Regression

More information

Vocabulary. 5-number summary Rule. Area principle. Bar chart. Boxplot. Categorical data condition. Categorical variable.

Vocabulary. 5-number summary Rule. Area principle. Bar chart. Boxplot. Categorical data condition. Categorical variable. 5-number summary 68-95-99.7 Rule Area principle Bar chart Bimodal Boxplot Case Categorical data Categorical variable Center Changing center and spread Conditional distribution Context Contingency table

More information

Monitoring of Manufacturing Process Conditions with Baseline Changes Using Generalized Regression

Monitoring of Manufacturing Process Conditions with Baseline Changes Using Generalized Regression Proceedings of the 10th International Conference on Frontiers of Design and Manufacturing June 10~1, 01 Monitoring of Manufacturing Process Conditions with Baseline Changes Using Generalized Regression

More information

A Modified Approach for Detection of Outliers

A Modified Approach for Detection of Outliers A Modified Approach for Detection of Outliers Iftikhar Hussain Adil Department of Economics School of Social Sciences and Humanities National University of Sciences and Technology Islamabad Iftikhar.adil@s3h.nust.edu.pk

More information

Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation

Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation November 2010 Nelson Shaw njd50@uclive.ac.nz Department of Computer Science and Software Engineering University of Canterbury,

More information

Analytical Techniques for Anomaly Detection Through Features, Signal-Noise Separation and Partial-Value Association

Analytical Techniques for Anomaly Detection Through Features, Signal-Noise Separation and Partial-Value Association Proceedings of Machine Learning Research 77:20 32, 2017 KDD 2017: Workshop on Anomaly Detection in Finance Analytical Techniques for Anomaly Detection Through Features, Signal-Noise Separation and Partial-Value

More information

3 Graphical Displays of Data

3 Graphical Displays of Data 3 Graphical Displays of Data Reading: SW Chapter 2, Sections 1-6 Summarizing and Displaying Qualitative Data The data below are from a study of thyroid cancer, using NMTR data. The investigators looked

More information

Acquisition Description Exploration Examination Understanding what data is collected. Characterizing properties of data.

Acquisition Description Exploration Examination Understanding what data is collected. Characterizing properties of data. Summary Statistics Acquisition Description Exploration Examination what data is collected Characterizing properties of data. Exploring the data distribution(s). Identifying data quality problems. Selecting

More information

Learner Expectations UNIT 1: GRAPICAL AND NUMERIC REPRESENTATIONS OF DATA. Sept. Fathom Lab: Distributions and Best Methods of Display

Learner Expectations UNIT 1: GRAPICAL AND NUMERIC REPRESENTATIONS OF DATA. Sept. Fathom Lab: Distributions and Best Methods of Display CURRICULUM MAP TEMPLATE Priority Standards = Approximately 70% Supporting Standards = Approximately 20% Additional Standards = Approximately 10% HONORS PROBABILITY AND STATISTICS Essential Questions &

More information

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines A popular method for moving beyond linearity 2. Basis expansion and regularization 1 Idea: Augment the vector inputs x with additional variables which are transformation of x use linear models in this

More information

Quality Control-2 The ASTA team

Quality Control-2 The ASTA team Quality Control-2 The ASTA team Contents 0.1 OC curves............................................... 1 0.2 OC curves - example......................................... 2 0.3 CUSUM chart.............................................

More information

3 Graphical Displays of Data

3 Graphical Displays of Data 3 Graphical Displays of Data Reading: SW Chapter 2, Sections 1-6 Summarizing and Displaying Qualitative Data The data below are from a study of thyroid cancer, using NMTR data. The investigators looked

More information

Chapter 1. Looking at Data-Distribution

Chapter 1. Looking at Data-Distribution Chapter 1. Looking at Data-Distribution Statistics is the scientific discipline that provides methods to draw right conclusions: 1)Collecting the data 2)Describing the data 3)Drawing the conclusions Raw

More information

Box-Cox Transformation

Box-Cox Transformation Chapter 190 Box-Cox Transformation Introduction This procedure finds the appropriate Box-Cox power transformation (1964) for a single batch of data. It is used to modify the distributional shape of a set

More information

UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication

UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication UvA-DARE (Digital Academic Repository) Memory-type control charts in statistical process control Abbas, N. Link to publication Citation for published version (APA): Abbas, N. (2012). Memory-type control

More information

Dealing with Categorical Data Types in a Designed Experiment

Dealing with Categorical Data Types in a Designed Experiment Dealing with Categorical Data Types in a Designed Experiment Part II: Sizing a Designed Experiment When Using a Binary Response Best Practice Authored by: Francisco Ortiz, PhD STAT T&E COE The goal of

More information

Unit 5: Estimating with Confidence

Unit 5: Estimating with Confidence Unit 5: Estimating with Confidence Section 8.3 The Practice of Statistics, 4 th edition For AP* STARNES, YATES, MOORE Unit 5 Estimating with Confidence 8.1 8.2 8.3 Confidence Intervals: The Basics Estimating

More information

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables Further Maths Notes Common Mistakes Read the bold words in the exam! Always check data entry Remember to interpret data with the multipliers specified (e.g. in thousands) Write equations in terms of variables

More information

Design of Fault Diagnosis System of FPSO Production Process Based on MSPCA

Design of Fault Diagnosis System of FPSO Production Process Based on MSPCA 2009 Fifth International Conference on Information Assurance and Security Design of Fault Diagnosis System of FPSO Production Process Based on MSPCA GAO Qiang, HAN Miao, HU Shu-liang, DONG Hai-jie ianjin

More information

Data Preprocessing. S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha

Data Preprocessing. S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha Data Preprocessing S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha 1 Why Data Preprocessing? Data in the real world is dirty incomplete: lacking attribute values, lacking

More information

Real-time models supporting resource management decisions in highly variable systems

Real-time models supporting resource management decisions in highly variable systems Real-time models supporting resource management decisions in highly variable systems Sara Casolari, Michele Colajanni, Stefania Tosi Department of Computer Engineering University of Modena and Reggio Emilia

More information

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time Today Lecture 4: We examine clustering in a little more detail; we went over it a somewhat quickly last time The CAD data will return and give us an opportunity to work with curves (!) We then examine

More information

STATISTICS (STAT) Statistics (STAT) 1

STATISTICS (STAT) Statistics (STAT) 1 Statistics (STAT) 1 STATISTICS (STAT) STAT 2013 Elementary Statistics (A) Prerequisites: MATH 1483 or MATH 1513, each with a grade of "C" or better; or an acceptable placement score (see placement.okstate.edu).

More information

Mean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242

Mean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242 Mean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242 Creation & Description of a Data Set * 4 Levels of Measurement * Nominal, ordinal, interval, ratio * Variable Types

More information

Statistics... Basic Descriptive Statistics

Statistics... Basic Descriptive Statistics Statistics... Statistics is the study of how best to 1. collect data "Probability" is the mathematical description of "chance" and is a tool for the above that must be part of the course... but it is not

More information

Statistics... Statistics is the study of how best to. 1. collect data. 2. summarize/describe data. 3. draw conclusions/inferences from data

Statistics... Statistics is the study of how best to. 1. collect data. 2. summarize/describe data. 3. draw conclusions/inferences from data Statistics... Statistics is the study of how best to 1. collect data 2. summarize/describe data 3. draw conclusions/inferences from data all in a framework that explicitly recognizes the reality and omnipresence

More information

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,

More information

The Cross-Entropy Method

The Cross-Entropy Method The Cross-Entropy Method Guy Weichenberg 7 September 2003 Introduction This report is a summary of the theory underlying the Cross-Entropy (CE) method, as discussed in the tutorial by de Boer, Kroese,

More information

MONITORING THE REPEATABILITY AND REPRODUCIBILTY OF A NATURAL GAS CALIBRATION FACILITY

MONITORING THE REPEATABILITY AND REPRODUCIBILTY OF A NATURAL GAS CALIBRATION FACILITY MONITORING THE REPEATABILITY AND REPRODUCIBILTY OF A NATURAL GAS CALIBRATION FACILITY T.M. Kegel and W.R. Johansen Colorado Engineering Experiment Station, Inc. (CEESI) 54043 WCR 37, Nunn, CO, 80648 USA

More information

Chapter 7: Dual Modeling in the Presence of Constant Variance

Chapter 7: Dual Modeling in the Presence of Constant Variance Chapter 7: Dual Modeling in the Presence of Constant Variance 7.A Introduction An underlying premise of regression analysis is that a given response variable changes systematically and smoothly due to

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING

IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING SECOND EDITION IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING ith Algorithms for ENVI/IDL Morton J. Canty с*' Q\ CRC Press Taylor &. Francis Group Boca Raton London New York CRC

More information

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis Submitted By: Amrita Mishra 11104163 Manoj C 11104059 Under the Guidance of Dr. Sumana Gupta Professor Department of Electrical

More information

Use of Extreme Value Statistics in Modeling Biometric Systems

Use of Extreme Value Statistics in Modeling Biometric Systems Use of Extreme Value Statistics in Modeling Biometric Systems Similarity Scores Two types of matching: Genuine sample Imposter sample Matching scores Enrolled sample 0.95 0.32 Probability Density Decision

More information

ADAPTIVE VIDEO STREAMING FOR BANDWIDTH VARIATION WITH OPTIMUM QUALITY

ADAPTIVE VIDEO STREAMING FOR BANDWIDTH VARIATION WITH OPTIMUM QUALITY ADAPTIVE VIDEO STREAMING FOR BANDWIDTH VARIATION WITH OPTIMUM QUALITY Joseph Michael Wijayantha Medagama (08/8015) Thesis Submitted in Partial Fulfillment of the Requirements for the Degree Master of Science

More information

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea Chapter 3 Bootstrap 3.1 Introduction The estimation of parameters in probability distributions is a basic problem in statistics that one tends to encounter already during the very first course on the subject.

More information

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes.

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes. Resources for statistical assistance Quantitative covariates and regression analysis Carolyn Taylor Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC January 24, 2017 Department

More information

Robotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.

Robotics. Lecture 5: Monte Carlo Localisation. See course website  for up to date information. Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:

More information

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Xavier Le Faucheur a, Brani Vidakovic b and Allen Tannenbaum a a School of Electrical and Computer Engineering, b Department of Biomedical

More information

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY Angela M. Kim and Richard C. Olsen Remote Sensing Center Naval Postgraduate School 1 University Circle Monterey, CA

More information

CHAPTER 1. Introduction. Statistics: Statistics is the science of collecting, organizing, analyzing, presenting and interpreting data.

CHAPTER 1. Introduction. Statistics: Statistics is the science of collecting, organizing, analyzing, presenting and interpreting data. 1 CHAPTER 1 Introduction Statistics: Statistics is the science of collecting, organizing, analyzing, presenting and interpreting data. Variable: Any characteristic of a person or thing that can be expressed

More information

SLStats.notebook. January 12, Statistics:

SLStats.notebook. January 12, Statistics: Statistics: 1 2 3 Ways to display data: 4 generic arithmetic mean sample 14A: Opener, #3,4 (Vocabulary, histograms, frequency tables, stem and leaf) 14B.1: #3,5,8,9,11,12,14,15,16 (Mean, median, mode,

More information

Data-driven fault detection with process topology for fault identification

Data-driven fault detection with process topology for fault identification Preprints of the 19th World Congress The International Federation of Automatic Control Data-driven fault detection with process topology for fault identification Brian S. Lindner*. Lidia Auret** * Department

More information

a. divided by the. 1) Always round!! a) Even if class width comes out to a, go up one.

a. divided by the. 1) Always round!! a) Even if class width comes out to a, go up one. Probability and Statistics Chapter 2 Notes I Section 2-1 A Steps to Constructing Frequency Distributions 1 Determine number of (may be given to you) a Should be between and classes 2 Find the Range a The

More information

Chapter 6: DESCRIPTIVE STATISTICS

Chapter 6: DESCRIPTIVE STATISTICS Chapter 6: DESCRIPTIVE STATISTICS Random Sampling Numerical Summaries Stem-n-Leaf plots Histograms, and Box plots Time Sequence Plots Normal Probability Plots Sections 6-1 to 6-5, and 6-7 Random Sampling

More information

Samuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR

Samuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR Detecting Missing and Spurious Edges in Large, Dense Networks Using Parallel Computing Samuel Coolidge, sam.r.coolidge@gmail.com Dan Simon, des480@nyu.edu Dennis Shasha, shasha@cims.nyu.edu Technical Report

More information

Continuous Improvement Toolkit. Normal Distribution. Continuous Improvement Toolkit.

Continuous Improvement Toolkit. Normal Distribution. Continuous Improvement Toolkit. Continuous Improvement Toolkit Normal Distribution The Continuous Improvement Map Managing Risk FMEA Understanding Performance** Check Sheets Data Collection PDPC RAID Log* Risk Analysis* Benchmarking***

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

Intro to ARMA models. FISH 507 Applied Time Series Analysis. Mark Scheuerell 15 Jan 2019

Intro to ARMA models. FISH 507 Applied Time Series Analysis. Mark Scheuerell 15 Jan 2019 Intro to ARMA models FISH 507 Applied Time Series Analysis Mark Scheuerell 15 Jan 2019 Topics for today Review White noise Random walks Autoregressive (AR) models Moving average (MA) models Autoregressive

More information

NCSS Statistical Software

NCSS Statistical Software Chapter 245 Introduction This procedure generates R control charts for variables. The format of the control charts is fully customizable. The data for the subgroups can be in a single column or in multiple

More information

STA 570 Spring Lecture 5 Tuesday, Feb 1

STA 570 Spring Lecture 5 Tuesday, Feb 1 STA 570 Spring 2011 Lecture 5 Tuesday, Feb 1 Descriptive Statistics Summarizing Univariate Data o Standard Deviation, Empirical Rule, IQR o Boxplots Summarizing Bivariate Data o Contingency Tables o Row

More information

Multichannel functional data decomposition and monitoring

Multichannel functional data decomposition and monitoring University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School 2005 Multichannel functional data decomposition and monitoring Hani Kababji University of South Florida Follow

More information

Spatial Outlier Detection

Spatial Outlier Detection Spatial Outlier Detection Chang-Tien Lu Department of Computer Science Northern Virginia Center Virginia Tech Joint work with Dechang Chen, Yufeng Kou, Jiang Zhao 1 Spatial Outlier A spatial data point

More information

Tutorial: Using Tina Vision s Quantitative Pattern Recognition Tool.

Tutorial: Using Tina Vision s Quantitative Pattern Recognition Tool. Tina Memo No. 2014-004 Internal Report Tutorial: Using Tina Vision s Quantitative Pattern Recognition Tool. P.D.Tar. Last updated 07 / 06 / 2014 ISBE, Medical School, University of Manchester, Stopford

More information

Response to API 1163 and Its Impact on Pipeline Integrity Management

Response to API 1163 and Its Impact on Pipeline Integrity Management ECNDT 2 - Tu.2.7.1 Response to API 3 and Its Impact on Pipeline Integrity Management Munendra S TOMAR, Martin FINGERHUT; RTD Quality Services, USA Abstract. Knowing the accuracy and reliability of ILI

More information

Descriptive Statistics, Standard Deviation and Standard Error

Descriptive Statistics, Standard Deviation and Standard Error AP Biology Calculations: Descriptive Statistics, Standard Deviation and Standard Error SBI4UP The Scientific Method & Experimental Design Scientific method is used to explore observations and answer questions.

More information

(X 1:n η) 1 θ e 1. i=1. Using the traditional MLE derivation technique, the penalized MLEs for η and θ are: = n. (X i η) = 0. i=1 = 1.

(X 1:n η) 1 θ e 1. i=1. Using the traditional MLE derivation technique, the penalized MLEs for η and θ are: = n. (X i η) = 0. i=1 = 1. EXAMINING THE PERFORMANCE OF A CONTROL CHART FOR THE SHIFTED EXPONENTIAL DISTRIBUTION USING PENALIZED MAXIMUM LIKELIHOOD ESTIMATORS: A SIMULATION STUDY USING SAS Austin Brown, M.S., University of Northern

More information

Statistical Methods. Instructor: Lingsong Zhang. Any questions, ask me during the office hour, or me, I will answer promptly.

Statistical Methods. Instructor: Lingsong Zhang. Any questions, ask me during the office hour, or  me, I will answer promptly. Statistical Methods Instructor: Lingsong Zhang 1 Issues before Class Statistical Methods Lingsong Zhang Office: Math 544 Email: lingsong@purdue.edu Phone: 765-494-7913 Office Hour: Monday 1:00 pm - 2:00

More information

SYS 6021 Linear Statistical Models

SYS 6021 Linear Statistical Models SYS 6021 Linear Statistical Models Project 2 Spam Filters Jinghe Zhang Summary The spambase data and time indexed counts of spams and hams are studied to develop accurate spam filters. Static models are

More information

Image Restoration and Reconstruction

Image Restoration and Reconstruction Image Restoration and Reconstruction Image restoration Objective process to improve an image Recover an image by using a priori knowledge of degradation phenomenon Exemplified by removal of blur by deblurring

More information

Lecture Series on Statistics -HSTC. Frequency Graphs " Dr. Bijaya Bhusan Nanda, Ph. D. (Stat.)

Lecture Series on Statistics -HSTC. Frequency Graphs  Dr. Bijaya Bhusan Nanda, Ph. D. (Stat.) Lecture Series on Statistics -HSTC Frequency Graphs " By Dr. Bijaya Bhusan Nanda, Ph. D. (Stat.) CONTENT Histogram Frequency polygon Smoothed frequency curve Cumulative frequency curve or ogives Learning

More information

Random Number Generation and Monte Carlo Methods

Random Number Generation and Monte Carlo Methods James E. Gentle Random Number Generation and Monte Carlo Methods With 30 Illustrations Springer Contents Preface vii 1 Simulating Random Numbers from a Uniform Distribution 1 1.1 Linear Congruential Generators

More information

Advanced Digital Signal Processing Adaptive Linear Prediction Filter (Using The RLS Algorithm)

Advanced Digital Signal Processing Adaptive Linear Prediction Filter (Using The RLS Algorithm) Advanced Digital Signal Processing Adaptive Linear Prediction Filter (Using The RLS Algorithm) Erick L. Oberstar 2001 Adaptive Linear Prediction Filter Using the RLS Algorithm A complete analysis/discussion

More information

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used. 1 4.12 Generalization In back-propagation learning, as many training examples as possible are typically used. It is hoped that the network so designed generalizes well. A network generalizes well when

More information

Detecting Salient Contours Using Orientation Energy Distribution. Part I: Thresholding Based on. Response Distribution

Detecting Salient Contours Using Orientation Energy Distribution. Part I: Thresholding Based on. Response Distribution Detecting Salient Contours Using Orientation Energy Distribution The Problem: How Does the Visual System Detect Salient Contours? CPSC 636 Slide12, Spring 212 Yoonsuck Choe Co-work with S. Sarma and H.-C.

More information

Application of Characteristic Function Method in Target Detection

Application of Characteristic Function Method in Target Detection Application of Characteristic Function Method in Target Detection Mohammad H Marhaban and Josef Kittler Centre for Vision, Speech and Signal Processing University of Surrey Surrey, GU2 7XH, UK eep5mm@ee.surrey.ac.uk

More information

Image Contrast Enhancement in Wavelet Domain

Image Contrast Enhancement in Wavelet Domain Advances in Computational Sciences and Technology ISSN 0973-6107 Volume 10, Number 6 (2017) pp. 1915-1922 Research India Publications http://www.ripublication.com Image Contrast Enhancement in Wavelet

More information

Data Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski

Data Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski Data Analysis and Solver Plugins for KSpread USER S MANUAL Tomasz Maliszewski tmaliszewski@wp.pl Table of Content CHAPTER 1: INTRODUCTION... 3 1.1. ABOUT DATA ANALYSIS PLUGIN... 3 1.3. ABOUT SOLVER PLUGIN...

More information

Long Term Analysis for the BAM device Donata Bonino and Daniele Gardiol INAF Osservatorio Astronomico di Torino

Long Term Analysis for the BAM device Donata Bonino and Daniele Gardiol INAF Osservatorio Astronomico di Torino Long Term Analysis for the BAM device Donata Bonino and Daniele Gardiol INAF Osservatorio Astronomico di Torino 1 Overview What is BAM Analysis in the time domain Analysis in the frequency domain Example

More information

Bootstrapping Method for 14 June 2016 R. Russell Rhinehart. Bootstrapping

Bootstrapping Method for  14 June 2016 R. Russell Rhinehart. Bootstrapping Bootstrapping Method for www.r3eda.com 14 June 2016 R. Russell Rhinehart Bootstrapping This is extracted from the book, Nonlinear Regression Modeling for Engineering Applications: Modeling, Model Validation,

More information

Outline. Topic 16 - Other Remedies. Ridge Regression. Ridge Regression. Ridge Regression. Robust Regression. Regression Trees. Piecewise Linear Model

Outline. Topic 16 - Other Remedies. Ridge Regression. Ridge Regression. Ridge Regression. Robust Regression. Regression Trees. Piecewise Linear Model Topic 16 - Other Remedies Ridge Regression Robust Regression Regression Trees Outline - Fall 2013 Piecewise Linear Model Bootstrapping Topic 16 2 Ridge Regression Modification of least squares that addresses

More information

Chapter 6: Examples 6.A Introduction

Chapter 6: Examples 6.A Introduction Chapter 6: Examples 6.A Introduction In Chapter 4, several approaches to the dual model regression problem were described and Chapter 5 provided expressions enabling one to compute the MSE of the mean

More information

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will Lectures Recent advances in Metamodel of Optimal Prognosis Thomas Most & Johannes Will presented at the Weimar Optimization and Stochastic Days 2010 Source: www.dynardo.de/en/library Recent advances in

More information

ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL

ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY BHARAT SIGINAM IN

More information

A Trimmed Translation-Invariant Denoising Estimator

A Trimmed Translation-Invariant Denoising Estimator A Trimmed Translation-Invariant Denoising Estimator Eric Chicken and Jordan Cuevas Florida State University, Tallahassee, FL 32306 Abstract A popular wavelet method for estimating jumps in functions is

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2012 http://ce.sharif.edu/courses/90-91/2/ce725-1/ Agenda Features and Patterns The Curse of Size and

More information

Robust Shape Retrieval Using Maximum Likelihood Theory

Robust Shape Retrieval Using Maximum Likelihood Theory Robust Shape Retrieval Using Maximum Likelihood Theory Naif Alajlan 1, Paul Fieguth 2, and Mohamed Kamel 1 1 PAMI Lab, E & CE Dept., UW, Waterloo, ON, N2L 3G1, Canada. naif, mkamel@pami.uwaterloo.ca 2

More information

Business Club. Decision Trees

Business Club. Decision Trees Business Club Decision Trees Business Club Analytics Team December 2017 Index 1. Motivation- A Case Study 2. The Trees a. What is a decision tree b. Representation 3. Regression v/s Classification 4. Building

More information

Resource Allocation Strategies for Multiple Job Classes

Resource Allocation Strategies for Multiple Job Classes Resource Allocation Strategies for Multiple Job Classes by Ye Hu A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Computer

More information

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER A.Shabbir 1, 2 and G.Verdoolaege 1, 3 1 Department of Applied Physics, Ghent University, B-9000 Ghent, Belgium 2 Max Planck Institute

More information

Statistics Lecture 6. Looking at data one variable

Statistics Lecture 6. Looking at data one variable Statistics 111 - Lecture 6 Looking at data one variable Chapter 1.1 Moore, McCabe and Craig Probability vs. Statistics Probability 1. We know the distribution of the random variable (Normal, Binomial)

More information

Regression III: Advanced Methods

Regression III: Advanced Methods Lecture 3: Distributions Regression III: Advanced Methods William G. Jacoby Michigan State University Goals of the lecture Examine data in graphical form Graphs for looking at univariate distributions

More information

Forecasting Video Analytics Sami Abu-El-Haija, Ooyala Inc

Forecasting Video Analytics Sami Abu-El-Haija, Ooyala Inc Forecasting Video Analytics Sami Abu-El-Haija, Ooyala Inc (haija@stanford.edu; sami@ooyala.com) 1. Introduction Ooyala Inc provides video Publishers an endto-end functionality for transcoding, storing,

More information

IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS /$ IEEE

IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS /$ IEEE IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS 1 Exploration of Heterogeneous FPGAs for Mapping Linear Projection Designs Christos-S. Bouganis, Member, IEEE, Iosifina Pournara, and Peter

More information

What We ll Do... Random

What We ll Do... Random What We ll Do... Random- number generation Random Number Generation Generating random variates Nonstationary Poisson processes Variance reduction Sequential sampling Designing and executing simulation

More information

A Topography-Preserving Latent Variable Model with Learning Metrics

A Topography-Preserving Latent Variable Model with Learning Metrics A Topography-Preserving Latent Variable Model with Learning Metrics Samuel Kaski and Janne Sinkkonen Helsinki University of Technology Neural Networks Research Centre P.O. Box 5400, FIN-02015 HUT, Finland

More information

You ve already read basics of simulation now I will be taking up method of simulation, that is Random Number Generation

You ve already read basics of simulation now I will be taking up method of simulation, that is Random Number Generation Unit 5 SIMULATION THEORY Lesson 39 Learning objective: To learn random number generation. Methods of simulation. Monte Carlo method of simulation You ve already read basics of simulation now I will be

More information

Diffusion Wavelets for Natural Image Analysis

Diffusion Wavelets for Natural Image Analysis Diffusion Wavelets for Natural Image Analysis Tyrus Berry December 16, 2011 Contents 1 Project Description 2 2 Introduction to Diffusion Wavelets 2 2.1 Diffusion Multiresolution............................

More information

CHAPTER 3. Preprocessing and Feature Extraction. Techniques

CHAPTER 3. Preprocessing and Feature Extraction. Techniques CHAPTER 3 Preprocessing and Feature Extraction Techniques CHAPTER 3 Preprocessing and Feature Extraction Techniques 3.1 Need for Preprocessing and Feature Extraction schemes for Pattern Recognition and

More information