A Sequential ESM Track Association Algorithm Based on the Use of Information Theoretic Criteria

Size: px
Start display at page:

Download "A Sequential ESM Track Association Algorithm Based on the Use of Information Theoretic Criteria"

Transcription

1 A Sequential ESM Track Association Algorithm Based on the Use of Information Theoretic Criteria Yifeng Zhou and Jim Mickeal Radar Electronic Warfare Section Defence R&D Canada - Ottawa 371 Carling Avenue, Ontario Canada K1A Z4 Abstract In this paper, a sequential track association algorithm for multiple Electronic Support Measures (ESM) sensors is proposed based on the application of clustering techniques. Association metrics are developed based on the use of the information theoretic criteria, in particular, the Akaike Information Criterion (AIC). The proposed association algorithm is able to handle track components of differing numbers of subcomponents including incomplete and erroneous ones. It has the advantage of not requiring any subjective threshold setting in deciding a nonassociation. Computer simulation and actual ESM processors are used to demonstrate the effectiveness and performance of the proposed association approach. Keywords: ESM, track association, radar, ELINT, theoretic information criteria, AIC, clustering I. INTRODUCTION ESM systems are designed to intercept radar signals and perform emitter classification and identification for electronic warfare (EW) purposes [1]. The extracted signal parameters can be used for threat warning, assisting with radar jamming in tactical situations and collection of electronic intelligence (ELINT) data in strategic situations. Recently, there has been increasing interest in combining information from multiple ESM sensors to provide passive geolocation capability and to enhance target identification performance for command and control systems [2]. The ESM processor output is usually in the form of streamline tracks, where each track corresponds to an emitter detected by the sensor and contains parameters of the emitter such as frequency, PRI and bearing etc.. Inorder to integrate multiple ESM sensor information, tracks produced by different ESM sensors that are from a same emitter must be associated. In this paper, we focus on the problem of associating track for multiple ESM sensors. Unlike radar track association problems where the association is mainly based on positional information of the targets, ESM track association is dependent on measured parameters and attribute information of the emitters, which usually includes declarations and statements about the emitters. There are several difficulties that are encountered in ESM track association. The main difficulty is due to the capability limitations of ESM sensors, which often fail to provide accurate emitter parameters and produce correct attribute information for the emitters, particularly emitters of complex characteristics. In the presence of complex emitters with radio frequency (RF) and pulse repetition interval (PRI) agility, ESM sensors tend to produce multiple tracks representing a single emitter or tracks with incomplete or erroneous RF and PRI components. Imperfect ESM sensors also produce imperfect tracks with low purity factors, which contain pulses from multiple emitters. In addition, since ESM tracks are of different types and contain different types of variables (some are declarations and others are numerical), it is important and difficult to define appropriate metrics for tracks. We will address these difficulties in the proposed track association algorithm. The problem of ESM track association has not received much attention and open literature publications are limited probably because of its classified nature in the past. One of the few techniques of ESM track association surveyed by the authors is the one by Wright [3], which was also discussed and commented on by Hall in his book [2]. This approach is an application of clustering techniques. In his approach, Wight used a measure of correlation, which is a combination of figure of merits for track components, to assess the similarity between a pair of tracks. This approach, however, can only handle tracks with same number of components. It has limited applications for complex signals with frequency and PRI agility. In this paper, we present a sequential ESM track association algorithm based on the use of clustering techniques and the information theoretic criteria. The Akaike Information Criteria (AIC), in particular, is used as the metrics for track association. In the algorithm, ESM tracks are defined to contain several components and each component consists of a number of subcomponents. A track component has a type parameter as well as numerical parameters associated with the type. Clusters are defined to represent the association results. When a track report arrives, matching scores are evaluated for the type parameters of the components for the track and existing clusters. Association metrics are then evaluated for the track and candidate clusters that have non-zero matching scores. Based on the association metrics, an association logic will decide whether the track be associated with one of the candidate clusters or not. The association metrics take into account the special states of tie and unknown of track components. They can handle tracks with differing number of components or subcomponents, a case which also occurs when tracks contain incomplete and erroneous components. Furthermore, the association metrics also have the advantage of not requiring any subjective threshold setting in distinguishing between an association and non-association. To demonstrate

2 the effectiveness and performance of the track association algorithm, a complex scenario that consists of hundreds of emitters and three ESM platforms is simulated. Simulated pulse descriptive words (PDWs) are simulated at the sites of the ESM platforms. Actual ESM processors are used to process the PDWs to generate streamline tracks for track association evaluation purpose. This paper is organized as follows. In Section II, the track association problem is formulated and the format of track reports is discussed. The ESM track association algorithm is described in Section III. In Section IV, we discuss the development of the track association metrics. Finally, in Section V, computer simulations are used to demonstrate the performance and effectiveness of the proposed algorithm. Comparisons with a baseline algorithm are also presented. ESM Sensor 1 track stream ESM Sensor 2 track stream ESM Sensor 3 track stream Figure 1. Cluster Manager cluster 1 cluster 2 cluster 3 new cluster Association Metrics Evaluation Association Logic Block diagram of the association algorithm. II. PROBLEM FORMULATION Assume that M (M = 3 is considered in this paper) distributed ESM sensors are deployed to observe a common signal environment. The environment consists of a large number of emitters. The platforms that carry the emitters include surface or airborne. The environment is mixed with friendly, unfriendly, threat and commercial platforms. The environment contains complex emitters that use agile RF/PRI and waveform modulation techniques. Many of the emitters may have similar signal characteristics such as frequency, PRI and pulse width etc.. The ESM sensors intercept emissions from the emitters, perform deinterleaving and pulse analysis, and produce streams of track reports and updates along the timeline. The objective of ESM track association is to associate track reports and updates from different ESM sensors into groups (or clusters) that represent the same emitters. A. Track Report The track report produced by an ESM processor usually consists of components such as RF, PRI, time-of-arrival (TOA) and pulse width (PW) and bearing of an emitter. It also contains emitter declarations such as track status, component type and antenna scan type etc.. The track report used in this study includes the following components. Identification (ID): a number assigned to a track by the ESM processor. Track State: a track state defines the status of the track. The track status includes active, inactive, gapped, merge and forced etc.. Time of Last Update: The time of last update is the timeof-arrival of the last pulse of the pulse trains on which the track report is based. It is synchronized to the time of arrival. Components: A track contains RF, PRI and PW component. Each component consists of a type parameter and a set of values. The RF type includes constant, agile, hopping unknown and tie while the PRI type can be one of constant, jittered, dwell, modulated, staggered, unknown and tie. A tie means that the processor can not distinguish the component type among the list of known types. The PW type can be either constant or unknown. Subcomponents: Each component may have differing numbers of subcomponents depending on the component type. For example, an RF component of constant type has only one subcomponent while the number of subcomponents of a hopping RF component is equal to the number of hops. The subcomponents are represented by their means, standard deviations and pulse counts that are used to compute the statistics. III. CLUSTERING BASED TRACK ASSOCIATION In this section, we describe the association algorithm. The detailed discussion of association metrics will be presented in the next section. Figure 1 is the block diagram of the association algorithm, where clusters are used to retain the associated tracks from different ESM sensors. Each cluster contains track reports and their updates that are considered to be a same emitter. It includes a list of tracks and a cluster center that represents the fused results of tracks from different ESM sensors. When a new track report from one of the ESM sensors arrives, the association algorithm first evaluates the matching scores for the component type parameters of the track against those of the existing clusters. Association metrics are then computed between the track and candidate clusters that have non-zero matching scores. Based on the association metrics, the association logic decides whether the track be assigned to one of the clusters or a new cluster be formed for it. The association metrics for a track and a cluster are a combination of association metrics for their components, i.e., the matching scores for their type parameters and the computation of similarity measures between the components. The matching scores for the component type parameters are computed based on a probability of agreement between the two types. For a component that has multiple subcomponents, the computation of similarity measures requires the process of subcomponent matching, which decides whether a subcomponent of the track component matches that of a cluster component. The matching is formulated as two-dimensional assignment problem [11] and

3 is solved using a computationally efficient modified auction algorithm. The general procedures of the proposed ESM track association algorithm can be summarized as follows. 1. A track report update is assigned to the cluster that contains track report with the same track ID. The cluster center is updated using the track report update. 2. If the track report is a new report, it will be compared with the existing clusters. First, the component type parameters of the track are compared against those of the existing clusters, and their matching scores are evaluated. If the maximum matching score is zero, then a new cluster is formed for the track; Otherwise, association metrics will be evaluated for the track and the candidate clusters that have non-zero matching scores. Non-association metrics are also calculated under the hypothesis that the track is not associated with any of the candidate clusters. The association logic will decide whether the track be associated with the cluster that has maximum association metric or a new cluster be formed for it. 3. When a track is associated with a cluster, it is applied to update the components of the cluster center. When a track and a track update is associated with a cluster, it is applied to update the associated cluster. The component type parameters of the cluster centers are replaced by those of the track update. For each component, subcomponent matching is performed. Note that the actual implementation of subcomponent matching may not be necessary because the process is already performed in the computation of the association metrics. For matched subcomponents, their means, standard deviations and pulse counts in the cluster center are updated. For subcomponents of the track report that do not match those of the cluster, they are inserted into the cluster center. The updating of the subcomponent of a cluster center is actually a process of recalculating the sample means and variances of each subcomponent. Let X and C k denote a pair of corresponding components of the associating track and cluster, respectively. Let { c k,j, σ 2 c,k,j,n c,k,j} and { x i, σ 2 x,i,n x,i} denote the sample means, variances and pulse counts of the ith and jth subcomponents of X and C k, respectively. Let I k denote the index set of matching subcomponents. The sample means and variances are updated by σ 2 c,k,j = 1 { n (nc,k,j σ 2 c,k,j + n x,i σ 2 x,i) c,k,j +n x,i ( x x,i c k,j) 2 + n c,k,j ( c k,j c k,j) 2}, (1) where the updated pulse count is given by IV. TRACK ASSOCIATION METRICS In this section, we discuss the association metrics for assigning a track to a cluster. The metrics are a combination of association metrics for all the components of the track. For a track component, the computation of association metrics includes the computation of matching scores for the type parameters, subcomponent matching and similarity measures between subcomponents. A. Matching Score for Track Components Matching scores are defined for the type parameters of a pair of track components to measure how close the two track components are in terms of their types. A track component type parameter is similar to a categorical or nominal variable in the context of clustering [4]. A categorical variable comprises a set of mutually exclusive discrete states such that it can only take one state at a time, and has no ordering on different states. However, the type parameters of track components are slightly different in that they contain, besides a set of mutually exclusive states, the special states of tie and unknown, where the state of tie is not mutually exclusive with the rest of the states, and the state of unknown is an imprecise state. There are several measures that can be used to measure the similarities between two nominal variables. They include the simple matching coefficient, the Jaccard s coefficient, the Gower s coefficient and the measure of agreement [4] [5]. However, these measures are only applicable to nominal variables with mutually exclusive states and are not capable of handling the non-exclusive states of tie and unknown. In the following, we propose a matching score that can take into account the special states of tie and unknown for measuring the similarities between the type parameters of a pair of track components. Let X denote a track component. The number of mutually exclusive states that the type parameter of X may fall into is denoted by s x. Define a state vector, η x, of length given by s x. When the type parameter of X falls into one of its exclusive states, the corresponding field of the state vector is set to 1; otherwise, it is zero. When the type parameter of X is tie, all the fields of η x are set to 1/s x, implying that the type parameter of X can take any of the mutually exclusive states equally. When the type parameter of X is unknown, all the fields of the state vector are set to zeros, meaning that the knowledge is imprecise and no decision on matching should be made. The values of the state vector can be interpreted as probabilities that the type parameter falls into the corresponding states. Define C k as a component of the kth cluster. The number of mutually exclusive states that the type parameter of C k mayfallintoisgivenbys c,k.letη c,k denote the state vector for the type parameter of C k. The matching score for X and C k is defined as P (X, C k )=η T x η c,k, (2) where {i, j} I k. n c,k,j = n c,k,j + n x,i which can be verified to lie in the range of [, 1]. We can interpret P (X, C k ) as the probability that the type parameters of X and C k fall into the same state. The overall matching

4 score for all the components of a pair of track and cluster can be evaluated as a product of the matching scores for each component by assuming that their type parameters are detected independently. A cluster and a track are said to match if the overall matching score is non-zero; otherwise, it is a nonmatch. B. Subcomponent Matching When a track matches a clusters, their components need to be compared and similarity measures be evaluated. However, when a component has multiple subcomponents, we need to match the subcomponents of the track with those of the cluster. As shown previously, subcomponent matching is also required when updating the cluster centers. Subcomponent matching is necessary because a track component may be incomplete contain erroneous subcomponents due to limitations of ESM processors. We formulate subcomponent matching as a two-dimensional assignment problem [11], where a subcomponent of a track component is assigned to either one subcomponent of the cluster component or a dummy according to a minimized cost criterion. The dummy is added to a cluster component to represent the state of no matching. A subcomponent assigned to the dummy means that it does not match any subcomponents. A modified auction algorithm is used for solving the assignment problem. The cost of matching a pair of subcomponents is defined based on the information theoretic criteria, in particular, the Akaike information criteria (AIC) [7]. The information theoretic criteria take into account both the goodness-of-fit (likelihood) of a model and the number of parameters used to achieve that fit. Denote I and J k as the numbers of subcomponents of a track and a cluster component, respectively. Let x i and c k,j be the ith and j th subcomponents of the track and the kth cluster, respectively. Let X = {x i ; i = 1, 2,...,I} and C k = {c k,j ; j =, 1, 2,...,J k },wherec k, denotes the dummy subcomponent. The pulse counts for x i and c k,j are denoted by n x,i and n c,k,j, respectively. Let { x i, σ x,i 2 } and { c k,j, σ c,k,j 2 } denote the sample means and variances of x i and c k,j, respectively. The cost that x i matches c k,j is defined as the AIC under the hypothesis that they share a common mean. Under the Gaussian assumption, it can be verified that the AIC is given by a ij = n c,k (ln 2π +lnˆσ 2 +1)+ξ ij (ln n c,k +1), (3) where ξ ij = J k +1 is the number of adjustable parameters and ˆσ 2 = 1 n c,k,j σ n c,k,j, 2 c,k j The same procedures are repeated for i =1, 2,...,I and j = 1, 2,...,J k to obtain all combinations of i and j. Thecost that x i matches the dummy in C k is given by the AIC under the hypothesis that x i does not match any subcomponents of C k, which can be computed using (3) by letting j =and using the variance estimate ˆσ 2 = 1 (n c,k,j σ c,k,j 2 + n x,i σ n x,i). 2 c,k j Note that the number of free adjustable parameters becomes ξ i = J k +2. The subcomponent matching is formulated as a 2-D assignment problem, which is to find the optimum α ij, i =1, 2,..., and j =, 1, 2,...,J k that minimizes the following min α ij α ij a ij, (4) ij subject to J k j= α ij =1,i=1, 2,...,I,whereα ij = {, 1} and a, =. The optimization (4) can be solved by using the modified auction algorithm described as follows. First, the matching costs are modified as a ij = a ij + a i for i =1, 2,...,J k and any subcomponent pairs with negative costs will be discarded. Define the subcomponents of the track and the cluster components as bidders and objects, respectively. Each bidder bids for a best object that provides the bidder with maximum profits. The auction algorithm iterates over unassigned bidders and terminates when either all the bidders are assigned or the objects have received bids. In the context of subcomponent matching, unassigned track subcomponents are assigned to the dummy of the cluster component, and, similarly, unassigned cluster subcomponents are assigned to the dummy of the track component. Unassigned cluster subcomponent means that either the cluster component has erroneous subcomponents or the track component is incomplete. If both the track and cluster components are singleton, i.e., both contain only one subcomponent, the matching process reduces to a simple test that compares the association and non-association costs. The subcomponents associate when the association cost is less than the non-association cost; otherwise, they are not associated. C. Association Metrics for Track Components After the subcomponents are matched, the association metrics for the components will be evaluated. The association metrics are based on the application of the AIC and are used to quantify the similarity measures between a pair of track components. Assume that there are K candidate clusters and a track is associated with the k th cluster. Let m k denote the number of matching subcomponents of C k. Under a Gaussian assumption and the hypothesis that x i is associated with C k, the AIC is given by AIC(X, C k )=N(ln 2π +lnˆσ 2 +1)+ξ k [ln N +1], (5) where

5 ˆσ 2 = 1 N + k,k k j,j I k n c,k,j σ c,k,j 2 + j i,i/ I k n x,i σ 2 x,i n c,k,j ˆσ2 c,k,j, (6) and ξ k = k j J k,j +(I m k )+1 is the number of free adjustable parameters under the hypothesis. The association metrics for a pair of track/cluster components combines the matching score 2lnP (X, C k ) AIC(X, C k ). (7) The non-association metric is evaluated under the hypothesis that X is not associated with any of candidate clusters. It can be verified that the non-association AIC is also given by (6) letting k =, where the variance is estimated by ˆσ 2 = 1 n z,k,j σ z,k,j 2 N + n x,i σ 2 x,i, k j i and ξ = j J k,j + I +1. D. Association Metrics for Track and Clusters The overall association metrics for associating a track and a candidate cluster are evaluated based on the metrics for each component, and are obtained by averaging the metrics for all the components. The non-association metric is also averaged for all the components. The association logic compares the AIC association and non-association metrics and searches for the minimum one. It associates the track with the cluster that has the minimum AIC association metric. If the nonassociation metrics turn out to be the minimum, the association logic will form a new cluster to contain the track. V. PERFORMANCE ANALYSIS Computer simulations are used to demonstrate the performance of the proposed ESM track association algorithm. The testbed consists of three ESM sensors: two shipborne and one airborne. In the scenario, about 3 emitters were simulated, which includes ships and aircraft. Both complex and simplex emitters were simulated with a large number of th emitters having similar parameters (navigation radars). Emitter PDWs are computed at the ESM platform sites and antenna/receiver models were applied. Isotropic antennas were used in the model. The receiver sensitivity was set to be 7 dbm. In the receiver model, Gaussian distributed noises with zero means and pre-set variances were added to the detected PDWs to simulate the receiver noise. The variances used by the receiver model are listed in Table I for each frequency band. Performance measures were developed for evaluating the effectiveness of the track association algorithms, which include cluster purity, collective score, detected emitter ratio and fragmentation ratio. Cluster Purity is the ratio of the peak to total pulse count of each cluster. The collective score is a Table I STANDARD DEVIATIONS FOR RECEIVER MODEL PARAMETERS RF Bearing TOA PW Power Band (MHz) (degrees) (ns) (ns) (dbm) 1 (1-2) GHz (2-6) GHz (6-12) GHz (12-18) GHz weighted sum of constructed track ratio, correct correlation, cluster purity and ambiguity (percent of pulses lost). The smaller the collective score, the better the performance. The detected emitter Ratio is the ratio of the number of correlated cluster to the total number of ground truth emitters. A value of 1 is ideal. Finally, the Fragmentation Ratio is defined as the ratio of the number of clusters that are not correlated to any ground truth cluster to the total number of clusters. The evaluation of the performance metrics requires a correlation of the clusters and the ground truth emitters. However, since real ESM processors with imperfect deinterleaving are used, the output track reports are not pure, i.e., each track report often contains pulses from more than one emitter. Obviously, impure track reports will result in impure clusters. In the presence of impure clusters, correlating a cluster with an emitter becomes a non-trivial task. We proposed an algorithm for correlating tracks and clusters with ground truth emitters based on a peak pulse count principle. Clusters that contain peak pulse counts corresponding to the same ground truth emitter are considered as fragmented. Among them, the cluster that has the maximum pulse count is correlated with the ground truth emitter, and the rest are considered not correlated with any of the ground truth emitters. A baseline algorithm based on the use of statistical distance measures is implemented for comparison [1]. Figure 2 shows the pulse count distribution of the emitter clusters versus the ground truth pulse index. For each cluster, the pulse counts of the ground truth emitters that are contained in the cluster is plotted. It can be seen that, due to the imperfectness of the ESM processors and the track association algorithm, the clusters contain pulses from more than one emitter, and pulses from one ground truth emitter are scattered into more than one cluster. Table II lists some of the performance metrics for the proposed and the baseline algorithm, which include cluster purity, the collective score, detected emitter ratio and fragmentation ratio. From the table, it can be seen that the proposed algorithm has better track purity and detected emitter ratio than the baseline algorithm. However, its fragmentation is slightly higher than that of the baseline algorithm, which in turn results in a slightly higher collective score. It should be the noted that the baseline algorithm requires subjective threshold settings, which is usually difficult in practice. Figure 3 is a cluster purity comparison for the proposed and the baseline algorithm. In the figure, the number of clusters that contain a specific number of ground truth emitters

6 3.5 x Proposed algorithm Alternate algorithm Pulse count Number of clusters Ground truth emitter index Cluster index Ground truth emitter count Figure 2. Cluster pulse count versus ground truth emitters. Figure 3. Distribution of clusters versus the number of ground truth emitters. Table II PERFORMANCE OF THE ASSOCIATION ALGORITHMS Proposed algorithm Baseline algorithm Cluster Collect. Detected Frag. Purity Score Track Ratio Ratio Proposed Baseline is counted, and plotted versus the number of ground truth emitters. The number of ground truth emitters corresponding to cluster count zero represents the number of emitters that are not detected by any ESM sensor. The proposed algorithm is shown to be able to produce more emitter clusters that contain one and only one ground truth emitter. Figure 4 is a track count histogram comparison. The figure shows the distribution of clusters that contain differing numbers of tracks for the two algorithms. It can be seen that the proposed algorithm produced more clusters that contain only one track, meaning that the proposed algorithm is more likely to form a new cluster to contain tracks. The total of number of clusters that contain two or three tracks are similar for both algorithms. VI. CONCLUSIONS In this paper, a sequential track association algorithm has been presented for multiple ESM sensors based clustering techniques and the application of the information theoretic criteria. The algorithm is capable of handling incomplete and erroneous track subcomponents and does not require any subjective threshold setting. A complex signal environment was simulated and actual ESM processors were used to produce track streams for demonstrating the performance of the proposed association algorithm. Performance metrics were proposed and it was shown that the proposed algorithm was able to produce reasonably good results when applied to imperfect tracks obtained from actual ESM processors with imperfect deinterleaving capabilities. The association results were also compared with a baseline algorithm based statistical distance measures. It was shown that the proposed algorithm Number of clusters Track count Figure 4. Track count histogram comparison. outperformed the baseline algorithm in terms of track purity. However, it had a slightly higher fragmentation track ratio than the baseline algorithm. It should be noted that the baseline algorithm requires subjective threshold settings in order to achieve the optimal performance. REFERENCES [1] R.G.Wiley,Electronic Intelligence: The Analysis of Radar Signals, 2nd Ed., Artech House, Norwood, MA, 1993 [2] D. L. Hall, Mathematical techniques in multisensor data fusion, Artech House, Norwood, MA, 1992 [3] F. L. Wright, The fusion of multi-source data, Signal, vol. no. pp , October 198 [4] M. R. Anderberg, Cluster Analysis for Application, Academic Press, New York, 1973 [5] A. D. Gordon, Classification: Method for the exploration analysis of multivariate data, Chapman and Hall, New York, 1981 [6] D. P. Bertsekas, Auction algorithms for network flow problem: A tutorial introduction, Computational Optimization and Applications, vol. 1, pp. 7-66, 1992 [7] H. Akaike, A new look at the statistical model identification, IEEE Trans. AC, vol. 6, pp , 1974 [8] J. Rissanen, Modeling by the shortest description, Automatica, vol. 14, pp , 1978

7 [9] G. Schwartz, Estimating the dimension of a model, Ann. Stat., vol. 6, pp , 1978 [1] J. Mickeal and Y. Zhou, Multiple ESM track association based on Clustering, in preparation, Technical Report, Defence R&D Canada, December 25 [11] D. P. Bertsekas, Auction algorithms for network flow problem: A tutorial introduction, Computational Optimization and Applications, vol. 1, pp. 7-66, 1992

3.2 Level 1 Processing

3.2 Level 1 Processing SENSOR AND DATA FUSION ARCHITECTURES AND ALGORITHMS 57 3.2 Level 1 Processing Level 1 processing is the low-level processing that results in target state estimation and target discrimination. 9 The term

More information

TELCOM2125: Network Science and Analysis

TELCOM2125: Network Science and Analysis School of Information Sciences University of Pittsburgh TELCOM2125: Network Science and Analysis Konstantinos Pelechrinis Spring 2015 Figures are taken from: M.E.J. Newman, Networks: An Introduction 2

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2008 CS 551, Spring 2008 c 2008, Selim Aksoy (Bilkent University)

More information

Lecture on Modeling Tools for Clustering & Regression

Lecture on Modeling Tools for Clustering & Regression Lecture on Modeling Tools for Clustering & Regression CS 590.21 Analysis and Modeling of Brain Networks Department of Computer Science University of Crete Data Clustering Overview Organizing data into

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Gene Clustering & Classification

Gene Clustering & Classification BINF, Introduction to Computational Biology Gene Clustering & Classification Young-Rae Cho Associate Professor Department of Computer Science Baylor University Overview Introduction to Gene Clustering

More information

Online Pattern Recognition in Multivariate Data Streams using Unsupervised Learning

Online Pattern Recognition in Multivariate Data Streams using Unsupervised Learning Online Pattern Recognition in Multivariate Data Streams using Unsupervised Learning Devina Desai ddevina1@csee.umbc.edu Tim Oates oates@csee.umbc.edu Vishal Shanbhag vshan1@csee.umbc.edu Machine Learning

More information

MATH3016: OPTIMIZATION

MATH3016: OPTIMIZATION MATH3016: OPTIMIZATION Lecturer: Dr Huifu Xu School of Mathematics University of Southampton Highfield SO17 1BJ Southampton Email: h.xu@soton.ac.uk 1 Introduction What is optimization? Optimization is

More information

INF 4300 Classification III Anne Solberg The agenda today:

INF 4300 Classification III Anne Solberg The agenda today: INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15

More information

Using Machine Learning to Optimize Storage Systems

Using Machine Learning to Optimize Storage Systems Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation

More information

Introduction to digital image classification

Introduction to digital image classification Introduction to digital image classification Dr. Norman Kerle, Wan Bakx MSc a.o. INTERNATIONAL INSTITUTE FOR GEO-INFORMATION SCIENCE AND EARTH OBSERVATION Purpose of lecture Main lecture topics Review

More information

Part I, Chapters 4 & 5. Data Tables and Data Analysis Statistics and Figures

Part I, Chapters 4 & 5. Data Tables and Data Analysis Statistics and Figures Part I, Chapters 4 & 5 Data Tables and Data Analysis Statistics and Figures Descriptive Statistics 1 Are data points clumped? (order variable / exp. variable) Concentrated around one value? Concentrated

More information

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Improving the Efficiency of Fast Using Semantic Similarity Algorithm International Journal of Scientific and Research Publications, Volume 4, Issue 1, January 2014 1 Improving the Efficiency of Fast Using Semantic Similarity Algorithm D.KARTHIKA 1, S. DIVAKAR 2 Final year

More information

Dynamic Thresholding for Image Analysis

Dynamic Thresholding for Image Analysis Dynamic Thresholding for Image Analysis Statistical Consulting Report for Edward Chan Clean Energy Research Center University of British Columbia by Libo Lu Department of Statistics University of British

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

Fusion of Radar and EO-sensors for Surveillance

Fusion of Radar and EO-sensors for Surveillance of Radar and EO-sensors for Surveillance L.J.H.M. Kester, A. Theil TNO Physics and Electronics Laboratory P.O. Box 96864, 2509 JG The Hague, The Netherlands kester@fel.tno.nl, theil@fel.tno.nl Abstract

More information

Multisensor Data Fusion Using Two-Stage Analysis on Pairs of Plots Graphs

Multisensor Data Fusion Using Two-Stage Analysis on Pairs of Plots Graphs Multisensor Data Fusion Using Two-Stage Analysis on Pairs of Plots Graphs Rogério Perroti Barbosa 1,2, Frederic Livernet 3, Beatriz S. L. P. de Lima 1, José Gomes de Carvalho Jr 2 1 - COPPE/ Federal University

More information

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Cluster Analysis Mu-Chun Su Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Introduction Cluster analysis is the formal study of algorithms and methods

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi Unsupervised Learning Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi Content Motivation Introduction Applications Types of clustering Clustering criterion functions Distance functions Normalization Which

More information

Expectation Maximization!

Expectation Maximization! Expectation Maximization! adapted from: Doug Downey and Bryan Pardo, Northwestern University and http://www.stanford.edu/class/cs276/handouts/lecture17-clustering.ppt Steps in Clustering Select Features

More information

The Curse of Dimensionality

The Curse of Dimensionality The Curse of Dimensionality ACAS 2002 p1/66 Curse of Dimensionality The basic idea of the curse of dimensionality is that high dimensional data is difficult to work with for several reasons: Adding more

More information

Application of Characteristic Function Method in Target Detection

Application of Characteristic Function Method in Target Detection Application of Characteristic Function Method in Target Detection Mohammad H Marhaban and Josef Kittler Centre for Vision, Speech and Signal Processing University of Surrey Surrey, GU2 7XH, UK eep5mm@ee.surrey.ac.uk

More information

Distributed Detection in Sensor Networks: Connectivity Graph and Small World Networks

Distributed Detection in Sensor Networks: Connectivity Graph and Small World Networks Distributed Detection in Sensor Networks: Connectivity Graph and Small World Networks SaeedA.AldosariandJoséM.F.Moura Electrical and Computer Engineering Department Carnegie Mellon University 5000 Forbes

More information

0.5 to 2.4GHz Instantaneous Frequency Measurement Unit

0.5 to 2.4GHz Instantaneous Frequency Measurement Unit Frequency Measurement Unit 0.5 to 2.4GHz Measurement Unit The uses proprietary Teledyne Defence & Space (TDS) Technology to provide state of the art performance in a package measuring 130mm x 90mm x 18mm.

More information

K-Means Clustering Using Localized Histogram Analysis

K-Means Clustering Using Localized Histogram Analysis K-Means Clustering Using Localized Histogram Analysis Michael Bryson University of South Carolina, Department of Computer Science Columbia, SC brysonm@cse.sc.edu Abstract. The first step required for many

More information

Dealing with Categorical Data Types in a Designed Experiment

Dealing with Categorical Data Types in a Designed Experiment Dealing with Categorical Data Types in a Designed Experiment Part II: Sizing a Designed Experiment When Using a Binary Response Best Practice Authored by: Francisco Ortiz, PhD STAT T&E COE The goal of

More information

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

The Comparative Study of Machine Learning Algorithms in Text Data Classification* The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification

More information

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response Authored by: Francisco Ortiz, PhD Version 2: 19 July 2018 Revised 18 October 2018 The goal of the STAT COE is to assist in

More information

Middle School Math Course 2

Middle School Math Course 2 Middle School Math Course 2 Correlation of the ALEKS course Middle School Math Course 2 to the Indiana Academic Standards for Mathematics Grade 7 (2014) 1: NUMBER SENSE = ALEKS course topic that addresses

More information

Probability Evaluation in MHT with a Product Set Representation of Hypotheses

Probability Evaluation in MHT with a Product Set Representation of Hypotheses Probability Evaluation in MHT with a Product Set Representation of Hypotheses Johannes Wintenby Ericsson Microwave Systems 431 84 Mölndal, Sweden johannes.wintenby@ericsson.com Abstract - Multiple Hypothesis

More information

3 Feature Selection & Feature Extraction

3 Feature Selection & Feature Extraction 3 Feature Selection & Feature Extraction Overview: 3.1 Introduction 3.2 Feature Extraction 3.3 Feature Selection 3.3.1 Max-Dependency, Max-Relevance, Min-Redundancy 3.3.2 Relevance Filter 3.3.3 Redundancy

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2012 http://ce.sharif.edu/courses/90-91/2/ce725-1/ Agenda Features and Patterns The Curse of Size and

More information

An ICA based Approach for Complex Color Scene Text Binarization

An ICA based Approach for Complex Color Scene Text Binarization An ICA based Approach for Complex Color Scene Text Binarization Siddharth Kherada IIIT-Hyderabad, India siddharth.kherada@research.iiit.ac.in Anoop M. Namboodiri IIIT-Hyderabad, India anoop@iiit.ac.in

More information

2016 Stat-Ease, Inc. & CAMO Software

2016 Stat-Ease, Inc. & CAMO Software Multivariate Analysis and Design of Experiments in practice using The Unscrambler X Frank Westad CAMO Software fw@camo.com Pat Whitcomb Stat-Ease pat@statease.com Agenda Goal: Part 1: Part 2: Show how

More information

MODEL BASED CLASSIFICATION USING MULTI-PING DATA

MODEL BASED CLASSIFICATION USING MULTI-PING DATA MODEL BASED CLASSIFICATION USING MULTI-PING DATA Christopher P. Carbone Naval Undersea Warfare Center Division Newport 76 Howell St. Newport, RI 084 (email: carbonecp@npt.nuwc.navy.mil) Steven M. Kay University

More information

1) Give decision trees to represent the following Boolean functions:

1) Give decision trees to represent the following Boolean functions: 1) Give decision trees to represent the following Boolean functions: 1) A B 2) A [B C] 3) A XOR B 4) [A B] [C Dl Answer: 1) A B 2) A [B C] 1 3) A XOR B = (A B) ( A B) 4) [A B] [C D] 2 2) Consider the following

More information

Metrics for Feature-Aided Track Association

Metrics for Feature-Aided Track Association Metrics for eature-aided rack Association Chee-Yee Chong BAE Systems Los Altos, CA, U.S.A. chee.chong@baesystems.com Shozo Mori BAE Systems Los Altos, CA, U.S.A. shozo.mori@baesystems.com Abstract: rack

More information

A New Online Clustering Approach for Data in Arbitrary Shaped Clusters

A New Online Clustering Approach for Data in Arbitrary Shaped Clusters A New Online Clustering Approach for Data in Arbitrary Shaped Clusters Richard Hyde, Plamen Angelov Data Science Group, School of Computing and Communications Lancaster University Lancaster, LA1 4WA, UK

More information

Radar Target Identification Using Spatial Matched Filters. L.M. Novak, G.J. Owirka, and C.M. Netishen MIT Lincoln Laboratory

Radar Target Identification Using Spatial Matched Filters. L.M. Novak, G.J. Owirka, and C.M. Netishen MIT Lincoln Laboratory Radar Target Identification Using Spatial Matched Filters L.M. Novak, G.J. Owirka, and C.M. Netishen MIT Lincoln Laboratory Abstract The application of spatial matched filter classifiers to the synthetic

More information

A Joint Histogram - 2D Correlation Measure for Incomplete Image Similarity

A Joint Histogram - 2D Correlation Measure for Incomplete Image Similarity A Joint Histogram - 2D Correlation for Incomplete Image Similarity Nisreen Ryadh Hamza MSc Candidate, Faculty of Computer Science & Mathematics, University of Kufa, Iraq. ORCID: 0000-0003-0607-3529 Hind

More information

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust

More information

Nearest Clustering Algorithm for Satellite Image Classification in Remote Sensing Applications

Nearest Clustering Algorithm for Satellite Image Classification in Remote Sensing Applications Nearest Clustering Algorithm for Satellite Image Classification in Remote Sensing Applications Anil K Goswami 1, Swati Sharma 2, Praveen Kumar 3 1 DRDO, New Delhi, India 2 PDM College of Engineering for

More information

Annotated multitree output

Annotated multitree output Annotated multitree output A simplified version of the two high-threshold (2HT) model, applied to two experimental conditions, is used as an example to illustrate the output provided by multitree (version

More information

A MATLAB TOOL FOR DEVELOPMENT AND TESTING OF TRACK INITIATION AND MULTIPLE TARGET TRACKING ALGORITHMS

A MATLAB TOOL FOR DEVELOPMENT AND TESTING OF TRACK INITIATION AND MULTIPLE TARGET TRACKING ALGORITHMS I&S A MATLAB TOOL FOR DEVELOPMENT AND TESTING OF TRACK INITIATION AND MULTIPLE TARGET TRACKING ALGORITHMS Kiril ALEXIEV 1. Introduction Digital computer simulation is a valuable tool, used for the design,

More information

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Exploratory data analysis tasks Examine the data, in search of structures

More information

Principal Component Image Interpretation A Logical and Statistical Approach

Principal Component Image Interpretation A Logical and Statistical Approach Principal Component Image Interpretation A Logical and Statistical Approach Md Shahid Latif M.Tech Student, Department of Remote Sensing, Birla Institute of Technology, Mesra Ranchi, Jharkhand-835215 Abstract

More information

A Neural Network for Real-Time Signal Processing

A Neural Network for Real-Time Signal Processing 248 MalkofT A Neural Network for Real-Time Signal Processing Donald B. Malkoff General Electric / Advanced Technology Laboratories Moorestown Corporate Center Building 145-2, Route 38 Moorestown, NJ 08057

More information

Directional Sensor Control for Maximizing Information Gain

Directional Sensor Control for Maximizing Information Gain Directional Sensor Control for Maximizing Information Gain Shankarachary Ragi a, Hans D. Mittelmann b, and Edwin K. P. Chong a a Department of Electrical and Computer Engineering, Colorado State University,

More information

Data Preprocessing. Why Data Preprocessing? MIT-652 Data Mining Applications. Chapter 3: Data Preprocessing. Multi-Dimensional Measure of Data Quality

Data Preprocessing. Why Data Preprocessing? MIT-652 Data Mining Applications. Chapter 3: Data Preprocessing. Multi-Dimensional Measure of Data Quality Why Data Preprocessing? Data in the real world is dirty incomplete: lacking attribute values, lacking certain attributes of interest, or containing only aggregate data e.g., occupation = noisy: containing

More information

Short Survey on Static Hand Gesture Recognition

Short Survey on Static Hand Gesture Recognition Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of

More information

Classification with Diffuse or Incomplete Information

Classification with Diffuse or Incomplete Information Classification with Diffuse or Incomplete Information AMAURY CABALLERO, KANG YEN Florida International University Abstract. In many different fields like finance, business, pattern recognition, communication

More information

CS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series

CS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series CS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series Jingyuan Chen //Department of Electrical Engineering, cjy2010@stanford.edu//

More information

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate

More information

Multisource Information Adaptive Fuzzy Logic Correlator for Recognized Maritime Picture

Multisource Information Adaptive Fuzzy Logic Correlator for Recognized Maritime Picture Multisource Information Adaptive Fuzzy Logic Correlator for Recognized Maritime Picture Eric Lefebvre Lockheed Martin Canada, 6111 Royalmount Avenue, Montréal, Québec, Canada, H4P 1K6 eric.lefebvre@lmco.com

More information

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY Angela M. Kim and Richard C. Olsen Remote Sensing Center Naval Postgraduate School 1 University Circle Monterey, CA

More information

A Study on Clustering Method by Self-Organizing Map and Information Criteria

A Study on Clustering Method by Self-Organizing Map and Information Criteria A Study on Clustering Method by Self-Organizing Map and Information Criteria Satoru Kato, Tadashi Horiuchi,andYoshioItoh Matsue College of Technology, 4-4 Nishi-ikuma, Matsue, Shimane 90-88, JAPAN, kato@matsue-ct.ac.jp

More information

(Refer Slide Time: 0:51)

(Refer Slide Time: 0:51) Introduction to Remote Sensing Dr. Arun K Saraf Department of Earth Sciences Indian Institute of Technology Roorkee Lecture 16 Image Classification Techniques Hello everyone welcome to 16th lecture in

More information

The Perils of Unfettered In-Sample Backtesting

The Perils of Unfettered In-Sample Backtesting The Perils of Unfettered In-Sample Backtesting Tyler Yeats June 8, 2015 Abstract When testing a financial investment strategy, it is common to use what is known as a backtest, or a simulation of how well

More information

University of Florida CISE department Gator Engineering. Clustering Part 5

University of Florida CISE department Gator Engineering. Clustering Part 5 Clustering Part 5 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville SNN Approach to Clustering Ordinary distance measures have problems Euclidean

More information

Data Mining. ❷Chapter 2 Basic Statistics. Asso.Prof.Dr. Xiao-dong Zhu. Business School, University of Shanghai for Science & Technology

Data Mining. ❷Chapter 2 Basic Statistics. Asso.Prof.Dr. Xiao-dong Zhu. Business School, University of Shanghai for Science & Technology ❷Chapter 2 Basic Statistics Business School, University of Shanghai for Science & Technology 2016-2017 2nd Semester, Spring2017 Contents of chapter 1 1 recording data using computers 2 3 4 5 6 some famous

More information

November 1998 doc.: IEEE /378 IEEE P Wireless LANs Extension of Bluetooth and Direct Sequence Interference Model.

November 1998 doc.: IEEE /378 IEEE P Wireless LANs Extension of Bluetooth and Direct Sequence Interference Model. IEEE P802.11 Wireless LANs Extension of Bluetooth and 802.11 Direct Sequence Interference Model Date: November 11, 1998 Author: Jim Zyren Harris Semiconductor Melbourne, FL, USA Phone: (407)729-4177 Fax:

More information

Chapter DM:II. II. Cluster Analysis

Chapter DM:II. II. Cluster Analysis Chapter DM:II II. Cluster Analysis Cluster Analysis Basics Hierarchical Cluster Analysis Iterative Cluster Analysis Density-Based Cluster Analysis Cluster Evaluation Constrained Cluster Analysis DM:II-1

More information

Decision Trees Dr. G. Bharadwaja Kumar VIT Chennai

Decision Trees Dr. G. Bharadwaja Kumar VIT Chennai Decision Trees Decision Tree Decision Trees (DTs) are a nonparametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

Machine Learning. Topic 4: Linear Regression Models

Machine Learning. Topic 4: Linear Regression Models Machine Learning Topic 4: Linear Regression Models (contains ideas and a few images from wikipedia and books by Alpaydin, Duda/Hart/ Stork, and Bishop. Updated Fall 205) Regression Learning Task There

More information

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK FOR IMAGE RESTORATION

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK FOR IMAGE RESTORATION 135 CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK FOR IMAGE RESTORATION 6.1 INTRODUCTION Neural networks have high fault tolerance and potential for adaptive training. A Full Counter Propagation Neural

More information

Multidimensional Image Registered Scanner using MDPSO (Multi-objective Discrete Particle Swarm Optimization)

Multidimensional Image Registered Scanner using MDPSO (Multi-objective Discrete Particle Swarm Optimization) Multidimensional Image Registered Scanner using MDPSO (Multi-objective Discrete Particle Swarm Optimization) Rishiganesh V 1, Swaruba P 2 PG Scholar M.Tech-Multimedia Technology, Department of CSE, K.S.R.

More information

6. Dicretization methods 6.1 The purpose of discretization

6. Dicretization methods 6.1 The purpose of discretization 6. Dicretization methods 6.1 The purpose of discretization Often data are given in the form of continuous values. If their number is huge, model building for such data can be difficult. Moreover, many

More information

Gaussian Mixture Models method and applications

Gaussian Mixture Models method and applications Gaussian Mixture Models method and applications Jesús Zambrano PostDoctoral Researcher School of Business, Society and Engineering www.mdh.se FUDIPO project. Machine Learning course. Oct.-Dec. 2017 Outline

More information

Clustering Analysis based on Data Mining Applications Xuedong Fan

Clustering Analysis based on Data Mining Applications Xuedong Fan Applied Mechanics and Materials Online: 203-02-3 ISSN: 662-7482, Vols. 303-306, pp 026-029 doi:0.4028/www.scientific.net/amm.303-306.026 203 Trans Tech Publications, Switzerland Clustering Analysis based

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning

More information

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,

More information

Thomas R Kronhamn Ericsson Microwave Systems AB Mölndal, Sweden

Thomas R Kronhamn Ericsson Microwave Systems AB Mölndal, Sweden 6HQVRU,QWHU1HW:RUNV Thomas R Kronhamn Ericsson Microwave Systems AB Mölndal, Sweden Thomas.Kronhamn@emw.ericsson.se $EVWUDFW It is shown how distributed sensors, networked by Internet techniques, can interact

More information

METAL OXIDE VARISTORS

METAL OXIDE VARISTORS POWERCET CORPORATION METAL OXIDE VARISTORS PROTECTIVE LEVELS, CURRENT AND ENERGY RATINGS OF PARALLEL VARISTORS PREPARED FOR EFI ELECTRONICS CORPORATION SALT LAKE CITY, UTAH METAL OXIDE VARISTORS PROTECTIVE

More information

Memorandum. Clint Slatton Prof. Brian Evans Term project idea for Multidimensional Signal Processing (EE381k)

Memorandum. Clint Slatton Prof. Brian Evans Term project idea for Multidimensional Signal Processing (EE381k) Memorandum From: To: Subject: Date : Clint Slatton Prof. Brian Evans Term project idea for Multidimensional Signal Processing (EE381k) 16-Sep-98 Project title: Minimizing segmentation discontinuities in

More information

Fast or furious? - User analysis of SF Express Inc

Fast or furious? - User analysis of SF Express Inc CS 229 PROJECT, DEC. 2017 1 Fast or furious? - User analysis of SF Express Inc Gege Wen@gegewen, Yiyuan Zhang@yiyuan12, Kezhen Zhao@zkz I. MOTIVATION The motivation of this project is to predict the likelihood

More information

Optimal designs for comparing curves

Optimal designs for comparing curves Optimal designs for comparing curves Holger Dette, Ruhr-Universität Bochum Maria Konstantinou, Ruhr-Universität Bochum Kirsten Schorning, Ruhr-Universität Bochum FP7 HEALTH 2013-602552 Outline 1 Motivation

More information

CHAPTER 6 QUANTITATIVE PERFORMANCE ANALYSIS OF THE PROPOSED COLOR TEXTURE SEGMENTATION ALGORITHMS

CHAPTER 6 QUANTITATIVE PERFORMANCE ANALYSIS OF THE PROPOSED COLOR TEXTURE SEGMENTATION ALGORITHMS 145 CHAPTER 6 QUANTITATIVE PERFORMANCE ANALYSIS OF THE PROPOSED COLOR TEXTURE SEGMENTATION ALGORITHMS 6.1 INTRODUCTION This chapter analyzes the performance of the three proposed colortexture segmentation

More information

Automatic Grayscale Classification using Histogram Clustering for Active Contour Models

Automatic Grayscale Classification using Histogram Clustering for Active Contour Models Research Article International Journal of Current Engineering and Technology ISSN 2277-4106 2013 INPRESSCO. All Rights Reserved. Available at http://inpressco.com/category/ijcet Automatic Grayscale Classification

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Unsupervised learning Until now, we have assumed our training samples are labeled by their category membership. Methods that use labeled samples are said to be supervised. However,

More information

Ranking Clustered Data with Pairwise Comparisons

Ranking Clustered Data with Pairwise Comparisons Ranking Clustered Data with Pairwise Comparisons Kevin Kowalski nargle@cs.wisc.edu 1. INTRODUCTION Background. Machine learning often relies heavily on being able to rank instances in a large set of data

More information

Adaptive Doppler centroid estimation algorithm of airborne SAR

Adaptive Doppler centroid estimation algorithm of airborne SAR Adaptive Doppler centroid estimation algorithm of airborne SAR Jian Yang 1,2a), Chang Liu 1, and Yanfei Wang 1 1 Institute of Electronics, Chinese Academy of Sciences 19 North Sihuan Road, Haidian, Beijing

More information

Scanner Parameter Estimation Using Bilevel Scans of Star Charts

Scanner Parameter Estimation Using Bilevel Scans of Star Charts ICDAR, Seattle WA September Scanner Parameter Estimation Using Bilevel Scans of Star Charts Elisa H. Barney Smith Electrical and Computer Engineering Department Boise State University, Boise, Idaho 8375

More information

SOCIAL MEDIA MINING. Data Mining Essentials

SOCIAL MEDIA MINING. Data Mining Essentials SOCIAL MEDIA MINING Data Mining Essentials Dear instructors/users of these slides: Please feel free to include these slides in your own material, or modify them as you see fit. If you decide to incorporate

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2013 http://ce.sharif.edu/courses/91-92/2/ce725-1/ Agenda Features and Patterns The Curse of Size and

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised

More information

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness Visible and Long-Wave Infrared Image Fusion Schemes for Situational Awareness Multi-Dimensional Digital Signal Processing Literature Survey Nathaniel Walker The University of Texas at Austin nathaniel.walker@baesystems.com

More information

Spatial Outlier Detection

Spatial Outlier Detection Spatial Outlier Detection Chang-Tien Lu Department of Computer Science Northern Virginia Center Virginia Tech Joint work with Dechang Chen, Yufeng Kou, Jiang Zhao 1 Spatial Outlier A spatial data point

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Features and Patterns The Curse of Size and

More information

A Self-Organizing Binary System*

A Self-Organizing Binary System* 212 1959 PROCEEDINGS OF THE EASTERN JOINT COMPUTER CONFERENCE A Self-Organizing Binary System* RICHARD L. MATTSONt INTRODUCTION ANY STIMULUS to a system such as described in this paper can be coded into

More information

AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION

AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION WILLIAM ROBSON SCHWARTZ University of Maryland, Department of Computer Science College Park, MD, USA, 20742-327, schwartz@cs.umd.edu RICARDO

More information

Air Force Test Center 412 th Test Wing, Electronic Warfare Group

Air Force Test Center 412 th Test Wing, Electronic Warfare Group Air Force Test Center 412 th Test Wing, Electronic Warfare Group I n t e g r i t y - S e r v i c e - E x c e l l e n c e Data Association Techniques for Open Air Range Testing Daniel Carlson EW Technical

More information

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016 edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract

More information

Cluster Analysis. Angela Montanari and Laura Anderlucci

Cluster Analysis. Angela Montanari and Laura Anderlucci Cluster Analysis Angela Montanari and Laura Anderlucci 1 Introduction Clustering a set of n objects into k groups is usually moved by the aim of identifying internally homogenous groups according to a

More information

Data Preprocessing. S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha

Data Preprocessing. S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha Data Preprocessing S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha 1 Why Data Preprocessing? Data in the real world is dirty incomplete: lacking attribute values, lacking

More information

Structural and Syntactic Pattern Recognition

Structural and Syntactic Pattern Recognition Structural and Syntactic Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent

More information

Acquisition Description Exploration Examination Understanding what data is collected. Characterizing properties of data.

Acquisition Description Exploration Examination Understanding what data is collected. Characterizing properties of data. Summary Statistics Acquisition Description Exploration Examination what data is collected Characterizing properties of data. Exploring the data distribution(s). Identifying data quality problems. Selecting

More information

A Non-Iterative Approach to Frequency Estimation of a Complex Exponential in Noise by Interpolation of Fourier Coefficients

A Non-Iterative Approach to Frequency Estimation of a Complex Exponential in Noise by Interpolation of Fourier Coefficients A on-iterative Approach to Frequency Estimation of a Complex Exponential in oise by Interpolation of Fourier Coefficients Shahab Faiz Minhas* School of Electrical and Electronics Engineering University

More information