A Comparative Study of Supervised and Unsupervised Learning Schemes for Intrusion Detection. NIS Research Group Reza Sadoddin, Farnaz Gharibian, and
|
|
- Donna Clark
- 5 years ago
- Views:
Transcription
1 A Comparative Study of Supervised and Unsupervised Learning Schemes for Intrusion Detection NIS Research Group Reza Sadoddin, Farnaz Gharibian, and
2 Agenda Brief Overview Machine Learning Techniques Clustering/Classification Techniques Dataset & Data Preparation Simulated Attacks Experimental Results Conclusions 2
3 Introduction Shortcomings of traditional techniques to intrusion detection Difficulties with specification of normal or attack behavior Expert knowledge Time-consuming Concept drift is a serious issue How to adapt to environment changes? 3
4 Security Intelligence Six Steps to Producing Security Intelligence Designate Data Log entries, raw or formatted measure of activity in an environment Model Analyst Expertise Weights, centers, and pertinent event knowledge comprise the analytic or data mining model Train Model Baseline of activity that is typical Generate Knowledge Live or offline data is compared against the baseline/classifier Teach Model User-supervision and infusing of expert knowledge Leverage Model 4
5 5 Introduction Advantages of machine learning and data mining Ability to learn and discover Adaptation can be done automatically
6 6 Goals of this study A blind comparison between different supervised and unsupervised techniques Overall accuracy in detecting attacks Performance in detecting different attack categories Sensitivity of techniques to distribution of training and test datasets
7 7 Machine Learning of Techniques Machine Learning Techniques Unsupervised Supervised Distance-based Clustering Unsupervised SVM K-Means C-Means EM SOM Y-Means ICLN Naïve Bayes Decision Tree Random Forest Support Vector Machine Gaussian
8 8 K-Means Clustering Basics Grouping the objects into k clusters Assumption: the number of clusters is given Objective: To minimize the total intra-cluster variance k i= 1 j S i x j μ i 2
9 9 K-Means Algorithm Initial k centroids Assign each each object to to its its closest centroid Calculate the the mean vector of of each each cluster Shift Shift centroids to to their their means No Are all centroids stable? Yes End End
10 K-Means Clustering Pros Simplicity Quick convergence Cons No guarantee to find the global optimum Its performance depends on the initial seeds (cluster centroids) What is the suitable K? 10
11 Fuzzy C-Means C Clustering Basics 11 Each point has a degree of belonging to clusters Center of clusters k = 1 Number of Number of Points All the points contribute Clusters with a degree u k ( x) = 1 Minimizing the following objective function J m ( X, V ) = N k j= 1 i= 1 ( u ij ) m d 2 ( X j, V Degree of membership i ) c k = x x u k u ( x) k m ( x) m x Distance Function
12 12 Fuzzy C-Means C Clustering Algorithm Choose initial number of clusters. Assign random coefficients for being in the clusters to each point (i.e., u ij ) Repeat until m u x k ( x) x new ck = Compute the centroid for each cluster. m u ( x) For each point, compute its coefficients of being in the clusters x k u new ij 1 = C xi c j ( ) 2 k = 1 m 1 xi ck
13 Fuzzy C-Means C Clustering Pros Useful when hard classification of data is not possible Cons Suffers from all the problems mentioned about K-Means initial memberships are required as well Convergence is slower than K-Means 13
14 14 EM (Expectation-Maximization) Basics A model-based approach to clustering Assumption: data is generated by K Gaussian distributions N( μ 1, σ 1 2 ) Provides a soft clustering as apposed to K-Means N( μ 2, σ 2 2 )
15 15 Y-Means Y-means A new classification method based on K- means. A dynamic clustering method without supervision. It overcomes three shortcomings of K- means: Degeneracy Dependency on the number of clusters Dependency on the initial states
16 16 Y-Means Algorithm Start normalize training data run K-means Is there degeneracy? False True remove empty clusters split clusters End link clusters True Are the False clusters stable?
17 17 SOM (Self-Organizing Maps) Extremely A Useful competitive for fast extracting learning (compared structural techniques to other information techniques) from Suitable data and for large visualization datasets in low dimensions Kohonen Layer Input Data
18 18 Algorithm Initialization Map is initialized with a specified topology and neighbor function Assignment & Relocation Presenting an input data to the map Updating the winner neuron and its neighbors based on the following update rule w ( t + 1) = w α( t):learning rate at time t h i ct i ( t) + α( t) h ( t)[ x( t) w ( t)] ( t) :neighbour kernelaround winner unit ci i c
19 19 Improved Competitive Learning Network (ICLN) Standard Competitive Learning Network (SCLN) Input layer Single layer of output neurons
20 20 ICLN Improved Competitive Learning Network (ICLN) What if this happen?
21 21 ICLN ICLN Algorithm η Weight update: β 2 : learning rate Euclidean Weight update: distance: : update factor d( xw, c= ) w= + η1( ( x i w c) 2 i 1 : learning rate e d η w = w η2β ( x w) 2 ) 0 2
22 Support Vector Machines SVM Powerful State-of-the-art classifier Supported by strong mathematical foundations Vapnik-Chervonenkis (VC) theory Good generalization to novel data Ability to classify non-linearly separable data 22
23 23 SVM : Conceptual Simplicity SVM model defines a hyperplane in the feature space in terms of coefficients (w) bias term (b) Prediction: ( w b) f = sign x + d - d +
24 Support Vector Machine : Multi- class 24
25 25 K-Nearest Neighbor (KNN) Describe instance x as feature vector Euclidean distance d ( x i < a x), a ( x), K, a ( ) 1( 2 n x Hamming distance, x j ) n r = distance between two values is 0 if they are the same, 1 if different. 1 ( a r ( x i ) > a r ( x j )) 2
26 26 The inductive bias of k-nnk Assumption that the classification of an instance x will be most similar to the classification of other instances that are nearby in Euclidian distance.
27 27 Voronoi diagram x q + + -
28 28 K-Nearest Neighbor (KNN) KNN-based outlier indices Kappa κ( x) = x z k z k is the kth nearest neighbor of x Gamma γ ( x) = 1 k k j= 1 x z k
29 29 Naive Bayes Classifier Probabilistic classifier based on Bayes Theorem Naive Bayes probabilistic model C = Class variable F 1,,F n = Features Is based on simplifying the assumption that the attribute values are conditionally independent given target value. Z = Scaling factor dependent on F 1,,F n
30 30 C4.5 (Decision Tree) A decision tree consists of nodes, leaves, and edges. A node of a decision tree specifies an attribute by which the data is to be partitioned. Each node bas a number of edges which are labeled according to a possible value of the attribute in the parent node. An edge connects either two nodes or a node and a leaf. Leaves are labeled with a decision value for categorization of the data.
31 31 Random Forest Classifier Generates many classification trees Each tree gives a vote that indicates the tree s decision about the class of the object. The forest chooses the class with the most votes for the object.
32 Dataset KDD99 Used in the third international Knowledge Discovery and Data Mining Tools Competition Acquired from a live network traffic by Lincoln LAB Attacks are simulated on top of background traffic Accepted as a standard dataset for evaluating IDS s Comes with both training (attack-free) and test dataset (contains both attack and normal data) 32
33 33 Dataset Simulation Network for collecting data
34 Simulated Attacks Category Probe DoS U2R R2L Description scan a network of computers to gather information or find known vulnerabilities Excessive consumption of resources that denies legitimate requests from system Successful execution of attacks results in normal user getting root privileges Attacker having no account gains a legal user account on the victim machine by sending packets over the network Example IPsweep, Saint Satan DDoS, Pingflood SYNflood Mailbomb Eject, Fdformat Loadmodule Perl Dictionary FTP-write Sendmail 34
35 KDD Features Feature Type Basic Content Time-based Description Common to all connections Based on information from local hosts Connections with respect to current one within a 2 second time window Example Duration of connection Service requested Bytes transferred Number of failed logins Number of root accesses Number of file creation operations # of connections that have SYN errors # of connections to the same service Connectionbased 35 Past 100 Connections with respect to current # of connections to the current host that have S0 error
36 Data Preparation Feature types (41 features in total) 38 continuous features 3 Discrete features (Protocol type, Flag, Service type) Converting discrete features to continuous features Using frequency instead of the initial values of discrete features 36
37 37 Data Preparation Necessity of normalization Features of different natures Large variance in maximum and minimum values Without normalization, large scales features dominate the low scale ones Normalization formula NewVali = normalize(ln( val X i Mini normalize( X i ) = Max Min wherei stands for i value of a i i + 1)) record on i th feature
38 38 Data Preparation Selecting datasets with different relative populations for train and test Normal-Attack Training Test
39 39 Labeling Heuristics for Clustering Techniques Why labeling heuristics are required? Practiced labeling heuristics Count-based : Label sparse clusters as anomalous (based on a threshold) Distance-based: Label the distant clusters as anomalous (based on a threshold) Inter-cluster Distance : ICD i = 1 C 1 C j= 1, i distance( c i, c j )
40 Performance Criteria Performance Criteria d Detection Rate = d+ c a b Normal False Positive Rate = a b + b c d Attack Actual Normal Attack Predicted 40
41 Different Experiments Per-Technique experiments Comparison between performance of count-based and distance-based labeling for each clustering technique Comparison between performance of each clustering technique in two different modes Direct application to test dataset Application of trained clusters to test dataset 41
42 42 Different Experiments Comparison between different techniques In direct application to test dataset In application of trained models to test dataset In detecting different attack categories (Probe, DoS, ) Average Cluster Purities (Clustering techniques only)
43 43 Count vs. Distance (EM, k = 50)
44 44 Count vs. Distance (ICLN, Init k = 50)
45 45 Count vs. Distance (K-Means, k = 50)
46 46 Count vs. Distance (SOM, k = 50)
47 47 Count vs. Distance (C-Means, k = 50)
48 48 Count vs. Distance (Y-Means, k = 50)
49 49 Count vs. Distance (Y-Means, Initial k = 50)
50 Different Experiments Per-Technique experiments Comparison between performance of count-based and distance-based labeling for each clustering technique Comparison between performance of each clustering technique in two different modes Direct application to test dataset Application of trained clusters to test dataset 50
51 51 Direct vs. Indirect (EM, 50, Test_9604)
52 52 Direct vs. Indirect (EM, 50, Test_8020)
53 53 Direct vs. Indirect (ICLN, 50, Test_8020)
54 54 Direct vs. Indirect (K-Means, 50, Test_9604)
55 55 Direct vs. Indirect (SOM, 50, Test_9604)
56 56 Direct vs. Indirect (SOM, 50, Test_8020)
57 57 Direct vs. Indirect (Y-Means, 50, Test_8020)
58 58 Direct vs. Indirect (C-Means, 50, Test_8020)
59 59 Different Experiments Comparison between different techniques In application of trained models to test dataset In direct application to test dataset In detecting different attack categories (Probe, DoS, ) Average Cluster Purities (Clustering techniques only)
60 60 Experimental Results (8020->8020)
61 61 Experimental Results (8020->9604)
62 62 Experimental Results (9604->8020)
63 63 Experimental Results (9604->9604)
64 64 Different Experiments Comparison between different techniques In application of trained models to test dataset In direct application to test dataset In detecting different attack categories (Probe, DoS, ) Average Cluster Purities (Clustering techniques only)
65 65 Direct application of techniques to Test_9604
66 66 Direct application of techniques to Test_8020
67 67 Different Experiments Comparison between different techniques In application of trained models to test dataset In direct application to test dataset In detecting different attack categories (Probe, DoS, ) Average Cluster Purities (Clustering techniques only)
68 68 Attack Category Detection (Train_8020, Test_8020)
69 69 Attack Category Detection (Train_8020, Test_9604)
70 70 Attack Category Detection (Train_9604, Test_8020)
71 71 Attack Category Detection (Train_9604, Test_9604)
72 72 Different Experiments Comparison between different techniques In application of trained models to test dataset In direct application to test dataset In detecting different attack categories (Probe, DoS, ) Average Cluster Purities (Clustering techniques only)
73 73 Experimental Results Cluster Purities Measurement : Information Entropy n H ( X ) = p( x )log2 i= 1 i p( x i ) Cluster impurity H ( C) = p( C p( C normal normal C ) = C normal )log p( C, p( C attack normal ) C ) = C p( C attack attack )log p( C attack )
74 74 Average Impurity of Clusters in Different Techniques Technique Impurity K-Means EM Y-Means SOM C-Means ICLN
75 Experimental Results (Supervised Schemes) Attack Detection Results FP DR : : : : : Gaussian Naïve Bayes C4.5 Random Forest SVM 75
76 Experimental Results (Supervised Schemes) 1.1 DoS Detection Results FP DR : : : : : Gaussian Naïve Bayes C4.5 Random Forest SVM 76
77 Experimental Results (Supervised Schemes) 1.1 Probe Detection Results FP DR 1: : : : : Gaussian Naïve Bayes C4.5 Random Forest SVM 77
78 Experimental Results (Supervised Schemes) 0.3 R2L Detection Results FP DR 0.2 1: : : : : Gaussian Naïve Bayes C4.5 Random Forest SVM 78
79 Experimental Results (Supervised Schemes) 0.7 U2R Datection Results FP DR : : : : : Gaussian Naïve Bayes C4.5 Random Forest SVM 79
80 Lessons Learned Distance-based labeling provides a more robust results to those of count-based labeling For most of the techniques, it is clearly dominant as well Direct application of clustering techniques performs as good as a two-step process Clustering techniques vs. other outlier detection schemes 80
81 81 Lessons Learned (Cont d) Most of the techniques are good at detecting probe and DoS attacks Almost all of them are poor at detecting R2L attacks Unsupervised SVM and Y-Means are good at detecting U2R attacks
82 82 Future Works Looking for more intelligent heuristics Combination of count-based and distance-based labeling Considering other criteria such as cluster density Looking for more discriminative features Of special value to detecting U2R and R2L attacks Comparison of other learning schemes Semi-supervised, Active Learning, Designing hybrid detectors based on the results of this study
83 83 Questions?
84 Approximate Auto Regressive Modeling For Network Attack Detection Harshit Nayyar and Ali A. Ghorbani
85 85 Scope NETWORK ATTACK DETECTION ANOMALY BASED STATISTICAL ANALYSIS STATISTICAL SIGNAL PROCESSING WAVELET FILTERING SYSTEM IDENTIFICATION SIGNAL APPROXIMATION ARX MODELING APPROXIMATE AUTOREGRESSIVE MODELING
86 86 Introduction Usual Methodology: Thresholds Network Dependent. Different threshold for different times? No scientific basis for determining the threshold. Basis of our technique: Assumption: Unusual is unexpected. Obtain Predictable Component from Network Data Create a predictive model of Network Create a model for high frequency components/peaks Flag large and/or persistent deviations from created model.
87 87 Network Data at a Glance
88 88 Techniques 1: Wavelet Approximations Wavelet Transform Haar Wavelet :-
89 89 Need & Effect of Wavelet Filtering
90 90 Techniques 2: ARX Model System ID: ARX Model Auto Regressive with external input. A linear difference equation relating previous outputs (AR) & External Input to future values. A(q)Y(t) = B(q)U(t) + Error. Predictive model ignores the error ARX[P,Q,R] P = Number of past outputs. (2) Q = Number of past inputs. (2) R = Time Delay in the System. (2T) System Identification deals with identifying: A(q) & B(q) given P,Q,R. Get most optimal A(q) and B(q) i.e. A(q) and B(q) which minimizes prediction Error.
91 91 Framework Phase1 : ARX Model Training Training Time-series Obtain ARX Model External Input
92 92 Phase1 Results: Model Training Phase1: ARX Model Training
93 93 Phase1 Results: Predictions
94 94 Framework Phase 2 Find Limits of Normal Peaks (windowed max) Obtain Peak Model
95 95 Results: Phase2
96 96 Phase3 Anomaly Detection
97 97 Phase 3 Results : Operation
98 98 Phase 3 Results : Operation
99 99 Phase 3 Results : Operation
100 Phase 3 Results : Operation 100
101 Phase 3 Results : Operation 101
102 Phase 3 Results : Operation 102
103 Table of Attacks 103
104 104 Conclusions Contributions: A technique for network data modeling which can automatically detect anomalies caused by network attacks. Technique is: Portable: Learning phase ensures portability across networks. Also, usable with other network signals. Effective: In detecting network anomalies caused by attacks. Unsupervised: Minimal Human Intervention Online: Detects attacks before completion.
105 105 Future Work Experiments with real network data. Test Performance in real network. Issues: Data Collection, Attack identification. Experiments with longer term data. Retraining (How and when) External Input Modification Correlation of Anomalies Improve identification of attack type. Allow higher level correlation rules. Improve Predictions: Nonlinear Models. Other wavelet basis.
106 106 Questions?
107 107 EM (Expectation EM (Expectation-Maximization) Maximization) Algorithm Initialization: Initialize the model parameters Expectation: Estimate the posterior probability of model k Maximization: re-estimate model parameters },, { k k k α σ μ Λ = = Λ j k n j k n k n x p x p x k P ) ( ) ( ), ( λ α λ α Λ Λ = n n n n n new k x k P x x k P ), ( ), ( μ Λ Λ = n n n k n n new k x k P x x k P d ), ( ), ( 1 2 μ σ Λ = n n new k x k P N ), ( 1 α Prior probability Mean Variance
108 One-Class SVM (Unsupervised( Unsupervised) Outlier detection Typical cases vs. outliers Tradeoff between including all examples and smallest sphere around the data Outliers are supposed to be excluded Outliers 108
109 109 Kernel Classifiers 1. Transform data via non-linear mapping to an inner product feature space Gaussian, polynomial and RBF kernels 2. Train a linear machine in the new feature space
110 110 C4.5 (Decision Tree) root = (null, All Rules,, ) root All nodes are represented by a tuple (C, R, F, L) C = condition (feature, operator, value) R = set of candidate detection rules F = feature set (already used to decompose tree) L = set of detection rules matched at that node C4.5: decision tree construction algorithm
Network Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationBased on Raymond J. Mooney s slides
Instance Based Learning Based on Raymond J. Mooney s slides University of Texas at Austin 1 Example 2 Instance-Based Learning Unlike other learning algorithms, does not involve construction of an explicit
More informationContents. Preface to the Second Edition
Preface to the Second Edition v 1 Introduction 1 1.1 What Is Data Mining?....................... 4 1.2 Motivating Challenges....................... 5 1.3 The Origins of Data Mining....................
More informationSVM in Oracle Database 10g: Removing the Barriers to Widespread Adoption of Support Vector Machines
SVM in Oracle Database 10g: Removing the Barriers to Widespread Adoption of Support Vector Machines Boriana Milenova, Joseph Yarmus, Marcos Campos Data Mining Technologies Oracle Overview Support Vector
More informationBioinformatics - Lecture 07
Bioinformatics - Lecture 07 Bioinformatics Clusters and networks Martin Saturka http://www.bioplexity.org/lectures/ EBI version 0.4 Creative Commons Attribution-Share Alike 2.5 License Learning on profiles
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,
More informationCS570: Introduction to Data Mining
CS570: Introduction to Data Mining Classification Advanced Reading: Chapter 8 & 9 Han, Chapters 4 & 5 Tan Anca Doloc-Mihu, Ph.D. Slides courtesy of Li Xiong, Ph.D., 2011 Han, Kamber & Pei. Data Mining.
More informationOverview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010
INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,
More informationINF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering
INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering Erik Velldal University of Oslo Sept. 18, 2012 Topics for today 2 Classification Recap Evaluating classifiers Accuracy, precision,
More informationUnsupervised Learning
Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised
More informationSupervised vs unsupervised clustering
Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful
More information10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors
Dejan Sarka Anomaly Detection Sponsors About me SQL Server MVP (17 years) and MCT (20 years) 25 years working with SQL Server Authoring 16 th book Authoring many courses, articles Agenda Introduction Simple
More informationData Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University
Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Descriptive model A descriptive model presents the main features of the data
More informationINF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering
INF4820 Algorithms for AI and NLP Evaluating Classifiers Clustering Murhaf Fares & Stephan Oepen Language Technology Group (LTG) September 27, 2017 Today 2 Recap Evaluation of classifiers Unsupervised
More informationSemi-Supervised Clustering with Partial Background Information
Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject
More informationData Mining in Bioinformatics Day 1: Classification
Data Mining in Bioinformatics Day 1: Classification Karsten Borgwardt February 18 to March 1, 2013 Machine Learning & Computational Biology Research Group Max Planck Institute Tübingen and Eberhard Karls
More informationClassification Algorithms in Data Mining
August 9th, 2016 Suhas Mallesh Yash Thakkar Ashok Choudhary CIS660 Data Mining and Big Data Processing -Dr. Sunnie S. Chung Classification Algorithms in Data Mining Deciding on the classification algorithms
More informationPerformance Evaluation of Various Classification Algorithms
Performance Evaluation of Various Classification Algorithms Shafali Deora Amritsar College of Engineering & Technology, Punjab Technical University -----------------------------------------------------------***----------------------------------------------------------
More informationClustering & Classification (chapter 15)
Clustering & Classification (chapter 5) Kai Goebel Bill Cheetham RPI/GE Global Research goebel@cs.rpi.edu cheetham@cs.rpi.edu Outline k-means Fuzzy c-means Mountain Clustering knn Fuzzy knn Hierarchical
More informationNaïve Bayes for text classification
Road Map Basic concepts Decision tree induction Evaluation of classifiers Rule induction Classification using association rules Naïve Bayesian classification Naïve Bayes for text classification Support
More informationSOCIAL MEDIA MINING. Data Mining Essentials
SOCIAL MEDIA MINING Data Mining Essentials Dear instructors/users of these slides: Please feel free to include these slides in your own material, or modify them as you see fit. If you decide to incorporate
More informationA Dendrogram. Bioinformatics (Lec 17)
A Dendrogram 3/15/05 1 Hierarchical Clustering [Johnson, SC, 1967] Given n points in R d, compute the distance between every pair of points While (not done) Pick closest pair of points s i and s j and
More informationClustering CS 550: Machine Learning
Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf
More informationSupervised Learning Classification Algorithms Comparison
Supervised Learning Classification Algorithms Comparison Aditya Singh Rathore B.Tech, J.K. Lakshmipat University -------------------------------------------------------------***---------------------------------------------------------
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationInternational Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X
Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,
More informationUnsupervised Learning
Unsupervised Learning Unsupervised learning Until now, we have assumed our training samples are labeled by their category membership. Methods that use labeled samples are said to be supervised. However,
More informationStatistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte
Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,
More informationA NEW HYBRID APPROACH FOR NETWORK TRAFFIC CLASSIFICATION USING SVM AND NAÏVE BAYES ALGORITHM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 6.017 IJCSMC,
More informationClustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford
Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically
More informationLecture on Modeling Tools for Clustering & Regression
Lecture on Modeling Tools for Clustering & Regression CS 590.21 Analysis and Modeling of Brain Networks Department of Computer Science University of Crete Data Clustering Overview Organizing data into
More informationDATA MINING LECTURE 10B. Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines
DATA MINING LECTURE 10B Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines NEAREST NEIGHBOR CLASSIFICATION 10 10 Illustrating Classification Task Tid Attrib1
More informationINF 4300 Classification III Anne Solberg The agenda today:
INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15
More informationCOMP 551 Applied Machine Learning Lecture 13: Unsupervised learning
COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning Associate Instructor: Herke van Hoof (herke.vanhoof@mail.mcgill.ca) Slides mostly by: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationINF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22
INF4820 Clustering Erik Velldal University of Oslo Nov. 17, 2009 Erik Velldal INF4820 1 / 22 Topics for Today More on unsupervised machine learning for data-driven categorization: clustering. The task
More informationA Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis
A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis Meshal Shutaywi and Nezamoddin N. Kachouie Department of Mathematical Sciences, Florida Institute of Technology Abstract
More informationContent-based image and video analysis. Machine learning
Content-based image and video analysis Machine learning for multimedia retrieval 04.05.2009 What is machine learning? Some problems are very hard to solve by writing a computer program by hand Almost all
More informationINF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering
INF4820 Algorithms for AI and NLP Evaluating Classifiers Clustering Erik Velldal & Stephan Oepen Language Technology Group (LTG) September 23, 2015 Agenda Last week Supervised vs unsupervised learning.
More informationMulti-label classification using rule-based classifier systems
Multi-label classification using rule-based classifier systems Shabnam Nazmi (PhD candidate) Department of electrical and computer engineering North Carolina A&T state university Advisor: Dr. A. Homaifar
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationChapter 7: Competitive learning, clustering, and self-organizing maps
Chapter 7: Competitive learning, clustering, and self-organizing maps António R. C. Paiva EEL 6814 Spring 2008 Outline Competitive learning Clustering Self-Organizing Maps What is competition in neural
More informationSemi-supervised Learning
Semi-supervised Learning Piyush Rai CS5350/6350: Machine Learning November 8, 2011 Semi-supervised Learning Supervised Learning models require labeled data Learning a reliable model usually requires plenty
More informationPreface to the Second Edition. Preface to the First Edition. 1 Introduction 1
Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches
More informationCMPUT 391 Database Management Systems. Data Mining. Textbook: Chapter (without 17.10)
CMPUT 391 Database Management Systems Data Mining Textbook: Chapter 17.7-17.11 (without 17.10) University of Alberta 1 Overview Motivation KDD and Data Mining Association Rules Clustering Classification
More informationClassification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University
Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate
More informationExploratory Data Analysis using Self-Organizing Maps. Madhumanti Ray
Exploratory Data Analysis using Self-Organizing Maps Madhumanti Ray Content Introduction Data Analysis methods Self-Organizing Maps Conclusion Visualization of high-dimensional data items Exploratory data
More informationReview on Data Mining Techniques for Intrusion Detection System
Review on Data Mining Techniques for Intrusion Detection System Sandeep D 1, M. S. Chaudhari 2 Research Scholar, Dept. of Computer Science, P.B.C.E, Nagpur, India 1 HoD, Dept. of Computer Science, P.B.C.E,
More informationLecture 9: Support Vector Machines
Lecture 9: Support Vector Machines William Webber (william@williamwebber.com) COMP90042, 2014, Semester 1, Lecture 8 What we ll learn in this lecture Support Vector Machines (SVMs) a highly robust and
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationUnsupervised Learning : Clustering
Unsupervised Learning : Clustering Things to be Addressed Traditional Learning Models. Cluster Analysis K-means Clustering Algorithm Drawbacks of traditional clustering algorithms. Clustering as a complex
More informationPattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition
Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant
More informationCSE4334/5334 DATA MINING
CSE4334/5334 DATA MINING Lecture 4: Classification (1) CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai Li (Slides courtesy
More informationAssociative Cellular Learning Automata and its Applications
Associative Cellular Learning Automata and its Applications Meysam Ahangaran and Nasrin Taghizadeh and Hamid Beigy Department of Computer Engineering, Sharif University of Technology, Tehran, Iran ahangaran@iust.ac.ir,
More informationSupervised vs.unsupervised Learning
Supervised vs.unsupervised Learning In supervised learning we train algorithms with predefined concepts and functions based on labeled data D = { ( x, y ) x X, y {yes,no}. In unsupervised learning we are
More informationOutlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data
Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data Ms. Gayatri Attarde 1, Prof. Aarti Deshpande 2 M. E Student, Department of Computer Engineering, GHRCCEM, University
More informationMethods for Intelligent Systems
Methods for Intelligent Systems Lecture Notes on Clustering (II) Davide Eynard eynard@elet.polimi.it Department of Electronics and Information Politecnico di Milano Davide Eynard - Lecture Notes on Clustering
More informationA REVIEW ON VARIOUS APPROACHES OF CLUSTERING IN DATA MINING
A REVIEW ON VARIOUS APPROACHES OF CLUSTERING IN DATA MINING Abhinav Kathuria Email - abhinav.kathuria90@gmail.com Abstract: Data mining is the process of the extraction of the hidden pattern from the data
More informationCLUSTERING. CSE 634 Data Mining Prof. Anita Wasilewska TEAM 16
CLUSTERING CSE 634 Data Mining Prof. Anita Wasilewska TEAM 16 1. K-medoids: REFERENCES https://www.coursera.org/learn/cluster-analysis/lecture/nj0sb/3-4-the-k-medoids-clustering-method https://anuradhasrinivas.files.wordpress.com/2013/04/lesson8-clustering.pdf
More informationTour-Based Mode Choice Modeling: Using An Ensemble of (Un-) Conditional Data-Mining Classifiers
Tour-Based Mode Choice Modeling: Using An Ensemble of (Un-) Conditional Data-Mining Classifiers James P. Biagioni Piotr M. Szczurek Peter C. Nelson, Ph.D. Abolfazl Mohammadian, Ph.D. Agenda Background
More information6. Learning Partitions of a Set
6. Learning Partitions of a Set Also known as clustering! Usually, we partition sets into subsets with elements that are somewhat similar (and since similarity is often task dependent, different partitions
More informationCIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]
CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.
More informationClustering. Partition unlabeled examples into disjoint subsets of clusters, such that:
Text Clustering 1 Clustering Partition unlabeled examples into disjoint subsets of clusters, such that: Examples within a cluster are very similar Examples in different clusters are very different Discover
More informationMachine Learning: Think Big and Parallel
Day 1 Inderjit S. Dhillon Dept of Computer Science UT Austin CS395T: Topics in Multicore Programming Oct 1, 2013 Outline Scikit-learn: Machine Learning in Python Supervised Learning day1 Regression: Least
More informationRobot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning
Robot Learning 1 General Pipeline 1. Data acquisition (e.g., from 3D sensors) 2. Feature extraction and representation construction 3. Robot learning: e.g., classification (recognition) or clustering (knowledge
More informationECG782: Multidimensional Digital Signal Processing
ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationIBL and clustering. Relationship of IBL with CBR
IBL and clustering Distance based methods IBL and knn Clustering Distance based and hierarchical Probability-based Expectation Maximization (EM) Relationship of IBL with CBR + uses previously processed
More informationMulti-label Classification. Jingzhou Liu Dec
Multi-label Classification Jingzhou Liu Dec. 6 2016 Introduction Multi-class problem, Training data (x $, y $ ) ( ), x $ X R., y $ Y = 1,2,, L Learn a mapping f: X Y Each instance x $ is associated with
More informationMachine Learning Classifiers and Boosting
Machine Learning Classifiers and Boosting Reading Ch 18.6-18.12, 20.1-20.3.2 Outline Different types of learning problems Different types of learning algorithms Supervised learning Decision trees Naïve
More informationClustering. Chapter 10 in Introduction to statistical learning
Clustering Chapter 10 in Introduction to statistical learning 16 14 12 10 8 6 4 2 0 2 4 6 8 10 12 14 1 Clustering ² Clustering is the art of finding groups in data (Kaufman and Rousseeuw, 1990). ² What
More informationUnsupervised Learning
Outline Unsupervised Learning Basic concepts K-means algorithm Representation of clusters Hierarchical clustering Distance functions Which clustering algorithm to use? NN Supervised learning vs. unsupervised
More informationEnhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques
24 Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Ruxandra PETRE
More informationGene Clustering & Classification
BINF, Introduction to Computational Biology Gene Clustering & Classification Young-Rae Cho Associate Professor Department of Computer Science Baylor University Overview Introduction to Gene Clustering
More informationSUPERVISED LEARNING METHODS. Stanley Liang, PhD Candidate, Lassonde School of Engineering, York University Helix Science Engagement Programs 2018
SUPERVISED LEARNING METHODS Stanley Liang, PhD Candidate, Lassonde School of Engineering, York University Helix Science Engagement Programs 2018 2 CHOICE OF ML You cannot know which algorithm will work
More informationArtificial Intelligence. Programming Styles
Artificial Intelligence Intro to Machine Learning Programming Styles Standard CS: Explicitly program computer to do something Early AI: Derive a problem description (state) and use general algorithms to
More informationClassification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University
Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate
More informationCHAPTER V KDD CUP 99 DATASET. With the widespread use of computer networks, the number of attacks has grown
CHAPTER V KDD CUP 99 DATASET With the widespread use of computer networks, the number of attacks has grown extensively, and many new hacking tools and intrusive methods have appeared. Using an intrusion
More informationPython With Data Science
Course Overview This course covers theoretical and technical aspects of using Python in Applied Data Science projects and Data Logistics use cases. Who Should Attend Data Scientists, Software Developers,
More informationMachine Learning. Chao Lan
Machine Learning Chao Lan Machine Learning Prediction Models Regression Model - linear regression (least square, ridge regression, Lasso) Classification Model - naive Bayes, logistic regression, Gaussian
More information"GET /cgi-bin/purchase?itemid=109agfe111;ypcat%20passwd mail 200
128.111.41.15 "GET /cgi-bin/purchase? itemid=1a6f62e612&cc=mastercard" 200 128.111.43.24 "GET /cgi-bin/purchase?itemid=61d2b836c0&cc=visa" 200 128.111.48.69 "GET /cgi-bin/purchase? itemid=a625f27110&cc=mastercard"
More informationCSE 573: Artificial Intelligence Autumn 2010
CSE 573: Artificial Intelligence Autumn 2010 Lecture 16: Machine Learning Topics 12/7/2010 Luke Zettlemoyer Most slides over the course adapted from Dan Klein. 1 Announcements Syllabus revised Machine
More informationR (2) Data analysis case study using R for readily available data set using any one machine learning algorithm.
Assignment No. 4 Title: SD Module- Data Science with R Program R (2) C (4) V (2) T (2) Total (10) Dated Sign Data analysis case study using R for readily available data set using any one machine learning
More informationAdministrative. Machine learning code. Supervised learning (e.g. classification) Machine learning: Unsupervised learning" BANANAS APPLES
Administrative Machine learning: Unsupervised learning" Assignment 5 out soon David Kauchak cs311 Spring 2013 adapted from: http://www.stanford.edu/class/cs276/handouts/lecture17-clustering.ppt Machine
More informationData Mining and Analytics
Data Mining and Analytics Aik Choon Tan, Ph.D. Associate Professor of Bioinformatics Division of Medical Oncology Department of Medicine aikchoon.tan@ucdenver.edu 9/22/2017 http://tanlab.ucdenver.edu/labhomepage/teaching/bsbt6111/
More informationClassifiers and Detection. D.A. Forsyth
Classifiers and Detection D.A. Forsyth Classifiers Take a measurement x, predict a bit (yes/no; 1/-1; 1/0; etc) Detection with a classifier Search all windows at relevant scales Prepare features Classify
More informationArtificial Neural Networks MLP, RBF & GMDH
Artificial Neural Networks MLP, RBF & GMDH Jan Drchal drchajan@fel.cvut.cz Computational Intelligence Group Department of Computer Science and Engineering Faculty of Electrical Engineering Czech Technical
More informationComparison of different preprocessing techniques and feature selection algorithms in cancer datasets
Comparison of different preprocessing techniques and feature selection algorithms in cancer datasets Konstantinos Sechidis School of Computer Science University of Manchester sechidik@cs.man.ac.uk Abstract
More informationUnsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi
Unsupervised Learning Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi Content Motivation Introduction Applications Types of clustering Clustering criterion functions Distance functions Normalization Which
More informationFunction approximation using RBF network. 10 basis functions and 25 data points.
1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data
More informationAnomaly Intrusion Detection System Using Hierarchical Gaussian Mixture Model
264 IJCSNS International Journal of Computer Science and Network Security, VOL.8 No.8, August 2008 Anomaly Intrusion Detection System Using Hierarchical Gaussian Mixture Model M. Bahrololum and M. Khaleghi
More informationThe k-means Algorithm and Genetic Algorithm
The k-means Algorithm and Genetic Algorithm k-means algorithm Genetic algorithm Rough set approach Fuzzy set approaches Chapter 8 2 The K-Means Algorithm The K-Means algorithm is a simple yet effective
More informationClassification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska
Classification Lecture Notes cse352 Neural Networks Professor Anita Wasilewska Neural Networks Classification Introduction INPUT: classification data, i.e. it contains an classification (class) attribute
More informationUnsupervised Learning
Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover
More informationCLUSTERING. JELENA JOVANOVIĆ Web:
CLUSTERING JELENA JOVANOVIĆ Email: jeljov@gmail.com Web: http://jelenajovanovic.net OUTLINE What is clustering? Application domains K-Means clustering Understanding it through an example The K-Means algorithm
More informationCS6375: Machine Learning Gautam Kunapuli. Mid-Term Review
Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes
More informationHierarchical Adaptive FCM To Detect Attacks Using Layered Approach
Hierarchical Adaptive FCM To Detect Attacks Using Layered Approach J.Jensi Edith 1, Dr. A.Chandrasekar 1.Research Scholar,Sathyabama University, Chennai.. Prof, CSE DEPT, St.Joseph s College of Engg.,
More informationCHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES
CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES 6.1 INTRODUCTION The exploration of applications of ANN for image classification has yielded satisfactory results. But, the scope for improving
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 189 Fall 2015 Introduction to Machine Learning Final Please do not turn over the page before you are instructed to do so. You have 2 hours and 50 minutes. Please write your initials on the top-right
More informationPart I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a
Week 9 Based in part on slides from textbook, slides of Susan Holmes Part I December 2, 2012 Hierarchical Clustering 1 / 1 Produces a set of nested clusters organized as a Hierarchical hierarchical clustering
More information