Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 6

Similar documents
Chapter 6: Basic Concepts: Association Rules. Basic Concepts: Frequent Patterns. (absolute) support, or, support. (relative) support, s, is the

Chapter 4: Mining Frequent Patterns, Associations and Correlations

CSE 5243 INTRO. TO DATA MINING

CS570 Introduction to Data Mining

Basic Concepts: Association Rules. What Is Frequent Pattern Analysis? COMP 465: Data Mining Mining Frequent Patterns, Associations and Correlations

Data Mining: Concepts and Techniques. Chapter 5. SS Chung. April 5, 2013 Data Mining: Concepts and Techniques 1

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 6

CSE 5243 INTRO. TO DATA MINING

Lecture Topic Projects 1 Intro, schedule, and logistics 2 Data Science components and tasks 3 Data types Project #1 out 4 Introduction to R,

CSE 5243 INTRO. TO DATA MINING

Frequent Pattern Mining

BCB 713 Module Spring 2011

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 6

Association rules. Marco Saerens (UCL), with Christine Decaestecker (ULB)

Association Rule Mining. Entscheidungsunterstützungssysteme

Scalable Frequent Itemset Mining Methods

Roadmap. PCY Algorithm

Nesnelerin İnternetinde Veri Analizi

CS145: INTRODUCTION TO DATA MINING

Performance and Scalability: Apriori Implementa6on

D Data Mining: Concepts and and Tech Techniques

CS6220: DATA MINING TECHNIQUES

Fundamental Data Mining Algorithms

Information Management course

Apriori Algorithm. 1 Bread, Milk 2 Bread, Diaper, Beer, Eggs 3 Milk, Diaper, Beer, Coke 4 Bread, Milk, Diaper, Beer 5 Bread, Milk, Diaper, Coke

Product presentations can be more intelligently planned

Frequent Pattern Mining S L I D E S B Y : S H R E E J A S W A L

Decision Support Systems

Information Management course

Frequent Pattern Mining

Advance Association Analysis

ANU MLSS 2010: Data Mining. Part 2: Association rule mining

Chapter 7: Frequent Itemsets and Association Rules

Improved Frequent Pattern Mining Algorithm with Indexing

Frequent Pattern Mining. Based on: Introduction to Data Mining by Tan, Steinbach, Kumar

Mining Association Rules in Large Databases

Association Rules. A. Bellaachia Page: 1

Data Mining Part 3. Associations Rules

Effectiveness of Freq Pat Mining

Association Rules Apriori Algorithm

To Enhance Projection Scalability of Item Transactions by Parallel and Partition Projection using Dynamic Data Set

Knowledge Discovery in Databases

Model for Load Balancing on Processors in Parallel Mining of Frequent Itemsets

DATA MINING II - 1DL460

What Is Data Mining? CMPT 354: Database I -- Data Mining 2

A Taxonomy of Classical Frequent Item set Mining Algorithms

Mining Frequent Patterns without Candidate Generation

Chapter 4: Association analysis:

An Evolutionary Algorithm for Mining Association Rules Using Boolean Approach

CompSci 516 Data Intensive Computing Systems

Mining Frequent Patterns, Associations, and Correlations: Basic Concepts and Methods

Data Mining: Concepts and Techniques. Chapter 5

Mining High Average-Utility Itemsets

Association mining rules

Market baskets Frequent itemsets FP growth. Data mining. Frequent itemset Association&decision rule mining. University of Szeged.

Data Mining Techniques

Association Pattern Mining. Lijun Zhang

AN ENHANCED SEMI-APRIORI ALGORITHM FOR MINING ASSOCIATION RULES

A Literature Review of Modern Association Rule Mining Techniques

signicantly higher than it would be if items were placed at random into baskets. For example, we

An Algorithm for Frequent Pattern Mining Based On Apriori

Association Rule Mining (ARM) Komate AMPHAWAN

Association Rules. Juliana Freire. Modified from Jeff Ullman, Jure Lescovek, Bing Liu, Jiawei Han

Data Structures. Notes for Lecture 14 Techniques of Data Mining By Samaher Hussein Ali Association Rules: Basic Concepts and Application

Mining of Web Server Logs using Extended Apriori Algorithm

Optimized Frequent Pattern Mining for Classified Data Sets

Pattern Mining. Knowledge Discovery and Data Mining 1. Roman Kern KTI, TU Graz. Roman Kern (KTI, TU Graz) Pattern Mining / 42

Discovering interesting rules from financial data

Jarek Szlichta Acknowledgments: Jiawei Han, Micheline Kamber and Jian Pei, Data Mining - Concepts and Techniques

Data Mining: Concepts and Techniques. Mining Frequent Patterns, Associations, and Correlations. Chapter 5.4 & 5.6

Data Mining: Concepts and Techniques. Chapter 5

AC-Close: Efficiently Mining Approximate Closed Itemsets by Core Pattern Recovery

An Improved Apriori Algorithm for Association Rules

A Decremental Algorithm for Maintaining Frequent Itemsets in Dynamic Databases *

Association Rules: Past, Present & Future. Ramakrishnan Srikant.

DMSA TECHNIQUE FOR FINDING SIGNIFICANT PATTERNS IN LARGE DATABASE

Parallel Algorithms for Discovery of Association Rules

Tutorial on Assignment 3 in Data Mining 2009 Frequent Itemset and Association Rule Mining. Gyozo Gidofalvi Uppsala Database Laboratory

Data Mining: Concepts and Techniques. Chapter Mining sequence patterns in transactional databases

Data Mining: Concepts and Techniques

Mining Frequent Patterns Based on Data Characteristics

Data Mining: Concepts and Techniques. Chapter 5 Mining Frequent Patterns. Slide credits: Jiawei Han and Micheline Kamber George Kollios

CompSci 516 Data Intensive Computing Systems

Discovery of Association Rules in Temporal Databases 1

H-Mine: Hyper-Structure Mining of Frequent Patterns in Large Databases. Paper s goals. H-mine characteristics. Why a new algorithm?

CHAPTER 8. ITEMSET MINING 226

CLOSET+:Searching for the Best Strategies for Mining Frequent Closed Itemsets

Chapter 4 Data Mining A Short Introduction

PLT- Positional Lexicographic Tree: A New Structure for Mining Frequent Itemsets

CMPUT 391 Database Management Systems. Data Mining. Textbook: Chapter (without 17.10)

Data Mining for Knowledge Management. Association Rules

A Data Mining Framework for Extracting Product Sales Patterns in Retail Store Transactions Using Association Rules: A Case Study

Incremental Frequent Pattern Mining. Abstract

Mining Frequent Patterns without Candidate Generation: A Frequent-Pattern Tree Approach

Parallel Mining Association Rules in Calculation Grids

Parallel Mining of Maximal Frequent Itemsets in PC Clusters

DATA MINING II - 1DL460

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Chapter 7: Frequent Itemsets and Association Rules

DATA MINING II - 1DL460

Transcription:

Data Mining: Concepts and Techniques (3 rd ed.) Chapter 6 Jiawei Han, Micheline Kamber, and Jian Pei University of Illinois at Urbana-Champaign & Simon Fraser University 2013-2017 Han, Kamber & Pei. All rights reserved. 1 Chapter 6: Mining Frequent Patterns, Association and Correlations: Basic Concepts and Methods Basic Concepts Frequent Itemset Mining Methods Which Patterns Are Interesting? Pattern Evaluation Methods (next week) Summary 2 1

What Is Pattern Discovery? What are patterns? A set of items, subsequences, or substructures, that occur frequently together (or strongly correlated) in a data set Patterns represent intrinsic and important properties of datasets Pattern discovery Uncovering patterns from massive data sets Motivation: Finding inherent regularities in data What products were often purchased together? Beer and diapers?! What are the subsequent purchases after buying a notebook? What kinds of DNA are sensitive to this new drug? Can we automatically classify web documents? 3 Beer and Diapers A 25-year-old legend: beer and diaper sales spike between the hours of 5 p.m. and 7 p.m. Original NCR (now TeraData) study in 1992 by Thomas Blischok (MindMeld Inc.) for American retail chain Osco Drugs. examined 1.2 million market baskets in 25 stores NCR identified 30 different shopping experiences, such as a correlation between fruit juice and cough medication sales. Osco removed approximately 5,000 slow-moving items from its inventory by re-arranging merchandise, consumers actually thought that Osco s selection had increased. putting the right merchandise in the right quantities at the right time November 9, 2017 Data Mining: Concepts and Techniques 4 2

Why Is Pattern Discovery Important? Finding inherent regularities in a data set Foundation for many essential data mining tasks Association, correlation, and causality analysis Mining sequential, structural (e.g., sub-graph) patterns Pattern analysis in spatiotemporal, multimedia, time-series, and stream data Classification: discriminative pattern-based analysis Cluster analysis: pattern-subspace clustering Many Applications Basket data analysis, cross-marketing, catalog design, sale campaign analysis, Web log analysis, and biological sequence analysis 5 Basic Concepts: Frequent Patterns Tid Items bought 10 Beer, Nuts, Diaper 20 Beer, Coffee, Diaper 30 Beer, Diaper, Eggs 40 Nuts, Eggs, Milk 50 Nuts, Coffee, Diaper, Eggs, Milk Let minsup = 50% Freq. 1-itemsets: Beer: 3 (60%); Nuts: 3 (60%) Diaper: 4 (80%); Eggs: 3 (60%) Freq. 2-itemsets: {Beer, Diaper}: 3 (60%) itemset: A set of one or more items k-itemset X = {x 1,, x k } (absolute) support (count) of X: Frequency or the number of occurrences of an itemset X (relative) support, s: The fraction of transactions that contains X (i.e., the probability that a transaction contains X) An itemset X is frequent if the support of X is no less than a minsup threshold (σ) 6 3

Tid From Frequent Itemsets to Association Rules Items bought 10 Beer, Nuts, Diaper 20 Beer, Coffee, Diaper 30 Beer, Diaper, Eggs 40 Nuts, Eggs, Milk 50 Nuts, Coffee, Diaper, Eggs, Milk Beer Transactions Containing both {Beer} {Diaper} Transactions Containing beer Diaper Transactions Containing Diaper {Beer} {Diaper} = {Beer, Diaper} Note: Itemset: X Y, a subtle notation! Association rules: X Y (s, c) Support, s: The probability that a transaction contains X Y Confidence, c: The conditional probability that a transaction containing X also contains Y c = sup(x Y) / sup(x) Association rule mining: Find all of the rules, X Y, with minimum support and confidence Frequent itemsets: Let minsup = 50% Freq. 1-itemsets: Beer: 3, Nuts: 3, Diaper: 4, Eggs: 3 Freq. 2-itemsets: {Beer, Diaper}: 3 Association rules: Let minconf = 50% Beer Diaper (60%, 100%) Diaper Beer (60%, 75%) 7 Challenge: There Are Too Many Frequent Patterns! A long pattern contains a combinatorial number of sub-patterns How many frequent itemsets does the following TDB 1 contain? TDB 1: T 1 : {a 1,, a 50 }; T 2 : {a 1,, a 100 } Assuming (absolute) minsup = 1 Let s have a try 1-itemsets: {a 1 }: 2, {a 2 }: 2,, {a 50 }: 2, {a 51 }: 1,, {a 100 }: 1, 2-itemsets: {a 1, a 2 }: 2,, {a 1, a 50 }: 2, {a 1, a 51 }: 1,, {a 99, a 100 }: 1,,,, 99-itemsets: {a 1, a 2,, a 99 }: 1,, {a 2, a 3,, a 100 }: 1 100-itemset: {a 1, a 2,, a 100 }: 1 In total: ( 100 1) + ( 100 2) + + ( 1 1 0 0 00 ) = 2 100 1 sub-patterns! 8 4

Expressing Patterns in Compressed Form: Closed Patterns How to handle such a challenge? Solution 1: Closed patterns: A pattern (itemset) X is closed if X is frequent, and there exists no super-pattern Y כ X, with the same support as X Let Transaction DB TDB 1 : T 1 : {a 1,, a 50 }; T 2 : {a 1,, a 100 } Suppose minsup = 1. How many closed patterns does TDB 1 contain? Two: P 1 : {a 1,, a 50 }: 2 ; P 2 : {a 1,, a 100 }: 1 Closed pattern is a lossless compression of frequent patterns Reduces the # of patterns but does not lose the support information! You will still be able to say: {a 2,, a 40 }: 2, {a 5, a 51 }: 1 November 9, 2017 Data Mining: Concepts and Techniques 9 Expressing Patterns in Compressed Form: Max-Patterns Solution 2: Max-patterns: A pattern X is a max-pattern if X is frequent and there exists no frequent super-pattern Y כ X Difference from close-patterns? Do not capture the real support of the sub-patterns of a max-pattern Let Transaction DB TDB 1 : T 1 : {a 1,, a 50 }; T 2 : {a 1,, a 100 } Suppose minsup = 1. How many max-patterns does TDB 1 contain? One: P: {a 1,, a 100 }: 1 Max-pattern is a lossy compression! We only know {a 1,, a 40 } is frequent But we do not know the real support of {a 1,, a 40 },, any more! Thus in many applications, mining close-patterns is more desirable than mining max-patterns 10 5

Chapter 6: Mining Frequent Patterns, Association and Correlations: Basic Concepts and Methods Basic Concepts Frequent Itemset Mining Methods Which Patterns Are Interesting? Pattern Evaluation Methods (next week) Summary 11 Scalable Frequent Itemset Mining Methods The Downward Closure Property of Frequent Patterns The Apriori Algorithm Extensions or Improvements of Apriori Mining Frequent Patterns by Exploring Vertical Data Format FPGrowth: A Frequent Pattern-Growth Approach Mining Closed Patterns 12 6

The Downward Closure Property of Frequent Patterns Observation: From TDB 1: T 1 : {a 1,, a 50 }; T 2 : {a 1,, a 100 } We get a frequent itemset: {a 1,, a 50 } Also, its subsets are all frequent: {a 1 }, {a 2 },, {a 50 }, {a 1, a 2 },, {a 1,, a 49 }, There must be some hidden relationships among frequent patterns! The downward closure (also called Apriori ) property of frequent patterns If {beer, diaper, nuts} is frequent, so is {beer, diaper} Every transaction containing {beer, diaper, nuts} also contains {beer, diaper} Apriori: Any subset of a frequent itemset must be frequent Efficient mining methodology If any subset of an itemset S is infrequent, then there is no chance for S to be frequent why do we even have to consider S!? A sharp knife for pruning! Apriori Pruning and Scalable Mining Methods Apriori pruning principle: If there is any itemset which is infrequent, its superset should not even be generated! (Agrawal & Srikant @VLDB 94, Mannila, et al. @ KDD 94) Scalable mining Methods: Three major approaches Level-wise, join-based approach: Apriori (Agrawal & Srikant@VLDB 94) Vertical data format approach: Eclat (Zaki, Parthasarathy, Ogihara, Li @KDD 97) Frequent pattern projection and growth: FPgrowth (Han, Pei, Yin @SIGMOD 00) 7

Apriori: A Candidate Generation & Test Approach Apriori pruning principle: If there is any itemset which is infrequent, its superset should not be generated/tested! (Agrawal & Srikant @VLDB 94, Mannila, et al. @ KDD 94) Outline of Apriori (level-wise, candidate generation and test) Initially, scan DB once to get frequent 1-itemset Repeat Generate length-(k+1) candidate itemsets from length-k frequent itemsets Test the candidates against DB to find frequent (k+1)-itemsets Set k := k +1 Until no frequent or candidate set can be generated Return all the frequent itemsets derived 15 Database TDB The Apriori Algorithm An Example minsup = 2 Tid Items 10 A, C, D 20 B, C, E 30 A, B, C, E 40 B, E 1 st scan Itemset L C 1 1 {B} 3 C 2 C 2 {A, B} 1 L 2 Itemset sup 2 nd scan {A, C} 2 {B, C} 2 {B, E} 3 {C, E} 2 C Itemset 3 3 rd scan L 3 {B, C, E} Itemset sup {A} 2 {C} 3 {D} 1 {E} 3 sup {A, C} 2 {A, E} 1 {B, C} 2 {B, E} 3 {C, E} 2 Itemset sup {B, C, E} 2 Itemset sup {A} 2 {B} 3 {C} 3 {E} 3 Itemset {A, B} {A, C} {A, E} {B, C} {B, E} {C, E} 16 8

The Apriori Algorithm (Pseudo-Code) C k : Candidate itemset of size k F k : frequent itemset of size k k := 1; F 1 = {frequent items}; while ( F k!= ) do C k+1 = candidates generated from F k ; for each transaction t in database do increment the count of all candidates in C k+1 that are contained in t ; F k+1 = candidates in C k+1 with min_support k := k + 1; od return k F k ; 17 Implementation of Apriori How to generate candidates? Step 1: self-joining F k Step 2: pruning Example of Candidate-generation F 3 ={abc, abd, acd, ace, bcd} Self-joining: F 3 *F 3 abcd from abc and abd acde from acd and ace Pruning: acde is removed because ade is not in F 3 C 4 = {abcd} 18 9

How to Count Supports of Candidates? Why is counting supports of candidates a problem? The total number of candidates can be very huge One transaction may contain many candidates Method: Candidate itemsets are stored in a hash-tree Leaf node of hash-tree contains a list of itemsets and counts Interior node contains a hash table Subset function: finds all the candidates contained in a transaction 19 Counting Supports of Candidates Using Hash Tree Items: 1, 2, 3, 4, 5, 6, 7, 8, 9 Subset function 3,6,9 1,4,7 2,5,8 Transaction: 1 2 3 5 6 1 + 2 3 5 6 1 3 + 5 6 1 2 + 3 5 6 1 4 5 1 2 4 4 5 7 1 2 5 4 5 8 2 3 4 5 6 7 1 3 6 1 5 9 3 4 5 3 5 6 3 5 7 6 8 9 3 6 7 3 6 8 Leafs: Candidate itemsets 20 10

Candidate Generation: An SQL Implementation SQL Implementation of candidate generation Suppose the items in F k-1 are listed in an order Step 1: self-joining F k-1 insert into C k select p.item 1, p.item 2,, p.item k-1, q.item k-1 from F k-1 p, F k-1 q where p.item 1 =q.item 1,, p.item k-2 =q.item k-2, p.item k-1 < q.item k-1 Step 2: pruning forall itemsets c in C k do forall (k-1)-subsets s of c do if (s is not in F k-1 ) then delete c from C k Use object-relational extensions like UDFs, BLOBs, and Table functions for efficient implementation [S. Sarawagi, S. Thomas, and R. Agrawal. Integrating association rule mining with relational database systems: Alternatives and implications. SIGMOD 98] 21 Scalable Frequent Itemset Mining Methods The Downward Closure Property of Frequent Patterns The Apriori Algorithm Extensions or Improvements of Apriori Mining Frequent Patterns by Exploring Vertical Data Format FPGrowth: A Frequent Pattern-Growth Approach Mining Closed Patterns 22 11

Further Improvement of the Apriori Method Major computational challenges Multiple scans of transaction database Huge number of candidates Tedious workload of support counting for candidates 23 Apriori: Improvements and Alternatives Reduce passes of transaction database scans Partitioning (e.g., Savasere, et al., 1995) Dynamic itemset counting (Brin, et al., 1997) Shrink the number of candidates Hashing (e.g., DHP: Park, et al., 1995) Pruning by support lower bounding (e.g., Bayardo 1998) Sampling (e.g., Toivonen, 1996) Exploring special data structures Tree projection (Agarwal, et al., 2001) H-miner (Pei, et al., 2001) To be discussed in subsequent slides To be discussed in subsequent slides Hypercube decomposition (e.g., LCM: Uno, et al., 2004) 12

Partition: Scan Database Only Twice Any itemset that is potentially frequent in DB must be frequent in at least one of the partitions of DB Scan 1: partition database and find local frequent patterns Scan 2: consolidate global frequent patterns A. Savasere, E. Omiecinski and S. Navathe, VLDB 95 DB 1 + DB 2 + + DB k = DB sup 1 (X) < σdb 1 sup 2 (X) < σdb 2 sup k (X) < σdb k sup(x) < σdb DHP: Reduce the Number of Candidates A k-itemset whose corresponding hashing bucket count is below the support threshold cannot be frequent Candidates: a, b, c, d, e Itemsets Count {ab, ad, ae} 35 Hash entries {bd, be, de} 298 {ab, ad, ae} {bd, be, de} {yz, qs, wt} 58 Frequent 1-itemset: a, b, d, e Hash Table ab is not a candidate 2-itemset if the sum of count of {ab, ad, ae} is below the support threshold J. Park, M. Chen, and P. Yu. An effective hash-based algorithm for mining association rules. SIGMOD 95 (Direct Hashing and Pruning (DHP)) 26 13

Exploring Vertical Data Format: ECLAT ECLAT (Equivalence Class Transformation): A depth-first search algorithm using set intersection [Zaki et al. @KDD 97] Tid-List: List of transaction-ids containing the itemset(s) Vertical format: t(e) = {T 10, T 20, T 30 }; t(a) = {T 10, T 20 }; t(ae) = {T 10, T 20 } Properties of Tid-Lists t(x) = t(y): X and Y always happen together (e.g., t(ac} = t(d} ) t(x) t(y): transaction having X always has Y (e.g., t(ac) t(ce) ) Deriving frequent patterns based on vertical intersections Using diffset to accelerate mining Only keep track of differences of tids t(e) = {T 10, T 20, T 30 }, t(ce) = {T 10, T 30 } Diffset (ce, e) = {T 20 } A transaction DB in Horizontal Data Format Tid The transaction DB in Vertical Data Format Item Itemset 10 a, c, d, e 20 a, b, e 30 b, c, e Tid-List a 10, 20 b 20, 30 c 10, 30 d 10 e 10, 20, 30 Sampling for Frequent Patterns Select a sample of the original database, mine frequent patterns within the sample using Apriori Scan database once to verify frequent itemsets found in sample. Here only borders of closure of frequent patterns are checked: Example: check abcd instead of ab, ac,, etc. (why?) Scan database again to find missed frequent patterns. H. Toivonen. Sampling large databases for association rules. In VLDB 96 28 14

Ref: Apriori and Its Improvements R. Agrawal and R. Srikant. Fast algorithms for mining association rules. VLDB'94. H. Mannila, H. Toivonen, and A. I. Verkamo. Efficient algorithms for discovering association rules. KDD'94. A. Savasere, E. Omiecinski, and S. Navathe. An efficient algorithm for mining association rules in large databases. VLDB'95. J. S. Park, M. S. Chen, and P. S. Yu. An effective hash-based algorithm for mining association rules. SIGMOD'95. H. Toivonen. Sampling large databases for association rules. VLDB'96. S. Brin, R. Motwani, J. D. Ullman, and S. Tsur. Dynamic itemset counting and implication rules for market basket analysis. SIGMOD'97. S. Sarawagi, S. Thomas, and R. Agrawal. Integrating association rule mining with relational database systems: Alternatives and implications. SIGMOD'98. 29 References (II) Efficient Pattern Mining Methods R. Agrawal and R. Srikant, Fast algorithms for mining association rules, VLDB'94 A. Savasere, E. Omiecinski, and S. Navathe, An efficient algorithm for mining association rules in large databases, VLDB'95 J. S. Park, M. S. Chen, and P. S. Yu, An effective hash-based algorithm for mining association rules, SIGMOD'95 S. Sarawagi, S. Thomas, and R. Agrawal, Integrating association rule mining with relational database systems: Alternatives and implications, SIGMOD'98 M. J. Zaki, S. Parthasarathy, M. Ogihara, and W. Li, Parallel algorithm for discovery of association rules, Data Mining and Knowledge Discovery, 1997 J. Han, J. Pei, and Y. Yin, Mining frequent patterns without candidate generation, SIGMOD 00 M. J. Zaki and Hsiao, CHARM: An Efficient Algorithm for Closed Itemset Mining, SDM'02 J. Wang, J. Han, and J. Pei, CLOSET+: Searching for the Best Strategies for Mining Frequent Closed Itemsets, KDD'03 C. C. Aggarwal, M.A., Bhuiyan, M. A. Hasan, Frequent Pattern Mining Algorithms: A Survey, in Aggarwal and Han (eds.): Frequent Pattern Mining, Springer, 2014 15

References (III) Pattern Evaluation C. C. Aggarwal and P. S. Yu. A New Framework for Itemset Generation. PODS 98 S. Brin, R. Motwani, and C. Silverstein. Beyond market basket: Generalizing association rules to correlations. SIGMOD'97 M. Klemettinen, H. Mannila, P. Ronkainen, H. Toivonen, and A. I. Verkamo. Finding interesting rules from large sets of discovered association rules. CIKM'94 E. Omiecinski. Alternative Interest Measures for Mining Associations. TKDE 03 P.-N. Tan, V. Kumar, and J. Srivastava. Selecting the Right Interestingness Measure for Association Patterns. KDD'02 T. Wu, Y. Chen and J. Han, Re-Examination of Interestingness Measures in Pattern Mining: A Unified Framework, Data Mining and Knowledge Discovery, 21(3):371-397, 2010 16