A New Approach to Mine Frequent Pattern in Spatial Database using TFP-Tree

Similar documents
ISSN: (Online) Volume 2, Issue 5, May 2014 International Journal of Advance Research in Computer Science and Management Studies

AN INTELLIGENT APPROACH FOR MINING FREQUENT SPATIAL OBJECTS IN GEOGRAPHIC INFORMATION SYSTEM

PC Tree: Prime-Based and Compressed Tree for Maximal Frequent Patterns Mining

Data Mining Part 3. Associations Rules

Mining of Web Server Logs using Extended Apriori Algorithm

AN IMPROVISED FREQUENT PATTERN TREE BASED ASSOCIATION RULE MINING TECHNIQUE WITH MINING FREQUENT ITEM SETS ALGORITHM AND A MODIFIED HEADER TABLE

INFREQUENT WEIGHTED ITEM SET MINING USING NODE SET BASED ALGORITHM

Mining Rare Periodic-Frequent Patterns Using Multiple Minimum Supports

CS570 Introduction to Data Mining

Mining Frequent Patterns without Candidate Generation

A New Fast Vertical Method for Mining Frequent Patterns

Chapter 7: Frequent Itemsets and Association Rules

Comparing the Performance of Frequent Itemsets Mining Algorithms

Chapter 6: Basic Concepts: Association Rules. Basic Concepts: Frequent Patterns. (absolute) support, or, support. (relative) support, s, is the

Chapter 4: Mining Frequent Patterns, Associations and Correlations

Association Rule Mining: FP-Growth

An Efficient Algorithm for Finding the Support Count of Frequent 1-Itemsets in Frequent Pattern Mining

ASSOCIATION rules mining is a very popular data mining

Association Rule Mining

Adaption of Fast Modified Frequent Pattern Growth approach for frequent item sets mining in Telecommunication Industry

Available online at ScienceDirect. Procedia Computer Science 45 (2015 )

A Survey on Moving Towards Frequent Pattern Growth for Infrequent Weighted Itemset Mining

Keywords: Mining frequent itemsets, prime-block encoding, sparse data

ETP-Mine: An Efficient Method for Mining Transitional Patterns

FP-Growth algorithm in Data Compression frequent patterns

Novel Techniques to Reduce Search Space in Multiple Minimum Supports-Based Frequent Pattern Mining Algorithms

Improved Frequent Pattern Mining Algorithm with Indexing

Salah Alghyaline, Jun-Wei Hsieh, and Jim Z. C. Lai

Apriori Algorithm. 1 Bread, Milk 2 Bread, Diaper, Beer, Eggs 3 Milk, Diaper, Beer, Coke 4 Bread, Milk, Diaper, Beer 5 Bread, Milk, Diaper, Coke

CLOSET+:Searching for the Best Strategies for Mining Frequent Closed Itemsets

ALGORITHM FOR MINING TIME VARYING FREQUENT ITEMSETS

Data Mining Techniques

This paper proposes: Mining Frequent Patterns without Candidate Generation

WIP: mining Weighted Interesting Patterns with a strong weight and/or support affinity

Data Mining for Knowledge Management. Association Rules

Vertical Mining of Frequent Patterns from Uncertain Data

Appropriate Item Partition for Improving the Mining Performance

PTclose: A novel algorithm for generation of closed frequent itemsets from dense and sparse datasets

Lecture Topic Projects 1 Intro, schedule, and logistics 2 Data Science components and tasks 3 Data types Project #1 out 4 Introduction to R,

PLT- Positional Lexicographic Tree: A New Structure for Mining Frequent Itemsets

Pattern Mining. Knowledge Discovery and Data Mining 1. Roman Kern KTI, TU Graz. Roman Kern (KTI, TU Graz) Pattern Mining / 42

DATA MINING II - 1DL460

An Algorithm for Mining Large Sequences in Databases

Product presentations can be more intelligently planned

Performance and Scalability: Apriori Implementa6on

DATA MINING II - 1DL460

Memory issues in frequent itemset mining

Maintenance of the Prelarge Trees for Record Deletion

Data Structure for Association Rule Mining: T-Trees and P-Trees

Association Rule Mining. Introduction 46. Study core 46

Survey: Efficent tree based structure for mining frequent pattern from transactional databases

An Improved Frequent Pattern-growth Algorithm Based on Decomposition of the Transaction Database

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 6

A Graph-Based Approach for Mining Closed Large Itemsets

A Comparative Study of Association Rules Mining Algorithms

Chapter 4: Association analysis:

ANALYSIS OF DENSE AND SPARSE PATTERNS TO IMPROVE MINING EFFICIENCY

Tutorial on Assignment 3 in Data Mining 2009 Frequent Itemset and Association Rule Mining. Gyozo Gidofalvi Uppsala Database Laboratory

An Evolutionary Algorithm for Mining Association Rules Using Boolean Approach

OPTIMISING ASSOCIATION RULE ALGORITHMS USING ITEMSET ORDERING

Web Page Classification using FP Growth Algorithm Akansha Garg,Computer Science Department Swami Vivekanad Subharti University,Meerut, India

IJESRT. Scientific Journal Impact Factor: (ISRA), Impact Factor: [35] [Rana, 3(12): December, 2014] ISSN:

Implementation of Data Mining for Vehicle Theft Detection using Android Application

An Improved Apriori Algorithm for Association Rules

Basic Concepts: Association Rules. What Is Frequent Pattern Analysis? COMP 465: Data Mining Mining Frequent Patterns, Associations and Correlations

Discovery of Multi-level Association Rules from Primitive Level Frequent Patterns Tree

An Efficient Reduced Pattern Count Tree Method for Discovering Most Accurate Set of Frequent itemsets

Tutorial on Association Rule Mining

Effectiveness of Freq Pat Mining

Ascending Frequency Ordered Prefix-tree: Efficient Mining of Frequent Patterns

Discovery of Frequent Itemset and Promising Frequent Itemset Using Incremental Association Rule Mining Over Stream Data Mining

Frequent Pattern Mining. Based on: Introduction to Data Mining by Tan, Steinbach, Kumar

IMPLEMENTATION AND COMPARATIVE STUDY OF IMPROVED APRIORI ALGORITHM FOR ASSOCIATION PATTERN MINING

APPLYING BIT-VECTOR PROJECTION APPROACH FOR EFFICIENT MINING OF N-MOST INTERESTING FREQUENT ITEMSETS

Chapter 6: Association Rules

Performance Based Study of Association Rule Algorithms On Voter DB

Frequent Pattern Mining

A Taxonomy of Classical Frequent Item set Mining Algorithms

AC-Close: Efficiently Mining Approximate Closed Itemsets by Core Pattern Recovery

CHAPTER 8. ITEMSET MINING 226

An Algorithm for Mining Frequent Itemsets from Library Big Data

Privacy Preserving Frequent Itemset Mining Using SRD Technique in Retail Analysis

Mining Recent Frequent Itemsets in Data Streams with Optimistic Pruning

Graph Based Approach for Finding Frequent Itemsets to Discover Association Rules

Mining Frequent Patterns Based on Data Characteristics

HYPER METHOD BY USE ADVANCE MINING ASSOCIATION RULES ALGORITHM

An Efficient Algorithm for finding high utility itemsets from online sell

ISSN: (Online) Volume 2, Issue 7, July 2014 International Journal of Advance Research in Computer Science and Management Studies

CSE 5243 INTRO. TO DATA MINING

Keshavamurthy B.N., Mitesh Sharma and Durga Toshniwal

A novel algorithm for frequent itemset mining in data warehouses

Discovering Quasi-Periodic-Frequent Patterns in Transactional Databases

Finding the boundaries of attributes domains of quantitative association rules using abstraction- A Dynamic Approach

Efficient Algorithm for Frequent Itemset Generation in Big Data

Research of Improved FP-Growth (IFP) Algorithm in Association Rules Mining

Association Rule Mining

2 CONTENTS

A BETTER APPROACH TO MINE FREQUENT ITEMSETS USING APRIORI AND FP-TREE APPROACH

Optimized Frequent Pattern Mining for Classified Data Sets

An Approach for Finding Frequent Item Set Done By Comparison Based Technique

Transcription:

A New Approach to Mine Frequent Pattern in Spatial Database using TFP-Tree Kedar Nath Singh Jitendra Agrawal Sanjeev Sharma School of I. Technology School of Information technology, School of Information technology, U.T.D., R.G.P.V., Bhopal U.T.D., R.G.P.V., Bhopal (M.P.) U.T.D., R.G.P.V., Bhopal (M.P.) e-mail: kn_tech2005@yahoo.co.in e-mail: jitendra@rgtu.net e-mail: sanjeev@rgtu.net Abstract Today there are several skilled algorithms to mine frequent patterns. Frequent item set mining provides the associations and correlations among items in large transactional or relational database. In this paper a new approach to mine frequent pattern in spatial database using TFP-tree is proposed. The proposed approach generates a TFP-tree that specifies the generations of frequent patterns. Our analysis approach generates maximal frequent patterns and performs only minimal generalizations of frequent candidate sets. Keywords: Spatial Database, Ordered List, Maximal Frequent Pattern, Closed Frequent Pattern 1. Introduction Spatial database provisions a large amount of space related data, like as maps, preprocessed remote sensing or medical imaging, and VLSI chip lay out data. In different fields, there is a need to manage geometric, geographic, or spatial data. The main application that is driving research in spatial database systems is the technology for GISs. We are unable to find couched regularities, rules or knowledge hidden in the data. Storage and recovery of spatial data has therefore become a challenge. Because in contrast to mining in relational databases, spatial data mining algorithms have to consider the neighbors of objects in order to extract useful knowledge. This is necessary because the attributes of neighboring object of interest may have a significant influence on the object itself [1]. Frequent pattern mining was first proposed by Agrawal et al. (1993) for market basket analysis in the form of association rule mining. It analyses customer buying habits by finding associations between the different items that customers place in their shopping baskets [2]. A frequent pattern is generated from the given database with its frequency no less than user-specified minimum support count. When the database contains large number of long frequent itemsets, mining all frequent itemsets might not be a good idea. For example, if there is a frequent itemset with size L, then all 2 L nonempty subsets of the item set have to be generated[3].due to this problem, closed frequent pattern mining and maximal frequent pattern mining were proposed. A pattern X is a closed frequent pattern in a data set D if X is frequent in D and there exists no proper super-pattern Y such that Y has the same support as X in D. A pattern X is a maximal frequent pattern in set D if X is frequent, and there exists no super-pattern Y such that X Y and Y is frequent in D. For the same min_sup threshold, the set of closed frequent patterns contains the complete information regarding to its corresponding frequent patterns; whereas the set of max-patterns, though more compact, usually does not contain the complete support information regarding to its corresponding frequent patterns[2]. FP-tree is a highly compact structure which stores the information for frequent-pattern mining. Since a single path a1 a2 an in the a1-prefix subtree registers all the transactions whose maximal frequent set is in the form of a1 a2 ak for any 1 k n, the size of the FP-tree is substantially smaller than the size of the database and that of the candidate sets generated in the association rule mining[5]. In this paper we propose Transaction Frequent Pattern (TFT) tree it makes use of numerical representation and intersecting the ordered list of set of items in similar type. And finally from generated frequent patterns we generate maximal frequent pattern from the set of closed frequent pattern. 2. Related Works Significant works have been done for the mining of association rules. Most of them based on three fundamental frequent patterns mining methodologies are Apriori, FP-growth, and FP-tree. Apriori algorithm adopts layer by layer search iteration method to produce frequent itemsets, which is frequent k-1 itemset. L k-1 is used to search K itemset L k. However, it needs to scan database when L k is found. So Apriori algorithm will generate a large number of candidate sets. If there are 10 4 frequent 1-itemsets,then more than 10 7 candidate 2- itemsets are produced[4]. Different types of the Apriori algorithm have been developed, like as AprioriTid[6], ArioriHybrid[2], direct hashing and 1062

pruning (DHP), dynamic item sets counting (DIC)[8]. FP-growth was first proposed briefly in Han et al. (2000), F-Growth algorithm does not generate candidate itemsets, and only scans two times database. So its efficiency is larger 10 times than Apriori algorithm, but it needs excessive CPU consumption and storage cost[5]. Eclat[9] is the first algorithm to find frequent patterns by a depth-first search and also perform well. The Pincer-Search algorithm combines a bottom-up and a top down techniques to find the maximal frequent itemsets. The bottom up process finds frequent itemsets, and non frequent itemsets. Then non frequent itemsets are used by a top down process to refine a set of potential maximal frequent itemsets[10]. But, when the database is sparse, variation in transactions will drop its advantage. Now, we present a new approach in this paper that maps and compresses the both dense and sparse database by making a Transaction Frequent Pattern tree using numerical representation and its arithmetic operations. This approach executes just on generalizations of frequent candidate sets. So that, the number of valid frequent patterns generated and tested is minimized as compared to existing algorithms. The creation of maximal patterns is completed by intersecting the ordered list of related type. 3. Proposed framework In this paper we proposed a framework by modified the framework [1] to discover knowledge in conditions of maximal frequent patterns of spatial database objects. We divide spatial database into dense and sparse datasets. The Required feature of proposed framework is given in Figure1. The Framework can be divided into a six steps. In the 1 st step we find dense and sparse datasets from spatial database. In transaction database we assume on the basis of occurrence of higher ranked objects [10] in spatial database. In 2nd step we find frequent ordered list of each dense and sparse database. The 3 rd step we represent the frequent order lists by Fibonacci series value where each object in the transaction is represented as a unique Fibonacci series value and then calculate the product value of the Fibonacci database which compresses the transaction. 4 th step constructs a TFP tree using the numerical representation of each transaction database. 5 th step finds frequent patterns with their respective support count by intersecting its numerical ordered list. And finally we generate the set of maximal frequent patterns for dense and sparse database. We use a Fibonacci-based data transformation technique to condense the size of transaction database to improve the performance of mining algorithms. Figure: 1 Proposed Framework 3.1Structuring of a Sample Database In this Paper we consider a sample Spatial Database for finding frequent Pattern. Our analyses are based on frequent spatial objects from Geographic Map Database and find the link between those objects. These objects occur in sets and each set take place at least as frequently as a predetermined minimum support count. The frequent objects will play role in frequent patterns. We have considered 100 Indian cities for mainly seven spatial objects. They are represented in Table 1. Table 1 S.N. Spatial objects Spatial objects representation 1 Museum A 2 Lake B 3 Monument C 4 Zoo D 5 Forest E 6 River F 7 Hill G 1063

For each city we consider the occurrence of these spatial objects to form a real time database of spatial transactions. Any two cities together with its above spatial objects make a transaction within a mid-air distance of 25kms. We assume a sample of 31 cities to analyze proposed framework from the geo spatial database of 100 Indian cities.. While building each tuple that corresponds to a city we have assumed lexicographic ordering of spatial objects [1]. For example the first transaction (TID: 1) as shown in Table 2. The Allahabad city has the spatial objects such as Museum (A), Monument (C), Forest (E) and River (F). Therefore in Table 2 {A,C,E,F} is denoted corresponding to Allahabad. Similarly for each city the occurrence of spatial objects have taken into consideration and represented in Table 2. From 31 Indian cities a sample of 13 cities are considered to be dense and other 18 are considered to be sparse. We assume the minimum threshold support is 40%, and 30% for sparse and dense database. A database is said to be dense when most of the transactions is similar among them. They have contained mostly the same length and the same objects. A database is said to be sparse database when the length of the transaction differs a large number from one another and contains mostly different objects. Table 2 A Sample of Spatial Database TID Reference City Objects 1 Allahabad A,C,E,F 2 Darjeeling A,D,G 3 Ranchi A,B,G 4 Chennai A,B,E 5 Ernakulum A,C,E,F 6 Coimbatore A,C,F 7 Kanpur A,C,F 8 Amritsar B,C,E 9 Mumbai A,B,C,D,F,G 10 Bangalore A,B,C,D,E 11 Ajmer A,B,C,F 12 Bhubaneswar A,B,C,D,F,G 13 Chandigarh A,B,D,E 14 Trivandrum A,B,D 15 Delhi A,C,D,E,F 16 Ahmadabad A,B,C,D,F 17 Pune A,B,C,D,F,G 18 Mysore B,C,D,E,G 19 Nagpur A,B,C,D 20 Patna A,C,D,F 21 Bhopal A,B,C,D,E,G 22 Mussorrie A,B,E,G 23 Varanasi A,C,F 24 Ooty A,B,G 25 Kodaikanal A,B,E,G 26 Port blair A,C,F 27 Thekaddy B,E,D,G 28 Chamba A,B,E,G 29 Locknow B,C,D,F 30 Nashik A,C,F 31 Agra A,C,E,F In the above transaction Table 2 We are considered the TIDs {9,10,11,12,13,14,15,16,17,18,19,20,21} to be dense while TIDs {1,2,3,4,5,6,7,8,22,23,24,25,26, 27,28,29,30,31} are the sparse database. We apply the proposed framework on Sample of Spatial Database and produce the set of frequent patterns. To produce the set of frequent patterns we propose an algorithm in section 4 and further in section 5 perform the analysis of the framework using the sample spatial database. 4. Proposed Algorithms In this proposed work we have modified the method used by Animesh Tripathy et al[1] by using Fibonacci numbers instead of Prime numbers and then analyzed the result. This algorithm is based on the proposed framework to mine frequent patterns from spatial database. The innermost concept is based on the numerical representation of ordered list to create the TFP Tree for dense and sparse spatial objects separately. The proposed algorithm is used to obtain suitable frequent patterns. The algorithm is divided into three phases: Phase I: To Find Frequent Ordered List: Input: Spatial Database (SB), Minimum Support (MS) Output: Order list (OL) Procedure: Find Frequent Order List as follows Step 1: Scanning SB and consider object[i] from object lists. Step 2: Calculate Number of count for each object[i] for each Transaction If object[i] = = Exist in object lists than Insert into Temp_list Count [object[i] ]++ Step 3: For each object[i] in Temp_list Calculate the Support Count for each object 1064

Step 4: Find Order List (OL) If (Support Count (object[i]) < MS) Remove (object[i]) else Sorting ( object[i].list ) in descending order. Produce Order list(ol) Phase II: To Find TFP Tree: Input: Ordered List (OL), Set of Fibonacci Numbers (FN) Output: TFP Tree Procedure: TFP Tree creates as follows Step 1: Scanning each transaction T and its corresponding Order List (OL). Step 2: Assign Fibonacci Numbers (FN) to Order List OL //FN Start from 2. Scanning each object[k] in Order list[j] than Insert FN to each object[i] in Order list[j] in increasing order. Step 3: Calculate the required Product Sum (PS) Initialize PS=1 For each Order list[j] PS = PS + MULTIPLY ( FN(object[i]) ) Step 4: Constructing TFP Tree For PS in each transaction Temp = Compare (PS.transaction[t], PS.transaction[t+1]) If (Temp = = 0) Create a leaf node after than increment count[t] ++ else if PS.transaction [t+1]) MOD (PS.transaction[t] = 0 than Create new leaf as child of transaction[t] Assign count[t] =1 else Create new leaf and assign count[t] =1 Phase III: To Find TFP Frequent Pattern: Input: TFP Tree, Minimum Support (ms), Tree (vertex, edge) Output: Transaction Frequent Patterns (TFP) Procedure: TFP creation as follows Step 1: Scanning TFP Tree and consider each branch of the Tree [v, e] // Assign v, e for vertex and edge Step 2: For each set of nodes in Tree [v, e] List[i].Tree [v, e] = Fibonacci Numbers Object pattern[i] = list[i].tree [v, e] Step 3: Initialize Frequent Patterns = = {} For each branch in Tree [v, e] TFP[j] = object pattern[i] Object pattern [i+1] Frequent Pattern = TFP[j] U Frequent Pattern Increment object pattern [i] Increment list[i].tree [v, e] Create TF 5. Analysis of Proposed framework To prove the proposed framework we have considered the transaction list given in Table 2 for our analysis. The spatial objects are associated to each other in space given in Table 1 and ranking report between these objects like as the first object is either ranked higher than, ranked lower than or ranked equal to the second object as per their frequency. The ranking relationship can be find from total number of frequency of an object occurred in database. After that we think two different types of objects demonstration i.e. dense and sparse objects on the basis of happening of higher ranked objects in the spatial database. We independently consider the dense and sparse spatial database for analysis. We make use of an analysis process in to Six Step as per the framework described in the earlier section: Step 1: We independently achieve Dense & Sparse Transaction List. Step 2: Construct Ordered list of objects in descending order of their frequencies occurrence. Step3: Mapping Ordered List in form on numerical value of Fibonacci series. Step 4: Making a TFP Tree by using numerical value. Step 5: Find frequent patterns and confirm it against their respective support count. Step 6: Create the set of Maximal Frequent Patterns 5.1Transaction Mapping Technique of spatial database In this part we have measured the TIDs {9,10,11,12,13,14,15,16,17,18,19,20,21} to form the dense database and remaining are sparse database. The spatial objects are arranged in the ordered list in decreasing order of their occurrence frequency in the complete dense database. The lower occurring objects having frequency less than 40%, and 30% for sparse and dense are pruned. The dense and sparse object list generated is shown in Table 3 & 4. The spatial objects of dense and sparse database are set in downward order of their frequencies. The obtained ordered list of each transaction is mapped by Fibonacci-based data transformation technique to reduce the size of transaction database. The details of it are shown in Table 3 & 4. We calculate the product value of each Order Lists. Example: We assume a sample dense database of TID #1 where Mumbai is a reference city and its corresponding spatial objects are {A,D,B,C,F }. Using the Fibonacci numbers in increasing order for representation such as [(A:2),(D:3),(B:5),(C:8),(F:13)]. So the ordered list can be mapped as {2,3,5,8,13} for numerical order list as product value of Fibonacci {A,D,B,C,F}. 1065

Further we find the product of the numbers i.e. (3120=2*3*5*8*13) and finally the both dense and sparse transaction is arranged in descending order of their product value. We create a TFP-tree as shown in Figure 2 and 3. Table 3 Numerical Representation of Dense Database TID # Reference City Ordered Dense Objects Fibonacci No.Represensation Product Value 1 Mumbai A,D,B,C,F 2,3,5,8,13 3120 TID # 2 Bhubaneswar A,D,B,C,F 2,3,5,8,13 3120 3 Ahmadabad A,D,B,C,F 2,3,5,8,13 3120 4 Pune A,D,B,C,F 2,3,5,8,13 3120 5 Ajmer A,B,C,F 2,5,8,13 1024 6 Delhi A,D,C,F 2,3,8,13 624 5 Patna A,D,C,F 2,3,8,13 624 8 Bangalore A,D,B,C 2,3,5,8 240 9 Nagpur A,D,B,C 2,3,5,8 240 10 Bhopal A,D,B,C 2,3,5,8 240 11 Mysore D,B,C 3,5,8 120 12 Chandigarh A,D,B 2,3,5 30 13 Trivandrum A,D,B 2,3,5 30 Table 4 Numerical Representation of Sparse Database References City Ordered Dense Object Fibonacci No. Rep resentation Produ ct Value 1 Munsorrie A,E,B,G 2,5,13,21 2730 2 Kodaikanal A,E,B,G 2,5,13,21 2730 3 Chamba A,E,B,G 2,5,13,21 2730 4 Thekaddy E,B,G 5,13,21 1365 5 Ooty A,B,G 2,13,21 546 6 Ranchi A,B,G 2,13,21 546 7 Amritsar C,F,B 3,8,13 312 8 Locknow C,F,B 3,8,13 312 9 Allahabad A,C,E,F 2,3,5,8 240 10 Ernakulam A,C,E,F 2,3,5,8 240 11 Agra A,C,E,F 2,3,5,8 240 12 Channai A,F,B 2,8,13 208 13 Coimbatore A,C,F 2,3,8 48 14 Varanshi A,C,F 2,3,8 48 15 Port blair A,C,F 2,3,8 48 16 Nashik A,C,F 2,3,8 48 17 Darjeeling A,G 2,21 42 18 Kanpur A,C,E 2,3,5 30 5.2Construction of TFP-Tree The TFP-Tree is based on Fibonacci number uniqueness for compressing data and attractive the efficiency of pattern generation. A TFP-Tree includes a root node and a child node that forms a sub tree as children of the root where each child stores product value of each transaction this node. While insertion of a new node takes place, the first and second transaction product values are compared and if the product value of two transactions is not divisible then it creates new descendants node. In case, it is divisible then it is inserted as a child of the existing node. If product value is equal to the current node only the count of the current node is increased by 1. The final TFP tree for dense and sparse datasets is shown in Figure 2 & 3 respectively. Suppose we take #1, #2, and #3 of sparse dataset (Figure 3) where the product values of transaction are equal so count value of node is increased by 3. More when we take #7 of sparse database, it is not divisible by 2730 so 312 create a new descendant node. Similarly for every successive insertion of a new node its product value is examined against existing nodes. The construction of TFP tree stops after considering every transaction present in the data set. The node under Root node in Figure 2 denotes {3120:4} where {3120} is the product value and {4} represents the frequency of occurrence of the product in similar transaction lists. Thus one can observe and verify that the tree generated in TFP is much simple with respect to level and complexity as compared to similar tree generated using FP-Growth or other successive algorithm. Figure 2: TFP Tree for Dense Database 1066

2730 1365 TFP-1 2,5,13,21 5,13,21 2730 546 42 TFP-2 2,513,21 2,13,21 2,21 Figure 3: TFP-Tree for Spatial Database 5.3 Generation of Frequent Patterns To mine the frequent pattern from TFP-tree we traverse from root to each branch of tree. Taking each branch as a new entry to an array list we find the set of Fibonacci series numbers for each product value for a given node.. Then by intersecting of Fibonacci number set of each node s product value, we find the frequent patterns. Suppose we consider Figure 2 and take (3120, 1040) branch as a new entry in an array list (say TFP-1) which is shown in Fig 4. Then by intersecting its Fibonacci-based representation of {2,3,5,8,13} with {2,5,8,13} we finally get {2,5,8,13} as its factor. corresponding to the factor {2,5,8,13},the possible frequent patterns are (2,5),(2,5,8) and(2,5,8,13). Each Fibonacci number set represents a spatial object pattern like {Museum, Lake}, {Museum, Lake, Monument}, {Museum, Lake, Monument, River} which are correlated with each other. To generate frequent pattern we assume the lexicographic ordering of spatial objects. Similarly for every successive insertion of a new array its pattern is examined against existing branch values to generate the consolidated list of frequent patterns. 3120 1040 TFP-1 2,3,5,8,13 2,5,8,13 TFP-2 TFP-3 TFP-4 3120 624 2,3,5,8,13 2,3,8,13 3120 240 120 2,3,5,8,13 2,3,5,8 3,5,8 3120 240 30 2,3,5,8,13 2,3,5,8 2,3,5 Figure 4: Dense Pattern in TFP Tree TFP-3 TFP-4 TFP-5 TFP-6 312 3,8,13 240 48 2,3,5,8 2,3,8 240 30 2,3,5,8 2,3,5 208 2,8,13 Figure 5: Sparse Patterns in TFP Tree 5.4Computing MFP from CFP Now from the set of generated frequent patterns we find the set of maximal frequent pattern (MFP) from a set of close frequent pattern (CFP). As discuss previous CFP can be defined as closed if it has no superset with the same frequency (support). In the previous part we find the intersecting patterns as {(2,5,8,13), (2,3,8,13), (2,3,8), (2,3,5)}. Due to this set of intersecting patterns we find the set of frequent patterns as {(2,5), (2,5,8), (2,5,8,13), (2,3), (2,3,8), (2,3,8,13), (2,3,5)} these support counts as (10,8,5,11,9,6,9). Each of these patterns is considered to be a set of closed frequent pattern (CFP). From this set of CFP we compute the set of maximal frequent pattern(mfp) as {(2,5,8,13), (2,3,8,13), (2,3,8),(2,3,5) }. Similarly the MFP for sparse is computed as ({5,13,21}, {2,21}, {3,8,13}, {2,3,8}, {2,3,5}, {2,8,13}). Furthermore, it proved all the An Improved Association Rule Based Algorithmic Approach to Mine Frequent Pattern in Spatial Database 370 maximal patterns is unique to each other, while all the frequent closed patterns can be related by subset or superset relationship with respect to each other. Each of this MFP is able of generating the full set of frequent patters which is demonstrated from analytical results. 6. Experimental Results In this paper, we explain our experimental result on the basis of efficiency and scalability in comparison with Apriori and FP-growth. We come across that TFP best performs from both Apriori and 1067

FP-growth and is efficient and highly scalable for mining very large databases. TFP was implemented by using Visual C ++ 6.0, while the version of FPgrowth and Apriori that we used is available at http://fizzy.cs.uni-magdeburg.de/~borglet/. All reports of iteration and number of rules created as shown in figure 6. Our results explain that TFP better forms the existing algorithms and No. of frequent pattern generate for finding MFP. 40 30 20 10 0 1 2 3 4 5 6 TFP FP-growth Apriori Figure 6: No. of Iterations Vs No. of Rules created 7. Conclusion In this paper we proposed a mathematical approach to find frequent patterns using TFP Tree. TFP-tree compresses both dense and sparse datasets by using numerical value representation. In this method we consider Fibonacci number characteristics to find CFP (closed frequent pattern) and then the MFP (maximal frequent pattern) our approach is efficient on both dense and sparse database.the algorithm will positively enhance the efficiency of judgment the relationship between spatial objects and further can be used in association analysis. The creation of maximal frequent patterns is done by intersecting the ordered list (OL) of similar type which reduces the search space. [3] Jun TAN, Yingyong BU and Bo YANG, An Efficient Close Frequent Pattern Mining Algorithm 978-0- 7695-3804-4/09 $26.00 2009 IEEE [4] Jian Hu, Xiang Yang-Li, A New Fast Frequent Itemsets Mining Algorithm Based on Forest, 978-0- 7695-3305-6/08 $25.00 2008 IEEE. [5] P. Shenoy, J. R. Haritsa, S. Sudarshan, G. Bhalotia, M. Bawa, and D. Shah, Turbo-charging Vertical Mining of Large Databases, Procedings of ACM SIGMOD International Conference on Management of Data ACM Press, 2000. [6] Burdick D., M. Calimlim and J. Gehrke, Mafia: A Maximal Frequent Itemset Algorithm for Transactional Databases, 17th International Conference on Data Engineering, pp. 443-452, 2001. [7] Han J., J. Pei, Y. Yin, and R. Mao, (2004), Mining Frequent Patterns without Candidate Generation: A Frequent-Pattern Tree Approach, Data Mining and Knowledge Discovery, 8, pp. 53-87, 2004. [8] Norwati Mustapha, Mohammad-Hossein Nadimi- Shahraki, Ali B Mamat, Md. Nasir B Sulaiman, A Numerical Method For Frequent Patterns Mining, JATIT, pp92-98, 2009. [9] P. Shenoy, J. R. Haritsa, S. Sudarshan, G. Bhalotia, M. Bawa, and D. Shah, Turbo-charging Vertical Mining of Large Databases, Procedings of ACM SIGMOD International Conference on Management of Data ACM Press, Dallas, Texas, pp. 22-23, 2000. [10] Nadimi-Shahraki M.H., N. Mustapha, M. N. B. Sulaiman, and A. B. Mamat, A New Method for Mining Maximal Frequent Itemsets, Presented at International IEEE Symposium on Information Technology, Malaysia, pp. 309-312, 2008. References: [1] Animesh Tripathy, Subhalaxmi Das, Prashanta Kumar Patra, An Association Rule Based Algorithmic Approach to Mine Frequent Pattern in Spatial Database System, International Journal of Computer Science & Communication Vol.1, No. 2, pp. 357-363, July-December 2010. [2] Han J., H. Cheng, D. Xin, and X. Yan, Frequent Pattern Mining: Current Status and Future Directions, Data Mining and Knowledge Discovery, 15, pp. 55-86 (2007). 1068