The Knowledge Mining Suite (KMS)

Similar documents
Principal Components Analysis for Trapezoidal Fuzzy Numbers

Analysis of Symbolic Data

self-organizing maps and symbolic data

Principal Component Analysis of Interval Data: a Symbolic Data Analysis Approach 1

complex data Edwin Diday, CEREMADE, Beijing 2011

8. MINITAB COMMANDS WEEK-BY-WEEK

A Self Organizing Map for dissimilarity data 0

2. (a) Briefly discuss the forms of Data preprocessing with neat diagram. (b) Explain about concept hierarchy generation for categorical data.

THE ENSEMBLE CONCEPTUAL CLUSTERING OF SYMBOLIC DATA FOR CUSTOMER LOYALTY ANALYSIS

Table Of Contents: xix Foreword to Second Edition

JMP Book Descriptions

Exporting symbolic objects to databases

Applied Regression Modeling: A Business Approach

Contents. Foreword to Second Edition. Acknowledgments About the Authors

1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda

Data Mining. CS57300 Purdue University. Bruno Ribeiro. February 1st, 2018

Statistical Pattern Recognition

Statistical Pattern Recognition

Getting Started. What is SAS/SPECTRAVIEW Software? CHAPTER 1

Supervised Variable Clustering for Classification of NIR Spectra

Analytical model A structure and process for analyzing a dataset. For example, a decision tree is a model for the classification of a dataset.

Workload Characterization Techniques

DATA WAREHOUING UNIT I

3. Data Preprocessing. 3.1 Introduction

R07. FirstRanker. 7. a) What is text mining? Describe about basic measures for text retrieval. b) Briefly describe document cluster analysis.

Data Mining and Analytics. Introduction

Scaling Techniques in Political Science

Points Lines Connected points X-Y Scatter. X-Y Matrix Star Plot Histogram Box Plot. Bar Group Bar Stacked H-Bar Grouped H-Bar Stacked

University of Florida CISE department Gator Engineering. Data Preprocessing. Dr. Sanjay Ranka

Data Preprocessing. S1 Teknik Informatika Fakultas Teknologi Informasi Universitas Kristen Maranatha

Decision Trees Dr. G. Bharadwaja Kumar VIT Chennai

Preprocessing Short Lecture Notes cse352. Professor Anita Wasilewska

Data Preprocessing. Data Preprocessing

UNIT 4. Research Methods in Business

Statistical Pattern Recognition

Data Science. Data Analyst. Data Scientist. Data Architect

Data Preprocessing. Why Data Preprocessing? MIT-652 Data Mining Applications. Chapter 3: Data Preprocessing. Multi-Dimensional Measure of Data Quality

Feature clustering and mutual information for the selection of variables in spectral data

2. Data Preprocessing

Data Preprocessing. Slides by: Shree Jaswal

Step-by-step data transformation

Computational Databases: Inspirations from Statistical Software. Linnea Passing, Technical University of Munich

Part I. Graphical exploratory data analysis. Graphical summaries of data. Graphical summaries of data

Distributed Objects with Sense of Direction

Data Mining Overview. CHAPTER 1 Introduction to SAS Enterprise Miner Software

Chapter 1: Introduction

Cambridge IGCSE mapping

SAS High-Performance Analytics Products

Advanced Applied Multivariate Analysis

K236: Basis of Data Science

A Pretopological Approach for Clustering ( )

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

Application of the topological gradient method to color image restoration and classification

Predict Outcomes and Reveal Relationships in Categorical Data

Applied Regression Modeling: A Business Approach

UNIT 2 Data Preprocessing

UNIT 2. DATA PREPROCESSING AND ASSOCIATION RULES

Learning Objectives for Data Concept and Visualization

COMPUTER AND ROBOT VISION

ECLT 5810 Clustering

ECLT 5810 Data Preprocessing. Prof. Wai Lam

Chemometrics. Description of Pirouette Algorithms. Technical Note. Abstract

CLUSTER ANALYSIS. V. K. Bhatia I.A.S.R.I., Library Avenue, New Delhi

CS425 Fall 2016 Boris Glavic Chapter 1: Introduction

ECLT 5810 Clustering

Detecting Novel Associations in Large Data Sets

Some questions of consensus building using co-association

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Graphes: Manipulations de base et parcours

Tree Models of Similarity and Association. Clustering and Classification Lecture 5

KS4 3 Year scheme of Work Year 10 Higher

Some Other Applications of the SOM algorithm : how to use the Kohonen algorithm for forecasting

PESIT- Bangalore South Campus Hosur Road (1km Before Electronic city) Bangalore

Co-clustering for differentially private synthetic data generation

Brief Contents. Foreword by Sarah Frostenson...xvii. Acknowledgments... Introduction... xxiii. Chapter 1: Creating Your First Database and Table...

An Incremental Hierarchical Clustering

Analysis of sandwich beams using mixed finite element

Exploratory Data Analysis EDA

Data Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1395

M359 Block5 - Lecture12 Eng/ Waleed Omar

Getting to Know Your Data

Product Catalog. AcaStat. Software

Minitab 17 commands Prepared by Jeffrey S. Simonoff

Kernel Methods and Visualization for Interval Data Mining

BIG DATA SCIENTIST Certification. Big Data Scientist

Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005

DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING

STA 570 Spring Lecture 5 Tuesday, Feb 1

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

Data Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1394

Section A. 1. a) Explain the evolution of information systems into today s complex information ecosystems and its consequences.

JAVA PROGRAMMING. Unit-3 :Creating Gui Using The Abstract Windowing Toolkit:

CS6220: DATA MINING TECHNIQUES

Introduction to BEST Viewpoints

Data Set. What is Data Mining? Data Mining (Big Data Analytics) Illustrative Applications. What is Knowledge Discovery?

Time: 3 hours. Full Marks: 70. The figures in the margin indicate full marks. Answers from all the Groups as directed. Group A.

Themes in the Texas CCRS - Mathematics

Visualization and text mining of patent and non-patent data

Dynamic Optimization of Generalized SQL Queries with Horizontal Aggregations Using K-Means Clustering

Transcription:

The Knowledge Mining Suite (KMS) Oldemar Rodríguez 1 University of Costa Rica and Predisoft International S.A., San José Costa Rica. oldemar.rodriguez@predisoft.com Abstract. The Knowledge Mining Suite (KMS) is an integrated data mining software for end-users, researchers and developers, implemented by the author and his collaborators. The KMS is oriented to mining both relational databases and data warehouses and it is based on the symbolic data analysis. In this paper a summary of the main characteristics of the KMS is presented, we also introduce some important features regarding extensible Markup Language (XML) and its application in representing, transmitting and storing symbolic data in the Knowledge Mining Suite. 1 XML as an Alternative for Data Representation in Knowledge Mining Suite XML stands for extensible Markup Language. It is a meta-language, characterized by the flexible and scalable use of markers. These markers, known as tags, act as delimiters for enclosing elements in a hierarchical way. XML is supported by the World Wide Web Consortium (W3C), becoming a recommended standard for data representation in web applications. As part of the open source resources, there are many projects for developing new tools and technologies related to XML. XML is a super set of HTML, the more used tagging system for internet pages. Then, one idea behind developing XML was to obtain a more flexible scheme for marking text. XML has evolved into a language for structuring knowledge in a file. It can be applied for representing, transmiting and storing data. Moreover, several tools have been built to reasoning about data in XML files. There are many advantages by using XML in software systems: It can be used as a data place holder, replacing databases or other data storages. The set of tags is customizable, making it suitable for different applications. The GUI (Graphic User Interface) is extracted, changes to the display don t mean changes in the data. Searching data is easy and efficient, engines can take advantage of tag structure rather than muddling in the data. Code is more legible, even for a new user. Complex relations can be represented due to its hierarchical nature. There is a large basis of users and support tools around XML world. The entire document or even parts of it can be analyzed or transmitted, making information sharing easier.

Documents can be validated according to some definition, offering more security concerning data consistency. Although XML implies an evolution over HTML, our aim for using XML was to overcome the limitations of flat files (plain ASCIII code). Figure 1 shows example codes for both approaches. The text on the left represents a typical flat file content with 2 individuals. On the right side we can observe the same information with some metadata regarding the individuals and the structure where they are embedded. Initially, data was stored in flat files for the Knowledge Mining Suite. Parsing, validating and analyzing data was difficult, dangerous and slow. With XML, however, all these activities are improved substantially. There are many alternative tools for parsing and representing XML data and anyone can have them for free. The time for developing and maintaining some parser is reduced to a few hours in learning how to use these tools. With XML schemas, documents can be validated automatically, providing consistency with expected structures. Schemas can be used also for building compliant files by external users. Besides this, data analysis process is enriched by several methods for searching information. 2 <matrix length= 2 > DaVinci <individual average= 5.0 > 5.0 DaVinci Rembrandt <individual/> 3.5 <individual average= 3.0 > Rembrandt <individual/> <matrix/> Fig. 1. Comparison between flat files and XML file contents 2 The Visual Symbolic Object Generator (VSOG) VSOG is a visual tool to generate symbolic data tables, in XML format, just by clicking the mouse. This tool allows the users to create symbolic data tables starting from relational databases such as SQLServer, ORACLE, Access, etc. You can also convert a SODAS file (*.sds) into a XML file for the Knowledge Mining Suite (*.kms). In Figure 2 VSOG user interface is shown. VSOG, as its name says, builds symbolic objects using a completely visual procedure, it is done with a wizard that guides the user by requesting the information while the system is running. Contrary to DB2SO, VSOG users don t have to know the SQL syntax. This is a great advantage, since it enlarges the quantity of users that can use the system. VSOG supports the first four SQL queries of the five types ([3, Bock and Diday (2000)]) supported by DB2SO. The fifth query type in the DB2SO (taxonomies) is not still supported by the version 1.0 of VSOG.

Fig. 2. VSOG user interface. The symbolic data table generated can be exported to a XML file, it offers a total flexibility for the manipulation of those symbolic data. This output will be the input for all the modules of KMS. The user can save the VSOG session, it allows the generation of the XML symbolic data tables without executing all steps of the wizard. Simultaneously, depending on the user requirements, he can be working in several sessions at the same time and with the same relational database. 3 Symbolic data analysis modules In this section a brief description of the symbolic methods supported by version 1.0 of the Knowledge Mining Suite (KMS) is presented. Each one of these modules, besides the graphics outputs, also generate as an output a XML file that can be used as an input for another module, so we can have a symbolic propagation of the analysis. 3.1 Interval Principal Components Analysis In this module we use a duality theorem for Centers Principal Components Analysis [4, Cazes, Chouakria, Diday and Schektman (1997)] which is an extension of classical principal component analysis, PCA, to interval data. Using the duality theorem this module used an improved algorithm for Centers PCA. Furthermore, this module uses the duality theorem to develop an efficient algorithm to compute the minimum and the

Fig. 3. Symbolic 3D principal cube to oils and fats example. maximum correlation r jk and r jk between the j th variable and the r th principal component. In [12, Rodríguez (2000)], you can find a complete presentation of the method. In this module we use a 3D representation for symbolic objects. This feature was implemented using OpenGL C++ library. OpenGL is a software interface to graphics hardware. This interface consists of about 250 distinct commands (about 200 in the core OpenGL and another 50 in the OpenGL utility library) that you use to specify the objects and operations needed to produce interactive three dimensional applications [15, Woo (1999)]. In Figure 3 the graphic output of this Interval PCA module of the, very well-known, oils and fats example is shown. 3.2 Histogram Principal Components Analysis In this module, we represent each histogram individual by a succession of k interval individuals (the first one included in the second one, the second one included in the third one and so on) where k is the maximum number of modalities taken by some variable in the input symbolic data table. Instead of representing the histograms in the factorial plane, we are going to represent the Empirical Distribution Function F Y defined in [3, Bock and Diday (2000)] associated with each histogram. In other words if we have a histogram variable Y on a set E = {a 1, a 2,...} of objects with domain Y represented by the mapping Y (a) = (U(a), π a ), for a E, where π a is a frequency distribution, then in the algorithm we will use the function F (x) = π i instead of the histogram. In [13, i / π i x

Fig. 4. Symbolic 3D principal cube to consop.sds example. Rodríguez, Diday and Winsberg (2000)], you can find a complete presentation of the method. The graphic output of this Histogram PCA module for the very well-known consop.sds example is shown in Figure 4. In this three-dimensional representation for histogram data type each individual is represented by a succession of cubes, one inside the other one, which represent the empirical distribution function. 3.3 Correspondence Factorial Analysis for Symbolic Multi Valued Variables One of the objectives of the Correspondence Factorial Analysis (CFA) is to explore the relations between the modalities of two qualitative variables by two dimensional representations. The objective for the module to Correspondence Factorial Analysis between two symbolic multi valued variables (CFASym) is the same one, but also there is some kind of uncertainty in the input data that will be reflected in the two dimensional or three dimensional representations, since each modality will be represented with a rectangle or a cube instead of a point, as usual. A symbolic variable Y is called multi valued one if its values Y (k) are all finite subsets [3, [Bock and Diday (2000)]. For the case of multi valued variables there will be individuals whose information regarding the assumed modality is diffuse. Example 1. Let X be the qualitative variable eyes-color with 3 modalities: green, blue and brown; it could be that the eyes color of the first individual is green or blue (but not both), that is to say X(1) =green or X(1) =blue. Let Y be the qualitative variable hair-color with 2 modalities blond and black, there could also be an individual whose hair color is not completely well defined, for instance, for the third individual, it might also be Y (3) =blond or Y (3) =black but not both). In this way there are two possible disjunctive complete tables for the variable X and two for Y, wich are:

1 0 0 0 1 0 0 1 0 1 0 0 1 X 1 = 1 0 0 0 0 1, X 0 0 1 2 = 1 0 0 0 0 1, Y 0 1 1 = 1 0 1 0, Y 0 1 2 = 0 1 1 0. 1 0 0 1 0 0 1 0 1 0 Using this information there will be 4 possible contingency tables among the variable X and Y, wich are: K 1 = X1Y t 1 = 2 1 0 0, K 2 = X2Y t 1 = 2 0 0 1, 1 1 1 1 K 3 = X1Y t 2 = 1 2 0 0, K 4 = X2Y t 2 = 1 1 0 1. 1 1 1 1 Taking the minimum and the maximum of the components of these 4 matrices, a contingency data table of interval type is obtained: [1, 2] [0, 2] K = [0, 0] [0, 1]. [1, 1] [1, 1] The main idea of the method used in this module is to carry out an interval factorial correspondence analysis on the matrix K, working with a similar idea just like in the Centers Method in principal component analysis for interval data. But there are some problems: The construction of the matrix K requires too many calculations. In fact K should be generated taking the minimum and the maximum of p m n m matrices. That is to say, p m n m products of matrices of sizes p m and m p should be made, where the variable X has p modalities taken in m individuals and the variable Y has n modalities and observed in m individuals. In a similar way that in centers method in PCA for interval data, all the tops of the hypercubes are projected in supplementary on the plane, then we choose the minimum and the maximum. The difference in this case is that the rows and columns should be projected in the same plane (simultaneous representation). Also, in this case the rows and columns of K are profiles, so the way the projections are made is not exactly the same that in a PCA. The centers of the matrix K are not integer, however the important thing is that the sum makes statistical sense, like in classical FCA. These problems are solved in [14, Rodríguez, Castillo, Diday and González (2003)].

3.4 Multidimensional Scaling for Interval Data: INTERSCAL Standard multidimensional scaling takes as input a dissimilarity matrix of general term δ ij which is a numerical value. In this module we input δ ij = [δ ij, δ ij ] where δ ij and δ ij are the lower bound and the upper bound of the dissimilarity between the stimulus/object S i and the stimulus/object S j respectively. As output instead of representing each stimulus/object on a factorial plane by a point, as in other multidimensional scaling methods, in the method used in this module each stimulus/object is visualized by a rectangle (or a cube in 3D representation), in order to represent dissimilarity variation. INTERSCAL generalizes the classical scaling method of Torgenson (1958) and Gower (1966) by looking for a method that produces results similar to the Tops Method in Principal Component Analysis proposed by [4, Cazes, Chouakria, Diday and Schektman (1997)]. 3.5 Symbolic Pyramidal Clustering Diday in [6, Diday (1984)] proposes the algorithm CAP to build numeric pyramids. Algorithms are also presented with this purpose in [1, Bertrand y Diday (1990)], [10, Gil (1998)] and [11, Mfoumoune (1998)]. Paula Brito in [2, Brito (1991)] proposes an algorithm that generalizes the algorithm to build numeric pyramids proposed by Diday to the symbolic case. This module uses two algorithm designed to build symbolic pyramids (CAPS and CAPSO), that is to say, a pyramid in which each node is a symbolic object. These algorithms also calculate the extension of each one of these symbolic objects and verify its completeness. The details of these two algorithms can be consulted in [12, Rodríguez (2000)]. 4 Classic data analysis modules The version 1.0 of the Knowledge Mining Suite (KMS) also has the following modules of classic data analysis. These classic modules also have three-dimensional graphic output. For example, the correlation circle of a classical PCA is shown in Figure 5. Principal Components Analysis: This module can be used to analyze the interrelations between a large number of variables to explain them in terms of their underlying common dimensions (factors). The objective is to find a way of condensing the information contained in a number of original variables into a smaller group of variables (factors) with minimum information loss. Hierarchical classification: This module can be used to find significant subgroups of individuals or objects. The specific objective is to classify a sample of entities (individuals or objects) based on their similarities, into a small number of groups that are mutually excluding. Correspondence Factorial Analysis: This correspondence analysis module can be apply in a contingency table containing a crossed tabulation of two category variables. The non-metrical data are later transformed into metrical to apply a dimensional reduction, similar to the reduction in the Principal components analysis.

Fig. 5. 3D correlation circle. Discriminant Analysis: The module of discriminant analysis is used to explain a qualitative variable (for example, a payer) based on a number of quantitative and/or qualitative variables known as explaining or predicting variables (salary, age, among others). Multiple Correspondence Analysis: The module of multiple factorial analysis that is a generalization of the factorial correspondence analysis can be applied to more than two qualitative variables. Univariate and Bivariable Data Analysis: This module allows for univariate or bivariate data analysis, estimating indexes such as the mean, median, variance, correlation, etc. In addition, it makes basic statistical graphs such as histograms and stem-leaf graphs, among others, as well as linear regressions. This module is used to make variable analyses prior to multi-variable analyses. Dynamic Clouds: The module of dynamic clouds can be used as an analytical technique to find significant subgroups of individuals or objects through the analysis of very large groups of data. Evolving Factorial analysis - STATIS Method: The Statis Method (statistical structure of a three-indexed table) is based on a group of data tables (individual-variables) obtained from different situations. A common application of this module is to use the same observations in different moments to create several data tables for the individuals and analyze their evolution.

Time series analysis: A time series is a set of observations x t, each one being recorded at a specified time t. A discrete time series (the type to which this module was primarily implemented) is one in which the set T 0 of time at which observations are made is a discrete set. References 1. Bertrand P et Diday E. Une generalisation des arbres hierarchiques: Les representations pyramidales, Statistique Appliquee, (3), 53-78, 1990. 2. Brito P. Analyse de donnees symboliques: Pyramides d heritage, Thèse de doctorat, Universite Paris 9 Dauphine, 1991. 3. Bock H-H. and Diday E. (eds.) Analysis of Symbolic Data. Exploratory methods for extracting statistical information from complex data. Springer Verlag, Heidelberg, 425 pages, ISBN 3-540-66619-2, 2000. 4. Cazes P., Chouakria A., Diday E. et Schektman Y. Extension de l analyse en composantes principales à des données de type intervalle, Rev. Statistique Appliquée, Vol. XLV Num. 3 pag. 5-24, Francia, 1997. 5. Chouakria A. Extension des méthodes d analysis factorialle à des données de type intervalle, Thèse de doctorat, Université Paris IX Dauphine, 1998. 6. Diday E. Une representation visuelle des classes empietantes. Rapport INRIA n. 291. Rocquencourt 78150, France, 1984. 7. Diday E. Introduction l approche symbolique en Analyse des Donnés. Première Journées Symbolique-Numérique. Université Paris IX Dauphine. Décembre 1987. 8. Diday E. L Analyse des Données Symboliques: un cadre théorique et des outils. Cahiers du CEREMADE, 1998. 9. Diday E. and Rodríguez, O. (eds.) Workshop on Symbolic Data Analysis. PKDD Lyon, 2000. 10. Gil A., Capdevila C. and Arcas A. On the efficiency and sensitivity of a pyramidal classification algorithm, Economics working paper 270, Barcelona, 1998. 11. Mfoumoune E. Les aspects algorithmiques de la classification ascendante pyramidale et incrementale. Thèse de doctorat, Universite Paris 9 Dauphine, 1998. 12. Rodríguez R. Classification et Modèles Linéaires en Analyse des Données Symboliques, Thèse de doctorat, Université Paris IX Dauphine, 2000. 13. Rodríguez O., Diday E. and Winsberg S., Generalization of the Principal Components Analysis to Histogram Data, PKDD2000, Lyon, 2000. 14. Rodríguez O., Castillo W., Diday E. and González J., Correspondence Factorial Analysis for Symbolic Multi Valued Variables, subjected for publication in Journal of Symbolic Data Analysis, 2003. 15. Woo M., Neider J., Davis T. and Shreiner D. OpenGL Programming Guide, Third edition, Addison Wesley, 1999.