Clustering in Ratemaking: Applications in Territories Clustering

Similar documents
University of Florida CISE department Gator Engineering. Clustering Part 4

Clustering Part 4 DBSCAN

Clustering in Data Mining

A Review on Cluster Based Approach in Data Mining

Unsupervised Learning

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm

CSE 5243 INTRO. TO DATA MINING

Clustering part II 1

Unsupervised Learning. Andrea G. B. Tettamanzi I3S Laboratory SPARKS Team

CSE 5243 INTRO. TO DATA MINING

Lesson 3. Prof. Enza Messina

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining: Concepts and Techniques. Chapter March 8, 2007 Data Mining: Concepts and Techniques 1

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

CHAPTER 4: CLUSTER ANALYSIS

Data Mining Cluster Analysis: Advanced Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining, 2 nd Edition

CSE 5243 INTRO. TO DATA MINING

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

Working with Unlabeled Data Clustering Analysis. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Clustering Algorithm (DBSCAN) VISHAL BHARTI Computer Science Dept. GC, CUNY

PAM algorithm. Types of Data in Cluster Analysis. A Categorization of Major Clustering Methods. Partitioning i Methods. Hierarchical Methods

Knowledge Discovery in Databases

Hierarchy. No or very little supervision Some heuristic quality guidances on the quality of the hierarchy. Jian Pei: CMPT 459/741 Clustering (2) 1

Clustering CS 550: Machine Learning

A Comparative Study of Various Clustering Algorithms in Data Mining

Clustering Techniques

Kapitel 4: Clustering

Clustering Algorithms In Data Mining

d(2,1) d(3,1 ) d (3,2) 0 ( n, ) ( n ,2)......

Unsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing

Cluster Analysis. Ying Shen, SSE, Tongji University

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017)

UNIT V CLUSTERING, APPLICATIONS AND TRENDS IN DATA MINING. Clustering is unsupervised classification: no predefined classes

CS Introduction to Data Mining Instructor: Abdullah Mueen

What is Cluster Analysis? COMP 465: Data Mining Clustering Basics. Applications of Cluster Analysis. Clustering: Application Examples 3/17/2015

Acknowledgements First of all, my thanks go to my supervisor Dr. Osmar R. Za ane for his guidance and funding. Thanks to Jörg Sander who reviewed this

Lecture-17: Clustering with K-Means (Contd: DT + Random Forest)

CS 2750 Machine Learning. Lecture 19. Clustering. CS 2750 Machine Learning. Clustering. Groups together similar instances in the data sample

Keywords Clustering, Goals of clustering, clustering techniques, clustering algorithms.

Clustering. Chapter 10 in Introduction to statistical learning

Cluster Analysis. Outline. Motivation. Examples Applications. Han and Kamber, ch 8

CS145: INTRODUCTION TO DATA MINING

COMP 465: Data Mining Still More on Clustering

Unsupervised Learning and Clustering

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Machine Learning (BSMC-GA 4439) Wenke Liu

Network Traffic Measurements and Analysis

Hierarchical Clustering

Data Mining. Clustering. Hamid Beigy. Sharif University of Technology. Fall 1394

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/09/2018)

Applied Clustering Techniques. Jing Dong

Course Content. What is Classification? Chapter 6 Objectives

Data Mining Algorithms

Data Informatics. Seon Ho Kim, Ph.D.

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Review of Spatial Clustering Methods

Cluster Analysis. Summer School on Geocomputation. 27 June July 2011 Vysoké Pole

Sponsored by AIAT.or.th and KINDML, SIIT

Lecture 7 Cluster Analysis: Part A

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Data Clustering Hierarchical Clustering, Density based clustering Grid based clustering

Olmo S. Zavala Romero. Clustering Hierarchical Distance Group Dist. K-means. Center of Atmospheric Sciences, UNAM.

Unsupervised Learning Partitioning Methods

Course Content. Classification = Learning a Model. What is Classification?

Unsupervised Learning and Clustering

DATA MINING AND WAREHOUSING

A Review: Techniques for Clustering of Web Usage Mining

数据挖掘 Introduction to Data Mining

Unsupervised Learning Hierarchical Methods

Clustering. Content. Typical Applications. Clustering: Unsupervised data mining technique

Applications. Foreground / background segmentation Finding skin-colored regions. Finding the moving objects. Intelligent scissors

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic

Analysis and Extensions of Popular Clustering Algorithms

CS570: Introduction to Data Mining

University of Florida CISE department Gator Engineering. Clustering Part 2

Analyzing Outlier Detection Techniques with Hybrid Method

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

Data Mining 4. Cluster Analysis

Clustering algorithms

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

DBSCAN. Presented by: Garrett Poppe

! Introduction. ! Partitioning methods. ! Hierarchical methods. ! Model-based methods. ! Density-based methods. ! Scalability

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

Data Mining and Data Warehousing Henryk Maciejewski Data Mining Clustering

CS 2750: Machine Learning. Clustering. Prof. Adriana Kovashka University of Pittsburgh January 17, 2017

Data Mining. Dr. Raed Ibraheem Hamed. University of Human Development, College of Science and Technology Department of Computer Science

Clustering Algorithms for Spatial Databases: A Survey

Efficient Parallel DBSCAN algorithms for Bigdata using MapReduce

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Clustering Algorithms for High Dimensional Data Literature Review

Supervised vs. Unsupervised Learning

Clustering Lecture 3: Hierarchical Methods

Scalable Varied Density Clustering Algorithm for Large Datasets

ISSN: [Saurkar* et al., 6(4): April, 2017] Impact Factor: 4.116

CS Data Mining Techniques Instructor: Abdullah Mueen

Based on Raymond J. Mooney s slides

DS504/CS586: Big Data Analytics Big Data Clustering II

Segmentation Computer Vision Spring 2018, Lecture 27

Transcription:

Clustering in Ratemaking: Applications in Territories Clustering Ji Yao, PhD FIA ASTIN 13th-16th July 2008

INTRODUCTION Structure of talk Quickly introduce clustering and its application in insurance ratemaking Review clustering methods and their applicability (problem) in insurance ratemaking Propose a clustering method (Exposure Adjusted Hybrid-EAH) and illustrates this method step-by-step using U.K. motor data Discuss some other considerations in clustering Conclusions and Questions Zurich - 2008 2

OVERVIEW OF CLUSTERING Definition of clustering clustering is the process of grouping a set of data objects into a cluster or clusters so that the data objects within the cluster are very similar to one another, but are dissimilar to objects in other clusters. Zurich - 2008 3

OVERVIEW OF CLUSTERING Purpose of Clustering in Insurance Better understand the data/trends Appropriate grouping Reduce the volatility of data and to make the rates stable over time Reduce the number of levels in rating factors Rates for vehicle Make the rate are reasonable and smooth the rates Rates of adjacent area Use of Clustering in Insurance Geographic Occupation/Trade Vehicle Product list Zurich - 2008 4

OVERVIEW OF CLUSTERING Nature of Insurance Dataset critical in choosing clustering method numerical vs. non-numerical Geographic, occupation, vehicle multi-dimensional Claim experience, and rating factors large noise not well-separated Conventional clustering method usually applied to well separated data the change between clusters could be gradual Zurich - 2008 5

OVERVIEW OF CLUSTERING Little Noise, Well separated, Clear Border Zurich - 2008 6

CLUSTERING METHODS Partitioning methods Hierarchical methods Density-based methods Grid-based methods Kernel methods Spectral methods Zurich - 2008 7

CLUSTERING METHODS -Partitioning Methods Partitioning Methods Broadly, this method organizes the data objects into a required number of clusters that optimizes certain similarity measure. Narrowly this is implemented by an iterative algorithm where the similarity measure is based on the distance between data objects. Generally, the algorithm of partitioning methods is as follows: i) choose initial data objects randomly as a center or a representation of clusters; ii) calculate the membership of each data object according to the present center or a representation of clusters; iii) update the center or representation of clusters that optimizes the total similarity measure; iv) repeat step (ii) if there is a change in the center or representation of clusters; otherwise stop. Zurich - 2008 8

CLUSTERING METHODS-Partitioning Methods K-Means Method The center of cluster, m, is defined as the mean of each cluster,, that is, Similarity function is the square-error function Example Looking for 3 clusters f i m k i 1 = x i= x C 1 i n i x C = x m i i 2 C i Zurich - 2008 9

CLUSTERING METHODS-Partitioning Methods Zurich - 2008 10

CLUSTERING METHODS-Partitioning Methods Zurich - 2008 11

CLUSTERING METHODS-Partitioning Methods Advantage easy to understand and apply time complexity of is lower than most other methods most widely used Disadvantage sensitive to noise and outliers difficult to choose the appropriate number of clusters tend to be sphere-shaped affected by the initial setting only converge to local optimal Zurich - 2008 12

CLUSTERING METHODS-Partitioning Methods K-Medoids Method defines the most centrally located data object of cluster Ci as the cluster center to calculate the squared-error function Advantage less sensitive to noise and outliers Disadvantage much higher run time to find the most centrally located data Other similar problem with k-means method Zurich - 2008 13

CLUSTERING METHODS-Partitioning Methods Expectation Maximization (EM) represents each cluster by a probability distribution Advantage time complexity is lower than K-Medoids method Disadvantage most of the problem K-Means suffers choice of probability distribution Zurich - 2008 14

CLUSTERING METHODS-Hierarchical Methods AGglomerative NESting (AGNES) clustering starts from sub-clusters that each includes only one data object. The distances between any two sub-clusters are then calculated and the two nearest sub-clusters are combined. This is done recursively until all sub-clusters are merged into one cluster that includes all data objects. Need to define the cluster-to-cluster similarity measure. Common ones are 1. Min distance 2. Max distance 3. Average distance Example Zurich - 2008 15

CLUSTERING METHODS-Hierarchical Methods Zurich - 2008 16

CLUSTERING METHODS-Hierarchical Methods Iteration Zurich - 2008 17

CLUSTERING METHODS-Hierarchical Methods The result is a dendrogram, looks like this Expectation Maximization (EM) Zurich - 2008 18

CLUSTERING METHODS-Hierarchical Methods DIvisia ANAlysis (DIANA) reverse to AGNES clustering starts from one cluster that includes all data objects. Then it iteratively chooses the appropriate border to split one cluster into two smaller sub-clusters that are least similar. result is also presented in dendrogram Zurich - 2008 19

CLUSTERING METHODS-Hierarchical Methods Advantage easy to understand and apply are less sphere-shaped than partitioning methods number of clusters is also chosen at a later stage Disadvantage the over-simplified similarity measure often gives erroneous clustering results irreversible high complexity of time Zurich - 2008 20

CLUSTERING METHODS-Hierarchical Methods Balanced Iterative Reducing and Clustering using Hierarchies (BIRTH) Key idea: compress the data objects into small sub-clusters in first stage and then perform clustering with these sub-clusters in the second stage The exact method is much more complex and use graph theory advantage greatly reduces the effective number of data objects that need to cluster reduces the time complexity. Disadvantage spherical shape clustering Zurich - 2008 21

CLUSTERING METHODS-Hierarchical Methods Clustering Using REpresentatives (CURE) use a fixed number of well-scattered data objects to represent each cluster and shrink these selected data objects towards their cluster centers at a specified rate. Advantage robust to outliers and has a better performance when clusters have non-spherical shape Disadvantage all parameters, such as number of representative data points of a cluster and shrinking speed, have a significant impact on the results Zurich - 2008 22

CLUSTERING METHODS-Hierarchical Methods CHAMELEON method more sophisticated measures of similarity such as interconnectivity and closeness are used uses a special graph partitioning algorithm to recursively partition the whole data objects into many small unconnected sub-clusters. Expectation Maximization (EM) Advantage more efficient than CURE in discovering arbitrarily shaped clusters of varying density Disadvantage the time complexity is quite high Zurich - 2008 23

CLUSTERING METHODS-Density-Based Methods Density-Based Spatial Clustering of Applications with Noise (DBSCAN) Basic idea: Example 1. Defines the density of a data object as the number of data objects within a certain distance of the data object. 2. If the density of a data object is larger then a threshold, this object is termed core. 3. Expand every cluster as far as the neighboring data object is a core object. 4. Outliers are discarded and not grouped to any clusters Threshold=3 Zurich - 2008 24

CLUSTERING METHODS-Density-Based Methods Zurich - 2008 25

CLUSTERING METHODS-Density-Based Methods Zurich - 2008 26

CLUSTERING METHODS-Density-Based Methods Advantage could find arbitrary shape of clusters Disadvantage efficiency of this method largely depends on parameters chosen by the user Expectation Maximization (EM) not work very well for a large or high-dimensional dataset Zurich - 2008 27

CLUSTERING METHODS- Density-Based Methods Ordering Points To Identify the Clustering Structure (OPTICS) This method produces a cluster ordering for a wide range of parameter settings Key Idea For each data object, find distance to the nearest core object, i.e. find the minimum distance that this data object could be clustered rather than discarded as noise. Ordering the data object from the minimum distance Advantage Solves the problem of dependency on parameters as in DBSCAN Zurich - 2008 28

CLUSTERING METHODS- Density-Based Methods DENsity-based CLUstEring (DENCLUE) This method is efficient for large datasets and high-dimensional noisy datasets; Many parameters to set and it may be difficult for the non-expert to apply; Zurich - 2008 29

CLUSTERING METHODS- Grid-Based Methods These methods quantize the space into a finite number of cells that form a grid structure on which all of the clustering operations are performed. Some features of cells are then used for clustering Combined with other methods Example Zurich - 2008 30

CLUSTERING METHODS- Grid-Based Methods Advantage fast processing time Disadvantage shape of the cluster is limited by the shape of grid ->smaller grid Advanced methods STING: explores statistical information WaveCluster: uses wavelet transform to store the information CLIQUE: discovers sub-clusters using the a priori principle Zurich - 2008 31

CLUSTERING METHODS- Kernel and Spectral Methods Kernel and Spectral Methods relatively new methods not easy for the non-expert to use and understand give no more advantages than other methods in actuarial application Expectation Maximization (EM) Zurich - 2008 32

CLUSTERING METHODS - Conclusions Simple method doesn t work well K-Mean, K-Medoids, AGNES, DIANA Complex methods Difficult to understand and apply Choice of parameters (no clear best answer for insurance dataset) Proposed method: adjust simple method to alleviate some problems Use the idea of BIRTH Zurich - 2008 33

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Other than the problem mentioned for BIRTH, there are other questions to ask: What to cluster? Claim frequency, severity, burning cost What data to cluster? ->GLM Volatility in data; Adjusted to exposure How to combine geographic and claim experience? Weighted distance measure Zurich - 2008 34

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD The whole procedure is: 1. Use the generalized linear model (GLM) technique to model the claim experience; 2. Calculate the residual of the GLM results as the pure effect of territory; 3. Use the partitioning method to generate small sub-clusters that contain highly similar data points; 4. Use the hierarchical method to derive the dendrogram clustering tree; 5. Choose an appropriate number of clusters and get corresponding clusters; 6. Repeat steps 3-5 with different initial setting to find a relatively consistent pattern in clusters; 7. Use the territory clustering results to re-run GLM and compare the results with that of Step 1. If there is large difference in the resulting relativities from GLM, then start again from Step 1; otherwise stop. Zurich - 2008 35

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Exposure adjusted distance measure ( μ,, μ, ) f E E 1 1 2 2 = 2 ( μ1 μ2) ( 1 E + 1 E ) based on Normal distributed assumption Geographic information Euclidean distance 1 2 2 2 (,,, ) = ( ) + ( ) g x y x y x x y y i i j j i j i j Haversine formula to take account of curve of earth surface Zurich - 2008 36

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Weighted distance measure g + w f ( ) ( ) Higher weight means more emphasis on claim history Other reasonable measures are possible Zurich - 2008 37

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Case study Use modified UK motor data for illustration purpose only The left graph show the adjusted claim experience by GLM Zurich - 2008 38

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Results of K-Means method: different initial setting Zurich - 2008 39

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Results of K-Means method: different weight Expectation Maximization (EM) W=0.1 W=1 W=10 Zurich - 2008 40

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Results of EAH method Zurich - 2008 41

EXPOSURE-ADJUSTED HYBRID (EAH) CLUSTERING METHOD Results of EAH method-different initial setting Expectation Maximization (EM) Zurich - 2008 42

MORE CONSIDERATION AND CONCLUSION Other consideration Existence of obstacles and constraints in clustering Change distance measure if severity or burning cost are used Validation of clustering results Conclusions Common clustering methods are introduced and compared: almost all methods have some problems if applied directly in insurance ratemaking; A new method is proposed: do not solve the problem thoroughly, but looks better than previous methods. Zurich - 2008 43

Questions? Zurich - 2008 44

Thank You Zurich - 2008 45