Hybridization of Expectation-Maximization and K-Means Algorithms for Better Clustering Performance

Size: px
Start display at page:

Download "Hybridization of Expectation-Maximization and K-Means Algorithms for Better Clustering Performance"

Transcription

1 BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 2 Sofa 2016 Prnt ISSN: ; Onlne ISSN: DOI: /cat Hybrdzaton of Expectaton-Maxmzaton and K-Means Algorthms for Better Clusterng Performance D. Raa Kshor 1, N. B. Venkateswarlu 2 1 Dept. of CSE, JNTU, Hyderabad, Telangana, Inda 2 Dept. of CSE, AITAM, Tekkal, Andhra Pradesh, Inda Emals: raakshor@gmal.com venkat_rtch@yahoo.com Abstract: The present work proposes hybrdzaton of Expectaton-Maxmzaton (EM) and K-means technques as an attempt to speed-up the clusterng process. Even though both the K-means and EM technques look nto dfferent areas, K-means can be vewed as an approxmate way to obtan maxmum lkelhood estmates for the means. Along wth the proposed algorthm for hybrdzaton, the present work also experments wth the Standard EM algorthm. Sx dfferent datasets, three of whch synthetc datasets, are used for the experments. Clusterng ftness and Sum of Squared Errors (SSE) are computed for measurng the clusterng performance. In all the experments t s observed that the proposed algorthm for hybrdzaton of EM and K-means technques s consstently takng less executon tme wth acceptable Clusterng Ftness value and less SSE than the standard EM algorthm. It s also observed that the proposed algorthm s producng better clusterng results than the Cluster package of Purdue Unversty. Keywords: Hybrdzaton, clusterng, K-means, mxture models, expectaton maxmzaton, clusterng ftness, sum of squared errors. 1. Introducton The Expectaton Maxmzaton (EM) algorthm s a model-based clusterng technque, whch attempts to optmze the ft between the gven data and some mathematcal model. Such methods are often based on the assumpton that the data s generated by a mxture of underlyng probablty dstrbutons [1]. The EM s an effectve, popular technque for estmatng mxture model parameters (lke cluster weghts and means) [7-9]. When compared to other clusterng algorthms, the EM algorthm demands more computatonal efforts although t produces exceptonally good results [20-22]. Many researchers 16

2 expermented on some varants (lke Generalzed EM (GEM), Expectaton Condtonal Maxmzaton (ECM), Sparse EM (SpEM), Lazy EM (LEM), Expectaton-Condtonal Maxmzaton Ether (ECME) algorthm and the Space Alternatng Generalzed Expectaton (SAGE) maxmzaton algorthms) n order to reduce the executon tme of EM algorthm [17, 18]. In [19], the use of Wnograd s algorthm s proposed to reduce the computatonal efforts of E-step and M-step of the standard EM algorthm. In [15], the use of mult-crtera models s proposed to desgn clusters wth the am of mproved clusterng performance. All ther experments amed at the speed-up of the EM algorthm by yeldng the same results as the Standard EM algorthm or better results wthout sacrfcng ts smplcty and stablty. As an attempt to speed up the clusterng process, the present work proposes the hybrdzaton of EM and K-means algorthms. The K-means algorthm s a very popular algorthm for data clusterng, whch ams at the local mnmum of the dstorton [2, 23]. EM s a model based approach, whch ams at fndng clusters such that maxmum lkelhood of each cluster s parameters s obtaned. In EM, each observaton belongs to each cluster wth a certan probablty [2]. The K-means algorthm s the 2nd domnantly used data mnng algorthm and the EM algorthm s the 5th domnantly used data mnng algorthm [3, 4, 24]. Though both K-means and EM technques look nto dfferent areas [2, 23], K-means can be vewed as an approxmate way to obtan maxmum lkelhood estmates for the means, whch s the goal of densty estmaton n EM [23, 24]. In the present work, along wth the proposed algorthm for hybrdzaton of EM and K-means technques, experments are carred out wth the standard EM algorthm. In all the experments, t s observed that the proposed algorthm for hybrdzaton of EM and K-means technques s consstently takng less executon tme to produce the clusterng results wth acceptable clusterng ftness value and less SSE n comparson to the standard EM algorthm. The proposed algorthm s also observed to produce clusterng results wth better performance than the Cluster Package of Purdue Unversty [26]. 2. The Standard EM (StEM) algorthm EM algorthm parttons the gven data by calculatng the maxmum a posteror prncple usng the condtonal probabltes [17]. Gven a guess for the parameter values, the EM algorthm calculates the probablty that each pont belongs to each dstrbuton and then uses these probabltes to compute a new estmate for the parameter. The EM algorthm teratvely refnes ntal mxture model parameter estmates to better ft the data and termnates at a locally optmal soluton. The standard EM [10, 11] for Gaussan Mxture Models (GMM) assumes that the algorthm wll estmate k class dstrbutons C, =1,, k. For each of the nput vectors X, =1,, N, the algorthm calculates the probablty P(C X ). The hghest probablty wll pont to the vector s class. 17

3 The EM algorthm works teratvely by applyng two steps: the Expectaton step (E-step) and the Maxmzaton step (M-step). Formally, ^ ( t ) { ( t ), ( t ), W ( t )}, 1,..., k, stands for successve parameter estmates. Gven a dataset of N, d-dmensonal vectors, the EM algorthm has to cluster them nto k groups. The mult-dmensonal Gaussan dstrbuton for the cluster C s parameterzed by the d-dmensonal mean column vector and d x d covarance matrx s gven as follows [10]: 1 ( ) T ( ) 1 1 X ( X ) 2 (1) P( X C ) e, d (2 ) where X s a sample column vector, the superscrpt T ndcates transpose of a column vector, s the determnant of, and ( ) 1 s ts matrx nverse of covarance matrx. The mxture model probablty densty functon [10] s (2) where W l s the weght of cluster C l Termnaton condton k P( X ) W P( X C ), l l1 As the termnaton condton, percentage change s computed usng the followng formula: t t 1 (3) Percentage change * 100, where t s the number of vectors assgned to new clusters n t-th teraton, and t+1 s the number of vectors assgned to new clusters n (t+1)-th teraton. The symbol * ndcates multplcaton. The algorthm termnates when the percentage change < 3. The EM algorthm for Gaussan Mxture Model [10] proceeds as follows: Step 1. Intalze mxture model parameters: set the current teraton t=0; set ntal weghts, W, to 1/k for all k clusters; select k vectors randomly from the dataset as the ntal cluster means, ; compute global covarance matrx for the dataset and set t to be the ntal covarance matrx,, for all clusters. Step 2 (E-step). Estmate the probablty of each class C, =1, 2,, k, gven a certan vector X, =1, 2,, N, for current teraton t usng the followng formula and assgn X to the cluster wth the maxmum probablty, (4) where X 1/2 WP( X C ) ( t) exp. W ( t) k PX ( ) 1/2 l l t Wl t l1 P( C ), ( ) exp. ( ) t 18

4 1 ( ( )) T 1 X t ( t )( X ( t )), 2 1 ( ( )) T 1 l X l t l ( t )( X l ( t )). 2 Each of the k clusters has ts mean ( ) and covarance ( ), = 1, 2,, k; W s the weght of -th cluster. Step 3 (M-step). Here, for -th cluster, update the parameter estmaton for the teraton t+1 as follows: P( C X ) X (5) 1 ( t 1) N, P( C X ) (6) (7) N 1 N T P( C X )( X ( t))( X ( t)) 1 ( t 1), N 1 P( C X ) N 1 W ( t 1) P( C X ). t N 1 Step 4. Compute percentage change usng (3). Step 5. Stop the process f the percentage change s < 3. Otherwse, set t=t+1 and repeat the Steps 2 up to 4 wth the updated parameters. 3. Hybrdzaton of EM and HbEMKM algorthms Though an effectvely used algorthm, the EM suffers from slow convergence as t requres heavly on computatonal efforts nvolved n repeated computaton of the many parameters lke covarance matrces, means and weghts of the clusters and repeated computaton of the nverses of covarance matrces of the clusters [3, 5, 24, 25]. On the other hand, the K-means algorthm can be used to smplfy the computaton and accelerate convergence as t requres only one parameter to compute,.e., cluster means [23, 24]. Whle assgnng ponts to the clusters, the EM maxmzes the lkelhood and the K-means mnmzes the dstorton wth respect to the clusters [23]. The algorthm for conventonal K-means s gven below [12]. Algorthm K-means Step 1. Select k vectors randomly from the dataset as the ntal cluster means,. Set the current teraton t=0. Step 2. Repeat. Step 3. Assgn each vector X from the dataset to ts closest cluster mean usng Eucldean dstance, (8) d 2 dst( X, ) ( xl l ), l1 19

5 where X s the -th vector n the dataset, s the mean of the cluster, and d s the number of dmensons of a data pont. Step 4. Re-compute the cluster means and set t=t+1. Step 5. Compute percentage change usng (3). Step 6. Untl percentage change s < 3. Step 7. End of K-means The present work, as an attempt to speed up the clusterng process, experments wth the Hybrdzaton of EM and K-Means algorthms (HbEMKM). Though both EM and K-means technques look nto dfferent areas [2, 23], K-means can be vewed as an approxmate way to obtan maxmum lkelhood estmates for the means, whch s the goal of densty estmaton n EM [23, 24]. Furthermore, K-means s formally equvalent to EM as K-means s a lmtng case of fttng data by a mxture of k Gaussans wth dentcal, sotropc covarance matrces ( = 2 I), when the soft assgnments of data ponts to mxture components are hardened to allocate each data pont solely to the most lkely component [3, 23]. A random space s sotropc f ts covarance functon depends on dstance alone [25]. In practce, there s often some conflaton of the two algorthms that K-means s sometmes used n densty estmaton applcatons due to ts more rapd convergence [23]. Also that selecton of ntal values s crtcal for EM, snce t most lkely converges to local maxma around the ntal values as EM uses maxmum lkelhood [2]. It may be a good practce, f the results of K-means are used as ntal parameter values for a subsequent executon of EM for the more exact computatons [23, 24]. The present work also experments on runnng the EM algorthm on the results of K-Means algorthm (KMEM). Along wth the proposed algorthm for hybrdzaton of EM and K-means technques, experments are carred out wth the standard EM algorthm and fnally performance comparson s made among the results of all experments. In all the experments same termnaton condton, dscussed Secton 2.1, s used. The pseudo code for the algorthm s gven below. Ths algorthm performs clusterng usng EM and K-means technques n the alternatve teratons tll termnaton. As part of the maxmzaton step for EM, cluster weghts, means and covarance matrces are calculated usng the results of K-means step. Algorthm HbEMKM N = number of samples n data n = number of samples n the -th cluster X = -th sample n data k = number of clusters W = weght of -th clusters = mean of -th cluster = covarance matrx of -th cluster Select k vectors randomly from the nput dataset as the ntal cluster means,. Frst, assgn each data vector X to the closest cluster wth mean, usng Eucldean dstance n the formula (8). Set sprogress = true 20

6 Repeat whle (sprogress == true) M-Step. Compute means and covarance matrces for = 1,, k, based on the results of K-Means step. Compute cluster weghts W = n /N for = 1,, k. E-Step. For each gven data vector X ( = 1, 2,, N), compute the cluster probablty P(C X ) for = 1,, k, usng (4). Assgn X to the cluster wth Max{ P( C X ); 1,..., k}. Compute percentage change usng (3). IF (percentage change >= 3) Compute cluster means for = 1,, k, usng (5). K-Means Step. Assgn each data vector X to the closest cluster wth mean, usng Eucldean dstance n the formula (8). Compute percentage change usng (3). IF (percentage change >= 3) Set sprogress = true ELSE Set sprogress = false End of nner IF ELSE Set sprogress = false End of outer IF End of Repeat Loop End of HbEMKM 4. Clusterng performance measure As a measure of clusterng performance, the Clusterng Ftness [13] s computed. The calculaton of Clusterng Ftness nvolves ntra-cluster smlarty, nter-cluster smlarty, and the experental knowledge,. The man obectve of any clusterng algorthm s to generate clusters wth hgher ntra-cluster smlarty and lower ntercluster smlarty [16]. So both the measures are taken nto consderaton for computng Clusterng Ftness. The computaton of Clusterng Ftness results n hgher values when the nter-cluster smlarty s low and results n lower values for when the nter-cluster smlarty s hgh. To make the computaton of Clusterng Ftness unbased, the value of s taken as 0.5 [13] Intracluster smlarty for the cluster C It can be quantfed va some functon of the recprocals of ntracluster rad wthn each of the resultng clusters. The ntracluster smlarty of a cluster C, 1 = = k, denoted as S tra(c ), s defned by 1 n (9) Stra ( C ). n 1 dst( IlCentrod) 1 21

7 Here, n s the number of tems n cluster C, 1 = l = n, I l s the l-th tem n cluster C l, and dst(i l, Centrod) calculates the dstance between I l and the centrod of C, whch s the ntracluster radus of C. To smooth the value of S tra(c ) and allow for possble sngleton clusters 1 s added to the denomnator and numerator Intracluster smlarty for one clusterng result C Denoted as S tra(c), Intracluster smlarty for one clusterng result C s defned by k S 1 tra ( C ) (10) Stra ( c). k Here, k s the number of resultng clusters n C Intercluster smlarty It can be quantfed va some functon of the recprocals of ntercluster rad of the clusterng centrods. The ntercluster smlarty for one of the possble clusterng results C, denoted as S ter(c), s defned by 1 n (11) Ster ( C). k 2 1 dst(centrod, Centrod ) 1 Here, k s the number of resultng clusters n C, 1 == k, Centrod s the centrod of the -th cluster n C, Centrod 2 s the centrod of all centrods of clusters n C. We compute ntercluster radus of Centrod by calculatng dst(centrod, Centrod 2 ), whch s dstance between Centrod, and Centrod 2. To smooth the value of S ter(c) and allow for possble all-nclusve clusterng result, 1 s added to the denomnator and the numerator Clusterng ftness The Clusterng Ftness (CF) for one of the possble clusterng results C s defned by 1 (12) CF * Stra. S Here, 0 < < 1 s an experental weght. The symbol * ndcates multplcaton. The present work consders = Sum of squared errors In the present work, Sum of Squared Errors (SSE) s also computed for all the clusterng results to measure the clusterng performance [6]. The clusterng performance s consdered to be good f the correspondng SSE s less when compared to the other clusterng technques. The SSE s computed usng the followng formula k 1 (13) SSE X. N 1 XC ter 22

8 Here, X s a vector from the dataset, s the means of the cluster C, k s the number of clusters and N s the number of vectors n the dataset. X denotes the dstance between X and. The obectve of clusterng s to mnmze the wthncluster sum of squared errors. The lesser the SSE, the better the goodness of ft s. 5. Experment and results Experments are carred out on the system wth Intel(R) Core K wth 3.50 GHz processor speed, 8GB RAM wth 1666FSB, Wndows 7 OS and usng JDK1.7.0_45. Separate modules are wrtten for each algorthm to observe the CPU tme for clusterng any dataset by keepng the same cluster seeds for all methods. I/O operatons are elmnated and tme observed s strctly for clusterng of the data (Table 1). Magc Gamma, Poker Hand, and Letter Recognton datasets are used for the present work from UCI ML dataset repostory [14]. An mportant ssue n evaluatng data analyss algorthms s the avalablty of representatve data. When real-lfe data s hard to obtan or when ts propertes are hard to modfy for testng varous algorthms, synthetc data becomes an appealng alternatve. The present work also uses three synthetc datasets that are generated by an algorthm for generatng multvarate normal random varables [27]. The frst synthetc dataset s generated assumng all clusters have dfferent means and dfferent covarance matrces. The second synthetc dataset s generated assumng some clusters have the same mean but dfferent covarance matrces. The thrd synthetc dataset s generated assumng some clusters have the same covarance matrx but dfferent means. Table 1. Datasets Data set Number of ponts Number of dmensons Letter Recognton Data 20, Magc Gamma data 19, Poker Hand data 1,025, Synthetc data-1 50, Synthetc data-2 50, Synthetc data-3 50, All the algorthms are studed by executng on each dataset by varyng number of clusters (.e., k=10, 11, 12, 13, 14, 15). The detals of executon tme, clusterng ftness and SSE of each algorthm are separately gven n the tables below for each dataset Observatons on letter recognton dataset The Tables 2, 3 and 4 consst of the executon tme, the cluster ftness and Sum of Squared Error (SSE), respectvely, of algorthms dscussed n sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Letter Recognton dataset. The observatons are also shown n the Fgs 1, 2 and 3. 23

9 Table 2. Executon tme of each clusterng method (s) Fg. 1. Letter recognton dataset: Executon tmes Table 3. Clusterng ftness of each clusterng method Fg. 2. Letter recognton dataset: Clusterng ftness Table 4. SSE of each clusterng method

10 Fg. 3. Letter recognton dataset: Sum of squared errors 5.2. Observatons on magc gamma dataset Tables 5, 6 and 7 consst of the executon tme, the cluster ftness and Sum of Squared Error (SSE), respectvely, of algorthms dscussed n Sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Magc Gamma dataset. The observatons are also shown n the Fgs 4, 5 and 6. Table 5. Executon tme of each clusterng method (s) Fg. 4. Magc gamma dataset: Executon tmes Table 6. Clusterng ftness of each clusterng method

11 Fg. 5. Magc gamma dataset: Clusterng ftness Table 7. SSE of each clusterng method Observatons on Poker hand dataset 26 Fg. 6. Magc gamma dataset: Sum of squared errors The tables 8, 9 and 10 consst of the executon tme, the cluster ftness and Sum of Squared Error (SSE), respectvely, of algorthms dscussed n Sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Poker Hand dataset. The observatons are also shown n Fgs 7, 8 and 9. Table 8. Executon tme of each clusterng method (s)

12 Fg. 7. Poker hand dataset: Executon tmes Table 9. Clusterng ftness of each clusterng method Fg. 8. Poker hand dataset: Clusterng ftness Table 10. SSE of each clusterng method

13 5.4. Observatons on Synthetc dataset-1 Fg. 9. Poker hand dataset: Sum of squared errors Tables 11, 12 and 13 consst of the executon tme, the cluster ftness and SSE, respectvely, of algorthms dscussed n Sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Synthetc dataset-1. The observatons are also shown n Fgs 10, 11 and 12. Table 11. Executon tme of each clusterng method (s) Fg. 10. Synthetc dataset-1: Executon tmes Table 12. Clusterng ftness of each clusterng method

14 Fg. 11. Synthetc dataset-1: Clusterng ftness Table 13. SSE of each clusterng method Observatons on synthetc dataset-2 Fg. 12. Synthetc dataset-1: Sum of squared errors Tables 14, 15 and 16 consst of the executon tme, the cluster ftness and Sum of Squared Error (SSE), respectvely, of algorthms dscussed n sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Synthetc dataset-2. The observatons are also shown n Fgs 13, 14 and 15. Table 14. Executon tme of each clusterng method (s)

15 Fg. 13. Synthetc dataset-2: Executon tmes Table 15. Clusterng ftness of each clusterng method Fg. 14. Synthetc dataset-2: Clusterng ftness Table 16. SSE of each clusterng method

16 5.6. Observatons on Synthetc dataset-3 Fg. 15. Synthetc dataset-2: Sum of squared errors Tables 17, 18 and 19 consst of the executon tme, the cluster ftness and SSE, respectvely, of algorthms dscussed n Sectons 2 and 3 and the Cluster Package of Purdue Unversty performed on Synthetc dataset-3. The observatons are also shown n Fgs 16, 17 and 18. Table 17. Executon tme of each clusterng method (s) Fg. 16. Synthetc dataset-3: Executon tmes Table 18. Clusterng ftness of each clusterng method

17 Fg. 17. Synthetc dataset-3: Clusterng ftness Table 19. SSE of each clusterng method Fg. 18. Synthetc dataset-3: Sum of squared errors 6. Concluson The proposed algorthm for Hybrdzaton of EM and K-means s consstently takng less computatonal tme wth all the tested datasets. The algorthm also takes less computatonal tme when compared to the Cluster package from Purdue Unversty. The proposed algorthm also produces the results wth hgher clusterng ftness values than the other algorthms ncludng Cluster It s also observed that the proposed algorthm produces the clusterng results wth lesser SSE values than the other algorthms ncludng the Cluster package. Therefore, the present work proposes Hybrdzaton of EM and K-means algorthms as a faster clusterng technque wth mproved performance. 32

18 R e f e r e n c e s 1. F r a l e y, C., A. E. R a f t e r y. Model-Based Clusterng, Dscrmnant Analyss, and Densty Estmaton. Journal of the Amercan Statstcal Assocaton, Vol. 97, 2002, No 458, p A d e b s, A. A., O. E. O l u s a y o, O. S. O l a t u n d e. An Exploratory Study of K-Means and Expectaton Maxmzaton Algorthms. Brtsh Journal of Mathematcs & Computer Scence, Vol. 2, 2012, No 2, pp W u, X., V. K u m a r, J. R. Q u n l a n, J. G h o s h, Q. Y a n g, H. M o t o d a, G. J. M c L a c h l a n, A. N g, B. Lu, P. S. Y u, Z.-H. Z h o u, M. S t e n b a c h, D. J. H a n d, D. S t e n b e r g. Survey Paper: Top 10 Algorthms n Data Mnng. Knowledge and Informaton Systems, Vol. 14, 2008, pp M a c Q u e e n, J. Some Methods for Classfcaton and Analyss of Multvarate Observatons. In: Proc. of 5th Berkeley Symposum on Mathematcs, Statstcs and Probablty, Vol. 1, 1967, pp M c L a c h l a n, G. J., T. K r s h n a n. The EM Algorthm and Extensons, 2/e. John Wley & Sons, Inc., Han, J., M. K a m b e r. Data Mnng Concepts and Technques, 2/e. New Delh, Inda, Elsever, Inc., Tan, P.-N., M. S t e n b a c h, V. K u m a r. Introducton to Data Mnng, 1/e. Pearson Educaton, Y e u n g, K. Y., C. F r a l e y, A. M u r u a, A. E. R a f t e r y, W. L. R u z z o. Model-Based Clusterng and Data Transformatons for Gene Expresson Data. Bonformatcs, Vol. 17, 2010, No 10, pp B r a d l e y, P. S., U. M. F a y y a d, C. A. R e n a. Scalng EM (Expectaton-Maxmzaton) Clusterng to Large Databases. Techncal Report, Mcrosoft Research, MSR-TR-98-35, K ö r t n g, T. S., L. V. D u t r a, L. M. G. F o n s e c a, G. J. E r t h a l. Assessment of a Modfed Verson of the EM Algorthm for Remote Sensng Data Classfcaton. In: Proc. of Iberoamercan Congress on Pattern Recognton (CIARP). São Paulo, Brazl, LNCS 6419, 2010, pp K ö r t n g, T. S., L. V. D u t r a, L. M. G. F o n s e c a, G. E r t h a l, F. C. da S l v a. Improvements to Expectaton-Maxmzaton Approach for Unsupervsed Classfcaton of Remote Sensng Data. GeoINFO, Campos do Jordão, SP, Brazl, A g g a r w a l, N., K. A g g a r w a l. A Md-Pont Based K-Mean Clusterng Algorthm for Data Mnng. Internatonal Journal on Computer Scence and Engneerng, Vol. 4, 2012, No 6, pp Han, X., T. Z h a o. Auto-K Dynamc Clusterng Algorthm. Journal of Anmal and Veternary Advances, Vol. 4, 2005, No 5, pp UCL Machne Learnng Repostory R a d e v a, I. Mult-Crtera Models for Clusters Desgn. Cybernetcs and Informaton Technology, Vol. 13, 2013, No 1, pp Rao, V. S. H., M. V. J o n n a l a g e d d a. Insurance Dynamcs A Data Mnng Approach for Customer Retenton n Health Care Insurance Industry. Cybernetcs and Informaton Technologes, Vol. 12, 2012, No 1, pp J o l l o s, F. X., M. N a d f. Speed-up for the Expectaton-Maxmzaton Algorthm for Clusterng Categorcal Data. J. Glob Optm, Vol. 37, 2007, pp M e n g, X.-L., D. V a n D y k. The EM Algorthm An Old Folk-Song Sung to a Fast New Tune. Journal of the Royal Statstcal Socety, Vol. 59, 1997, No 3, pp N a g e n d r a, K. D. J., J. V. R. M u r t h y, N. B. V e n k a t e s w a r l u. Fast Expectaton Maxmzaton Clusterng Algorthm. Internatonal Journal of Computatonal Intellgence Research, Vol. 8, 2012, No 2, pp J o l l o s, F.-X., M. N a d f. Speed-up for the Expectaton-Maxmzaton Algorthm for Clusterng Categorcal Data. Journal of Global Optmzaton, Vol. 37, 2007, No 4, pp

19 21. N e a l, R., G. E. H n t o n. A Vew of the EM Algorthm That Justfes Incremental, Sparse, and Other Varants, Learnng n Graphcal Models. MA, USA, Kluwer Academc Publshers, X u, R., D. W u n s c h II. Survey of Clusterng Algorthms. IEEE Transactons on Neural Networks, Vol. 16, 2005, No 3, pp K e a r n s, M., Y. M a n s o u r, A. N g. An Informaton-Theoretc Analyss of Hard and Soft Assgnment Methods for Clusterng, Uncertanty n Artfcal Intellgence. In: Proc. of 13th Conference Annual Conference on Uncertanty n Artfcal Intellgence (UAI-97), San Francsco, CA, Morgan Kaufmann, 1997, pp D u d a, R. O., P. E. H a r t, D. G. S t o r k. Pattern Classfcaton, 2/e. New Delh, Wley-Inda Edton, P o r c u, E m l o, M o n t e r o, J o s e-m a r t a, S c h l a t h e r, M a r t n. Advances and Challenges n Space-Tme Modellng of Natural Events. In: Lecture Notes n Statstcs. Vol Berln, Hedelberg, Sprnger-Verlag, Purdue Unversty Cluster Software. software/cluster/ 27. A m t a v a, G., H. S. W. K. P n n a d u w a. A FORTRAN Program for Generaton of Multvarate Normally Dstrbuted Random Varables. Computers & Geoscences, Vol. 13, 1987, No 3, pp

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Machine Learning. Topic 6: Clustering

Machine Learning. Topic 6: Clustering Machne Learnng Topc 6: lusterng lusterng Groupng data nto (hopefully useful) sets. Thngs on the left Thngs on the rght Applcatons of lusterng Hypothess Generaton lusters mght suggest natural groups. Hypothess

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

A Robust Method for Estimating the Fundamental Matrix

A Robust Method for Estimating the Fundamental Matrix Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

CMPS 10 Introduction to Computer Science Lecture Notes

CMPS 10 Introduction to Computer Science Lecture Notes CPS 0 Introducton to Computer Scence Lecture Notes Chapter : Algorthm Desgn How should we present algorthms? Natural languages lke Englsh, Spansh, or French whch are rch n nterpretaton and meanng are not

More information

An Efficient Genetic Algorithm with Fuzzy c-means Clustering for Traveling Salesman Problem

An Efficient Genetic Algorithm with Fuzzy c-means Clustering for Traveling Salesman Problem An Effcent Genetc Algorthm wth Fuzzy c-means Clusterng for Travelng Salesman Problem Jong-Won Yoon and Sung-Bae Cho Dept. of Computer Scence Yonse Unversty Seoul, Korea jwyoon@sclab.yonse.ac.r, sbcho@cs.yonse.ac.r

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Classifying Acoustic Transient Signals Using Artificial Intelligence

Classifying Acoustic Transient Signals Using Artificial Intelligence Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)

More information

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like: Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide Lobachevsky State Unversty of Nzhn Novgorod Polyhedron Quck Start Gude Nzhn Novgorod 2016 Contents Specfcaton of Polyhedron software... 3 Theoretcal background... 4 1. Interface of Polyhedron... 6 1.1.

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and

More information

Implementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status

Implementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status Internatonal Journal of Appled Busness and Informaton Systems ISSN: 2597-8993 Vol 1, No 2, September 2017, pp. 6-12 6 Implementaton Naïve Bayes Algorthm for Student Classfcaton Based on Graduaton Status

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

SIGGRAPH Interactive Image Cutout. Interactive Graph Cut. Interactive Graph Cut. Interactive Graph Cut. Hard Constraints. Lazy Snapping.

SIGGRAPH Interactive Image Cutout. Interactive Graph Cut. Interactive Graph Cut. Interactive Graph Cut. Hard Constraints. Lazy Snapping. SIGGRAPH 004 Interactve Image Cutout Lazy Snappng Yn L Jan Sun Ch-Keung Tang Heung-Yeung Shum Mcrosoft Research Asa Hong Kong Unversty Separate an object from ts background Compose the object on another

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

A fast algorithm for color image segmentation

A fast algorithm for color image segmentation Unersty of Wollongong Research Onlne Faculty of Informatcs - Papers (Arche) Faculty of Engneerng and Informaton Scences 006 A fast algorthm for color mage segmentaton L. Dong Unersty of Wollongong, lju@uow.edu.au

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

On Supporting Identification in a Hand-Based Biometric Framework

On Supporting Identification in a Hand-Based Biometric Framework On Supportng Identfcaton n a Hand-Based Bometrc Framework Pe-Fang Guo 1, Prabr Bhattacharya 2, and Nawwaf Kharma 1 1 Electrcal & Computer Engneerng, Concorda Unversty, 1455 de Masonneuve Blvd., Montreal,

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005 Exercses (Part 4) Introducton to R UCLA/CCPR John Fox, February 2005 1. A challengng problem: Iterated weghted least squares (IWLS) s a standard method of fttng generalzed lnear models to data. As descrbed

More information

EXTENDED BIC CRITERION FOR MODEL SELECTION

EXTENDED BIC CRITERION FOR MODEL SELECTION IDIAP RESEARCH REPORT EXTEDED BIC CRITERIO FOR ODEL SELECTIO Itshak Lapdot Andrew orrs IDIAP-RR-0-4 Dalle olle Insttute for Perceptual Artfcal Intellgence P.O.Box 59 artgny Valas Swtzerland phone +4 7

More information

Wavefront Reconstructor

Wavefront Reconstructor A Dstrbuted Smplex B-Splne Based Wavefront Reconstructor Coen de Vsser and Mchel Verhaegen 14-12-201212 2012 Delft Unversty of Technology Contents Introducton Wavefront reconstructon usng Smplex B-Splnes

More information

Detection of an Object by using Principal Component Analysis

Detection of an Object by using Principal Component Analysis Detecton of an Object by usng Prncpal Component Analyss 1. G. Nagaven, 2. Dr. T. Sreenvasulu Reddy 1. M.Tech, Department of EEE, SVUCE, Trupath, Inda. 2. Assoc. Professor, Department of ECE, SVUCE, Trupath,

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

Machine Learning. K-means Algorithm

Machine Learning. K-means Algorithm Macne Learnng CS 6375 --- Sprng 2015 Gaussan Mture Model GMM pectaton Mamzaton M Acknowledgement: some sldes adopted from Crstoper Bsop Vncent Ng. 1 K-means Algortm Specal case of M Goal: represent a data

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming CS 4/560 Desgn and Analyss of Algorthms Kent State Unversty Dept. of Math & Computer Scence LECT-6 Dynamc Programmng 2 Dynamc Programmng Dynamc Programmng, lke the dvde-and-conquer method, solves problems

More information

Understanding K-Means Non-hierarchical Clustering

Understanding K-Means Non-hierarchical Clustering SUNY Albany - Techncal Report 0- Understandng K-Means Non-herarchcal Clusterng Ian Davdson State Unversty of New York, 1400 Washngton Ave., Albany, 105. DAVIDSON@CS.ALBANY.EDU Abstract The K-means algorthm

More information

Hierarchical clustering for gene expression data analysis

Hierarchical clustering for gene expression data analysis Herarchcal clusterng for gene expresson data analyss Gorgo Valentn e-mal: valentn@ds.unm.t Clusterng of Mcroarray Data. Clusterng of gene expresson profles (rows) => dscovery of co-regulated and functonally

More information

A Post Randomization Framework for Privacy-Preserving Bayesian. Network Parameter Learning

A Post Randomization Framework for Privacy-Preserving Bayesian. Network Parameter Learning A Post Randomzaton Framework for Prvacy-Preservng Bayesan Network Parameter Learnng JIANJIE MA K.SIVAKUMAR School Electrcal Engneerng and Computer Scence, Washngton State Unversty Pullman, WA. 9964-75

More information

A Deflected Grid-based Algorithm for Clustering Analysis

A Deflected Grid-based Algorithm for Clustering Analysis A Deflected Grd-based Algorthm for Clusterng Analyss NANCY P. LIN, CHUNG-I CHANG, HAO-EN CHUEH, HUNG-JEN CHEN, WEI-HUA HAO Department of Computer Scence and Informaton Engneerng Tamkang Unversty 5 Yng-chuan

More information

Stability Region based Expectation Maximization for Model-based Clustering

Stability Region based Expectation Maximization for Model-based Clustering Stablty Regon based Expectaton Maxmzaton for Model-based Clusterng Chandan K. Reddy, Hsao-Dong Chang School of Electrcal and Computer Engneerng, Cornell Unversty, Ithaca, NY - 14853. Bala Rajaratnam Department

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

Clustering Algorithm Combining CPSO with K-Means Chunqin Gu 1, a, Qian Tao 2, b

Clustering Algorithm Combining CPSO with K-Means Chunqin Gu 1, a, Qian Tao 2, b Internatonal Conference on Advances n Mechancal Engneerng and Industral Informatcs (AMEII 05) Clusterng Algorthm Combnng CPSO wth K-Means Chunqn Gu, a, Qan Tao, b Department of Informaton Scence, Zhongka

More information

Assembler. Building a Modern Computer From First Principles.

Assembler. Building a Modern Computer From First Principles. Assembler Buldng a Modern Computer From Frst Prncples www.nand2tetrs.org Elements of Computng Systems, Nsan & Schocken, MIT Press, www.nand2tetrs.org, Chapter 6: Assembler slde Where we are at: Human Thought

More information

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research

More information

EVALUATION OF THE PERFORMANCES OF ARTIFICIAL BEE COLONY AND INVASIVE WEED OPTIMIZATION ALGORITHMS ON THE MODIFIED BENCHMARK FUNCTIONS

EVALUATION OF THE PERFORMANCES OF ARTIFICIAL BEE COLONY AND INVASIVE WEED OPTIMIZATION ALGORITHMS ON THE MODIFIED BENCHMARK FUNCTIONS Academc Research Internatonal ISS-L: 3-9553, ISS: 3-9944 Vol., o. 3, May 0 EVALUATIO OF THE PERFORMACES OF ARTIFICIAL BEE COLOY AD IVASIVE WEED OPTIMIZATIO ALGORITHMS O THE MODIFIED BECHMARK FUCTIOS Dlay

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

K-means Optimization Clustering Algorithm Based on Hybrid PSO/GA Optimization and CS validity index

K-means Optimization Clustering Algorithm Based on Hybrid PSO/GA Optimization and CS validity index Orgnal Artcle Prnt ISSN: 3-6379 Onlne ISSN: 3-595X DOI: 0.7354/jss/07/33 K-means Optmzaton Clusterng Algorthm Based on Hybrd PSO/GA Optmzaton and CS valdty ndex K Jahanbn *, F Rahmanan, H Rezae 3, Y Farhang

More information

Cell Count Method on a Network with SANET

Cell Count Method on a Network with SANET CSIS Dscusson Paper No.59 Cell Count Method on a Network wth SANET Atsuyuk Okabe* and Shno Shode** Center for Spatal Informaton Scence, Unversty of Tokyo 7-3-1, Hongo, Bunkyo-ku, Tokyo 113-8656, Japan

More information

Available online at ScienceDirect. Procedia Environmental Sciences 26 (2015 )

Available online at   ScienceDirect. Procedia Environmental Sciences 26 (2015 ) Avalable onlne at www.scencedrect.com ScenceDrect Proceda Envronmental Scences 26 (2015 ) 109 114 Spatal Statstcs 2015: Emergng Patterns Calbratng a Geographcally Weghted Regresson Model wth Parameter-Specfc

More information

An efficient method to build panoramic image mosaics

An efficient method to build panoramic image mosaics An effcent method to buld panoramc mage mosacs Pattern Recognton Letters vol. 4 003 Dae-Hyun Km Yong-In Yoon Jong-Soo Cho School of Electrcal Engneerng and Computer Scence Kyungpook Natonal Unv. Abstract

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Supervsed vs. Unsupervsed Learnng Up to now we consdered supervsed learnng scenaro, where we are gven 1. samples 1,, n 2. class labels for all samples 1,, n Ths s also

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE

SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE Dorna Purcaru Faculty of Automaton, Computers and Electroncs Unersty of Craoa 13 Al. I. Cuza Street, Craoa RO-1100 ROMANIA E-mal: dpurcaru@electroncs.uc.ro

More information

Load Balancing for Hex-Cell Interconnection Network

Load Balancing for Hex-Cell Interconnection Network Int. J. Communcatons, Network and System Scences,,, - Publshed Onlne Aprl n ScRes. http://www.scrp.org/journal/jcns http://dx.do.org/./jcns.. Load Balancng for Hex-Cell Interconnecton Network Saher Manaseer,

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article Avalable onlne www.jocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(6):2512-2520 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 Communty detecton model based on ncremental EM clusterng

More information

Query Clustering Using a Hybrid Query Similarity Measure

Query Clustering Using a Hybrid Query Similarity Measure Query clusterng usng a hybrd query smlarty measure Fu. L., Goh, D.H., & Foo, S. (2004). WSEAS Transacton on Computers, 3(3), 700-705. Query Clusterng Usng a Hybrd Query Smlarty Measure Ln Fu, Don Hoe-Lan

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Virtual Memory. Background. No. 10. Virtual Memory: concept. Logical Memory Space (review) Demand Paging(1) Virtual Memory

Virtual Memory. Background. No. 10. Virtual Memory: concept. Logical Memory Space (review) Demand Paging(1) Virtual Memory Background EECS. Operatng System Fundamentals No. Vrtual Memory Prof. Hu Jang Department of Electrcal Engneerng and Computer Scence, York Unversty Memory-management methods normally requres the entre process

More information

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier Some materal adapted from Mohamed Youns, UMBC CMSC 611 Spr 2003 course sldes Some materal adapted from Hennessy & Patterson / 2003 Elsever Scence Performance = 1 Executon tme Speedup = Performance (B)

More information

Maximum Variance Combined with Adaptive Genetic Algorithm for Infrared Image Segmentation

Maximum Variance Combined with Adaptive Genetic Algorithm for Infrared Image Segmentation Internatonal Conference on Logstcs Engneerng, Management and Computer Scence (LEMCS 5) Maxmum Varance Combned wth Adaptve Genetc Algorthm for Infrared Image Segmentaton Huxuan Fu College of Automaton Harbn

More information

BIN XIA et al: AN IMPROVED K-MEANS ALGORITHM BASED ON CLOUD PLATFORM FOR DATA MINING

BIN XIA et al: AN IMPROVED K-MEANS ALGORITHM BASED ON CLOUD PLATFORM FOR DATA MINING An Improved K-means Algorthm based on Cloud Platform for Data Mnng Bn Xa *, Yan Lu 2. School of nformaton and management scence, Henan Agrcultural Unversty, Zhengzhou, Henan 450002, P.R. Chna 2. College

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

The Shortest Path of Touring Lines given in the Plane

The Shortest Path of Touring Lines given in the Plane Send Orders for Reprnts to reprnts@benthamscence.ae 262 The Open Cybernetcs & Systemcs Journal, 2015, 9, 262-267 The Shortest Path of Tourng Lnes gven n the Plane Open Access Ljuan Wang 1,2, Dandan He

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

Intelligent Information Acquisition for Improved Clustering

Intelligent Information Acquisition for Improved Clustering Intellgent Informaton Acquston for Improved Clusterng Duy Vu Unversty of Texas at Austn duyvu@cs.utexas.edu Mkhal Blenko Mcrosoft Research mblenko@mcrosoft.com Prem Melvlle IBM T.J. Watson Research Center

More information

An Accurate Evaluation of Integrals in Convex and Non convex Polygonal Domain by Twelve Node Quadrilateral Finite Element Method

An Accurate Evaluation of Integrals in Convex and Non convex Polygonal Domain by Twelve Node Quadrilateral Finite Element Method Internatonal Journal of Computatonal and Appled Mathematcs. ISSN 89-4966 Volume, Number (07), pp. 33-4 Research Inda Publcatons http://www.rpublcaton.com An Accurate Evaluaton of Integrals n Convex and

More information

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices Steps for Computng the Dssmlarty, Entropy, Herfndahl-Hrschman and Accessblty (Gravty wth Competton) Indces I. Dssmlarty Index Measurement: The followng formula can be used to measure the evenness between

More information

Comparison of Heuristics for Scheduling Independent Tasks on Heterogeneous Distributed Environments

Comparison of Heuristics for Scheduling Independent Tasks on Heterogeneous Distributed Environments Comparson of Heurstcs for Schedulng Independent Tasks on Heterogeneous Dstrbuted Envronments Hesam Izakan¹, Ath Abraham², Senor Member, IEEE, Václav Snášel³ ¹ Islamc Azad Unversty, Ramsar Branch, Ramsar,

More information

Programming in Fortran 90 : 2017/2018

Programming in Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Exercse 1 : Evaluaton of functon dependng on nput Wrte a program who evaluate the functon f (x,y) for any two user specfed values

More information

Parameter estimation for incomplete bivariate longitudinal data in clinical trials

Parameter estimation for incomplete bivariate longitudinal data in clinical trials Parameter estmaton for ncomplete bvarate longtudnal data n clncal trals Naum M. Khutoryansky Novo Nordsk Pharmaceutcals, Inc., Prnceton, NJ ABSTRACT Bvarate models are useful when analyzng longtudnal data

More information