Linearized Cluster Assignment via Spectral Ordering

Size: px
Start display at page:

Download "Linearized Cluster Assignment via Spectral Ordering"

Transcription

1 Lnearzed Cluster Assgnment va Spectral Orderng Chrs Dng Computatonal Research Dvson, Lawrence Berkeley Natonal Laboratory, Berkeley, CA 9472 Xaofeng He Computatonal Research Dvson, Lawrence Berkeley Natonal Laboratory, Berkeley, CA 9472 Abstract Spectral clusterng uses egenvectors of the Laplacan of the smlarty matrx. They are most convenently appled to 2-way clusterng problems. When applyng to mult-way clusterng, ether the 2-way spectral clusterng s recursvely appled or an embeddng to spectral space s done and some other methods are used to cluster the ponts. Here we propose and study a K-way cluster assgnment method. The method transforms the problem to fnd valleys and peaks of a 1-D quantty called cluster crossng, whch measures the symmetrc cluster overlap across a cut pont along a lnear orderng of the data ponts. The method can ether determne K clusters n one shot or recursvely splt a current cluster nto several smaller ones. We show that a lnear orderng based on a dstance senstve objectve has a contnuous soluton whch s the egenvector of the Laplacan, showng the close relatonshp between clusterng and orderng. The method reles on the connectvty matrx constructed as the truncated spectral expanson of the smlarty matrx, useful for revealng cluster structure. The method s appled to newsgroups to llustrate ntroduced concepts; experments show t outperforms the recursve 2-way clusterng and the standard K-means clusterng. 1. Introducton In recent years spectral clusterng emerges as sold approach for data clusterng. Spectral clusterng ncludes a class of clusterng methods (Bach & Jordan, 23; Appearng n Proceedngs of the 21 st Internatonal Conference on Machne Learnng, Banff, Canada, 24. Copyrght 24 by the authors. Chan et al., 1994; Dng et al., 22; Hagen & Kahng, 1992; Mela & Xu, 23; Ng et al., 21; Sh & Malk, 2; Yu & Sh, 23) that use egenvectors of the Laplacan of the symmetrc matrx W = (w j ) contanng the parwse smlarty between data objects, j. Spectral clusterng has well-motvated clusterng objectve functons and many nterestng and useful propertes can be proved. Spectral clusterng s most convenently appled to 2-way clusterng problem usng a sngle egenvector. When applyng to mult-way (K-way) clusterng, there are two man approaches: (1) the 2-way spectral clusterng s recursvely appled or (2) an embeddng to spectral space usng several egenvectors s frst done and some other methods, such as K-means (Ng et al., 21; Zha et al., 22; Bach & Jordan, 23), are used to cluster the ponts. These cluster assgnment methods are ndrect (except perhaps a recent study (Yu & Sh, 23)). 2. Lnearzed cluster assgnment Here we propose and study a drect K-way cluster assgnment method. The method transforms the problem to one of fndng valleys and peaks of a 1-D quantty called cluster crossng, whch measures the cluster overlap across a cut pont along lnear orderng of data objects. In other words, the method lnearzes the clusterng assgnment problem. The lnearzed assgnment algorthm depends crucally on an algorthm for orderng objects based on a parwse smlarty metrc. The orderng s such that adjacent objects are smlar whle objects far away along the orderng are ds-smlar. We show that for such an orderng objectve functon the nverse ndex permutaton has a contnuous (relaxed) soluton whch s the egenvector of the Laplacan of the smlarty matrx. Ths spectral orderng approach has been prevously consdered for reducton of the envelope of a sparse

2 symmetrc matrx. (Barnard et al., 1993). Our contrbutons are (1) provdng a clear orderng objectve functon and a new dervaton, and (2) ntroducng a modfcaton that sgnfcantly mproves the orderng. Ths s dscussed n 3. The actual lnearzaton s performed va the cluster crossng, the sum of smlartes symmetrcally across a cut pont along the lnear orderng. Computatonally, ths s the sum along ant-dagonal drecton on W wthn a pre-specfed bandwdth. Detals are dscussed n 4. If the clusters n a dataset are well-separated,.e., the smlarty matrx W s nearly dsconnected, the clusterng crossng along the spectral orderng can easly detect the clusters. For datasets where clusters moderately or strongly overlap, cluster crossng drectly computed from the smlarty matrx W provdes weak sgnals for revealng cluster structure. The connectvty matrx (Dng et al., 22) provdes sharper cluster structure and s adoped n the lnearzed assgnment algorthm. Ths s brefly dscussed n 5. In summary, the lnearzed assgnment algorthm depends on three technques: () an orderng of the data objects, () cluster crossng, () the connectvty matrx. These are dscussed n the followng sectons. 3. Dstance senstve orderng Gven n objects and the smlartes between them W = (w j ), the objectve of orderng s to nsure that () adjacent objects are smlar () the larger the dstance between the objects, the less smlar the two objects are. The orderng s defned by the ndex permutaton π(1,2,,n) = (π 1,,π n ). For a vector x = (x 1,,x n ) T, the permuted vector s π(x) = (x π1,,x πn ) T. The permuted smlarty matrx s (πwπ T ) j = w π,π j. Let J l (π) = n l =1 w π,π +l represent the parwse smlartes between objects wth fxed dstance l on the permuted order. We defne the global orderng objectve as n 1 mn J(π), J(π) = l 2 J l (π) = π l=1 n 1 n l l 2 w π,π +l l=1 =1 (1) Here larger dstance smlartes are mnmzed more heavly than smaller dstances, to ensure that the larger the dstance between a par of objects, the less smlar these two objects. Let us compute the optmal π. Frst, let j = + l or l = j, J(π) can be rewrtten as J(π) = 1 ( j) 2 w π,π 2 j = 1 ( j) 2 w π,π 2 j π,π j,j Replacng π by n the summaton and notng that ndex s permuted to π 1, where π 1 s the nverse permutaton, we obtan = n2 8,j J(π) = 1 2,j (π 1 π 1 j ) 2 w j ( ) 2 π 1 (n + 1)/2 π 1 j (n + 1)/2 w j. n/2 n/2 For smplcty, we defne the shfted and rescaled nverse ndex permutaton q π 1 (n + 1)/2 { 1 n n/2 n, 3 n n,, n 1 n }, (2) whch satsfes q =, q 2 = 1. (3) where q s further scaled by q (n 3 /12 n/3) 1/2 q whch does not change the permutaton. Note that (q q j ) 2 w j = (q 2 +qj 2 2q q j )w j = 2q T (D W)q j j where D s a dagonal matrx wth each dagonal element beng the sum of the correspondng row (d = j w j). Therefore, we need to mnmze q T (D W)q for q takng those dscrete values of Eq.(2), subject to the constrants n Eq.(3). Usng a Lagrangan multpler for the second constrant n Eq.(3), mnmzaton of J(π) becomes mn q J1, J1 = qt (D W)q q T q (4) Fndng the optmal soluton for the dscrete values of q s a combnatoral optmzaton problem, and s lkely to have no polynomal-tme optmal algorthms. However a contnuous soluton for q can be computed. We relax the restrcton that q must take dscrete values of Eq.(2) n [ 1,1], and let q take contnuous values n [ 1,1]. Wth ths, J1 can be mnmzed by solvng an egenvalue problem. It s well-known that q s an egenvector of the equaton (D W)q = ζq. (5) Clearly q = 1 = (1,,1) T s an egenvector wth ζ =. All other egenvector are orthogonal to q,.e.,

3 the frst constrant n Eq.(3) s also satsfed. Therefore q1 s the desred contnuous soluton of the dstance senstve orderng. We note that earler work on sparse matrx envelope reducton (Barnard et al., 1993) based on dfferent motvaton, reaches the same egenvector soluton. Our contrbuton here s to ntroduce the dstance senstve objectve functon J(π), and provde the detaled dervaton usng shfted nverse permutaton vector π 1 to show that the soluton s q1. Thus π 1 can be unquely recovered from q1. A smple mplementaton to recover the permutatons s to sort the elements of q1 n ncreasng order. Ths sortng nduces the desred ndex permutaton π Now we make a crucal modfcaton on the above soluton whch (a) mproves the qualty of the soluton, and (b) makes a drect connecton to the scaled PCA and connectvty matrx n 5. The modfcaton s made on the constrants n Eq.(3); we weght each pont wth the degree d, the column sum of the smlarty matrx W. In graph theory (Chung, 1997), d s called the volume of node. The new constrants are X X q d =, q2 d = 1. (6) (more dscusson later). Wth these constrants, the mnmzaton problem of J(π) becomes mn J 2, q T q (D W )q J 2 =. qt Dq 3 (7) Relaxng q to contnuous values n ( 1, 1). the soluton for q satsfes the egenvalue equaton (D W )q = ζdq. (8) 15 2 Let q = D 1/2 z. Substtutng t nto Eq.(8), we obtan D 1/2 W D 1/2 z = λz, λ = 1 ζ. 3 (9) Ths s a standard egenvalue equaton. Thus the egenvectors zk and qk have the orthogonalty relaton ½ 1 f k = ℓ T T (1) zk zℓ = qk Dqℓ = δkℓ = f k 6= ℓ The trval egenvector s q = e wth ζ =. Thus the constrants n Eq.(6) are automatcally satsfed. Snce (D W ) s sempostve defnte, we have ζk, λk = 1 ζk 1. (11) In dstance-senstve orderng, we seek qk wth the smallest ζk, or the largest λk. The desred soluton for the permutaton π 1 s q1 () (the th element of q1 ), subject to the rescalng and a constant shft condton accordng to Eq.(2). Note that q1 () < q1 (j) = π 1 < πj (12) Fgure 1. The connectvty matrx of 5-newsgroup (see 5) s dsplayed usng the J 1 orderng of Eq.(4) (top), and usng the J 2 orderng of Eq.(7) (mddle). The orgnal cosne smlarty of the 5 newsgroups s shown n bottom. In Fgure 1, we show a matrx where J 1 orderng s compared to J 2 orderng. Clearly, J 2 orderng provdes a better dstance-senstve orderng. The values of the ntal orderng objectve J(π) are J(π)/hJ =.846, usng J 1 usng J 2 J(π)/hJ =.584, (13) J(π)/hJ =.949, usng random orderng P P where hj = ( j wj /n2 ) j ( j)2 s the expected

4 mean value for J(π). Another way to measure the effects of orderng s to use the bandwdth and envelope of a symmetrc sparse matrx C. The bandwdth b() at row s the largest dstance between the dagonal element and any nonzero element n row. The bandwdth of the entre matrx s the largest of b() and the envelope s the sum of b(). For the J 1 orderng of C, bandwdth = 495, envelope = 156,24. For the J 2 orderng of C, bandwdth = 32, envelope = 48, 65. Clearly, J2 orderng s better. J 2 orderng uses the weghted constrants of Eq.(6) whle J 1 uses the unweghted constrants of Eq.(3). To understand why the weghtng leads to better orderng, frst observe that objects wth large d wll get smaller q to balance out the equaton. Now we rewrte Eq.(2) as π 1 = q (n+1)/2 gnorng the overall scalng factor. Snce π 1 = {1,,n}, (n + 1)/2 s n the mddle. Smaller q ndcates π 1 s near the mddle, thus objects wth large d are more lkely to be permuted towards the mddle usng the weghted constrants of Eq.(6). Ths s favorable, snce objects wth large d are more lkely to have more edges; and movng these objects towards mddle decreases the dstances among these smlar objects, therefore mproves J(π). Connecton to spectral clusterng Note that egenvector of Eq.(5) s used n the Rato cut spectral clusterng (Hagen & Kahng, 1992) and egenvector of Eq.(8) s used n the normalzed cut (Sh & Malk, 2) and mn-max cut (Dng et al., 21) spectral clusterng. In dervng 2-way spectral clusterng, only the sgns of the cluster ndcator vector are useful and all objects n a cluster have the same magntude. Ths ndcator vector s then relaxed nto the egenvector. In our dervaton of spectral orderng, both the sgn and magntude of the scaled and shfted permutaton vector are useful, see Eq.(2). Snce an egenvector has both sgn and magntude, the relaxaton of the permutaton vector s therefore better qualty approxmaton than the relaxaton of cluster ndcator. From ths analyss, we beleve the better reason for the success of spectral clusterng s due to the orderng, nstead of relaxng the dscrete cluster ndcators. In fact, ths orderng perspectve s used n actual mplementaton of spectral clusterng (Hagen & Kahng, 1992; Sh & Malk, 2): one frst sort q 1 to provde a lnear orderng, then along ths orderng, search for the cut that optmzes the cluster objectve functon. Thus our orderng analyss provdes a deeper understandng of spectral clusterng. Our results ndcates that J 2 orderng s better than J 1 orderng. Ths s due to the weghtng of d, the node degree, n Eq.(6). Smlar motvaton s used n normalzed cut. Let s 12 be the cut between two subgraphs C 1,C 2. All three graph clusterng objectve functon can be wrtten as J = s 12 /a 1 + s 12 /a 2. For Rato cut, a k = C k 1. For normalzed cut, a k = C k d. Ths weghtng of the subgraph volume mproves upon the smple weghtng of the subgraph sze n rato cut. For mn-max cut, a k =,j C k w j, the sum of edge weghts nsde C k. That J 2 orderng s better than J 1 orderng mples that normalzed cut and mn-max cut n general provdes a better clusterng than rato cut. Ths fact s observed n experments (Sh & Malk, 2; Dng et al., 21). It s sometmes happens that there s a symmetry among several nodes n the graph,.e., G(W)G T = W, where G s a permutaton specfyng an element n the nvarant symmetry group. In ths case, an egenvector of W,D W,D 1/2 WD 1/2 wll have several nodes wth the same value. If ths happens, nether the orderng nor clusterng problems can be unquely determned. Ths s not necessarly a weakness of the spectral methods, although t become obvous from the perspectve of the egenvector. In practce, ths happens rarely for weghted graphs. 4. Cluster crossng and assgnment We start wth cluster overlap. Gven two clusters C k,c l, the cluster overlap can be defned as the sum of parwse assocatons between two clusters, s kl = C k,j C l w j. (14) In spectral clusterng, s kl are mnmzed. Cluster overlap nvolves all C k C l parwse smlartes. We defne cluster crossng as the sum of a small fracton of the parwse smlartes. Ths s aded by lnear orderng data ponts. Gven a lnear order o of all objects, at each ste of the order, we can sum over w j wthn a wndow sze 2m + 1 across the ste, ρ() = m j=1 w o( j),o(+j) Ths corresponds to sum along the ant-dagonal drectons n the smlarty matrx W wth a bandwdth m. There are 2n 1 ant-dagonals n a matrx, among them ρ() are n full-step ant-dagonals. We also utlze

5 the other n 1 half-step ant-dagonals,.e., ρ( ± 1/2) = m w o( j),o(+j±1). j=1 The fnal crossng s the weghted average ρ() = ρ( + 1/2)/4 + ρ()/2 + ρ( 1/2)/4. (When s close to the two ends,.e., n m or m, the sum should be properly weghted to reflect the fact that the number of smlartes n the sum s less than the normal case.) Clearly, cluster crossng ρ() should have a mnmum at the cluster boundary between C k,c l. As moves away from the boundary, ρ() ncreases. Ths form the bass of the lnearzed cluster assgnment. Ths approach works for K > 2 as well as for K = 2; In essence, t reformulate a problem of K-way clusterng wth parwse smlartes nto a K-way clusterng problem n 1-dmenson (a) (b) (c) Fgure 2. Crossng curves for dataset A. Top: computed based on the smlarty matrx W shown n Fg.1(bottom). Mddle: computed based on the connectvty matrx C usng J 2 orderng shown n Fg.1(mddle). Bottom: computed based on C usng J 1 orderng shown n Fg.1(top). To llustrates the basc deas n ths approach. we compute the crossng of the matrx C shown n Fgure 1(bottom). The crossng curves are shown n Fgure 2. For matrx C usng J 2 orderng, the crossng (mddle panel n Fgure 2) exhbt clearly the fve-cluster structure. In cluster crossng curve, the valleys are more mportant than the peaks. Ths s because between two consecutve valleys, there could be several overlappng clusters so that the peaks are not as pronounced as the valleys. Usng the valleys, we can clearly separate sets of clusters (composte cluster). Ths suggest a dvde-and-conquer approach,.e., recursvely apply the algorthm on each set of clusters, untl the total number of cluster reach the pre-specfed K, or some other crtera are met (lke the top-down dvsve clusterng approach). For example n Fgure 1 (mddle), the 2nd and 3rd clusters overlap slghtly, and the correspondng valley pont n the crossng curve (mddle panel n Fgure 2) s not as low as others (although unambguously clear). We mght consder these two cluster as one composte cluster. Thus we cut the crossng nto 4 clusters at present round and cut the composte cluster n next round. 5. Scaled prncpal components and connectvty matrx Gven a symmetrc smlarty matrx W, one can obtan a spectral decomposton to get the prncpal component analyss (PCA), a wdely used technque n multvarate statstcs. In scaled PCA proposed n (Dng et al., 22), one performs the followng spectral decomposton W = D 1/2 (D 1/2 WD 1/2 )D 1/2 K = D 1/2 ( z kˆλk z T k )D 1/2 k=1 here the spectral decomposton s performed on the scaled matrx Ŵ = D 1/2 WD 1/2. Clearly, the egenvectors z k are governed by Eq.(9). The magntudes of all egenvalues are less than 1, as n Eq.(11). The connectvty matrx C s obtaned by truncatng the PCA expanson at K terms and settng the egenvalues to unty, λ k = 1, as K K C = D 1/2 z k z T k D 1/2 = D q k q T k D (15) k=1 k=1 where q k = D 1/2 z k s called scaled prncpal components due to ts smlarty to the usual PCA. q s governed by Eq.(8), and s closely related to spectral clusterng. It s shown va a perturbaton analyss that C has a so-called self-aggregaton property that connectvtes (matrx elements n C) between dfferent clusters are suppressed whle connectvtes wthn clusters are enhanced. Thus C s useful for revealng cluster structure. Connectvty matrx approach nvolves a nose reducton procedure. The probablty that two objects, j

6 belong to the same cluster s p j = C j /C 1/2 C 1/2 jj. To reduce nose one set C j = f p j < β, (16) where β =.8. For a range of problems, β =.5.9 leads to very smlar results. As an llustraton, the orgnal smlarty matrx (based on 5 newsgroups n 6) s shown n Fgure 1(bottom) usng J 2 orderng, where cluster structure s not apparent. Connectvty matrx (shown n Fgure 1) constructed based on ths smlarty matrx has clear cluster structure. 6. Complete assgnment algorthm Prespecfy K as the numnber of clusters, and set bandwdth m = n/k (or the expected largest cluster sze). The complete algorthm s as follows: (1) Compute connectve matrx C; (2) Compute the J 2 orderng of C; (3) Compute the crossng of C based on J 2 orderng. (4) Locate valley ponts n the crossng curve. Assgn each regon sandwched between two valley ponts or ends to one composte cluster. (5) If the total number of current composte cluster s less than K, recursvely apply the algorthm to the largest to further splt t. Note that ths recursve clusterng algorthm dffers from the usual recursve 2-way clusterng n that, a current composte cluster s parttoned nto several clusters dependng on the crossng curve, not restrcted to 2 clusters. It s possble that all K clusters are dentfed usng one crossng curve as n 4. Thus the total number of recurson s less than or equal to K-1, whch s requred by the usual recursve 2-way clusterng. In step (5), the choce of next cluster to splt s based on the largest (sze) cluster. Ths smple choce s more orented towards cluster balance. More refned choces for cluster splt s dscussed n (Dng & He, 22). The man advantage of ths approach s that clusters are formed consstently. In 2-way recursve clusterng, each current cluster s formed va a certan clusterng objectve functon whch s correctly motvated for only true clusters, not for composte clusters. For example n normalzed cut, the cluster objectve functon for K-way clusterng can not be recursvely constructed from the 2-way clusterng objectve. Therefore the 2- way recursve procedure s only a heurstc for K-way clusterng. In our lnearzed assgnment, cluster are assgn based on the crtera that the connectvty between clusters are small, whch s vald for both true clusters and composte clusters. 7. Experments The lnearzed cluster assgnment method s appled to Internet newsgroup artcles. A 2-newsgroup dataset s from 11/www/nave-bayes.html. Word-document matrx s frst constructed. 1 words are selected accordng to the mutual nformaton between words and documents n unsupervsed manner. Standard tf.df term weghtng s used. Each document s normalzed to 1. We focus on two sets of 5-newsgroup combnatons. The choce of K = 5 s to have some varety n the recursve steps (we avod K = 4,8). These two newsgroup combnatons are lsted below: A B NG2: comp.graphcs NG2: comp.graphcs NG9: rec.motorcycles NG3: comp.os.ms-wndows NG1: rec.sport.baseball NG8: rec.autos NG15: sc.space NG13: sc.electroncs NG18: talk.poltcs.mdeast NG19: talk.poltcs.msc Datasets wth well separated clusters are easy to handle; we are nterested n clusterng medum and large overlappng clusters. To measure cluster separaton, we compute s kl and defne the symmetrcally scaled cluster overlap as the cluster separaton ndex between clusters C k,c l as µ kl = s kl / s kk s ll. The over-all separaton s defned as 2 µ = K(K 1) k,l;k l µ kl. Clearly µ 1 and µ kk = 1. For a complete graph µ kl = 1 and µ = 1. Cluster overlap s kl and separaton ndex µ kl can be convenently stored n a matrx S, where S(upper-rght trangle ncludng dagonals) = s kl and S(lower-left trangle)= µ kl. For datasets A,B, ther overlap-separatons are S(A) = S(B) = Ther average separaton are µ(a) =.695, µ(b) =.755. These nformaton are useful. For example, for dataset B, NG18 (mdeast) s a coherent cluster because s 55 = 996 s relatvely large, whereas NG13(electroncs) s less coherent because s 44 = 472 s relatvely small. The overlap between NG2 (graphcs)

7 and NG3(wndows OS) s relatvely large: µ 12 =.578 whle the overlap between NG8 (auto) and NG13 (electroncs) s also large: µ 34 =.483. Overall, dataset A s moderately overlappng and dataset B s strongly overlappng. The above s based on a random sample of documents from the newsgroups. Each cluster has 1 documents. To accumulate suffcent statstcs, for each newsgroup combnaton, we generate 5 samples and average ther performance. Dataset A The cosne smlarty matrx W among documents shown n Fgure 1(bottom). The cluster structure s not clear from the smlarty matrx W. The connectvty matrx C s dsplayed n Fgure 1(top) whch exhbts the cluster structure. Clusterng crossngs based on W and C are shown n Fgure 2. The crossng for C based on J 2 orderng (mddle panel) shows clear cluster structure, whereas crossng for C based on J 1 orderng (bottom pannel) shows less clear cluster structure. Crossng for W (top pannel) show even less cluster structure. Based on the crossng of C usng J 2 orderng, local mnma n the valleys are dentfed usng a smple smoothng procedure, where the new smoothed value on each pont s the average of old values on 5 nearest ponts. Ths smoothng procedure overcomes the local abrupt changes and automatcally compute the more stable or consensus valley ponts, although the dfference wth un-smoothed one s often small. Data ponts between two valleys or ends are assgned to one cluster. Thus all ponts are assgned nto 5 clusters n one shot. Each of newsgroup artcle s cluster label s known (although not necessarly perfect). Usng ths, the confuson matrx T = (t kl ) for the clusterng results are computed, where t kl = number of ponts belongng to cluster k but clustered to cluster l. Based on the lnearzed cluster assgnment results, T s computed as T = For ths results, the clusterng accuracy, Q = k t kk/n = 9.5%. The clusterng experment s repeated for 5 dfferent random samples. The accuracy for ths lnearzed orderng approach s lsted n Table 1. For the same Table 1. Clusterng accuracy as of dfferent methods on the 5-newsgroup datasets. LA: lnearzed assgnment; R2W: recursve 2-way clusterng; usng MnMaxCut wth cluster choce for splt based on largest sze cluster; K-means. Method LA R2W K-means Data A 89.% 82.8% 75.1% Data B 75.7% 67.2% 56.4% dataset, the cluster accuracy usng recursve 2-way spectral clusterng and standard K-means are also lsted n Table 1. One see that the lnearzed assgnment outperform slghtly over the recursve 2-way clusterng and sgnfcantly over the K-means. Dataset B The cosne-smlarty matrx W s shown n Fgure 3 (top panel, usng J 2 orderng). The connectvty matrx C of ths dataset s shown n Fgure 4. The overlap between NG2 (computer graphcs) and NG3 (Wndows OS) s large; the overlap between NG8 (autos) and NG13 (electroncs) s large as well. These are expected from the cluster separaton ndexes S(B), and also can be confrmed by nspectng the cosnesmlarty W shown n Fgure 3 (bottom panel), usng J 2 orderng based on C. The crossng based on C s shown n Fgure 5 (top panel). The lower panels show the crossng curves after four successve applcatons of smoothng. Based on ths crossng, we can dentfy three composte clusters by two clear and low-lyng valley ponts. The two large composte clusters are further clustered usng the same lnearzed algorthm. Repeatng the experments on 5 random samples from dataset B, the clusterng accuracy s lsted n Table 1. The lnearzed assgnment outperforms the recursve 2-way spectral clusterng and the standard K-means. 8. Summary In summary, we propose and study a drect K-way cluster assgnment method that lnearze the clusterng problem nto 1-D clusterng crossng curve. The method depends on an effectve lnear orderng provded by the spectral orderng. We prove a clear dervaton of the dstance senstve orderng and show the shfted and scaled ndex permutaton vector s relaxed nto egenvectors of the Laplacan of the smlarty matrx. Our results provdes a deeper nsghts to spectral clusterng as well. Ths work s supported by U.S. Department of Energy,

8 (a) Fgure 4. The connectvty matrx of dataset B. (b) Fgure 3. The cosne smlarty matrx W of dataset B, dsplayed usng (a) the J 2 orderng based on W, (b) the J 2 orderng based on the connectvty matrx C n Fg.(4). Offce of Scence, Offce of Laboratory Polcy and Infrastructure, through an LBNL LDRD, under contract DE-AC3-76SF98. References Bach, F. R., & Jordan, M. I. (23). Learnng spectral clusterng. Neural Info. Processng Systems 16 (NIPS 23). Barnard, S. T., Pothen, A., & Smon, H. D. (1993). A spectral algorthm for envelope reducton of sparse matrces. Proc. Supercomputng 93, IEEE, Chan, P., M.Schlag, & Zen, J. (1994). Spectral k-way rato-cut parttonng and clusterng. IEEE Trans. CADIntegrated Crcuts and Systems, 13, Chung, F. (1997). Spectral graph theory. Amer. Math. Socety. Dng, C., & He, X. (22). Cluster merge and splt n herarchcal clusterng. Proc. IEEE Int l Conf. Data Mnng, Dng, C., He, X., Zha, H., Gu, M., & Smon, H. (21). A mn-max cut algorthm for graph parttonng and data clusterng. Proc. IEEE Int l Conf. Data Mnng. Dng, C., He, X., Zha, H., & Smon, H. (22). Unsupervsed learnng: self-aggregaton n scaled prncpal component space. Proc. 6th European Conf. Prncples of Fgure 5. Crossngs computed based on the connectvty matrx n Fg.(4). Frst curve s the crossng. Lower curves show the smoothng of the crossngs. The two ponts n bottom curve ndcate the cut ponts. Data Mnng and Knowledge Dscovery (PDKK 22), Hagen, L., & Kahng, A. (1992). New spectral methods for rato cut parttonng and clusterng. IEEE. Trans. on Computed Aded Desgn, 11, Mela, M., & Xu, L. (23). Multway cuts and spectral clusterng. U. Washngton Tech Report. Ng, A., Jordan, M., & Wess, Y. (21). On spectral clusterng: Analyss and an algorthm. Proc. Neural Info. Processng Systems (NIPS 21). Sh, J., & Malk, J. (2). Normalzed cuts and mage segmentaton. IEEE. Trans. on Pattern Analyss and Machne Intellgence, 22, Yu, S. X., & Sh, J. (23). Multclass spectral clusterng. Int l Conf. on Computer Vson. Zha, H., Dng, C., Gu, M., He, X., & Smon, H. (22). Spectral relaxaton for K-means clusterng. Advances n Neural Informaton Processng Systems 14 (NIPS 1),

Graph-based Clustering

Graph-based Clustering Graphbased Clusterng Transform the data nto a graph representaton ertces are the data ponts to be clustered Edges are eghted based on smlarty beteen data ponts Graph parttonng Þ Each connected component

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Hierarchical clustering for gene expression data analysis

Hierarchical clustering for gene expression data analysis Herarchcal clusterng for gene expresson data analyss Gorgo Valentn e-mal: valentn@ds.unm.t Clusterng of Mcroarray Data. Clusterng of gene expresson profles (rows) => dscovery of co-regulated and functonally

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Chapter 6 Programmng the fnte element method Inow turn to the man subject of ths book: The mplementaton of the fnte element algorthm n computer programs. In order to make my dscusson as straghtforward

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 5 Luca Trevisan September 7, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 5 Luca Trevisan September 7, 2017 U.C. Bereley CS294: Beyond Worst-Case Analyss Handout 5 Luca Trevsan September 7, 207 Scrbed by Haars Khan Last modfed 0/3/207 Lecture 5 In whch we study the SDP relaxaton of Max Cut n random graphs. Quc

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

EXTENDED BIC CRITERION FOR MODEL SELECTION

EXTENDED BIC CRITERION FOR MODEL SELECTION IDIAP RESEARCH REPORT EXTEDED BIC CRITERIO FOR ODEL SELECTIO Itshak Lapdot Andrew orrs IDIAP-RR-0-4 Dalle olle Insttute for Perceptual Artfcal Intellgence P.O.Box 59 artgny Valas Swtzerland phone +4 7

More information

New l 1 -Norm Relaxations and Optimizations for Graph Clustering

New l 1 -Norm Relaxations and Optimizations for Graph Clustering Proceedngs of the Thrteth AAAI Conference on Artfcal Intellgence (AAAI-6) New l -Norm Relaxatons and Optmzatons for Graph Clusterng Fepng Ne, Hua Wang, Cheng Deng 3, Xnbo Gao 3, Xuelong L 4, Heng Huang

More information

Sorting Review. Sorting. Comparison Sorting. CSE 680 Prof. Roger Crawfis. Assumptions

Sorting Review. Sorting. Comparison Sorting. CSE 680 Prof. Roger Crawfis. Assumptions Sortng Revew Introducton to Algorthms Qucksort CSE 680 Prof. Roger Crawfs Inserton Sort T(n) = Θ(n 2 ) In-place Merge Sort T(n) = Θ(n lg(n)) Not n-place Selecton Sort (from homework) T(n) = Θ(n 2 ) In-place

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like: Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

Simulation Based Analysis of FAST TCP using OMNET++

Simulation Based Analysis of FAST TCP using OMNET++ Smulaton Based Analyss of FAST TCP usng OMNET++ Umar ul Hassan 04030038@lums.edu.pk Md Term Report CS678 Topcs n Internet Research Sprng, 2006 Introducton Internet traffc s doublng roughly every 3 months

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Accounting for the Use of Different Length Scale Factors in x, y and z Directions

Accounting for the Use of Different Length Scale Factors in x, y and z Directions 1 Accountng for the Use of Dfferent Length Scale Factors n x, y and z Drectons Taha Soch (taha.soch@kcl.ac.uk) Imagng Scences & Bomedcal Engneerng, Kng s College London, The Rayne Insttute, St Thomas Hosptal,

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

Face Recognition University at Buffalo CSE666 Lecture Slides Resources:

Face Recognition University at Buffalo CSE666 Lecture Slides Resources: Face Recognton Unversty at Buffalo CSE666 Lecture Sldes Resources: http://www.face-rec.org/algorthms/ Overvew of face recognton algorthms Correlaton - Pxel based correspondence between two face mages Structural

More information

GSLM Operations Research II Fall 13/14

GSLM Operations Research II Fall 13/14 GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are

More information

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming CS 4/560 Desgn and Analyss of Algorthms Kent State Unversty Dept. of Math & Computer Scence LECT-6 Dynamc Programmng 2 Dynamc Programmng Dynamc Programmng, lke the dvde-and-conquer method, solves problems

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Supervsed vs. Unsupervsed Learnng Up to now we consdered supervsed learnng scenaro, where we are gven 1. samples 1,, n 2. class labels for all samples 1,, n Ths s also

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming Optzaton Methods: Integer Prograng Integer Lnear Prograng Module Lecture Notes Integer Lnear Prograng Introducton In all the prevous lectures n lnear prograng dscussed so far, the desgn varables consdered

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

Query Clustering Using a Hybrid Query Similarity Measure

Query Clustering Using a Hybrid Query Similarity Measure Query clusterng usng a hybrd query smlarty measure Fu. L., Goh, D.H., & Foo, S. (2004). WSEAS Transacton on Computers, 3(3), 700-705. Query Clusterng Usng a Hybrd Query Smlarty Measure Ln Fu, Don Hoe-Lan

More information

5 The Primal-Dual Method

5 The Primal-Dual Method 5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton

More information

An Image Fusion Approach Based on Segmentation Region

An Image Fusion Approach Based on Segmentation Region Rong Wang, L-Qun Gao, Shu Yang, Yu-Hua Cha, and Yan-Chun Lu An Image Fuson Approach Based On Segmentaton Regon An Image Fuson Approach Based on Segmentaton Regon Rong Wang, L-Qun Gao, Shu Yang 3, Yu-Hua

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

Machine Learning. Topic 6: Clustering

Machine Learning. Topic 6: Clustering Machne Learnng Topc 6: lusterng lusterng Groupng data nto (hopefully useful) sets. Thngs on the left Thngs on the rght Applcatons of lusterng Hypothess Generaton lusters mght suggest natural groups. Hypothess

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

cos(a, b) = at b a b. To get a distance measure, subtract the cosine similarity from one. dist(a, b) =1 cos(a, b)

cos(a, b) = at b a b. To get a distance measure, subtract the cosine similarity from one. dist(a, b) =1 cos(a, b) 8 Clusterng 8.1 Some Clusterng Examples Clusterng comes up n many contexts. For example, one mght want to cluster journal artcles nto clusters of artcles on related topcs. In dong ths, one frst represents

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

High-Boost Mesh Filtering for 3-D Shape Enhancement

High-Boost Mesh Filtering for 3-D Shape Enhancement Hgh-Boost Mesh Flterng for 3-D Shape Enhancement Hrokazu Yagou Λ Alexander Belyaev y Damng We z Λ y z ; ; Shape Modelng Laboratory, Unversty of Azu, Azu-Wakamatsu 965-8580 Japan y Computer Graphcs Group,

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following.

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following. Complex Numbers The last topc n ths secton s not really related to most of what we ve done n ths chapter, although t s somewhat related to the radcals secton as we wll see. We also won t need the materal

More information

Laplacian Eigenmap for Image Retrieval

Laplacian Eigenmap for Image Retrieval Laplacan Egenmap for Image Retreval Xaofe He Partha Nyog Department of Computer Scence The Unversty of Chcago, 1100 E 58 th Street, Chcago, IL 60637 ABSTRACT Dmensonalty reducton has been receved much

More information

Wavefront Reconstructor

Wavefront Reconstructor A Dstrbuted Smplex B-Splne Based Wavefront Reconstructor Coen de Vsser and Mchel Verhaegen 14-12-201212 2012 Delft Unversty of Technology Contents Introducton Wavefront reconstructon usng Smplex B-Splnes

More information

Small Network Segmentation with Template Guidance

Small Network Segmentation with Template Guidance Small Network Segmentaton wth Template Gudance Krstn Dane Lu Department of Mathematcs Unversty of Calforna, Davs Davs, CA 95616 kdlu@math.ucdavs.edu Ian Davdson Department of Computer Scence Unversty of

More information

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements Module 3: Element Propertes Lecture : Lagrange and Serendpty Elements 5 In last lecture note, the nterpolaton functons are derved on the bass of assumed polynomal from Pascal s trangle for the fled varable.

More information

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints TPL-ware Dsplacement-drven Detaled Placement Refnement wth Colorng Constrants Tao Ln Iowa State Unversty tln@astate.edu Chrs Chu Iowa State Unversty cnchu@astate.edu BSTRCT To mnmze the effect of process

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Analysis of Continuous Beams in General

Analysis of Continuous Beams in General Analyss of Contnuous Beams n General Contnuous beams consdered here are prsmatc, rgdly connected to each beam segment and supported at varous ponts along the beam. onts are selected at ponts of support,

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

Insertion Sort. Divide and Conquer Sorting. Divide and Conquer. Mergesort. Mergesort Example. Auxiliary Array

Insertion Sort. Divide and Conquer Sorting. Divide and Conquer. Mergesort. Mergesort Example. Auxiliary Array Inserton Sort Dvde and Conquer Sortng CSE 6 Data Structures Lecture 18 What f frst k elements of array are already sorted? 4, 7, 1, 5, 1, 16 We can shft the tal of the sorted elements lst down and then

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

CSE 326: Data Structures Quicksort Comparison Sorting Bound

CSE 326: Data Structures Quicksort Comparison Sorting Bound CSE 326: Data Structures Qucksort Comparson Sortng Bound Bran Curless Sprng 2008 Announcements (5/14/08) Homework due at begnnng of class on Frday. Secton tomorrow: Graded homeworks returned More dscusson

More information

Very simple computational domains can be discretized using boundary-fitted structured meshes (also called grids)

Very simple computational domains can be discretized using boundary-fitted structured meshes (also called grids) Structured meshes Very smple computatonal domans can be dscretzed usng boundary-ftted structured meshes (also called grds) The grd lnes of a Cartesan mesh are parallel to one another Structured meshes

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

Non-Split Restrained Dominating Set of an Interval Graph Using an Algorithm

Non-Split Restrained Dominating Set of an Interval Graph Using an Algorithm Internatonal Journal of Advancements n Research & Technology, Volume, Issue, July- ISS - on-splt Restraned Domnatng Set of an Interval Graph Usng an Algorthm ABSTRACT Dr.A.Sudhakaraah *, E. Gnana Deepka,

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

CSE 326: Data Structures Quicksort Comparison Sorting Bound

CSE 326: Data Structures Quicksort Comparison Sorting Bound CSE 326: Data Structures Qucksort Comparson Sortng Bound Steve Setz Wnter 2009 Qucksort Qucksort uses a dvde and conquer strategy, but does not requre the O(N) extra space that MergeSort does. Here s the

More information

Intra-Parametric Analysis of a Fuzzy MOLP

Intra-Parametric Analysis of a Fuzzy MOLP Intra-Parametrc Analyss of a Fuzzy MOLP a MIAO-LING WANG a Department of Industral Engneerng and Management a Mnghsn Insttute of Technology and Hsnchu Tawan, ROC b HSIAO-FAN WANG b Insttute of Industral

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT 3. - 5. 5., Brno, Czech Republc, EU APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT Abstract Josef TOŠENOVSKÝ ) Lenka MONSPORTOVÁ ) Flp TOŠENOVSKÝ

More information

Out-of-Sample Extensions for LLE, Isomap, MDS, Eigenmaps, and Spectral Clustering

Out-of-Sample Extensions for LLE, Isomap, MDS, Eigenmaps, and Spectral Clustering Out-of-Sample Extensons for LLE, Isomap, MDS, Egenmaps, and Spectral Clusterng Yoshua Bengo, Jean-Franços Paement, Pascal Vncent Olver Delalleau, Ncolas Le Roux and Mare Oumet Département d Informatque

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Inverse-Polar Ray Projection for Recovering Projective Transformations

Inverse-Polar Ray Projection for Recovering Projective Transformations nverse-polar Ray Projecton for Recoverng Projectve Transformatons Yun Zhang The Center for Advanced Computer Studes Unversty of Lousana at Lafayette yxz646@lousana.edu Henry Chu The Center for Advanced

More information

Load-Balanced Anycast Routing

Load-Balanced Anycast Routing Load-Balanced Anycast Routng Chng-Yu Ln, Jung-Hua Lo, and Sy-Yen Kuo Department of Electrcal Engneerng atonal Tawan Unversty, Tape, Tawan sykuo@cc.ee.ntu.edu.tw Abstract For fault-tolerance and load-balance

More information

Today s Outline. Sorting: The Big Picture. Why Sort? Selection Sort: Idea. Insertion Sort: Idea. Sorting Chapter 7 in Weiss.

Today s Outline. Sorting: The Big Picture. Why Sort? Selection Sort: Idea. Insertion Sort: Idea. Sorting Chapter 7 in Weiss. Today s Outlne Sortng Chapter 7 n Wess CSE 26 Data Structures Ruth Anderson Announcements Wrtten Homework #6 due Frday 2/26 at the begnnng of lecture Proect Code due Mon March 1 by 11pm Today s Topcs:

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information