Small Network Segmentation with Template Guidance

Size: px
Start display at page:

Download "Small Network Segmentation with Template Guidance"

Transcription

1 Small Network Segmentaton wth Template Gudance Krstn Dane Lu Department of Mathematcs Unversty of Calforna, Davs Davs, CA Ian Davdson Department of Computer Scence Unversty of Calforna, Davs Davs, CA ABSTRACT The mportance of segmentng socal networks has grown along wth the usage of socal network stes such as Facebook. Socal networks provde a wealth of nformaton about people ncludng frendshps and nterests. Informaton can be added to segmentng algorthms such as spectral clusterng as n the form of a sngle constrant. Takng advantage of multple nstances of nformaton, such as a template composton, for clusterng s not trval snce addng multple constrants to spectral clusterng technques s dffcult. We present an approach to ncorporate template constrants to the sem-defnte program formulaton of spectral clusterng. We descrbe dfferent types of template constrants and ther nterpretaton. Also, we present a randomze roundng scheme to ensure the rounded soluton to the SDP formulaton satsfes the template segmentaton. Keywords Spectral Clusterng, Semdefnte Programmng, Randomzed Roundng, Constraned Clusterng 1. INTRODUCTION Socal networks have grown n popularty n the last decade where webstes such as Facebook, Myspace, and Twtter have over 100 mllon users each. Wth growng use, t s mportant to have useful segmentatons of these networks. Segmentatons based just on graph topology (edge structure) have ther uses but n ths paper we shall look at segmentatons of ego networks wth template constrants that allow doman experts to encode segment expectatons as a template structure. Ths work s well suted to socal networks so as to allow users to descrbe themselves along lnes of nterests, locaton, and work afflaton, for example. Extendng segmentaton of socal networks to factor n these descrptons allows for a varety of drectons ncludng creatng segments of people who are frends and also share (or do not share) smlar nterests. Permsson to make dgtal or hard copes of all or part of ths work for personal or classroom use s granted wthout fee provded that copes are not made or dstrbuted for proft or commercal advantage and that copes bear ths notce and the full ctaton on the frst page. To copy otherwse, to republsh, to post on servers or to redstrbute to lsts, requres pror specfc permsson and/or a fee. MLG 11 San Dego, CA, USA Copyrght 2011 ACM $10.00 A key problem s then how to represent ths addtonal node nformaton, or background nformaton, n a meanngful way so that exstng clusterng algorthms can take advantage of t. Spectral clusterng s an mportant technque used to fnd segmentatons of a sngle graph [19][5][23]. The quadratc programmng formulaton (whch spectral clusterng s an example of) does not lend tself to easly addng more constrants on ths extra nformaton [26]. In ths earler work, we could only add one preference for the segmentaton. Sem-defnte programmng (SDP) s an alternatve convex programmng formulaton [27][3] (whose objectve functon s lnear, not quadratc) that easly allows the addton of many constrants. Usng the SDP formulaton, we can segment small networks wth templates. Templates are naturally modeled as postve sem-defnte matrces (kernels) that can be added as lnear constrants n the SDP formulaton. As a constrant, we can dctate whether the segmentaton wll have strct or loose adherence to the template composton. For example, we can conduct pragmatc operatons such as, Segment a network so we have n the same clusterng: all my close frends who have smlar musc tastes for strct adherence to the constrant or, Segment a network so we have n the same clusterng: some of my close frends who have smlar musc tastes for loose adherence to the constrant. More complex operatons could nvolve segmentng a network so that people n the same cluster have smlar musc tastes but dd not go to school together. In ths way, we are placng a compostonal restrcton on who can be n the same cluster together. Consder the example shown n Fgure (1). The left graph wth black nodes shows the segmentaton (ndcated by the dotted lne) from normalzed spectral clusterng [19]. Spectral clusterng s dscussed n detal n Secton 3.1. The rght graph shows the clusterng result from template clusterng. The red square nodes have dfferent labels from the black crcle nodes. The template clusterng formulaton yelds a segmentaton whch consders both the normalzed clusterng segmentaton and the gudance from the template constrants. Templates can take any number of forms whch nclude: 1. Topologcal Structures - preserve x% of a group afflaton such as frendshps or co-workers. 2. Characterstcs of People - preserve x% of people who

2 We present a novel approach usng a SDP formulaton for normalzed spectral clusterng to segment a socal network wth template compostons. We present lnear constrants that represent desred template compostons for segmentaton. We can place lmts on how strctly or loosely the templates are adhered to. Fgure 1: Example of normalzed spectral clusterng (left) and spectral clusterng wth a template constrant (rght). Normalzed spectral clusterng seeks a balanced cut of the graph. The template constrant seeks a clusterng that also ncludes known nodes wth the same label (shown as red squares). share a certan characterstc such as people who enjoy rock musc. 3. Alternatve Segmentatons - preserve x% of a known segmentaton of the graph [25][7]. 4. Multple Dstance Metrcs or Vews - Segment a network based on multple metrcs and vews An mportant contrbuton of our work s the ablty to add multple templates to the one clusterng. Examples of the dfferent types of template compostons and ther constrant forms are outlned n Secton 4. We base our work on the SDP formulaton of normalzed spectral clusterng smlar to Xng and Jordan [27] to segment a socal network wth templates based on background knowledge. Our SDP formulaton s shown n Secton 4. The SDP s easly mplementable n MATLAB by usng the CVX package developed by Grant and Boyd [11] for convex optmzaton. The SDP formulaton guarantees an optmal soluton to the relaxed objectve functon and template composton constrants can be added to the optmzaton problem as addtonal constrants. Another advantage of the SDP formulaton s that we are not lmted n the number of constrants that are used so multple sources of background knowledge can be used for segmentaton resultng n a desred template for clusterng. However, usng templates n an SDP formulaton has a sgnfcant challenge. The template constrants can be wrtten so that a certan quantty of the structure must be preserved, such as 80% of people wthn each segment must be from the same school. Whle the soluton from the SDP formulaton wll satsfy ths constrant, due to the relaxed formulaton of the membershp assgnments (whch allows fractonal values) the soluton obtaned wth roundng up and down membershp assgnments to ntegers may not satsfy the template. We present a randomzed roundng scheme n Secton 5 to ensure the template constrants are satsfed. For example, let there be a network wth four people: C = {w, x, y, z} and a desred template composton s to have nodes w, x, y together. A possble segmentaton s c = [0.7, 0.6, 0.08, 0.3] but roundng the vector c may result n ĉ = [1, 1, 1, 1] and the template constrant s no longer satsfed. Our contrbutons are: We present a randomzed roundng scheme to ensure the rounded partton soluton stll satsfes the template segmentaton. Our expermental results show the usefulness of our work. It should be noted that a well known lmtaton of SDP formulatons s handlng large problems. In future work we shall extend our work by consderng low rank SDP solvers [16][3]. 2. PREVIOUS WORK Prevous work that s closely related to our own ncludes modfyng graph Laplacans wth gudance from a pror knowledge by Tollver et al. [22] and addng must lnk and cannot lnk lnear constrants to an SDP for graph segmentaton by Heler et al. [12]. Tollver et al. performed mage segmentaton by modfyng the mage s normalzed laplacan matrx by ncorporatng external nformaton of the mage n order to perform the segmentaton. A pror knowledge s generally pxel-level nformaton of the mage such as the desred structure of the segmentaton. The Felder space s teratvely modfed by the vector representng the a pror knowledge untl the vectors are algned. Whle we also utlze a pror knowledge n the forms of labels and alternatve cuts of the graph, our work dffers n that we do not modfy the graph Laplacan. Heler et al. also ncorporates a pror knowledge, specfcally about class membershp nformaton, by addng lnear constrants to the SDP formulaton of spectral clusterng. The lnear constrants for nodes and j are of the form: trace ( (e e T j + e je T )X ) = 2, where e = (0,..., 0, 1, 0,..., 0) s the elementary vector wth 1 n the th poston. The constrant encodes a must-lnk constrant between nodes and j and can be changed to a cannot-lnk constrant by settng the constrant to 2 nstead. Moreover, multple ( constrants can be combned ) as a sngle constrant: trace (e e T j + e je T )X = 2 P k where P k s the set (,j) P k of nodes that are n the same class. Whle the constrants we propose are smlar n nature, the formulaton of the constrants allows for a more flexble nterpretaton of the types of membershps (known structures, class membershps, and alternatve segmentatons). Addtonally, there s more flexblty to the adherence to the a pror nformaton. For example, nodes and j that are known to have the same class membershp are not necessarly requred to be n the same segmentaton. The Heler et al. constrant forces the cut to put nodes and j n the same or dfferent clusters. Another method that seeks to ncorporate a pror knowledge s segmentatons wth multple graphs. A straghtfor-

3 ward approach to clusterng wth multple vews or representatons of a graph s to lnearly combne kernels [13][31] or convexly combne graph Laplacans [20][1] and apply exstng clusterng technques. Tang et al. proposed Lnked Matrx Factorzaton (LMF) for clusterng wth multple vews [21]. LMF seeks a common factor matrx P that s the low dmensonal embeddng of enttes characterzed by the multple graphs and matrx Λ that captures the characterstcs of each graph. None of the methods are able to utlze the extra nformaton provded by the multple representatons. Instead, they seek to average the multple vews so standard clusterng technques can be used. For template segmentaton, we want to use the extra nformaton to fnd nodes that satsfy a template. Thus, prevous work n segmentaton wth multple vews, or kernels, do not lend themselves to template segmentaton. 3. SEGMENTING GRAPHS 3.1 Spectral Formulaton Let G(V, E) be a weghted undrected graph wth node set V (G) and edge set E(G). Each edge has an edge weght a j such that a j > 0 f there s an edge between nodes and j and a j = 0 f there s no edge between nodes and j. The affnty matrx A s defned as the non-negatve symmetrc matrx A = {a j}. Let the degree of a node be the sum of the edge weghts of the node: d = j aj. Then let the degree matrx D be a dagonal matrx wth the node degrees on the dagonal. The graph partton problem seeks to partton graph G nto k dsjont sets (S 1, S 2,..., S k ) by mnmzng the total sum of the edge weghts removed to create the parttons. Ths total sum s also known as the cut of the graph whch for two sets s defned as: cut(s 1, S 2) = S 1,j S 2 a j. The drawback of the graph partton problem s that found parttons tend to favor separatng a sngle node from the graph. To prevent cuts of the graph whch are unfavorable, Sh and Malk developed the normalzed cut, or Ncut [19] whch s defned as Ncut(S 1,..., S k ) = k =1 cut(s, S ), (1) vol(s ) where vol(s) s the sum of the edge weghts n set S. The defnton ams to fnd balanced clusters based on the edge weghts n the cluster. Solvng ths normalzed partton problem s known to be NP hard. However, the problem can be relaxed to a normalzed spectral clusterng problem. Let the Laplacan matrx be defned as L = D A and the normalzed Laplacan as L = D 1/2 LD 1/2. Now equaton (1) can be solved by solvng the relaxed optmzaton problem: argmn z z T Lz z T z such that z T z = 1 whch s known as normalzed spectral clusterng. Ths problem s easly solved by fndng the k top egenvectors of L. These k egenvectors are then used to fnd the k clusters of the graph. It s dffcult to add constrants to the spectral clusterng formulaton. To add constrants, some methods alter the affnty matrx or the graph Laplacan [14][24][28] such as the method by Lu et al. [17] whch ncorporates par-wse constrants nto the affnty matrx. Other methods use the constrants to restrct the feasble soluton space [6][8][29][30]. Wang and Davdson provde a formulaton for flexble constraned spectral clusterng n whch a constrant s added to the objectve functon for spectral clusterng to create a novel constraned optmzaton problem. However, the formulaton s only able to ncorporate a sngle set of labels as a sngle constrant. 3.2 SDP Formulaton A sem-defnte program (SDP) s a convex optmzaton problem that seeks to mnmze (or maxmze) a lnear objectve functon over the convex cone of symmetrc and postve semdefnte matrces subject to lnear constrants. The canoncal form of a SDP s: argmn trace(kx) X such that trace(a X) = b for = 1,..., m X 0 where the constrant X 0 restrcts X to postve semdefnte matrces. The advantage of formulatng spectral clusterng problems as a SDP s that the relaxed soluton of the SDP formulaton s tghter than the relaxed soluton provded by spectral clusterng. Also, addtonal lnear constrants can be added to the SDP formulaton. Prevous work usng SDP to approxmate graph optmzaton problems ncludes Max-Cut and graph parttonng [9][10][15]. Peng reformulates the spectral clusterng problem as a 0-1 SDP [18] whch, as an nteger program, s NP hard but can also be relaxed. Normalzed spectral clusterng s stll more desrable because clusters are balanced and parttons of sngle nodes are avoded. Below s the relaxed SDP formulaton for normalzed spectral clusterng gven by Xng and Jordan [27]: argmn Z trace( LZ) such that Zdag(D 1/2 ) = dag(d 1/2 ) (2) Z 0 (elementwse) (3) trace (Z) = k (4) Z 0, Z = Z t (5) I Z 0 (6) L s the normalzed Laplacan defned earler. Z R n n s a postve semdefnte co-occurrence matrx and the soluton to the relaxed SDP formulaton. Constrant (2) ensures that each node s assgned to one cluster; (3) and (5) are from relaxng the SDP formulaton; (4) corresponds to the k clusters desred for segmentaton; and (6) s from the defnton of Y. Z s decomposed nto Z = Y Y T where Y R n k yelds the k clusters of the graph. The factor Y has the followng propertes: Y T Y = I and Y = D 1/2 X(X T DX) 1/2. Y s obtaned by applyng a sngular value decomposton to Z and usng the top k sngular vectors. We then apply

4 the randomzed roundng scheme wth restarts to obtan the segmentaton. A dsadvantage of usng the SDP formulatons s that the formulaton cannot handle large matrces. However, low-rank SDP formulatons do exst [16][3]. Another dsadvantage s that the soluton Z from the relaxed formulaton takes on real values n the nterval [-1,1] nstead of ntegers {0, 1}. We present a roundng scheme n Secton 5 that allows us to round the real values of the solutons to more nterpretable ntegers for segmentaton. 4. ADDING TEMPLATES TO SEGMENTING Background knowledge can take the form of par-wse smlartes va a Laplacan or a co-occurrence matrx. We can use the background knowledge we have to create templates for desred clusters. For example, we may want to fnd a segmentaton of the graph that ncludes the template of coworkers who enjoy the same outdoor actvtes. Background knowledge can be added to the SDP formulaton for normalzed spectral clusterng as addtonal lnear constrants. The template constrants can smply be added to the SDP formulaton: argmn Z such that trace( LZ) normalzed spectral clusterng constrants template constrant 1 template constrant 2 The template constrants are of the form trace(kz) β. The value of the constant β determnes whether the segmentaton follows strct or loose adherence to the template constrant. Examples of dfferent types of matrx K for template constrants and the effect of strct and loose adherence to the constrants on segmentatons are shown Secton 4.1. Explanaton and nterpretaton of the constant β for template constrants s n Secton Types of Templates We present four types of template constrants: topologcal groups, characterstcs of people, alternatve segmentatons, and multple metrcs. In each secton, we offer an llustratve example to explan the affect of each template constrant on the segmentaton. Whle the SDP formulaton for spectral clusterng can fnd a segmentaton of a graph for k-clusters, our examples wll be for k = 2. The same graph s used for all examples and the mage of the graph and ts normalze spectral clusterng segmentaton s shown n Fgure (1) Topologcal Groups/Structures It may be desrable to preserve known relatonshps n a socal network such as famly members, co-workers, or rval sports teams. Such relatonshps can be encoded n a matrx. Let the matrx M = {m j}, where m j { 1, 1}, be the n n matrx that holds the together and apart relatons. The template constrant s then trace(mz) β, where β > 0 s some constant (nterpretaton of trace(mz) and β are n Secton 4.2). Consder the networks n Fgure (2). The red square nodes ndcate nodes whch are known to have a relatonshp that the user would lke to preserve. The graph on the left shows strong adherence to the template constrant snce all the red square nodes are n the same cluster. The graph on the rght shows a looser adherence to the template constrant snce one red square node s n a dfferent cluster. Note that the rght graph (wth the looser adherence) has a segmentaton that s close to the normalzed segmentaton. Fgure 2: Example of template clusterng usng nformaton from known relatonshps between a few nodes. (Secton 4.1.1). The left graph shows a strong adherence to the template constrant. The rght graph shows a looser adherence. The red square nodes are assocated wth the template constrant for groups Characterstcs of People We can encode k bnary labels n k ndcator vectors. For ndcator vector v, v = 1 f node has the label and v = 1 f the node does not have the label. The labels may encode the lkes and dslkes of people n a socal network such as whch people enjoy mountan bkng, rock musc, or cookng. We can mplement the labels as a rank-1 matrx M = {m j}, where m j {0, 1}, whch s created from the outer product of the ndcator matrx. The template constrant s then trace(mz) β, where β > 0 s some constant. Consder the two graphs n Fgure (3). The red square nodes are nodes wth labels on them whle the labels for the black crcle nodes are unknown. The graph on the left shows the resultng segmentaton from strong adherence to the template constrant wth all the red square nodes n the same cluster. The rght network shows a looser adherence to the template constrant as red square and black crcle nodes are clustered together nstead of separately. Fgure 3: Example of template clusterng usng nformaton from labels on nodes (Secton 4.1.2). The left graph shows a strong adherence to the constrant. The rght graph shows a looser adherence. The red square nodes are assocated wth the template constrant for labels Alternatve Segmentatons There may be cases where we have a cut of a graph based on another kernel. We can then use the nformaton of the cut as a template constrant. Let M = {m j}, where m j { 1, 1}, be the co-occurrence matrx that encodes the other segmentaton. The template constrant s then trace(mz) β, where β > 0 s some constant. The hgher value of the β, the stronger the soluton Z wll adhere to the constrant. Consder the graphs n Fgure (4). The blue

5 trangle nodes represent an alternatve clusterng. The left graph shows strong adherence to the template constrant: the blue trangle nodes are clustered together, separate from the rest of the graph. The rght graph shows a looser adherence to the template constrant: the blue nodes are clustered together but a black crcle node s allowed as part of the cluster. Fgure 4: Example of template clusterng usng nformaton from another segmentaton (Secton 4.1.3). The left graph shows strong adherence to the template constrant. The rght graph shows a looser adherence. The blue trangle nodes are assocated wth the template constrant for an alternatve segmentaton Multple Dstance Metrcs or Vews The afflatons of a socal network can also be measured wth dfferent metrcs. For example, one metrc may represent work afflatons whle another metrc represents frendshps. We may wsh to fnd a segmentaton that partally adheres to the secondary metrc. The template constrant s trace(mz) β, where β > 0 s some constant. Consder the graphs n Fgure (5). The graph on the left represents the same set of nodes as the graph on the rght but the edges are determned by a some other metrc. The graph on the rght shows the resultng segmentaton when the secondary metrc s ncorporated as a template constrant. produce the normalze cut of the graph. Conversely, large values of β wll favor the template constrant. For example, the graphs n Fgure (3) have nodes wth known labels (shown as red square nodes). The graph on the left shows strct adherence to the template constrant by clusterng all the red square nodes together. The graph on the rght shows the same formulaton but wth a smaller β value for the template constrant. A majorty of the red square nodes are clustered together and the normalzed cut s more closely followed. However, the cut devates from the normalzed cut to cluster three of the red square nodes together. 4.3 Combnng Templates The SDP formulaton allows for the addton of more than one constrant. Thus, a user could construct a template composton of several dfferent nstances of each type of template constrant presented n Secton 4.1. However, one can do more than add multple nstances of one type of constrant such as usng a characterstc constrant wth a groupng constrant. Consder the followng example. Two template constrants are used: Constrant 1 for labels on nodes represented by red square nodes (Secton 4.1.2) wth bound β 1 and Constrant 2 for alternatve segmentatons represented by blue trangle nodes (Secton 4.1.3) wth bound β 2. The affect of changng the values for β 1 and β 2 are shown n Fgure (6). The graph on the left s the result of enforcng a stronger adherence to the alternatve clusterng template constrant va a larger β 2 value. The graph on the rght s the result of enforcng a stronger adherence to the labels on nodes template constrant va a larger β 1 value. Fgure 5: Example of template clusterng usng nformaton from another metrc for the same graph (Secton 4.1.4). The left graph represents the same graph but wth a dfferent metrc. The rght graph s the resultng cut. 4.2 Interpretng constrants/constants The four types of template constrants we have presented take the form of trace(kz) β where K s a matrx representng the addtonal nformaton of the graph, such as known relatonshps or labels, and β > 0 s a constant. An nterpretaton of ths constrant s that we want the segmentaton suggested by Z to match the assgnments gven by K at least β tmes. That s, at least β nodes of the graph are clustered together and match the nformaton gven by K. Note that β [0, m] where m = max n Kj. Note that for small values of β, the segmentaton wll favor the normalzed spectral clusterng. That s, the template constrant wll be loosely adhered to. In fact, β = 0 wll Fgure 6: Example of template clusterng usng nformaton from two types of constrants: alternatve segmentaton and labels. The left graph shows a stronger adherence to the template constrant for the blue trangle nodes. The rght graph shows a stronger adherence to the template constrant for red square nodes. 5. MAKING TEMPLATES INTERPRETABLE 5.1 Roundng Schemes The soluton to the relaxed SDP formulaton for spectral clusterng must be rounded from real values to ntegers for cluster assgnments. Roundng schemes are used to recover the feasble soluton partton matrx X = {x j} where x j {0, 1} A drawback of roundng s that the rounded soluton may no longer satsfy the template constrants. Some roundng schemes nclude: 1. Drectonal cosne method [4]: Orthogonal projecton from n-dmensons to k dmensons

6 2. Randomzed Projecton Heurstc Method [9]: Take rows of Z and project them to a random lower dmensonal space. 3. Roundng by Clusterng: Whle treatng each row of Z as a vertex, use a clusterng method, such as k-mean clusterng, to obtan X. However, none of these roundng schemes provde bounds on the qualty of the rounded soluton. We now present, to our knowledge, some of the only work that attempts to do ths. We present a randomzed roundng scheme for recoverng the soluton X that satsfes the un-relaxed normalzed spectral clusterng problem wth template constrants. Let u : u j [ 1, 1] be the soluton to the relaxed SDP for normalzed spectral clusterng. We can apply the followng randomzed roundng scheme to vector u to get soluton v, v { 1, 1} where v s determned probablstcally as a functon of u : { 1 wth probablty (1 + u)/2 v = 1 wth probablty (1 u )/2 We note the followng well known results: E[v ] = u V ar[v ] = 1 u 2 We now wsh to determne bounds to answer two questons: 1. If we apply the roundng scheme how wll the qualty of the soluton be affected? 2. How many tmes should the roundng scheme be appled n order to satsfy the bounds. One nsght we shall emprcally valdate s that for queston (1) the mean qualty of the soluton after applyng the roundng scheme wll be the same as the qualty of the soluton whch s found, as shown below. [ ] E[trace(KV )] = E l v T = [ ] l E v T = l u T = trace(ku) where K s some kernel However, an answer to queston (2) s not so straght forward. As mentoned before, trace(kv ) = n =1 K V s essentally nterpreted as the agreement between the labelng (kernel K) and segmentaton (soluton V ) schemes. Ths can be broken down nto n summatons whose total value s greater than β. Consder the nstance gvng the worst such agreement whch we know wll have numercal value less than β/n, that s argmn jk j V j < β/n. Let q be the ndex that mnmzes ths expresson. We wsh to determne how many tmes we should apply our randomzed roundng scheme so that ts value s greater than β/n. Let ɛ = K q V q β/n. Note that ɛ cannot exceed one snce β s at most the number of nodes havng the label whch s always less than n and the least that K q V q can be s 0. We wsh our randomzed roundng scheme, when appled to each pont, to have a sum greater than ɛ. Gven that our data set conssts of n ponts, we requre that 1 n K V > ɛ: that s, the roundng scheme wll on average need to exceed the worst possble case. We can now defne the Chernoff nequalty to determne the chance our roundng scheme wll satsfy the template constrant nequalty: [ ] P θ 1 u > ɛ e 2nɛ2 n [ ] P K q V q 1 tr (LU ) > ɛ e 2nɛ2 (7) n Where n s the number of nstances n the data set. Ths gves us the chance that one applcaton of the roundng scheme wll satsfy the bound. We wsh to obtan at least one such success (we can easly emprcally verfy the success) whch s gven by the bnomal dstrbuton 1 f(0; n, p) = ( n 0) p 0 (1 p) n 0 = (1 p) n. We can then determne the number of tmes one must use the randomze roundng scheme so that V satsfes template constrant wth a certan confdence. 6. EXPERIMENTAL RESULTS The purpose of ths expermental secton s two fold: To verfy the correctness of our randomzed roundng scheme. Emprcally verfy the bound n Equaton (7). The frst purpose s to show that the composton of the clusterng does ndeed satsfy the rght hand sde of the bound. For example, for the constrant trace(mv ) β, where matrx M encodes the characterstcs of people (Secton 4.1.2), we are statng our clusterng must contan a cluster wth at least β people wth the characterstc represented by M. The second purpose s requred to determne the tghtness of our bound n Equaton (7). Snce ths s an upper bound we wsh to test how tght the bound s. In ths paper, we use a subset of a Facebook data set 1 (analyss of data set n [2]) consstng of 92 ndvduals. Each ndvdual has a number of possble labels assocated wth hm/her: current locaton, undergraduate unversty, graduate unversty, and current employer. For our frst experment, we label those people who lve n Washngton, DC. We wsh to cluster the network accordng to the frendshp lnks but also requre all nne people who lve n Washngton, DC to be n the same cluster. To acheve ths we created a ndcator label vector l of 0 s and 1 s to dentfy those people who lve n Washngton, DC. We then take the outer product M = ll T to obtan a matrx that s postve sem-defnte. Snce we want all nne people to be n the same cluser we enforce the template constrant: trace(mv ) β 9. Accordng to the bound n Equaton (7), we can generate the probablty of the randomzed roundng scheme satsfyng the requrement that all 9 people 1 Obtaned from the Davs Socal Lnks project ucdavs.edu/lab_webste/)

7 Label set From Went to Wash., DC UVA Upper bound of roundng scheme 82.81% 88.63% success from Eqn (7) Emprcally how often the roundng scheme 65% 43% satsfed trace(mv ) β Emprcally how often roundng scheme clustered 3% 25% labeled nodes together Table 1: Results of usng our randomze roundng scheme on a Facebook data set wth labels. 100 trals of the rounded scheme were used and successes were measured by whether the rounded soluton satsfed trace(mb) β and f the labeled nodes were clustered together. from Vrgna are n the same cluster. We can then emprcally compare what proporton of tmes the roundng scheme when appled satsfed the constrant to verfy the correctness of the bound. For each tme the roundng scheme s successful we can then verfy the composton of the clusterng does ndeed have the nne people from Vrgna n the same cluster. Results are shown n Table 1. We repeat ths experment for another label, people who went to the Unversty of Vrgna (UVA) of whch there are 10 such people n our data set. trace(mv ) β 10. The template constrant s Expermental results are shown n Table 1. Based on our experments, we can conclude that for small β, the bound s loose. For the frst case where the number of labeled nodes s nne, the predcted upper bound of success s 82.81% and was emprcally verfed to be 65%. For the second case where the number of labeled nodes s 10, the predted upper bound of success s 88.63% but was verfed emprcally to be 43%. Furthermore, we emprcally verfed than when the bound was satsfed after randomzed roundng that ndeed all 9 or 10 people from Washngton, D.C. and who went to UVA respectvely were n the same cluster. 7. CONCLUSION AND FUTURE WORK Templates allows us to mpose our background expectatons on the clusterng of a socal network. It allows us to specfy the composton of the segmentatons wth respect to the propertes of the people n the network not just on ther frendshp structure. However, addng multple constrants to spectral clusterng s dffcult. Methods whch utlze the nformaton of other kernels of a network can only do so by takng a lnear combnaton of those kernels and then applyng spectral clusterng technques to the new kernel. We presented a novel approach to clusterng small socal networks wth template compostons usng a relaxed SDP formulaton for normalzed spectral clusterng. The SDP formulaton allows the addton of multple constrants and provdes solutons that satsfy template compostons. We showed the possble types of template constrants, the ways n whch they are mplemented n the SDP, and the ways n whch the user can enforce strct or loose adherence to the constrant. We also addressed the drawback of the SDP formulaton, whch s that the rounded soluton may no longer satsfy the template constrants, by presentng a randomzed roundng scheme. We showed how the SDP for spectral clusterng wth template constrants and the randomzed roundng scheme can be appled to a small socal network. Future work ncludes mplementng other types of template constrants such as groupng constrants and usng alternatve cut nformaton and also mplementng combnatons of two or more of dfferent template constrant types such as a constrant for characterstcs of people and for alternatve cut nformaton. Fnally, applyng the template constrants wth a fast SDP or low-rank SDP formulatons to test constrants for larger data sets. 8. REFERENCES [1] A. Argyrou, M. Herbster, M. Herbster, M. Pontl, and M. Pontl. Combnng graph laplacans for sem-supervsed learnng. In Advances n Neural Informaton Processng Systems 18, pages MIT Press, [2] P. Bhattacharyya, A. Garg, and S. F. Wu. Analyss of user keyword n onlne socal networks. In In Socal Networks Analyss and Mnng Journal (by Sprnger), [3] T. D. Be, N. Crstann, P. Bennett, and E. Parrado-hernández. Fast sdp relaxatons of graph cut clusterng, transducton, and other combnatoral problems. JMLR, 7: , [4] P. K. Chan, M. D. F. Schlag, and J. Y. Zen. Spectral k-way rato-cut parttonng and clusterng. In Proceedngs of the 30th nternatonal Desgn Automaton Conference, DAC 93, pages , New York, NY, USA, ACM. [5] F. R. K. Chung. Spectral Graph Theory (CBMS Regonal Conference Seres n Mathematcs, No. 92). Amercan Mathematcal Socety, Feb [6] T. Coleman, J. Saunderson, and A. Wrth. Spectral clusterng wth nconsstent advce. In Proceedngs of the 25th nternatonal conference on Machne learnng, ICML 08, pages ACM, [7] I. Davdson and Z. Q. Fndng alternatve clusterngs usng constrants. In ICDM 08, pages , [8] T. De Be, J. Suykens, and B. D. Moor. Learnng from general label constrants. In SSPR/SPR, pages , [9] A. M. Freze and M. Jerrum. Improved approxmaton algorthms for max k-cut and max bsecton. In Proceedngs of the 4th Internatonal IPCO Conference on Integer Programmng and Combnatoral Optmzaton, pages 1 13, London, UK, Sprnger-Verlag. [10] M. X. Goemans and D. P. Wllamson. Improved approxmaton algorthms for maxmum cut and satsfablty problems usng semdefnte programmng. JACM, 42: , November [11] M. Grant and S. Boyd. CVX: Matlab software for dscplned convex programmng, verson Apr [12] M. Heler, J. Keuchel, and C. Schnörr. Semdefnte clusterng for mage segmentaton wth a-pror knowledge. In DAGM-Symposum, pages ,

8 2005. [13] T. Joachms, T. J. De, N. Crstann, and N. R. A. Uk. Composte kernels for hypertext categorsaton. In In Proceedngs of the Internatonal Conference on Machne Learnng (ICML, pages Morgan Kaufmann Publshers, [14] S. Kamvar, D. Klen, and C. Mannng. Spectral learnng. In In IJCAI, pages , [15] S. E. Karsch and F. Rendl. Semdefnte programmng and graph equpartton. In In Topcs n Semdefnte and Interor-Pont Methods, pages AMS, [16] B. Kuls. Fast low-rank semdefnte programmng for embeddng and clusterng. In n Eleventh Internatonal Conference on Artfcal Intellgence and Statstcs, AISTATS 2007, page 2007, [17] Z. Lu and M. Carrera-Perpnan. Constraned spectral clusterng through affnty propagaton. In Computer Vson and Pattern Recognton, CVPR IEEE Conference on, pages 1 8, june [18] J. Peng and Y. We. Approxmatng k-means-type clusterng va semdefnte programmng. SIAM J. on Optmzaton, 18: , February [19] J. Sh and J. Malk. Normalzed cuts and mage segmentaton. In Proceedngs of the 1997 Conference on Computer Vson and Pattern Recognton (CVPR 97), CVPR 97, pages 731, Washngton, DC, USA, IEEE Computer Socety. [20] V. Sndhwan and P. Nyog. A co-regularzed approach to sem-supervsed learnng wth multple vews. In Proceedngs of the ICML Workshop on Learnng wth Multple Vews, [21] W. Tang, Z. Lu, and I. S. Dhllon. Clusterng wth multple graphs. In Proceedngs of the 2009 Nnth IEEE Internatonal Conference on Data Mnng, ICDM 09, pages , Washngton, DC, USA, IEEE Computer Socety. [22] D. Tollver, G. L. Mller, and R. Collns. Corrected laplacans: closer cuts and segmentaton wth shape prors. In Computer Vson and Pattern Recognton, CVPR, volume 2, pages IEEE, June [23] U. von Luxburg. A tutoral on spectral clusterng. Statstcs and Computng, 17: , December [24] F. Wang, C. Dng, and T. Lz. Integrated kl (k-means - laplacan) clusterng: A new clusterng approach by combnng attrbute data and parwse relatons. SDM, pages 38 48, [25] X. Wang and I. Davdson. Dscoverng contexts and contextual outlers usng random walks n graphs. In ICDM 09, pages , [26] X. Wang and I. Davdson. Flexble constraned spectral clusterng. In Proceedngs of the 16th ACM SIGKDD Internatonal Conference on Knowledge Dscovery and Data Mnng, KDD 10, pages , New York, NY, USA, ACM. [27] E. P. Xng and M. I. Jordan. On semdefnte relaxaton for normalzed k-cut and connectons to spectral clusterng. Techncal Report UCB/CSD , EECS Department, Unversty of Calforna, Berkeley, Jun [28] Q. Xu, M. desjardns, and K. Wagstaff. Constraned spectral clusterng under a local proxmty structure assumpton. In FLAIRS Conference, pages , [29] S. X. Yu and J. Sh. Groupng wth bas. In In Advances n Neural Informaton Processng Systems. MIT Press, [30] S. X. Yu, I. C. Socety, J. Sh, and I. C. Socety. Segmentaton gven partal groupng constrants. IEEE Transactons on Pattern Analyss and Machne Intellgence, 26: , [31] T. Zhang, A. Popescul, and B. Dom. Lnear predcton models wth graph regularzaton for web-page categorzaton. In Proceedngs of the 12th ACM SIGKDD nternatonal conference on Knowledge dscovery and data mnng, KDD 06, pages , New York, NY, USA, ACM.

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Graph-based Clustering

Graph-based Clustering Graphbased Clusterng Transform the data nto a graph representaton ertces are the data ponts to be clustered Edges are eghted based on smlarty beteen data ponts Graph parttonng Þ Each connected component

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

GSLM Operations Research II Fall 13/14

GSLM Operations Research II Fall 13/14 GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 5 Luca Trevisan September 7, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 5 Luca Trevisan September 7, 2017 U.C. Bereley CS294: Beyond Worst-Case Analyss Handout 5 Luca Trevsan September 7, 207 Scrbed by Haars Khan Last modfed 0/3/207 Lecture 5 In whch we study the SDP relaxaton of Max Cut n random graphs. Quc

More information

Chapter 6 Programmng the fnte element method Inow turn to the man subject of ths book: The mplementaton of the fnte element algorthm n computer programs. In order to make my dscusson as straghtforward

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

5 The Primal-Dual Method

5 The Primal-Dual Method 5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

A Facet Generation Procedure. for solving 0/1 integer programs

A Facet Generation Procedure. for solving 0/1 integer programs A Facet Generaton Procedure for solvng 0/ nteger programs by Gyana R. Parja IBM Corporaton, Poughkeepse, NY 260 Radu Gaddov Emery Worldwde Arlnes, Vandala, Oho 45377 and Wlbert E. Wlhelm Teas A&M Unversty,

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints TPL-ware Dsplacement-drven Detaled Placement Refnement wth Colorng Constrants Tao Ln Iowa State Unversty tln@astate.edu Chrs Chu Iowa State Unversty cnchu@astate.edu BSTRCT To mnmze the effect of process

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

y and the total sum of

y and the total sum of Lnear regresson Testng for non-lnearty In analytcal chemstry, lnear regresson s commonly used n the constructon of calbraton functons requred for analytcal technques such as gas chromatography, atomc absorpton

More information

Private Information Retrieval (PIR)

Private Information Retrieval (PIR) 2 Levente Buttyán Problem formulaton Alce wants to obtan nformaton from a database, but she does not want the database to learn whch nformaton she wanted e.g., Alce s an nvestor queryng a stock-market

More information

Face Recognition University at Buffalo CSE666 Lecture Slides Resources:

Face Recognition University at Buffalo CSE666 Lecture Slides Resources: Face Recognton Unversty at Buffalo CSE666 Lecture Sldes Resources: http://www.face-rec.org/algorthms/ Overvew of face recognton algorthms Correlaton - Pxel based correspondence between two face mages Structural

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following.

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following. Complex Numbers The last topc n ths secton s not really related to most of what we ve done n ths chapter, although t s somewhat related to the radcals secton as we wll see. We also won t need the materal

More information

CHAPTER 2 DECOMPOSITION OF GRAPHS

CHAPTER 2 DECOMPOSITION OF GRAPHS CHAPTER DECOMPOSITION OF GRAPHS. INTRODUCTION A graph H s called a Supersubdvson of a graph G f H s obtaned from G by replacng every edge uv of G by a bpartte graph,m (m may vary for each edge by dentfyng

More information

Overlapping Clustering with Sparseness Constraints

Overlapping Clustering with Sparseness Constraints 2012 IEEE 12th Internatonal Conference on Data Mnng Workshops Overlappng Clusterng wth Sparseness Constrants Habng Lu OMIS, Santa Clara Unversty hlu@scu.edu Yuan Hong MSIS, Rutgers Unversty yhong@cmc.rutgers.edu

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

Analysis of Continuous Beams in General

Analysis of Continuous Beams in General Analyss of Contnuous Beams n General Contnuous beams consdered here are prsmatc, rgdly connected to each beam segment and supported at varous ponts along the beam. onts are selected at ponts of support,

More information

Positive Semi-definite Programming Localization in Wireless Sensor Networks

Positive Semi-definite Programming Localization in Wireless Sensor Networks Postve Sem-defnte Programmng Localzaton n Wreless Sensor etworks Shengdong Xe 1,, Jn Wang, Aqun Hu 1, Yunl Gu, Jang Xu, 1 School of Informaton Scence and Engneerng, Southeast Unversty, 10096, anjng Computer

More information

Machine Learning. Topic 6: Clustering

Machine Learning. Topic 6: Clustering Machne Learnng Topc 6: lusterng lusterng Groupng data nto (hopefully useful) sets. Thngs on the left Thngs on the rght Applcatons of lusterng Hypothess Generaton lusters mght suggest natural groups. Hypothess

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

New Appearance Models for Natural Image Matting

New Appearance Models for Natural Image Matting New Appearance Models for Natural Image Mattng Dheeraj Sngaraju Johns Hopkns Unversty Baltmore, MD, USA. dheeraj@cs.jhu.edu Carsten Rother Mcrosoft Research Cambrdge, UK. carrot@mcrosoft.com Chrstoph Rhemann

More information

11. APPROXIMATION ALGORITHMS

11. APPROXIMATION ALGORITHMS Copng wth NP-completeness 11. APPROXIMATION ALGORITHMS load balancng center selecton prcng method: vertex cover LP roundng: vertex cover generalzed load balancng knapsack problem Q. Suppose I need to solve

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

UNIT 2 : INEQUALITIES AND CONVEX SETS

UNIT 2 : INEQUALITIES AND CONVEX SETS UNT 2 : NEQUALTES AND CONVEX SETS ' Structure 2. ntroducton Objectves, nequaltes and ther Graphs Convex Sets and ther Geometry Noton of Convex Sets Extreme Ponts of Convex Set Hyper Planes and Half Spaces

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

New Appearance Models for Natural Image Matting

New Appearance Models for Natural Image Matting New Appearance Models for Natural Image Mattng Dheeraj Sngaraju Johns Hopkns Unversty Baltmore, MD, USA. dheeraj@cs.jhu.edu Carsten Rother Mcrost Research Cambrdge, UK. carrot@mcrost.com Chrstoph Rhemann

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Ecient Computation of the Most Probable Motion from Fuzzy. Moshe Ben-Ezra Shmuel Peleg Michael Werman. The Hebrew University of Jerusalem

Ecient Computation of the Most Probable Motion from Fuzzy. Moshe Ben-Ezra Shmuel Peleg Michael Werman. The Hebrew University of Jerusalem Ecent Computaton of the Most Probable Moton from Fuzzy Correspondences Moshe Ben-Ezra Shmuel Peleg Mchael Werman Insttute of Computer Scence The Hebrew Unversty of Jerusalem 91904 Jerusalem, Israel Emal:

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

Fitting: Deformable contours April 26 th, 2018

Fitting: Deformable contours April 26 th, 2018 4/6/08 Fttng: Deformable contours Aprl 6 th, 08 Yong Jae Lee UC Davs Recap so far: Groupng and Fttng Goal: move from array of pxel values (or flter outputs) to a collecton of regons, objects, and shapes.

More information

An Image Fusion Approach Based on Segmentation Region

An Image Fusion Approach Based on Segmentation Region Rong Wang, L-Qun Gao, Shu Yang, Yu-Hua Cha, and Yan-Chun Lu An Image Fuson Approach Based On Segmentaton Regon An Image Fuson Approach Based on Segmentaton Regon Rong Wang, L-Qun Gao, Shu Yang 3, Yu-Hua

More information

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming Optzaton Methods: Integer Prograng Integer Lnear Prograng Module Lecture Notes Integer Lnear Prograng Introducton In all the prevous lectures n lnear prograng dscussed so far, the desgn varables consdered

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

EXTENDED BIC CRITERION FOR MODEL SELECTION

EXTENDED BIC CRITERION FOR MODEL SELECTION IDIAP RESEARCH REPORT EXTEDED BIC CRITERIO FOR ODEL SELECTIO Itshak Lapdot Andrew orrs IDIAP-RR-0-4 Dalle olle Insttute for Perceptual Artfcal Intellgence P.O.Box 59 artgny Valas Swtzerland phone +4 7

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Discriminative classifiers for object classification. Last time

Discriminative classifiers for object classification. Last time Dscrmnatve classfers for object classfcaton Thursday, Nov 12 Krsten Grauman UT Austn Last tme Supervsed classfcaton Loss and rsk, kbayes rule Skn color detecton example Sldng ndo detecton Classfers, boostng

More information

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK L-qng Qu, Yong-quan Lang 2, Jng-Chen 3, 2 College of Informaton Scence and Technology, Shandong Unversty of Scence and Technology,

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Querying by sketch geographical databases. Yu Han 1, a *

Querying by sketch geographical databases. Yu Han 1, a * 4th Internatonal Conference on Sensors, Measurement and Intellgent Materals (ICSMIM 2015) Queryng by sketch geographcal databases Yu Han 1, a * 1 Department of Basc Courses, Shenyang Insttute of Artllery,

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

Feature-Based Matrix Factorization

Feature-Based Matrix Factorization Feature-Based Matrx Factorzaton arxv:1109.2271v3 [cs.ai] 29 Dec 2011 Tanq Chen, Zhao Zheng, Quxa Lu, Wenan Zhang, Yong Yu {tqchen,zhengzhao,luquxa,wnzhang,yyu}@apex.stu.edu.cn Apex Data & Knowledge Management

More information

Understanding K-Means Non-hierarchical Clustering

Understanding K-Means Non-hierarchical Clustering SUNY Albany - Techncal Report 0- Understandng K-Means Non-herarchcal Clusterng Ian Davdson State Unversty of New York, 1400 Washngton Ave., Albany, 105. DAVIDSON@CS.ALBANY.EDU Abstract The K-means algorthm

More information

MATHEMATICS FORM ONE SCHEME OF WORK 2004

MATHEMATICS FORM ONE SCHEME OF WORK 2004 MATHEMATICS FORM ONE SCHEME OF WORK 2004 WEEK TOPICS/SUBTOPICS LEARNING OBJECTIVES LEARNING OUTCOMES VALUES CREATIVE & CRITICAL THINKING 1 WHOLE NUMBER Students wll be able to: GENERICS 1 1.1 Concept of

More information

Optimal Workload-based Weighted Wavelet Synopses

Optimal Workload-based Weighted Wavelet Synopses Optmal Workload-based Weghted Wavelet Synopses Yoss Matas School of Computer Scence Tel Avv Unversty Tel Avv 69978, Israel matas@tau.ac.l Danel Urel School of Computer Scence Tel Avv Unversty Tel Avv 69978,

More information

Type-2 Fuzzy Non-uniform Rational B-spline Model with Type-2 Fuzzy Data

Type-2 Fuzzy Non-uniform Rational B-spline Model with Type-2 Fuzzy Data Malaysan Journal of Mathematcal Scences 11(S) Aprl : 35 46 (2017) Specal Issue: The 2nd Internatonal Conference and Workshop on Mathematcal Analyss (ICWOMA 2016) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES

More information

LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS. Matthew R. O Shaughnessy and Mark A. Davenport

LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS. Matthew R. O Shaughnessy and Mark A. Davenport 2016 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 13 16, 2016, SALERNO, ITALY LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS Matthew R. O Shaughnessy and Mark A. Davenport

More information

Potential Malicious Users Discrimination with Time Series Behavior Analysis

Potential Malicious Users Discrimination with Time Series Behavior Analysis Murat Semerc MURAT.SEMERCI@BOUN.EDU.TR Al Taylan Cemgl TAYLAN.CEMGIL@BOUN.EDU.TR Department of Computer Engneerng, Bogazc Unversty, 34342, Bebek, Istanbul, Turkey Bulent Sankur BULENT.SANKUR@BOUN.EDU.TR

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

Determining Fuzzy Sets for Quantitative Attributes in Data Mining Problems

Determining Fuzzy Sets for Quantitative Attributes in Data Mining Problems Determnng Fuzzy Sets for Quanttatve Attrbutes n Data Mnng Problems ATTILA GYENESEI Turku Centre for Computer Scence (TUCS) Unversty of Turku, Department of Computer Scence Lemmnkäsenkatu 4A, FIN-5 Turku

More information

New l 1 -Norm Relaxations and Optimizations for Graph Clustering

New l 1 -Norm Relaxations and Optimizations for Graph Clustering Proceedngs of the Thrteth AAAI Conference on Artfcal Intellgence (AAAI-6) New l -Norm Relaxatons and Optmzatons for Graph Clusterng Fepng Ne, Hua Wang, Cheng Deng 3, Xnbo Gao 3, Xuelong L 4, Heng Huang

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

From Comparing Clusterings to Combining Clusterings

From Comparing Clusterings to Combining Clusterings Proceedngs of the Twenty-Thrd AAAI Conference on Artfcal Intellgence (008 From Comparng Clusterngs to Combnng Clusterngs Zhwu Lu and Yuxn Peng and Janguo Xao Insttute of Computer Scence and Technology,

More information

Enhancement of Infrequent Purchased Product Recommendation Using Data Mining Techniques

Enhancement of Infrequent Purchased Product Recommendation Using Data Mining Techniques Enhancement of Infrequent Purchased Product Recommendaton Usng Data Mnng Technques Noraswalza Abdullah, Yue Xu, Shlomo Geva, and Mark Loo Dscplne of Computer Scence Faculty of Scence and Technology Queensland

More information

3D vector computer graphics

3D vector computer graphics 3D vector computer graphcs Paolo Varagnolo: freelance engneer Padova Aprl 2016 Prvate Practce ----------------------------------- 1. Introducton Vector 3D model representaton n computer graphcs requres

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Fuzzy C-Means Initialized by Fixed Threshold Clustering for Improving Image Retrieval

Fuzzy C-Means Initialized by Fixed Threshold Clustering for Improving Image Retrieval Fuzzy -Means Intalzed by Fxed Threshold lusterng for Improvng Image Retreval NAWARA HANSIRI, SIRIPORN SUPRATID,HOM KIMPAN 3 Faculty of Informaton Technology Rangst Unversty Muang-Ake, Paholyotn Road, Patumtan,

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information