Decentralized Collaborative Learning of Personalized Models over Networks

Size: px
Start display at page:

Download "Decentralized Collaborative Learning of Personalized Models over Networks"

Transcription

1 Decentralzed Collaboratve Learnng of Personalzed Models over Networks Paul Vanhaesebrouck Aurélen Bellet Marc Tommas INRIA INRIA Unversté de Llle Abstract We consder a set of learnng agents n a collaboratve peer-to-peer network, where each agent learns a personalzed model accordng to ts own learnng objectve. The queston addressed n ths paper s: how can agents mprove upon ther locally traned model by communcatng wth other agents that have smlar objectves? We ntroduce and analyze two asynchronous gossp algorthms runnng n a fully decentralzed manner. Our frst approach, nspred from label propagaton, ams to smooth pre-traned local models over the network whle accountng for the confdence that each agent has n ts ntal model. In our second approach, agents jontly learn and propagate ther model by makng teratve updates based on both ther local dataset and the behavor of ther neghbors. To optmze ths challengng objectve, our decentralzed algorthm s based on ADMM. 1 Introducton Increasng amounts of data are beng produced by nterconnected devces such as moble phones, connected objects, sensors, etc. For nstance, hstory logs are generated when a smartphone user browses the web, gves product ratngs and executes varous applcatons. The currently domnant approach to extract useful nformaton from such data s to collect all users personal data on a server (or a tghtly coupled system hosted n a data center) and apply centralzed machne learnng and data mnng technques. However, ths centralzaton poses a number of ssues, such as the need for users to surrender ther personal data Proceedngs of the 20 th Internatonal Conference on Artfcal Intellgence and Statstcs (AISTATS) 2017, Fort Lauderdale, Florda, USA. JMLR: W&CP volume 54. Copyrght 2017 by the author(s). to the servce provder wthout much control on how the data wll be used, whle ncurrng potentally hgh bandwdth and devce battery costs. Even when the learnng algorthm can be dstrbuted n a way that keeps data on users devces, a central entty s often stll requred for aggregaton and coordnaton (see e.g., McMahan et al., 2016). In ths paper, we envson an alternatve settng where many users (agents) wth local datasets collaborate to learn models by engagng n a fully decentralzed peer-to-peer network. Unlke exstng work focusng on problems where agents seek to agree on a global consensus model (see e.g., Nedc and Ozdaglar, 2009; We and Ozdaglar, 2012; Duch et al., 2012), we study the case where each agent learns a personalzed model accordng to ts own learnng objectve. We assume that the network graph s gven and reflects a noton of smlarty between agents (two agents are neghbors n the network f they have a smlar learnng objectve), but each agent s only aware of ts drect neghbors. An agent can then learn a model from ts (typcally scarce) personal data but also from nteractons wth ts neghborhood. As a motvatng example, consder a decentralzed recommender system (Boutet et al., 2013, 2014) n whch each user rates a small number of moves on a smartphone applcaton and expects personalzed recommendatons of new moves. In order to tran a relable recommender for each user, one should rely on the lmted user s data but also on nformaton brought by users wth smlar taste/profle. The peer-to-peer communcaton graph could be establshed when some users go the same move theater or attend the same cultural event, and some smlarty weghts between users could be computed based on hstorcal data (e.g., countng how many tmes people have met n such locatons). Our contrbutons are as follows. After formalzng the problem of nterest, we propose two asynchronous and fully decentralzed algorthms for collaboratve learnng of personalzed models. They belong to the famly of gossp algorthms (Shah, 2009; Dmaks et al., 2010): agents only communcate wth a sngle negh-

2 Decentralzed Collaboratve Learnng of Personalzed Models over Networks bor at a tme, whch makes our algorthms sutable for deployment n large peer-to-peer real networks. Our frst approach, called model propagaton, s nspred by the graph-based label propagaton technque of Zhou et al. (2004). In a frst phase, each agent learns a model based on ts local data only, wthout communcatng wth others. In a second phase, the model parameters are regularzed so as to be smooth over the network graph. We ntroduce some confdence values to account for potental dscrepances n the agents tranng set szes, and derve a novel asynchronous gossp algorthm whch s smple and effcent. We prove that ths algorthm converges to the optmal soluton of the problem. Our second approach, called collaboratve learnng, s more flexble as t nterweaves learnng and propagaton n a sngle process. Specfcally, t optmzes a trade-off between the smoothness of the model parameters over the network on the one hand, and the models accuracy on the local datasets on the other hand. For ths formulaton, we propose an asynchronous gossp algorthm based on a decentralzed verson of Alternatng Drecton Method of Multplers (ADMM) (Boyd et al., 2011). Fnally, we evaluate the performance of our methods on two synthetc collaboratve tasks: mean estmaton and lnear classfcaton. Our experments show the superorty of the proposed approaches over baselne strateges, and confrm the effcency of our decentralzed algorthms. The rest of the paper s organzed as follows. Secton 2 formally descrbes the problem of nterest and dscusses some related work. Our model propagaton approach s ntroduced n Secton 3, along wth our decentralzed algorthm. Secton 4 descrbes our collaboratve learnng approach, and derves an equvalent formulaton whch s amenable to optmzaton usng decentralzed ADMM. Fnally, Secton 5 shows our numercal results, and we conclude n Secton 6. 2 Prelmnares 2.1 Notatons and Problem Settng We consder a set of n agents V = n where n := {1,..., n}. Gven a convex loss functon l : R p X Y, the goal of agent s to learn a model θ R p whose expected loss E (x,y ) µ l(θ ; x, y ) s small wth respect to an unknown and fxed dstrbuton µ over X Y. Each agent has access to a set of m 0..d. tranng examples S = {(x j, yj )}m j=1 drawn from µ. We allow the tranng set sze to vary wdely across agents (some may even have no data at all). Ths s mportant n practce as some agents may be more actve than others, may have recently joned the servce, etc. In solaton, an agent can learn a soltary model θ sol by mnmzng the loss over ts local dataset S : θ sol arg mn θ R p L (θ) = m j=1 l(θ; x j, yj ). (1) The goal for the agents s to mprove upon ther soltary model by leveragng nformaton from other users n the network. Formally, we consder a weghted connected graph G = (V, E) over the set V of agents, where E V V s the set of undrected edges. We denote by W R n n the symmetrc nonnegatve weght matrx assocated wth G, where W j gves the weght of edge (, j) E and by conventon, W j = 0 f (, j) / E or = j. We assume that the weghts represent the underlyng smlarty between the agents objectves: W j should tend to be large (resp. small) when the objectves of agents and j are smlar (resp. dssmlar). Whle we assume n ths paper that the weghts are gven, n practcal scenaros one could for nstance use some auxlary nformaton such as users profles (when avalable) and/or predcton dsagreement to estmate the weghts. For notatonal convenence, we defne the dagonal matrx D R n n where D = n j=1 W j. We wll also denote by N = {j : W j > 0} the set of neghbors of agent. We assume that the agents only have a local vew of the network: they know ther neghbors and the assocated weghts, but not the global topology or how many agents partcpate n the network. Our goal s to propose decentralzed algorthms for agents to collaboratvely mprove upon ther soltary model by leveragng nformaton from ther neghbors. 2.2 Related Work Several peer-to-peer algorthms have been developed for decentralzed averagng (Kempe et al., 2003; Boyd et al., 2006; Coln et al., 2015) and optmzaton (Nedc and Ozdaglar, 2009; Ram et al., 2010; Duch et al., 2012; We and Ozdaglar, 2012, 2013; Iutzeler et al., 2013; Coln et al., 2016). These approaches solve a consensus problem of the form: mn θ R p L (θ), (2) =1 resultng n a global soluton common to all agents (e.g., a classfer mnmzng the predcton error over the unon of all datasets). Ths s unsutable for our settng, where all agents have personalzed objectves. Our problem s remnscent of Mult-Task Learnng (MTL) (Caruana, 1997), where one jontly learns models for related tasks. Yet, there are several dfferences wth our settng. In MTL, the number of tasks s often small, tranng sets are well-balanced across tasks, and

3 Paul Vanhaesebrouck, Aurélen Bellet, Marc Tommas all tasks are usually assumed to be postvely related (a popular assumpton s that all models share a common subspace). Lastly, the algorthms are centralzed, asde from the dstrbuted MTL of Wang et al. (2016) whch s synchronous and reles on a central server. 3 Model Propagaton In ths secton, we present our model propagaton approach. We frst ntroduce a global optmzaton problem, and then propose and analyze an asynchronous gossp algorthm to solve t. 3.1 Problem Formulaton In ths formulaton, we assume that each agent has learned a soltary model θ sol by mnmzng ts local loss, as n (1). Ths can be done wthout any communcaton between agents. Our goal here conssts n adaptng these models by makng them smoother over the network graph. In order account for the fact that the soltary models were learned on tranng sets of dfferent szes, we wll use c (0, 1] to denote the confdence we put n the model θ sol of user {1,..., n}. The c s should be proportonal to the number of tranng ponts m one may for nstance set c = m max j m j (plus some small constant n the case where m = 0). Denotng Θ = [θ 1 ;... ; θ n ] R n p, the objectve functon we am to mnmze s as follows: Q MP (Θ) = 1 ( n W j θ θ j <j µ =1 D c θ θ sol 2 ), (3) where µ > 0 s a trade-off parameter and denotes the Eucldean norm. The frst term n the rght hand sde of (3) s a classc quadratc form used to smooth the models wthn neghborhoods: the dstance between the new models of agents and j s encouraged to be small when the weght W j s large. The second term prevents models wth large confdence from dvergng too much from ther orgnal values so that they can propagate useful nformaton to ther neghborhood. On the other hand, models wth low confdence are allowed large devatons: n the extreme case where agent has very lttle or even no data (.e., c s neglgble), ts model s fully determned by the neghborng models. The presence of D n the second term s smply for normalzaton. We have the followng result (the proof s n the supplementary materal). Proposton 1 (Closed-form soluton). Let P = D 1 W be the stochastc smlarty matrx assocated wth the graph G and Θ sol = [θ sol 1 ;... ; θ sol n ] R n p. The soluton Θ = arg mn Θ R n p Q MP (Θ) s gven by Θ = ᾱ(i ᾱ(i C) αp ) 1 CΘ sol, (4) wth α (0, 1) such that µ = (1 α)/α, and ᾱ = 1 α. Our formulaton s a generalzaton of the semsupervsed label propagaton technque of (Zhou et al., 2004), whch can be recovered by settng C = I (same confdence for all nodes). Note that t s strctly more general: we can see from (4) that unless the confdence values are equal for all agents, the confdence nformaton cannot be ncorporated by usng dfferent soltary models Θ sol or by consderng a dfferent graph (because ᾱ α (I C) P s not stochastc). The asynchronous gossp algorthm we present below thus apples to label propagaton for whch, to the best of our knowledge, no such algorthm was prevously known. Computng the closed form soluton (4) requres the knowledge of the global network and of all soltary models, whch are unknown to the agents. Our startng pont for the dervaton of an asynchronous gossp algorthm s the followng teratve form: for any t 0, Θ(t + 1) = (αi + ᾱc) 1 ( αp Θ(t) + ᾱcθ sol), (5) The sequence (Θ(t)) t N can be shown to converge to (4) regardless of the choce of ntal value Θ(0), see supplementary materal for detals. An nterestng observaton about ths recurson s that t can be decomposed nto agent-centrc updates whch only nvolve neghborhoods. Indeed, for any agent and any t 0: θ (t + 1) = 1 ( α W ) j θ j (t) + ᾱc θ sol. α + ᾱc D j N The teraton (5) can thus be understood as a decentralzed but synchronous process where, at each step, every agent communcates wth all ts neghbors to collect ther current model parameters and uses ths nformaton to update ts model. Assumng that the agents do have access to a global clock to synchronze the updates (whch s unrealstc n many practcal scenaros), synchronzaton ncurs large delays snce all agents must fnsh the update at step t before anyone starts step t + 1. The fact that agents must contact all ther neghbors at each teraton further hnders the effcency of the algorthm. To avod these lmtatons, we propose below an asynchronous gossp algorthm. 3.2 Asynchronous Gossp Algorthm In the asynchronous settng, each agent has a local clock tckng at the tmes of a rate 1 Posson process, and wakes up when t tcks. As local clocks are..d.,

4 Decentralzed Collaboratve Learnng of Personalzed Models over Networks t s equvalent to actvatng a sngle node unformly at random at each tme step (Boyd et al., 2006). 1 The dea behnd our algorthm s the followng. At any tme t 0, each agent wll mantan a (possbly outdated) knowledge of ts neghbors models. For mathematcal convenence, we wll consder a matrx Θ (t) R n p where ts -th lne Θ (t) Rp s agent s model at tme t, and for j, ts j-th lne Θ j (t) Rp s agent s last knowledge of the model of agent j. For any j / N {} and any t 0, we wll mantan Θ j (t) = 0. Let Θ = [ Θ 1,..., Θ n ] R n2 p be the horzontal stackng of all the Θ s. If agent wakes up at tme step t, two consecutve actons are performed: communcaton step: agent selects a random neghbor j N wth prob. π j and both agents update ther knowledge of each other s model: Θ j (t + 1) = Θ j j (t) and Θ j(t + 1) = Θ (t), update step: agents and j update ther own models based on current knowledge. For l {, j}: Θ l l(t + 1) = (α + ᾱc l ) 1 (α W ) lk Θk D l (t + 1) + ᾱc l θl sol. (6) ll k N l All other varables n the network reman unchanged. In the communcaton step above, π j corresponds to the probablty that agent selects agent j. For any n, we have π [0, 1] n such that n j=1 πj = 1 and π j > 0 f and only f j N. Our algorthm belongs to the famly of gossp algorthms as each agent communcates wth at most one neghbor at a tme. Gossp algorthms are known to be very effectve for decentralzed computaton n peer-topeer networks (see Dmaks et al., 2010; Shah, 2009). Thanks to ts asynchronous updates, our algorthm has the potental to be much faster than a synchronous verson when executed n a large peer-to-peer network. The man result of ths secton shows that our algorthm converges to a state where all nodes have ther optmal model (and those of ther neghbors). Theorem 1 (Convergence). Let Θ(0) n2 p R be some arbtrary ntal value and ( Θ(t)) t N be the sequence generated by our algorthm. Let Θ = arg mn Θ R n p Q MP (Θ) be the optmal soluton to model propagaton. For any n, we have: ] lm [ Θj E t (t) = Θ j for j N {}. 1 Our analyss straghtforwardly extends to the case where agents have clocks tckng at dfferent rates. Sketch of proof. The frst step of the proof s to rewrte the algorthm as an equvalent random teratve process over Θ R n2 p of the form: Θ(t + 1) = A(t) Θ(t) + b(t), for any t 0. Then, we show that the spectral radus of E[A(t)] s smaller than 1, whch allows us to exhbt the convergence to the desred quantty. The proof can be found n the supplementary materal. 4 Collaboratve Learnng In the approach presented n the prevous secton, models are learned locally by each agent and then propagated through the graph. In ths secton, we allow the agents to smultaneously learn ther model and propagate t through the network. In other words, agents teratvely update ther models based on both ther local dataset and the behavor of ther neghbors. Whle n general ths s computatonally more costly than merely propagatng pre-traned models, we can expect sgnfcant mprovements n terms of accuracy. As n the case of model propagaton, we frst ntroduce the global objectve functon and then propose an asynchronous gossp algorthm, whch s based on the general paradgm of ADMM (Boyd et al., 2011). 4.1 Problem Formulaton In contrast to model propagaton, the objectve functon to mnmze here takes nto account the loss of each personal model on the local dataset, rather than smply the dstance to the soltary model: Q CL (Θ) = W j θ θ j 2 + µ D L (θ ), (7) <j =1 where µ > 0 s a trade-off parameter. The assocated optmzaton problem s Θ = arg mn Θ R n p Q CL (Θ). The frst term n the rght hand sde of (7) s the same as n the model propagaton objectve (3) and tends to favor models that are smooth on the graph. However, whle n model propagaton enforcng smoothness on the models may potentally translate nto a sgnfcant decrease of accuracy on the local datasets (even for relatvely small changes n parameter values wth respect to the soltary models), here the second term prevents ths. It allows more flexblty n settngs where very dfferent parameter values defne models whch actually gve very smlar predctons. Note that the confdence s bult n the second term as L s a sum over the local dataset of agent. In general, there s no closed-form expresson for Θ, but we can solve the problem wth a decentralzed teratve algorthm, as shown n the rest of ths secton.

5 Paul Vanhaesebrouck, Aurélen Bellet, Marc Tommas 4.2 Asynchronous Gossp Algorthm We propose an asynchronous decentralzed algorthm for mnmzng (7) based on the Alternatve Drecton Method of Multplers (ADMM). Ths general method s a popular way to solve consensus problems of the form (2) n the dstrbuted and decentralzed settngs (see e.g., Boyd et al., 2011; We and Ozdaglar, 2012, 2013; Iutzeler et al., 2013). In our settng, we do not seek a consensus n the classc sense of (2) snce our goal s to learn a personalzed model for each agent. However, we show below that we can reformulate (7) as an equvalent partal consensus problem whch s amenable to decentralzed optmzaton wth ADMM. Problem reformulaton. Let Θ be the set of N +1 varables θ j R p for j N {}, and denote θ j by Θ j. Ths s smlar to the notatons used n Secton 3, except that here we consder Θ as lvng n R ( N +1) p. We now defne Q CL(Θ ) = 1 W j θ θ j 2 + µd L (θ ), 2 j N so that we can rewrte our problem (7) as mn Θ R n p n =1 Q CL (Θ ). In ths formulaton, the objectve functons assocated wth the agents are dependent as they share some decson varables n Θ. In order to apply decentralzed ADMM, we need to decouple the objectves. The dea s to ntroduce a local copy Θ R ( N +1) p of the decson varables Θ for each agent and to mpose equalty constrants on the varables Θ = Θ j for all n, j N. Ths partal consensus can be seen as requrng that two neghborng agents agree on each other s personalzed model. We further ntroduce 4 secondary varables Ze, Z ej, Zj e and Zj ej for each edge e = (, j), whch can be vewed as estmates of the models Θ and Θ j known by each end of e and wll allow an effcent decomposton of the ADMM updates. Formally, denotng Θ = [ Θ 1,..., Θ n ] R (2 E +n) p and Z R 4 E p, we ntroduce the formulaton mn Θ R (2 E +n) p Z C E s.t. e = (, j) E, Q CL( Θ ) =1 { Z e = Θ, Z j e = Θ j Z j ej = Θ j j, Z ej = Θ j, (8) where C E = {Z R 4 E p Ze = Z ej, Zj ej = Zj e for all e = (, j) E}. It s easy to see that Problem (8) s equvalent to the orgnal problem (7) n the followng sense: the mnmzer Θ of (8) satsfes ( Θ ) j = Θ j for all n and j N {}. Further observe that the set of constrants nvolvng Θ can be wrtten D Θ+HZ = 0 where H = I of dmenson 4 E 4 E s dagonal nvertble and D of dmenson 4 E (2 E + n) contans exactly one entry of 1 n each row. The assumptons of We and Ozdaglar (2013) are thus met and we can apply asynchronous decentralzed ADMM. Before presentng the algorthm, we derve the augmented Lagrangan assocated wth Problem (8). Let Λ j e be dual varables assocated wth constrants nvolvng Θ n (8). For convenence, we denote by Z R 2 N the set of secondary varables {{Ze } {Zj e }} e=(,j) E assocated wth agent. Smlarly, we denote by Λ R 2 N the set of dual varables {{Λ e } {Λj e }} e=(,j) E. The augmented Lagrangan s gven by: L ρ ( Θ, Z, Λ) = L ρ( Θ, Z, Λ ), =1 where ρ > 0 s a penalty parameter, Z C E and L ρ( Θ, Z, Λ ) = Q CL( Θ )+ +Λ j e ( Θ j Zj e )+ ρ 2 j:e=(,j) E [ Λ e( Θ Z e) ( Θ Z e 2 + Θ j Zj e 2) ]. Algorthm. ADMM conssts n approxmately mnmzng the augmented Lagrangan L ρ ( Θ, Z, Λ) by alternatng mnmzaton wth respect to the prmal varable Θ and the secondary varable Z, together wth an teratve update of the dual varable Λ. We frst brefly dscuss how to nstantate the ntal values Θ(0), Z(0) and Λ(0). The only constrant on these ntal values s to have Z(0) C E, so a smple opton s to ntalze all varables to 0. That sad, t s typcally advantageous to use a warm-start strategy. For nstance, each agent can send ts soltary model to ts neghbors, and then set Θ = θsol for all j N, Ze = Θ, Zj e = Θ j θ sol, Θ j = θsol j for all e = (, j) E, and Λ(0) = 0. Alternatvely, one can ntalze the algorthm wth the model propagaton soluton obtaned usng the method of Secton 3. Recall from Secton 3.2 that n the asynchronous settng, a sngle agent wakes up at each tme step and selects one of ts neghbors. Assume that agent wakes up at some teraton t 0 and selects j N. Denotng e = (, j), the teraton goes as follows: 1. Agent updates ts prmal varables: Θ (t + 1) = arg mn L ρ(θ, Z (t), Λ (t)), Θ R ( N +1) p and sends Θ (t + 1), Θ j (t + 1), Λ e (t), Λj e (t) to agent j. Agent j executes the same steps w.r.t..

6 Decentralzed Collaboratve Learnng of Personalzed Models over Networks 2. Usng Θ j j (t + 1), Θ j (t + 1), Λj ej (t), Λ ej (t) receved from j, agent updates ts secondary varables: Ze(t + 1) = 1 [ 1 ( Λ 2 ρ e (t) + Λ ej(t) ) + Θ (t + 1) + Θ ] j(t + 1), Z j e (t + 1) = 1 [ 1 ( Λ j ej 2 ρ (t) + Λj e (t)) + Θ j j (t + 1) + Θ j ]. (t + 1) Agent j updates ts secondary varables symmetrcally, so by constructon we have Z(t + 1) C E. 3. Agent updates ts dual varables: Λ e(t + 1) = Λ e(t) + ρ ( Θ (t + 1) Z e(t + 1) ), Λ j e (t + 1) = Λj e (t) + ρ( Θj (t + 1) Zj e (t + 1)). Agent j updates ts dual varables symmetrcally. All other varables n the network reman unchanged. Step 1 has a smple soluton for some loss functons commonly used n machne learnng (such as quadratc and L 1 loss), and when t s not the case ADMM s typcally robust to approxmate solutons to the correspondng subproblems (obtaned for nstance after a few steps of gradent descent), see Boyd et al. (2011) for examples and further practcal consderatons. Asynchronous ADMM converges almost surely to an optmal soluton at a rate of O(1/t) for convex objectve functons (see We and Ozdaglar, 2013). 5 Experments In ths secton, we provde numercal experments to evaluate the performance of our decentralzed algorthms wth respect to accuracy, convergence rate and the amount of communcaton. To ths end, we ntroduce two synthetc collaboratve tasks: mean estmaton and lnear classfcaton. 5.1 Collaboratve Mean Estmaton We frst ntroduce a smple task n whch the goal of each agent s to estmate the mean of a 1D dstrbuton. To ths end, we adapt the two ntertwnng moons dataset popular n sem-supervsed learnng (Zhou et al., 2004). We consder a set of 300 agents, together wth auxlary nformaton about each agent n the form of a vector v R 2. The true dstrbuton µ of an agent s ether N (1, 40) or N ( 1, 40) dependng on whether v belongs to the upper or lower moon, see Fgure 1(a). Each agent receves m samples x 1,..., xm R from ts dstrbuton µ. Its soltary model s then gven by θ sol = 1 m m j=1 xj, whch corresponds to the use of the quadratc loss functon l(θ; x ) = θ x 2. Fnally, the graph over agents s the complete graph where the weght between agents and j s gven by a Gaussan kernel on the agents auxlary nformaton W j = exp( v v j 2 /2σ 2 ), wth σ = 0.1 for approprate scalng. In all experments, the parameter α of model propagaton was set to 0.99, whch gave the best results on a held-out set of random problem nstances. We frst use ths mean estmaton task to llustrate the mportance of consderng confdence values n our model propagaton formulaton, and then to evaluate the effcency of our asynchronous decentralzed algorthm. Relevance of confdence values. Our goal here s to show that ntroducng confdence values nto the model propagaton approach can sgnfcantly mprove the overall accuracy, especally when the agents receve unbalanced amounts of data. In ths experment, we only compare model propagaton wth and wthout confdence values, so we compute the optmal solutons drectly usng the closed-form soluton (4). We generate several problem nstances wth varyng standard devaton for the confdence values c s. More precsely, we sample c for each agent from a unform dstrbuton centered at 1/2 wth wdth ɛ [0, 1]. The number of samples m gven to agent s then set to m = c 100. The larger ɛ, the more varance n the sze of the local datasets. Fgures 1(b)-1(d) gve a vsualzaton of the models before and after propagaton on a problem nstance for the hardest settng ɛ = 1. Fgure 2 (left-mddle) shows results averaged over 1000 random problem nstances for several values of ɛ. As expected, when the local dataset szes are well-balanced (small ɛ), model propagaton performs the same wth or wthout the use of confdence values. Indeed, both have smlar L 2 error wth respect to the target mean, and the wn rato s about 0.5. However, the performance gap n favor of usng confdence values ncreases sharply wth ɛ. For ɛ = 1, the wn rato n favor of usng confdence values s about Strkngly, the error of model propagaton wth confdence values remans constant as ɛ ncreases. These results emprcally confrm the relevance of ntroducng confdence values nto the objectve functon. Asynchronous algorthm. In ths second experment, we compare asynchronous model propagaton wth the synchronous varant gven by (5). We are nterested n the average L 2 error of the models as a functon of the number of parwse communcatons (number of exchanges from one agent to another). Note that a sngle teraton of the synchronous (resp. asynchronous) algorthm corresponds to 2 E (resp. 2) communcatons. For the asynchronous algorthm, we set the neghbor selecton dstrbuton π of agent

7 Paul Vanhaesebrouck, Aurélen Bellet, Marc Tommas (a) Ground models (b) Soltary models (c) MP wthout confdence (d) MP wth confdence Fgure 1: Illustraton of the collaboratve mean estmaton task, where each pont represents an agent and ts 2D coordnates the assocated auxlary nformaton. Fgure 1(a) shows the ground truth models (blue for mean 1 and red for mean -1). Fgure 1(b) shows the soltary models (local averages) for an nstance where ɛ = 1. Fgures 1(c)-1(d) show the models after propagaton, wthout/wth the use of confdence values. L2 error MP wthout confdence MP wth confdence Wdth ε Wn rato Wdth ε L2 error Sync. MP Async. MP Number of communcatons 105 Fgure 2: Results on the mean estmaton task. (Left-mddle) Model propagaton wth and wthout confdence values w.r.t. the unbalancedness of the local datasets. The left fgure shows the L 2 errors, whle the mddle one shows the percentage of wns n favor of usng confdence values. (Rght) L 2 error of the synchronous and asynchronous model propagaton algorthms wth respect to the number of parwse communcatons. n to be unform over the set of neghbors N. Fgure 2 (rght) shows the results on a problem nstance generated as n the prevous experment (wth ε = 1). Snce the asynchronous algorthm s randomzed, we average ts results on 100 random runs. We see that our asynchronous algorthm acheves an accuracy/communcaton trade-off whch s almost as good as that of the synchronous one, wthout requrng any synchronzaton. It s thus expected to be much faster than the synchronous algorthm on large decentralzed networks wth communcaton delays and/or wthout effcent global synchronzaton. 5.2 Collaboratve Lnear Classfcaton In the prevous mean estmaton task, the squared dstance between two model parameters (.e., estmated means) translates nto the same dfference n L 2 error wth respect to the target mean. Therefore, our collaboratve learnng formulaton s essentally equvalent to our model propagaton approach. To show the benefts that can be brought by collaboratve learnng, we now consder a lnear classfcaton task. Snce two lnear separators wth sgnfcantly dfferent parameters can lead to smlar predctons on a gven dataset, ncorporatng the local errors nto the objectve functon rather than smply the dstances between parameters should lead to more accurate models. We consder a set of 100 agents whose goal s to perform lnear classfcaton n R p. For ease of vsualzaton, the target (true) model of each agent les n a 2-dmensonal subspace: we represent t as a vector n R p wth the frst two entres drawn from a normal dstrbuton centered at the orgn and the remanng ones equal to 0. We consder the smlarty graph where the weght between two agents and j s a Gaussan kernel on the dstance between target models, where the dstance here refers to the length of the chord of the angle φ j between target models projected on a unt crcle. More formally, W,j = exp((cos(φ,j ) 1)/σ) wth σ = 0.1 for approprate scalng. Edges wth neglgble weghts are gnored to speed up computaton. We refer the reader to the supplementary materal for a 2D vsualzaton of the target models and the lnks between them. Every agent receves a random number of tranng ponts drawn unformly between 1 and 20. Each tranng pont (n R p ) s drawn unformly around the orgn, and the bnary label s gven by the predcton of the target lnear separator. We then add some label nose by randomly flppng each label wth probablty The loss functon used by the agents s the hnge loss, gven by l(θ; (x, y )) = max ( ) 0, 1 y θ x. As n the prevous experment, for each algorthm we tune the value of α on a held-out set of random prob-

8 Decentralzed Collaboratve Learnng of Personalzed Models over Networks Test accuracy Model P. Collaboratve L. Soltary models Consensus model Dmenson p Test accuracy Collaboratve Learnng Model Propagaton Soltary models Number of tranng ponts Test accuracy Sync. CL 0.65 Async. CL Async. MP Number of communcatons 10 5 Fgure 3: Results on the lnear classfcaton task. (Left) Test accuracy of model propagaton and collaboratve learnng wth varyng feature space dmenson. (Mddle) Average test accuracy of model propagaton and collaboratve learnng wth respect to the number of tranng ponts avalable to the agent (feature dmenson p = 50). (Rght) Test accuracy of synchronous and asynchronous collaboratve learnng and asynchronous model propagaton wth respect to the number of parwse communcatons (lnear classfcaton task, p = 50). lem nstances. Fnally, we wll evaluate the qualty of the learned model of each agent by computng the accuracy on a separate sample of 100 test ponts drawn from the same dstrbuton as the tranng set. In the followng, we use ths lnear classfcaton task to compare the performance of collaboratve learnng aganst model propagaton, and to evaluate the effcency of our asynchronous algorthms. MP vs. CL. In ths frst experment, we compare the accuracy of the models learned by model propagaton and collaboratve learnng wth feature space dmenson p rangng from 2 to 100. Fgure 3 (left) shows the results averaged over 10 randomly generated problem nstances for each value of p. As baselnes, we also plot the average accuracy of the soltary models and of the global consensus model mnmzng (2). The accuracy of all models decreases wth the feature space dmenson, whch comes from the fact that the expected number of tranng samples remans constant. As expected, the consensus model acheves very poor performance snce agents have very dfferent objectves. On the other hand, both model propagaton and collaboratve learnng are able to mprove very sgnfcantly over the soltary models, even n hgher dmensons where on average these ntal models barely outperform a random guess. Furthermore, collaboratve learnng always outperforms model propagaton. We further analyze these results by plottng the accuracy wth respect to the sze of the local tranng set (Fgure 3, mddle). As expected, the accuracy of the soltary models s hgher for larger tranng sets. Furthermore, collaboratve learnng converges to models whch have smlar accuracy regardless of the tranng sze, effectvely correctng for the ntal unbalancedness. Whle model propagaton also performs well, t s consstently outperformed by collaboratve learnng on all tranng szes. Ths gap s larger for agents wth more tranng data: n model propagaton, the large confdence values assocated wth these agents prevent them from devatng much from ther soltary model, thereby lmtng ther own gan n accuracy. Asynchronous algorthms. Ths second experment compares our asynchronous collaboratve learnng algorthm wth a synchronous varant also based on ADMM (see supplementary materal for detals) n terms of number parwse of communcatons. Fgure 3 (rght) shows that our asynchronous algorthm performs as good as ts synchronous counterpart and should thus be largely preferred for deployment n real peer-to-peer networks. It s also worth notng that asynchronous model propagaton converges an order of magntude faster than collaboratve learnng, as t only propagates models that are pre-traned locally. Model propagaton can thus provde a valuable warmstart ntalzaton for collaboratve learnng. Scalablty. We also observe expermentally that the number of teratons needed by our decentralzed algorthms to converge scales favorably wth the sze of the network (see supplementary materal for detals). 6 Concluson We proposed, analyzed and evaluated two asynchronous peer-to-peer algorthms for the novel settng of decentralzed collaboratve learnng of personalzed models. Ths work opens up nterestng perspectves. The lnk between the smlarty graph and the generalzaton performance of the resultng models should be formally analyzed. Ths could n turn gude the desgn of generc methods to estmate the graph weghts, makng our approaches wdely applcable. Other drectons of nterest nclude the development of prvacypreservng algorthms as well as extensons to tmeevolvng networks and sequental arrval of data. Acknowledgments. Ths work was partally supported by grant ANR-16-CE and by a grant from CPER Nord-Pas de Calas/FEDER DATA Advanced data scence and technologes

9 Paul Vanhaesebrouck, Aurélen Bellet, Marc Tommas References Boutet, A., Frey, D., Guerraou, R., Jégou, A., and Kermarrec, A.-M. (2013). WHATSUP: A Decentralzed Instant News Recommender. In Proceedngs of the 27th IEEE Internatonal Symposum on Parallel and Dstrbuted Processng (IPDPS), pages Boutet, A., Frey, D., Guerraou, R., Kermarrec, A.-M., and Patra, R. (2014). Hyrec: leveragng browsers for scalable recommenders. In Proceedngs of the 15th Internatonal Mddleware Conference, pages Boyd, S., Ghosh, A., Prabhakar, B., and Shah, D. (2006). Randomzed gossp algorthms. IEEE/ACM Transactons on Networkng (TON), 14(SI): Boyd, S., Parkh, N., Chu, E., Peleato, B., and Ecksten, J. (2011). Dstrbuted optmzaton and statstcal learnng va the alternatng drecton method of multplers. Foundatons and Trends R n Machne Learnng, 3(1): Caruana, R. (1997). Multtask Learnng. Machne Learnng, 28(1): Coln, I., Bellet, A., Salmon, J., and Clémençon, S. (2015). Extendng Gossp Algorthms to Dstrbuted Estmaton of U-statstcs. In Proceedngs of the Annual Conference on Neural Informaton Processng Systems (NIPS). Coln, I., Bellet, A., Salmon, J., and Clémençon, S. (2016). Gossp Dual Averagng for Decentralzed Optmzaton of Parwse Functons. In Proceedngs of the 33rd Internatonal Conference on Machne Learnng (ICML). Dmaks, A. G., Kar, S., Moura, J. M. F., Rabbat, M. G., and Scaglone, A. (2010). Gossp Algorthms for Dstrbuted Sgnal Processng. Proceedngs of the IEEE, 98(11): Duch, J. C., Agarwal, A., and Wanwrght, M. J. (2012). Dual Averagng for Dstrbuted Optmzaton: Convergence Analyss and Network Scalng. IEEE Transactons on Automatc Control, 57(3): Iutzeler, F., Banch, P., Cblat, P., and Hachem, W. (2013). Asynchronous Dstrbuted Optmzaton usng a Randomzed Alternatng Drecton Method of Multplers. In Proceedngs of the 52nd IEEE Conference on Decson and Control (CDC), pages Kempe, D., Dobra, A., and Gehrke, J. (2003). Gossp- Based Computaton of Aggregate Informaton. In Proceedngs of the 44th Annual IEEE Symposum on Foundatons of Computer Scence (FOCS), pages McMahan, H. B., Moore, E., Ramage, D., and Agüera y Arcas, B. (2016). Federated Learnng of Deep Networks usng Model Averagng. Techncal report, arxv: Nedc, A. and Ozdaglar, A. E. (2009). Dstrbuted Subgradent Methods for Mult-Agent Optmzaton. IEEE Transactons on Automatc Control, 54(1): Ram, S. S., Nedc, A., and Veeravall, V. V. (2010). Dstrbuted Stochastc Subgradent Projecton Algorthms for Convex Optmzaton. Journal of Optmzaton Theory and Applcatons, 147(3): Shah, D. (2009). Gossp Algorthms. Foundatons and Trends n Networkng, 3(1): Wang, J., Kolar, M., and Srebro, N. (2016). Dstrbuted Mult-Task Learnng. In Proceedngs of the 19th Internatonal Conference on Artfcal Intellgence and Statstcs (AISTATS), pages We, E. and Ozdaglar, A. E. (2012). Dstrbuted Alternatng Drecton Method of Multplers. In Proceedngs of the 51th IEEE Conference on Decson and Control (CDC), pages We, E. and Ozdaglar, A. E. (2013). On the O(1/k) Convergence of Asynchronous Dstrbuted Alternatng Drecton Method of Multplers. In IEEE Global Conference on Sgnal and Informaton Processng (GlobalSIP). Zhou, D., Bousquet, O., Lal, T. N., Weston, J., and Schölkopf, B. (2004). Learnng wth local and global consstency. In Proceedngs of the Annual Conference on Neural Informaton Processng Systems (NIPS), volume 16, pages

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

y and the total sum of

y and the total sum of Lnear regresson Testng for non-lnearty In analytcal chemstry, lnear regresson s commonly used n the constructon of calbraton functons requred for analytcal technques such as gas chromatography, atomc absorpton

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

Efficient Distributed Linear Classification Algorithms via the Alternating Direction Method of Multipliers

Efficient Distributed Linear Classification Algorithms via the Alternating Direction Method of Multipliers Effcent Dstrbuted Lnear Classfcaton Algorthms va the Alternatng Drecton Method of Multplers Caoxe Zhang Honglak Lee Kang G. Shn Department of EECS Unversty of Mchgan Ann Arbor, MI 48109, USA caoxezh@umch.edu

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Cost-efficient deployment of distributed software services

Cost-efficient deployment of distributed software services 1/30 Cost-effcent deployment of dstrbuted software servces csorba@tem.ntnu.no 2/30 Short ntroducton & contents Cost-effcent deployment of dstrbuted software servces Cost functons Bo-nspred decentralzed

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Control strategies for network efficiency and resilience with route choice

Control strategies for network efficiency and resilience with route choice Control strateges for networ effcency and reslence wth route choce Andy Chow Ru Sha Centre for Transport Studes Unversty College London, UK Centralsed strateges UK 1 Centralsed strateges Some effectve

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

Related-Mode Attacks on CTR Encryption Mode

Related-Mode Attacks on CTR Encryption Mode Internatonal Journal of Network Securty, Vol.4, No.3, PP.282 287, May 2007 282 Related-Mode Attacks on CTR Encrypton Mode Dayn Wang, Dongda Ln, and Wenlng Wu (Correspondng author: Dayn Wang) Key Laboratory

More information

Alternating Direction Method of Multipliers Implementation Using Apache Spark

Alternating Direction Method of Multipliers Implementation Using Apache Spark Alternatng Drecton Method of Multplers Implementaton Usng Apache Spark Deterch Lawson June 4, 2014 1 Introducton Many applcaton areas n optmzaton have benefted from recent trends towards massve datasets.

More information

High-Boost Mesh Filtering for 3-D Shape Enhancement

High-Boost Mesh Filtering for 3-D Shape Enhancement Hgh-Boost Mesh Flterng for 3-D Shape Enhancement Hrokazu Yagou Λ Alexander Belyaev y Damng We z Λ y z ; ; Shape Modelng Laboratory, Unversty of Azu, Azu-Wakamatsu 965-8580 Japan y Computer Graphcs Group,

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Proceedngs of the Wnter Smulaton Conference M E Kuhl, N M Steger, F B Armstrong, and J A Jones, eds A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Mark W Brantley Chun-Hung

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

Private Information Retrieval (PIR)

Private Information Retrieval (PIR) 2 Levente Buttyán Problem formulaton Alce wants to obtan nformaton from a database, but she does not want the database to learn whch nformaton she wanted e.g., Alce s an nvestor queryng a stock-market

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

Learning-Based Top-N Selection Query Evaluation over Relational Databases

Learning-Based Top-N Selection Query Evaluation over Relational Databases Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **

More information

Wavefront Reconstructor

Wavefront Reconstructor A Dstrbuted Smplex B-Splne Based Wavefront Reconstructor Coen de Vsser and Mchel Verhaegen 14-12-201212 2012 Delft Unversty of Technology Contents Introducton Wavefront reconstructon usng Smplex B-Splnes

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

Positive Semi-definite Programming Localization in Wireless Sensor Networks

Positive Semi-definite Programming Localization in Wireless Sensor Networks Postve Sem-defnte Programmng Localzaton n Wreless Sensor etworks Shengdong Xe 1,, Jn Wang, Aqun Hu 1, Yunl Gu, Jang Xu, 1 School of Informaton Scence and Engneerng, Southeast Unversty, 10096, anjng Computer

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems Taxonomy of Large Margn Prncple Algorthms for Ordnal Regresson Problems Amnon Shashua Computer Scence Department Stanford Unversty Stanford, CA 94305 emal: shashua@cs.stanford.edu Anat Levn School of Computer

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

Hierarchical clustering for gene expression data analysis

Hierarchical clustering for gene expression data analysis Herarchcal clusterng for gene expresson data analyss Gorgo Valentn e-mal: valentn@ds.unm.t Clusterng of Mcroarray Data. Clusterng of gene expresson profles (rows) => dscovery of co-regulated and functonally

More information

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK L-qng Qu, Yong-quan Lang 2, Jng-Chen 3, 2 College of Informaton Scence and Technology, Shandong Unversty of Scence and Technology,

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints TPL-ware Dsplacement-drven Detaled Placement Refnement wth Colorng Constrants Tao Ln Iowa State Unversty tln@astate.edu Chrs Chu Iowa State Unversty cnchu@astate.edu BSTRCT To mnmze the effect of process

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

Fixing Max-Product: Convergent Message Passing Algorithms for MAP LP-Relaxations

Fixing Max-Product: Convergent Message Passing Algorithms for MAP LP-Relaxations Fxng Max-Product: Convergent Message Passng Algorthms for MAP LP-Relaxatons Amr Globerson Tomm Jaakkola Computer Scence and Artfcal Intellgence Laboratory Massachusetts Insttute of Technology Cambrdge,

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Data-dependent Hashing Based on p-stable Distribution

Data-dependent Hashing Based on p-stable Distribution Data-depent Hashng Based on p-stable Dstrbuton Author Ba, Xao, Yang, Hachuan, Zhou, Jun, Ren, Peng, Cheng, Jan Publshed 24 Journal Ttle IEEE Transactons on Image Processng DOI https://do.org/.9/tip.24.2352458

More information

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated.

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated. Some Advanced SP Tools 1. umulatve Sum ontrol (usum) hart For the data shown n Table 9-1, the x chart can be generated. However, the shft taken place at sample #21 s not apparent. 92 For ths set samples,

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

A Post Randomization Framework for Privacy-Preserving Bayesian. Network Parameter Learning

A Post Randomization Framework for Privacy-Preserving Bayesian. Network Parameter Learning A Post Randomzaton Framework for Prvacy-Preservng Bayesan Network Parameter Learnng JIANJIE MA K.SIVAKUMAR School Electrcal Engneerng and Computer Scence, Washngton State Unversty Pullman, WA. 9964-75

More information

Simulation Based Analysis of FAST TCP using OMNET++

Simulation Based Analysis of FAST TCP using OMNET++ Smulaton Based Analyss of FAST TCP usng OMNET++ Umar ul Hassan 04030038@lums.edu.pk Md Term Report CS678 Topcs n Internet Research Sprng, 2006 Introducton Internet traffc s doublng roughly every 3 months

More information

Data Mining: Model Evaluation

Data Mining: Model Evaluation Data Mnng: Model Evaluaton Aprl 16, 2013 1 Issues: Evaluatng Classfcaton Methods Accurac classfer accurac: predctng class label predctor accurac: guessng value of predcted attrbutes Speed tme to construct

More information

A Robust Method for Estimating the Fundamental Matrix

A Robust Method for Estimating the Fundamental Matrix Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.

More information

Load-Balanced Anycast Routing

Load-Balanced Anycast Routing Load-Balanced Anycast Routng Chng-Yu Ln, Jung-Hua Lo, and Sy-Yen Kuo Department of Electrcal Engneerng atonal Tawan Unversty, Tape, Tawan sykuo@cc.ee.ntu.edu.tw Abstract For fault-tolerance and load-balance

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

5 The Primal-Dual Method

5 The Primal-Dual Method 5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton

More information

Overview. Basic Setup [9] Motivation and Tasks. Modularization 2008/2/20 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION

Overview. Basic Setup [9] Motivation and Tasks. Modularization 2008/2/20 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION Overvew 2 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION Introducton Mult- Smulator MASIM Theoretcal Work and Smulaton Results Concluson Jay Wagenpfel, Adran Trachte Motvaton and Tasks Basc Setup

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

DISTRIBUTED POSE AVERAGING IN CAMERA SENSOR NETWORKS USING CONSENSUS ON MANIFOLDS

DISTRIBUTED POSE AVERAGING IN CAMERA SENSOR NETWORKS USING CONSENSUS ON MANIFOLDS DISTRIBUTED POSE AVERAGING IN CAMERA SENSOR NETWORKS USING CONSENSUS ON MANIFOLDS Roberto Tron, René Vdal Johns Hopns Unversty Center for Imagng Scence 32B Clar Hall, 34 N. Charles St., Baltmore MD 21218,

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

A Topology-aware Random Walk

A Topology-aware Random Walk A Topology-aware Random Walk Inkwan Yu, Rchard Newman Dept. of CISE, Unversty of Florda, Ganesvlle, Florda, USA Abstract When a graph can be decomposed nto clusters of well connected subgraphs, t s possble

More information

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide Lobachevsky State Unversty of Nzhn Novgorod Polyhedron Quck Start Gude Nzhn Novgorod 2016 Contents Specfcaton of Polyhedron software... 3 Theoretcal background... 4 1. Interface of Polyhedron... 6 1.1.

More information

Analysis of Continuous Beams in General

Analysis of Continuous Beams in General Analyss of Contnuous Beams n General Contnuous beams consdered here are prsmatc, rgdly connected to each beam segment and supported at varous ponts along the beam. onts are selected at ponts of support,

More information

11. HARMS How To: CSV Import

11. HARMS How To: CSV Import and Rsk System 11. How To: CSV Import Preparng the spreadsheet for CSV Import Refer to the spreadsheet template to ad algnng spreadsheet columns wth Data Felds. The spreadsheet s shown n the Appendx, an

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information