Collaborative Filtering Ensemble for Ranking

Size: px
Start display at page:

Download "Collaborative Filtering Ensemble for Ranking"

Transcription

1 Collaboratve Flterng Ensemble for Rankng ABSTRACT Mchael Jahrer commo research & consultng 8580 Köflach, Austra Ths paper provdes the soluton of the team commo on the Track2 dataset of the KDD Cup 20 [3]. Yahoo Labs provdes a snapshot of ther musc-ratng database as dataset for the competton, consstng of approxmately 62 mllon ratngs from 250k users on 300k tems. The dataset ncludes herachcal nformaton about the tems. The goal of the competton s to dstngush beteen Hgh rated and Not rated tems of a user. The ratng scale s dscrete and ranges from 0 to 00, whle a Hgh ratng s a ratng 80. The error measure s the percent of false rated tracks over all users, known as the fractons of msclassfcatons. The task s to mnmze ths error rate, hence the rankng should be optmzed. Our fnal submsson s a bl of dfferent collaboratve flterng algorthms enhanced, wth basc statstcs. The algorthms are traned consecutvely and they are bled together wth a neural network. Each of the algorthms optmzes a rank error measure. Categores and Subect Descrptors H.2.8 [Database Management]: Database Applcatons- Data Mnng General Terms Applcaton Keywords Collaboratve Flterng, Ensemble Learnng, KDD Cup, Ratng Predcton, Rankng, Musc Ratngs. INTRODUCTION In the KDD Cup 20 challenge we work wth the Yahoo! Musc dataset [3]. The datasets of both tracks are dsont, sowe havedfferentusers on thetrack2. TheTrack2dataset s sgnfcantly smaller to the one n Track.. Notaton We consder the dataset as sparse matrx R = [r u], where we use the letter u for users and for tems durng ths wrteup. Bold letters are used for matrces and vectors, non bold for scalars. The set of tems s denoted by I and I stands for the total number of tems, I =296,. The set ofusersswrttenas U, whle U s thetotal numberofusers, U =249,02. Predctons for user/tem pars are denoted as ˆr u. We have no nformaton about the ratng date. The set of tems, whch s rated by user u s I(u). The set of users, whch rate tem s U(). Addtonal to the ratng data there s taxonomy nformaton avalable. Taxonomy means the relatonshp between Andreas Töscher commo research & consultng 8580 Köflach, Austra andreas.toescher@commo.at the tems wthn the musc doman. For example track x belongs to album y, and album y belongs to artst z. For a more smple and holstc vew we calculate tem-to-parent and tem-to chld tables. We use the notaton of P() for the parents of tem and C() for the chldren of tem. Parents and chldren are tself tems. The total number of chldren s,203,596 (same number of parents). For the Track2 the goal s to mnmze the error rate E(θ) on the test set. The letter θ ndents the tranable weghts/parameter of the predcton model. E(θ) = 00% 0, r u 80 AND ˆr u(θ) = H 0, unrated AND ˆr u(θ) = Lo R test u, R test, else () The error assessment s done as follows. For each user we should predct ratngs for 6 tems. Ths can be any numerc score, scalng does not matter. Then, the evaluaton machne sorts the scores and assgns the 3-hghest scores to class Hgh and the others to Low. Equaton shows how ths s calculated. If we predct all tem scores for the partcular user correctly, t would result n a 0% error rate. 0% error mples that we separate the Hgh rated tems perfectly from all others. Here we can fnd a sketch of the notaton for model parameters. p u s a user-depent feature and q s an temdepent feature. All features are stored n matrces, e.g. all the user features p u are stored n the matrx p R Fx U and the columns are the F-dmensonal features. The feature values are typcally ntalzed wth small random numbers and they are traned wth stochastc gradent descent from the data r u. As learnng rate we use the constant η and λ for L2-regularzaton. In a basc matrx factorzaton model a predcton ˆr u s calculated by the dot product of the user and the tem feature ˆr u = p T uq..2 The Dataset The provded dataset for Track2 [3] conssts of 62,55,438 ratngs from 249,02 users on 296, tems. The range of the ratng value s between 0 and 00. An tem can be a track, an album, an artst or a genre. Most of the ratngs are gven to tracks (44%) followed by artsts (3%), albums (9%) and genres (6%). The dstrbuton of types wthn all tems are gven by 224,04 tracks, 52,829 albums, 8,674 artsts and 567 genres. By consderng the data as sparse user x tem matrx R = [r u] the fll-rate of 0.08% s very low value. For example n the Netflx Prze dataset [] the fll-rate of the ratng matrx s about %..3 The Tranng of the Models Our soluton s a bled set of predctors. We tran each

2 predctor n two steps. In the frst step we remove a valdaton set from thedataset andtran the model. Then we store the predctons of the valdaton set. In the second step we tran on all avalable data wth the same meta parameters as n the frst step, such as learnng rate η, regularzaton λ, number of epochs etc. and store the predctons of the test set. The valdaton predctons from the frst step are used to tran the bler. The target of all models s the mnmzaton of the error rate on the valdaton set..4 Rankng and Samplng In the KDD Cup 20 Track2, the key dea for collaboratve flterng s to ntroduce samplng. As mentoned n the ntroducton, the error measure s the fracton of msclassfcatons. The ratngs are n the range from 0 to 00. In the competton, the Hgh ratng s defned as ratng 80. For a partcular user we should dstngush between Hgh rated tems and unrated ones. In the test set, the task s to predct scores for 6 tems per user. We know that 3 of these tems are rated Hgh, the others are Not rated by the user, but sampled wth the Item-Hgh probablty. Item-Hgh probablty means that the drawn tem occurs proportonal wth ts number of Hgh-ratngs. It s gven n equaton 2. prob H() = R 80 u U() {, r u 80 0, else In contrast to the hgh-probablty, we can sample an tem accordng to ther popularty. Equaton 3 denotes ths probablty. prob() = U() R.5 Drawng the Valdaton Set The valdaton set of KDD Cup 20 Track s gven by a separate text fle. For Track2 there s no pre-defned valdaton set avalable. Therefore, we generate one by ourselves. The generaton procedure s based on the publc knowledge of the test set generaton. We terate over all users and pck those users, who have 3 Hgh ratngs on the tranng set. For those users, 3 random Hgh ratngs and 3 un-rated tems wth the probablty of prob H() are chosen. We up wth 09,360 users and 656,60 ratngs beng n the valdaton set. For the fnal blng stage t s mportant to use the same valdaton set for every traned predctor. For all models n chapter 2 we report error rate values on the valdaton set. In table we lst three baselne error rates. A low error rate of a sngle predctor s not necessarly needed n order to bl well wthn the ensemble. An average user has 250 ratngs, an average tem 20 ratngs. 2. ALGORITHMS Here we lst the set of collaboratve flterng models. All models are traned on the ratng data drectly. All dfferent models get lsted here wth ther correspondng explanaton. The formula for predctng one user/tem par s also added to each of the algorthms. For completeness we add the error rate value on the valdaton set of each of the models wth the used meta-parameters. The (2) (3) predctor error rate global mean 50.0% tem count 42.38% user count 50.0% Table : Baselne predctons. The error rate of global mean and user count s random (50%), because there s no dfference n the predcted score for a user. Item count s the number of ratngs per tem n the dataset and t performs a lttle better than random. lsted numbers are taken from a predctor whch s actually n our fnal bl. 2. Basc Taxonomy The dataset contans taxonomy nformaton. Items consst of tracks, albums, artsts and genres. The nformaton whch track belongs to whch album, the connecton from the albums to the artsts and from artsts to genres was gven. The dea of the basc taxonomy models s to produce predctors whch bl well by usng taxonomy nformaton whch s not captured by the followng matrx factorzaton models. In total we use 4 dfferent basc taxonomy predctors:. r u = the number of users whch gave a hgh ratng to tem (42.42%) 2. r u = the number of users whch rated tem (42.36%) 3. r u = the number of hgh ratngs the user u has gven to tracks ncluded n the album of the gven track (4.05%) 4. r u = the number of hgh ratngs the user u has gven to tracks belongng to the artst(s) of the gven track (36.93%) 5. r u = the number of hgh ratngs the user u has gven to tracks belongng to the genre(s) of the gven track (25.33%) 6. r u = the number of ratngs the user u has gven to tracks ncluded n the album of the gven track (38.50%) 7. r u = the number of ratngs the user u has gven to tracks belongng to the artst(s) of the gven track (33.45%) 8. r u = the number of ratngs the user u has gven to tracks belongng to the genre(s) of the gven track (25.23%) 9. r u = the number of hgh ratngs the user u has gven to a parent (album, artst, genre) of the gven track (20.24%) 0. r u = the number of ratngs the user u has gven to a parent (album, artst, genre) of the gven track (8.43%). r u = the number of hgh ratngs the user u has gven to parents (album, artst, genre) and all chldren of those parents (dstance 2) of track (24.27%) 2. r u = the number of ratngs the user u has gven to parents (album, artst, genre) and all chldren of those parents (dstance 2) of track (25.55%) 3. r u = the number of hgh ratngs to tem and parents of tem (47.29%)

3 4. r u = the number of ratngs to tem and parents of tem (47.29%) The fnal predctons are gven by addng n order to avod a zero and usng the logarthm, whch leads to: ˆr u = log( r u +) (4) 2.2 Matrx Factorzaton Models Ths secton s about dfferent matrx factorzaton technques. The dea of matrx factorzaton s to proect users and tems to a low-dmensonal space. Soeach user and each tem s represented by a latent feature vector R F. The dot product of them p T uq s the predcton SVD (regresson) The SVD model s a smple matrx factorzaton of the ratng matrx. The tranng s based on regresson wth a quadratc error functon and therefore, t does not drectly optmze the rankng score. Predctons of a user/tem par are gven by the followng formula: ˆr u = p T uq (5) Ths s one of the most popular models n collaboratve flterng snce the Netflx Prze [] n When usng gradent descent as learnng algorthm, tranng tme grows lnearly wth the number of ratngs R, predcton tme of one user/tem par can be done n constant tme O(), hence very fast. Wth F =500, η=0.0006, λ=.5 and 20 teraton epochs we get error rate=5.0% on the valdaton set SVD The SVD explaned here s the same model as the SVD (regresson) beforehand. The predcton s a dot product, gven by ˆr u = p T uq (6) The dfference s how to do stochastc gradent descent by smultaneously optmzng the rankng error. Our frst approach was to use zero-samplng, the dea s to tran a SVD (regresson) on the ratng data and on some artfcal zero-ratngs for randomly sampled tems. Ths was less effectve, but better than the regresson one. However, after a few addtonal experments we found an effectve way to drectly learn the rankng from the data. The dea s to learn the ratng dfference of two tems for a gven user u. The frst tem s the known ratng, the other tem 0 s drawn from all tems I, not rated by u. Hence we have two target values r u and r u0. The value of r u0 can be any constant, we set t for example to 0. The complete algorthm for tranng the SVD, whch optmzed the rankng, s shown n Algorthm. The algorthm we use s very smlar to the one n [7], whch s rankng optmzaton usng the ordnal regresson loss. Agan, we terate over all elements of the sparse ratng matrx R (lne 3). The man dfference to standard stochastc gradent descent s that thetranng s done on pars, here are tem pars (, 0). The second tem 0 (not rated by user u) s sampled accordng to the Item-Hgh probablty from all tems, lne 4 n Algorthm. The next steps are to predct the ratngs ˆr u, ˆr u0 (lne 5, lne 6). The target of the sampled tem s set to 0 (lne 7). The most mportant step s to learn the dfference between the targets and the actual predctons. The error s shown n lne 8. Updates of the features are done feature-wse (lne 0 to lne 3) Input: Sparse ratng matrx R R U x I = [r u] Tunable: Learnng rate η, feature sze F Intalze user weghts p R Fx U and tem weghts q R Fx I wth small random values whle error on valdaton set decreases do forall u, R do draw an tem 0, not rated by u, from all tems I accordng to the probablty prob H() ˆr u p T uq ˆr u0 p T uq 0 r u0 0 e (ˆr u ˆr u0 ) (r u r u0 ) for k =...F do c p uk p uk p uk η (e (q k q 0 k) q k q k η (e c) q 0 k q 0 k η (e ( c)) Algorthm : Pseudo code for tranng a SVD for rankng on ratng data. Wth F=000, η= and 25 teraton epochs we get error rate=5.84% on the valdaton set wth the SVD model AFM Ths model s called asymmetrc factor model because t uses only tem-depent parameters. It was frst mentoned by Paterek n [5]. It learns a dfferent knd of data varablty compared to the smple SVD model, hence t bls well wthn the ensemble. The followng equaton s the predcton of a user/tem par. In a smplfed vew, the predcton s agan a dot product of the tem feature q and the user feature (the rght sde). But the user feature s gven by the set of tems I(u) that user u has rated. Therefore the user s expressed purely by hs rated tems. The features whch are learned from the data are bult together n a dfferent way compared to the pure SVD model, hence a lnear combnaton of an SVD and AFM results mostly n a more accurate model. ˆr u = µ u +µ +q T q (0) I(u) (7) The two matrces q and q (0) have both the sze I x F. The term s used for normalzaton purpose. User and tem bases are µ u and µ. Tranng of the AFM wth stochastc gradent descent s more effectve by teratng over all users u U and make a batch update on the asymmetrc part q 0. Algorthm 2 sketches the tranng process. In lne 4 we ntalze the update vector x R F wth zero. Lne 6 sums the updates and n lne 25 the update s appled to the asymmetrc features q (0). The target value α of the sampled tem 0 can be any number.

4 Input: Sparse ratng matrx R R U x I = [r u] Tunable: Learnng rate η, sampled target constant α, feature sze F Intalze tem weghts q R Fx I and asymmetrc tem weghts q (0) R Fx I wth small random values whle error on valdaton set decreases do forall u U do x 0 y vrtual user feat I(u) q(0) forall I(u) do draw an tem 0, not rated by u, from all tems I accordng to prob H ˆr u µ u +µ +y T q ˆr u0 µ u +µ 0 +y T q 0 r u0 α e (ˆr u ˆr u0 ) (r u r u0 ) µ u µ u η e µ µ η e µ 0 µ 0 η e for k =...F do x k x k +e (q k q 0 k) for k =...F do q k q k η (e y k ) q 0 k q 0 k η (e ( y k )) forall I(u) do for k =...F do q (0) k q(0) k η x k Algorthm 2: Pseudo code for gradent descent tranng an AFM for rankng on ratng data. Wth F =000, η= , α=-50 and 30 teraton epochs we get error rate=6.64% on the valdaton set AFM flpped Agan the same asymmetrc dea, but flpped to the tem sde. We have a fxed user feature, but the tem s expressed by the set of users U() whch rated ths tem. Thus, the tem feature s buld from a normalzed sum of rated user features. User and tem bases are µ u and µ. ˆr u = µ u +µ +p T u U() U() p (0) (8) The two matrces p and p (0) have both the sze U x F. Accordng to Algorthm 2 we use the same schema, but flpped to the other sde (tran tem-wse, etc.). Wth F =50, η= , α=-50 and 50 teraton epochs we get an error rate=7.050% on the valdaton set ASVD The ASVD model combnes the basc SVD approach wth the dea from the AFM. The user s descrbed by the userdepentfeaturep u andthenormalzed sumofratedtems. Ths model s exactly thesame as SVD++from Y.Koren[4]. ˆr u = µ u +µ +q T p u + I(u) q (0) (9) The two matrces q and q (0) have both the sze I x F, matrx p has the sze U x F. Stochastc gradent descent s appled n order to learn the features from the data. Tranng s smlar to Algorthm 2, but the samplng of the second tem 0 s done by pckng a random tem. Wth F=000, η= , α=0 and 35 teraton epochs we get an error rate=5.949% on the valdaton set ASVD flpped As n the secton above, we apply the asymmetrc dea from the ASVD model to the tem sde. Now, the user s descrbedbyp u andwe haveatthetemsde(rghtbrackets) the tem-depent feature q and the normalzed sum of rated user features. ˆr u = µ u +µ +p T u q + U() U() p (0) (0) The two matrces p and p (0) have both the sze U x F, matrx q has the sze I x F. As n Algorthm 2 we use the same schema, but flpped to the other sde (tran tem-wse, etc.). Wth F =2000, η= , α=-50 and 30 teraton epochs we get an error rate=5.63% on the valdaton set ASVD flpped taxonomy The ASVD flpped model can be further mproved by usng the taxonomy nformaton. Items are categorzed n 4 dfferent types: tracks, albums, artsts and genres. Herarchcal relatonshps of them are gven by the taxonomy nfo, whch comes wth the dataset. From the taxonomy we buld chld and parent tables. C() s the set of related chldren of tem. P() s the set of parents of the tem. In the model we ext the tem part wth two new features, the chld features q () and the parent features q (2). We descrbe the tem by ts bag of chldren and by the bag of parents n the same way as we dd t n the asymmetrc part q (0). Here s shown the predcton formula for a user/tem par: ˆr u = µ u +µ +p T u(q + P() P() U() U() q () + C() C() p (0) + q (2) ) () The matrces q, q (0), q (), q (2) have the sze of I x F, matrx p has the sze U x F. The parameters are learned accordng to Algorthm 2, but flpped to the other sde (tran tem-wse, etc.). Wth F =3000, η= , α=-50 and 40 teraton epochs we get an error rate=4.78% on the valdaton set. Ths s our most accurate sngle model ASVD flpped rated taxonomy Taxonomy nformaton was very helpful n the ASVD flpped taxonomy model, furthermore we ntroduce a new parameter q (3) whch learns structure from rated parents.

5 The set P() s the set of parents, now we use the set P u() as parents rated by user u. ˆr u = µ u +µ +p T u(q + P() Pu() P() P u() U() U() q () + C() q (3) ) C() p (0) + q (2) + (2) Wth F =000, η= , α=-63 and 45 teraton epochs we get an error rate=4.277% on the valdaton set. 2.3 Neghborhood models 2.3. Item-Item KNN In order to predct the ratng ˆr u, a k-nearest neghbor algorthm selects K ratngs wth the hghest correlatons to the tem. Therefore, we ntroduce theset of tems I(u,,K) whch conssts of rated tems from user u wth the K-hghest correlatons to tem. I(u,,K) I(u). Here we have the fnal predcton formula for the model Item-Item KNN model. ˆr u = I(u,,K) ruĉ I(u,,K) ĉ (3) Our tem-tem neghborhood model uses a sgmod functon to map the correlatons c to ĉ by ntroducng two new parameters σ (scale) and γ (offset). ĉ = 0.5 tanh(σ c +γ)+0.5 (4) In general we use the Pearson correlaton between two tems, by dong ths, the model s called Item-Item KNN Pearson. Another, more effectve way of computng the correlaton between two tems s to use the feature dstance. The equaton below denotes the correlaton c between two tems and. We dub the model Item-Item KNN SVD feat. c = F k= (q k q k ) 2 F k= q k F k= q k 2 (5) Two models of the Item-Item KNN wth SVD features are n the fnal bl. For those, we calculate the KNN on the resduals of the SVD. We get good results, by usng an SVD wth F=00 and σ=0.62 γ=-2.6 K=27 the error rate on the valdaton set s 5.66% User-User KNN The User-User KNN s the flpped verson of the Item- Item KNN. The only dfference s that users and tems are changed. Snce we have approxmately the same number of tems and users, the User-User KNN has smlar computatonal complexty to the tem one. The only model, whch s n the bl s based on Pearson correlatons between users. Hence the model s called User-User KNN Pearson. Wth metaparameters σ=23.3 γ=-4. and K=507 the error rate on the valdaton set s 8.35%. 2.4 Restrcted Boltzmann Machnes The use of Condtonal Restrcted Boltzmann Machnes for collaboratve flterng s descrbed n [6]. We use a very smlar setup, but adopted the methodology for optmzng the gven rankng error measure. The set of tems rated by user u s denoted as I(u). Ths leads tothe set of unrated tems I I(u) for user u. For every user we choose a subset of the unrated tems Ĩ(u) I I(u) wth the same sze as the set of rated tems Ĩ(u) =. One mportant thng s to draw the tems from the set of unrated tems Ĩ(u) accordng to the probablty of an tem beng rated hgh. Every tem s represented va two bnary vsble unts v () and v (2). The unt v (2) s, f the user rated the tem I(u), otherwse t s 0. The unt v () represents the unrated tems and s, f Ĩ(u) otherwse t s 0. For the condtonng of the hdden states we use the tems rated wth an hgh ratng, ths set of tems s denoted as Î(u) and s a subset Î(u) I(u) of all ratngs rated by user u. The used RBM topology can be seen n Fgure. The probabltes of the vsble and hdden unts for beng hgh are gven by: p(v (l) = h) = p(h = v) = σ b + exp(a (l) + hw(l) ) 2 l= exp(a(l) + hw(l) ) (6) 2 I(u) Ĩ(u) l= σ(x) = w (l) v(l) + d Î(u) (7) +e x (8) We ntalze the bases of the vsble unts a (l) to the logarthm of the probabltes of the correspondng unt of beng hgh. Thebases ofthehddenuntsb are ntalzed tozero. The weghts w and d are ntalzed to values drawn from a zero mean Gaussan dstrbuton wth a small standard devaton of Wth 300 hdden unts we acheve an error rate=0.% on the valdaton set. For every tranng case we actvate only a small set of vsble unts, the set of rated I(u) and the sampled set of unrated tems Ĩ(u). Every vsble unt belongng to an tem not ncluded n the set Î(u) = Ĩ(u) I(u) must not be reconstructed for the tranng of user u. Both sets have the same sze Ĩ(u) = and every tem s represented va 2 vsble unts. Hence there s the need to reconstruct 2 Î(u) vsble unts, whch s less than the total number of tems I. It s also possble to use a sngle vsble unt for every tem beng for rated and 0 for unrated. The problem s that ths setup would requre a reconstructon of every vsble unt, whch equals I reconstructed vsble unts for every tranng user u. On the gven dataset, where every user only rated a small fracton of all tems, ths ncreases the computatonal costs. By usng two vsble unts per tem t s not needed to reconstruct all tems, only the rated tems and a subset of unrated tems have to be reconstructed.

6 Fgure : A Condtonal Restrcted Boltzmann Machne. The hdden unts are condtoned on the tems rated hgh by the gven user. 3. BLENDING Fortrack2therewas nopredefnedholdoutset. InSecton.5 we descrbe how we choose our fxed hould out set. Ths holdout set s desgned to resemble the dsclosed test set and was fxed for all traned models. In order to produce one predctor, the holdout set s removed from the data. The algorthm trans the model on the remanng data. Then predctons for the holdout set are produced. Afterwards we nsert the holdout set n the tranng set, whch results n the orgnal tranng set. Now the model s traned agan on the orgnal tranng set and produces predctons for the test set. Hence every predctor p n = [p () n p (2) n ] consts of predctons on the holdout set p () n and of predctons for the test set p (2) n. For the holdout set the true rankng s known. Ths leads to a supervsed learnng problem. Weuse asngle layer neural network wth sgmod unts on the hdden layer and a lnear readout. The network structure can be seen n Fgure 2. The network s traned usng parwse rankngs, the dea for ths methodology orgnates n [2]. The neural network used s vsualzed n Fgure 2. The output s gven by: o = k h k = tanh w (2) k h k = k ( n w () kn pn ) w (2) k tanh ( n w () kn pn ) (9) (20) The dfference between two outputs o and o s denoted as o = ζ(o o ) (2) where ζ s a meta parameter for scalng. In order to tran the weghts we consder parwse rankngs. We choose two ndces and, where ndex s ranked hgher than. The cost functon C = o +log(+e o ) (22) s descrbed n [2]. It s low, f the outputs of the network predct the correct order, meanng that o s hgher than o. The dervatves of the parwse error between the tranng sample and, where should be ranked hgher than s gven by: ( C e o ) = ζ(h w (2) k h k ) +e o (23) k Fgure 2: The structure of the neural network used for blng. The N nputs p n are normalzed to mean zero and standard devaton. The last nput s constantly. The hdden layer conssts of K sgmod unts, whle the readout s purely lnear. C w () kn = ζw (2) ( ) ( ) k ( h 2 k )p n ( h 2 e o k)p n +e o (24) The tranng s done usng stochastc gradent descent and 32 fold cross valdaton. The weghts are ntalzed to random values drawn unformly between 0.4 and 0.4. The parameter ζ s set n the range from to 2, for the fnal bls we used ζ =.69. The number of hdden unts was set to 30. Table 2 contans all predctors used n the fnal bl, whch produced an error of % on the leaderboard. 4. GENERAL THOUGHTS Samplng s the key dea when optmze a rank error measure wthn a collaboratve flterng algorthm. Compared to pure stochastc gradent descent t works much better. We found out that regularzaton s not that mportant, hence we skp t for most of the models. Durng the competton, our goal was to try many dfferent models, whch can explan the varablty of the data. The accuracy of matrx factorzaton models can be easly mproved by addng more features. We have a much hgher relatve mprovement n the error rate compared to RMSE when doublng the number of features. Hence, some of our best models act wth over 000 features. We dd experments wth samplng tems of the same type, e.g. f processng a track then also a track s sampled. However the results only got worse. In general, taxonomy nformaton helps n gettng an accurate collaboratve flterng model. Our opnon s that the ratng date would also help n further mprovng the model. Our approach on Track shows the effectveness of ntegratng the ratng date n the models. 5. REFERENCES [] J. Bennett, S. Lannng, and N. Netflx. The Netflx Prze. In KDD Cup and Workshop n conuncton wth KDD, [2] C. Burges, T. Shaked, E. Renshaw, A. Lazer, M. Deeds, N. Hamlton, and G. Huller. Learnng to rank usng gradent descent. In Proceedngs of the 22nd

7 nternatonal conference on Machne learnng, ICML 05, pages 89 96, New York, NY, USA, ACM. [3] G. Dror, Y. Koren, and M. Wemer. Yahoo! Musc Ratng Dataset for the KDD Cup 20 Track2. [4] Y. Koren. Factorzaton meets the neghborhood: a multfaceted collaboratve flterng model. In KDD 08: Proceedng of the 4th ACM SIGKDD nternatonal conference on Knowledge dscovery and data mnng, pages ACM, [5] A. Paterek. Improvng regularzed sngular value decomposton for collaboratve flterng. Proceedngs of KDD Cup and Workshop, [6] R. Salakhutdnov, A. Mnh, and G. E. Hnton. Restrcted boltzmann machnes for collaboratve flterng. In ICML, pages , [7] M. Wemer, A. Karatzoglou, and A. J. Smola. Improvng maxmum margn matrx factorzaton. In W. Daelemans, B. Goethals, and K. Mork, edtors, Machne Learnng and Knowledge Dscovery n Databases, volume 52 of LNAI, pages 4 4. Sprnger, 2008.

8 no model error rate vald meta parameters AFM 5.877% F =3000 epochs=30 η= α=-50 2 AFM 6.64% F =000 epochs=30 η= α=-50 3 AFM flpped 7.050% F =50 epochs=50 η= α=-50 4 ASVD flpped 5.5% F =200 epochs=30 η= α=-50 5 ASVD flpped taxo 4.78% F =3000 epochs=40 η= α=-50 6 ASVD flpped rated taxo 4.65% F =200 epochs=50 η= α=-50 7 ASVD flpped rated taxo 4.258% F =000 epochs=45 η= α=-50 8 ASVD flpped rated taxo 4.280% F =3000 epochs=45 η= α=-63 9 ASVD flpped rated taxo 4.277% F =000 epochs=45 η= α=-63 0 ASVD flpped rated taxo 5.96% F =00 epochs=40 η= α=-63 ASVD flpped rated taxo 7.62% F =30 epochs=20 η= α=-63 2 ASVD flpped rated taxo 7.263% F =50 epochs=30 η= α=-50, same type n samplng 3 ASVD 6.00% F =000 epochs=5 η= α=0 4 ASVD 5.949% F =000 epochs=35 η= α=0 5 Basc Taxonomy (-4) 8% - 47% detals n Secton 2. 6 RBM dscrete 28.8% nhd=0 η=0.000 λ= RBM dscrete 28.29% nhd=0 η=0.000 λ= RBM dscrete 2.92% nhd=00 η= λ= RBM dscrete 0.% nhd=300 η=0.005 λ= Item-Item KNN Pearson 6.03% σ=0.5 γ=-8.2 K=29 2 Item-Item KNN Pearson 3.68% σ=9.2 γ=-3.2 K= Item-Item KNN SVD feat 5.66% σ=0.62 γ=-2.6 K=27 (SVD: F =00 η=0.000 epochs=30) 23 Item-Item KNN SVD feat 5.78% σ=.5 γ=-.66 K=67 (SVD: F =00 η= epochs=50) 24 User-User KNN Pearson 8.35% σ=23.3 γ=-4. K= SVD (regresson) 5.50% F =00 epochs=20 η= λ=.5 26 SVD (regresson) 5.0% F =500 epochs=20 η= λ=.5 27 SVD 6.88% F =00 epochs=30 η= SVD 5.84% F =000 epochs=25 η= Table 2: Predctors n the fnal bl. The Neural-Network bler acheves an error rate of 2.67% on the valdaton set and an error rate of % on the KDD Cup 20 Track2 leaderbord.

Feature-Based Matrix Factorization

Feature-Based Matrix Factorization Feature-Based Matrx Factorzaton arxv:1109.2271v3 [cs.ai] 29 Dec 2011 Tanq Chen, Zhao Zheng, Quxa Lu, Wenan Zhang, Yong Yu {tqchen,zhengzhao,luquxa,wnzhang,yyu}@apex.stu.edu.cn Apex Data & Knowledge Management

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like: Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Recommended Items Rating Prediction based on RBF Neural Network Optimized by PSO Algorithm

Recommended Items Rating Prediction based on RBF Neural Network Optimized by PSO Algorithm Recommended Items Ratng Predcton based on RBF Neural Network Optmzed by PSO Algorthm Chengfang Tan, Cayn Wang, Yuln L and Xx Q Abstract In order to mtgate the data sparsty and cold-start problems of recommendaton

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming CS 4/560 Desgn and Analyss of Algorthms Kent State Unversty Dept. of Math & Computer Scence LECT-6 Dynamc Programmng 2 Dynamc Programmng Dynamc Programmng, lke the dvde-and-conquer method, solves problems

More information

Signed Distance-based Deep Memory Recommender

Signed Distance-based Deep Memory Recommender Sgned Dstance-based Deep Memory Recommender ABSTRACT Personalzed recommendaton algorthms learn a user s preference for an tem, by measurng a dstance/smlarty between them. However, some of exstng recommendaton

More information

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY SSDH: Sem-supervsed Deep Hashng for Large Scale Image Retreval Jan Zhang, and Yuxn Peng arxv:607.08477v2 [cs.cv] 8 Jun 207 Abstract Hashng

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Parallel Numerics. 1 Preconditioning & Iterative Solvers (From 2016)

Parallel Numerics. 1 Preconditioning & Iterative Solvers (From 2016) Technsche Unverstät München WSe 6/7 Insttut für Informatk Prof. Dr. Thomas Huckle Dpl.-Math. Benjamn Uekermann Parallel Numercs Exercse : Prevous Exam Questons Precondtonng & Iteratve Solvers (From 6)

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

High-Boost Mesh Filtering for 3-D Shape Enhancement

High-Boost Mesh Filtering for 3-D Shape Enhancement Hgh-Boost Mesh Flterng for 3-D Shape Enhancement Hrokazu Yagou Λ Alexander Belyaev y Damng We z Λ y z ; ; Shape Modelng Laboratory, Unversty of Azu, Azu-Wakamatsu 965-8580 Japan y Computer Graphcs Group,

More information

CMPS 10 Introduction to Computer Science Lecture Notes

CMPS 10 Introduction to Computer Science Lecture Notes CPS 0 Introducton to Computer Scence Lecture Notes Chapter : Algorthm Desgn How should we present algorthms? Natural languages lke Englsh, Spansh, or French whch are rch n nterpretaton and meanng are not

More information

A Clustering Algorithm Solution to the Collaborative Filtering

A Clustering Algorithm Solution to the Collaborative Filtering Internatonal Journal of Scence Vol.4 No.8 017 ISSN: 1813-4890 A Clusterng Algorthm Soluton to the Collaboratve Flterng Yongl Yang 1, a, Fe Xue, b, Yongquan Ca 1, c Zhenhu Nng 1, d,* Hafeng Lu 3, e 1 Faculty

More information

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following.

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following. Complex Numbers The last topc n ths secton s not really related to most of what we ve done n ths chapter, although t s somewhat related to the radcals secton as we wll see. We also won t need the materal

More information

Private Information Retrieval (PIR)

Private Information Retrieval (PIR) 2 Levente Buttyán Problem formulaton Alce wants to obtan nformaton from a database, but she does not want the database to learn whch nformaton she wanted e.g., Alce s an nvestor queryng a stock-market

More information

A Hybrid Collaborative Filtering Model with Deep Structure for Recommender Systems

A Hybrid Collaborative Filtering Model with Deep Structure for Recommender Systems Proceedngs of the Thrty-Frst AAAI Conference on Artfcal Intellgence (AAAI-17) A Hybrd Collaboratve Flterng Model wth Deep Structure for Recommender Systems Xn Dong, Le Yu, Zhonghuo Wu, Yuxa Sun, Lngfeng

More information

Announcements. Supervised Learning

Announcements. Supervised Learning Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

Comparison Study of Textural Descriptors for Training Neural Network Classifiers

Comparison Study of Textural Descriptors for Training Neural Network Classifiers Comparson Study of Textural Descrptors for Tranng Neural Network Classfers G.D. MAGOULAS (1) S.A. KARKANIS (1) D.A. KARRAS () and M.N. VRAHATIS (3) (1) Department of Informatcs Unversty of Athens GR-157.84

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

A Parallel and Efficient Algorithm for Learning to Match

A Parallel and Efficient Algorithm for Learning to Match A Parallel and Effcent Algorthm for Learnng to Match Jngbo Shang 1,, Tanq Chen, Hang L 3, Zhengdong Lu 3, Yong Yu 1 Unversty of Illnos at Urbana Champagn, IL, USA shang7@llnos.edu Unversty of Washngton,

More information

IN recent years, recommender systems, which help users discover

IN recent years, recommender systems, which help users discover Heterogeneous Informaton Network Embeddng for Recommendaton Chuan Sh, Member, IEEE, Bnbn Hu, Wayne Xn Zhao Member, IEEE and Phlp S. Yu, Fellow, IEEE 1 arxv:1711.10730v1 [cs.si] 29 Nov 2017 Abstract Due

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Proceedngs of the Wnter Smulaton Conference M E Kuhl, N M Steger, F B Armstrong, and J A Jones, eds A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Mark W Brantley Chun-Hung

More information

Machine Learning. Topic 6: Clustering

Machine Learning. Topic 6: Clustering Machne Learnng Topc 6: lusterng lusterng Groupng data nto (hopefully useful) sets. Thngs on the left Thngs on the rght Applcatons of lusterng Hypothess Generaton lusters mght suggest natural groups. Hypothess

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

Programming in Fortran 90 : 2017/2018

Programming in Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Exercse 1 : Evaluaton of functon dependng on nput Wrte a program who evaluate the functon f (x,y) for any two user specfed values

More information

Your Neighbors Affect Your Ratings: On Geographical Neighborhood Influence to Rating Prediction

Your Neighbors Affect Your Ratings: On Geographical Neighborhood Influence to Rating Prediction Your Neghbors Affect Your Ratngs: On Geographcal Neghborhood Influence to Ratng Predcton Longke Hu lhu003@e.ntu.edu.sg Axn Sun axsun@ntu.edu.sg Yong Lu luy0054@e.ntu.edu.sg School of Computer Engneerng,

More information

Load Balancing for Hex-Cell Interconnection Network

Load Balancing for Hex-Cell Interconnection Network Int. J. Communcatons, Network and System Scences,,, - Publshed Onlne Aprl n ScRes. http://www.scrp.org/journal/jcns http://dx.do.org/./jcns.. Load Balancng for Hex-Cell Interconnecton Network Saher Manaseer,

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

Application of Maximum Entropy Markov Models on the Protein Secondary Structure Predictions

Application of Maximum Entropy Markov Models on the Protein Secondary Structure Predictions Applcaton of Maxmum Entropy Markov Models on the Proten Secondary Structure Predctons Yohan Km Department of Chemstry and Bochemstry Unversty of Calforna, San Dego La Jolla, CA 92093 ykm@ucsd.edu Abstract

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Learning to Project in Multi-Objective Binary Linear Programming

Learning to Project in Multi-Objective Binary Linear Programming Learnng to Project n Mult-Objectve Bnary Lnear Programmng Alvaro Serra-Altamranda Department of Industral and Management System Engneerng, Unversty of South Florda, Tampa, FL, 33620 USA, amserra@mal.usf.edu,

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines (IJCSIS) Internatonal Journal of Computer Scence and Informaton Securty, Herarchcal Web Page Classfcaton Based on a Topc Model and Neghborng Pages Integraton Wongkot Srura Phayung Meesad Choochart Haruechayasak

More information

CSCI 5417 Information Retrieval Systems Jim Martin!

CSCI 5417 Information Retrieval Systems Jim Martin! CSCI 5417 Informaton Retreval Systems Jm Martn! Lecture 11 9/29/2011 Today 9/29 Classfcaton Naïve Bayes classfcaton Ungram LM 1 Where we are... Bascs of ad hoc retreval Indexng Term weghtng/scorng Cosne

More information

Optimal Workload-based Weighted Wavelet Synopses

Optimal Workload-based Weighted Wavelet Synopses Optmal Workload-based Weghted Wavelet Synopses Yoss Matas School of Computer Scence Tel Avv Unversty Tel Avv 69978, Israel matas@tau.ac.l Danel Urel School of Computer Scence Tel Avv Unversty Tel Avv 69978,

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

ON SOME ENTERTAINING APPLICATIONS OF THE CONCEPT OF SET IN COMPUTER SCIENCE COURSE

ON SOME ENTERTAINING APPLICATIONS OF THE CONCEPT OF SET IN COMPUTER SCIENCE COURSE Yordzhev K., Kostadnova H. Інформаційні технології в освіті ON SOME ENTERTAINING APPLICATIONS OF THE CONCEPT OF SET IN COMPUTER SCIENCE COURSE Yordzhev K., Kostadnova H. Some aspects of programmng educaton

More information

Recommender System based on Higherorder Logic Data Representation

Recommender System based on Higherorder Logic Data Representation Recommender System based on Hgherorder Logc Data Representaton Lnna L Bngru Yang Zhuo Chen School of Informaton Engneerng, Unversty Of Scence and Technology Bejng Abstract Collaboratve Flterng help users

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

Additive Groves of Regression Trees

Additive Groves of Regression Trees Addtve Groves of Regresson Trees Dara Sorokna, Rch Caruana, and Mrek Redewald Department of Computer Scence, Cornell Unversty, Ithaca, NY, USA {dara,caruana,mrek}@cs.cornell.edu Abstract. We present a

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

An Item-Targeted User Similarity Method for Data Service Recommendation

An Item-Targeted User Similarity Method for Data Service Recommendation 01 IEEE 16th Internatonal Enterprse Dstrbuted Object Computng Conference Workshops An Item-Targeted User Smlarty Method for Data Serce Recommendaton Cheng Zhang, Xaofang Zhao Insttute of Computng Technology

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Mult-stable Percepton Necker Cube Spnnng dancer lluson, Nobuuk Kaahara Fttng and Algnment Computer Vson Szelsk 6.1 James Has Acknowledgment: Man sldes from Derek Hoem, Lana Lazebnk, and Grauman&Lebe 2008

More information

Journal of Process Control

Journal of Process Control Journal of Process Control (0) 738 750 Contents lsts avalable at ScVerse ScenceDrect Journal of Process Control j ourna l ho me pag e: wwwelsevercom/locate/jprocont Decentralzed fault detecton and dagnoss

More information

On Some Entertaining Applications of the Concept of Set in Computer Science Course

On Some Entertaining Applications of the Concept of Set in Computer Science Course On Some Entertanng Applcatons of the Concept of Set n Computer Scence Course Krasmr Yordzhev *, Hrstna Kostadnova ** * Assocate Professor Krasmr Yordzhev, Ph.D., Faculty of Mathematcs and Natural Scences,

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

Artificial Intelligence (AI) methods are concerned with. Artificial Intelligence Techniques for Steam Generator Modelling

Artificial Intelligence (AI) methods are concerned with. Artificial Intelligence Techniques for Steam Generator Modelling Artfcal Intellgence Technques for Steam Generator Modellng Sarah Wrght and Tshldz Marwala Abstract Ths paper nvestgates the use of dfferent Artfcal Intellgence methods to predct the values of several contnuous

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

Learning physical Models of Robots

Learning physical Models of Robots Learnng physcal Models of Robots Jochen Mück Technsche Unverstät Darmstadt jochen.mueck@googlemal.com Abstract In robotcs good physcal models are needed to provde approprate moton control for dfferent

More information

Overview. Basic Setup [9] Motivation and Tasks. Modularization 2008/2/20 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION

Overview. Basic Setup [9] Motivation and Tasks. Modularization 2008/2/20 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION Overvew 2 IMPROVED COVERAGE CONTROL USING ONLY LOCAL INFORMATION Introducton Mult- Smulator MASIM Theoretcal Work and Smulaton Results Concluson Jay Wagenpfel, Adran Trachte Motvaton and Tasks Basc Setup

More information

Intelligent Information Acquisition for Improved Clustering

Intelligent Information Acquisition for Improved Clustering Intellgent Informaton Acquston for Improved Clusterng Duy Vu Unversty of Texas at Austn duyvu@cs.utexas.edu Mkhal Blenko Mcrosoft Research mblenko@mcrosoft.com Prem Melvlle IBM T.J. Watson Research Center

More information

Deep Classification in Large-scale Text Hierarchies

Deep Classification in Large-scale Text Hierarchies Deep Classfcaton n Large-scale Text Herarches Gu-Rong Xue Dkan Xng Qang Yang 2 Yong Yu Dept. of Computer Scence and Engneerng Shangha Jao-Tong Unversty {grxue, dkxng, yyu}@apex.sjtu.edu.cn 2 Hong Kong

More information

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010 Smulaton: Solvng Dynamc Models ABE 5646 Week Chapter 2, Sprng 200 Week Descrpton Readng Materal Mar 5- Mar 9 Evaluatng [Crop] Models Comparng a model wth data - Graphcal, errors - Measures of agreement

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS. Matthew R. O Shaughnessy and Mark A. Davenport

LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS. Matthew R. O Shaughnessy and Mark A. Davenport 2016 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 13 16, 2016, SALERNO, ITALY LOCALIZING USERS AND ITEMS FROM PAIRED COMPARISONS Matthew R. O Shaughnessy and Mark A. Davenport

More information

Speeding Up the Xbox Recommender System Using a Euclidean Transformation for Inner-Product Spaces

Speeding Up the Xbox Recommender System Using a Euclidean Transformation for Inner-Product Spaces Speedng Up the Xbox Recommender System Usng a Eucldean Transformaton for Inner-Product Spaces Ran Glad-Bachrach Mcrosoft Research Yoram Bachrach Mcrosoft Research Nr Nce Mcrosoft R&D Lran Katzr Computer

More information

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros.

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros. Fttng & Matchng Lecture 4 Prof. Bregler Sldes from: S. Lazebnk, S. Setz, M. Pollefeys, A. Effros. How do we buld panorama? We need to match (algn) mages Matchng wth Features Detect feature ponts n both

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information