Meta-Prediction for Collective Classification

Size: px
Start display at page:

Download "Meta-Prediction for Collective Classification"

Transcription

1 McDowell, L.., Gupta,.M., & Aha, D.W. (200). Meta-predcton n collectve classfcaton. To appear n Proceedngs of the Twenty- Thrd Florda Artfcal Intellgence Research Socety Conference. Daytona Beach, FL: AAAI Press. Meta-Predcton for Collectve Classfcaton Luke. McDowell, alyan Moy Gupta 2, and Davd W. Aha 3 Dept. of Computer Scence; U.S. Naval Academy; Annapols, MD nexus Research Corp.; Sprngfeld, VA Navy Center for Appled Research n Artfcal Intellgence; Naval Research Laboratory (Code 554); Washngton, DC lmcdowel@usna.edu kalyan.gupta@knexusresearch.com davd.aha@nrl.navy.ml Abstract When data nstances are nter-related, as are nodes n a socal network or hyperlnk graph, algorthms for collectve classfcaton (CC) can sgnfcantly mprove accuracy. Recently, an algorthm for CC named Cautous ICA (ICA C ) was shown to mprove accuracy compared to the popular ICA algorthm. ICA C mproves performance by ntally favorng ts more confdent predctons durng collectve nference. In ths paper, we ntroduce ICA MC, a new algorthm that outperforms ICA C when the attrbutes that descrbe each node are not hghly predctve. ICA MC learns a meta-classfer that dentfes whch node label predctons are most lkely to be correct. We show that ths approach sgnfcantly ncreases accuracy on a range of real and synthetc data sets. We also descrbe new features for the meta-classfer and demonstrate that a smple search can dentfy an effectve feature set that ncreases accuracy. Introducton In many classfcaton tasks, the nstances to be classfed (such as web pages or people n a socal network) are related n some way. Collectve classfcaton (CC) s a methodology that ontly classfes such nstances (or nodes). CC algorthms can attan hgher accuraces than non-collectve methods when nodes are nterrelated (Nevlle and Jensen 2000; Taskar, Abbeel, and oller 2002). Several CC algorthms have been studed, ncludng relaxaton labelng (Chakrabart, Dom, and Indyk 998), the Iteratve Classfcaton Algorthm (ICA) (Sen et al. 2008), loopy belef propagaton (LBP) (Taskar et al. 2002), and Gbbs samplng (Jensen, Nevlle, and Gallagher 2004). We focus on ICA because t s a popular and computatonally effcent algorthm that has good classfcaton performance (Sen et al. 2008). It makes ntal label predctons for each node v, then teratvely recomputes them based on the predctons for every node that lnks to v. Recently, a varant of ICA named Cautous ICA (ICA C ) (McDowell et al. 2007, 2009) was shown to often attan hgher accuraces than ICA. ICA C s based on the Copyrght 200, Assocaton for the Advancement of Artfcal Intellgence ( All rghts reserved. observaton that, snce some label predctons wll be ncorrect, ICA s use of all predctons may sometmes decrease accuracy. To counter ths effect, ICA C nstead ntally uses only some label predctons. By cautously choosng only those predctons that appear more lkely to be correct, ICA C can ncrease accuracy vs. ICA. In ths paper, we ntroduce Meta-Cautous ICA (ICA MC ), whch s exactly lke ICA C except n how t selects the set of predcted labels to use durng classfcaton. In partcular, ICA MC learns a meta-classfer to predct the lkelhood that a label predcton s correct. By carefully constructng a meta-tranng set from the orgnal tranng set, ICA MC can learn ths classfer and use t to select more relable predcted labels than ICA C, ncreasng accuracy. Our contrbutons are as follows. Frst, we present ICA MC, a novel algorthm that can sgnfcantly ncrease accuracy compared to ICA C, especally when the attrbutes that descrbe each node are not very predctve. Second, we ntroduce a technque to mprove accuracy by generatng more tranng examples for ICA MC s meta-classfer. Thrd, we descrbe new features for the meta-classfer and demonstrate that, whle the most effectve meta-features for ICA MC are task-dependent, a smple search dentfes an effectve set that ncreases accuracy. Emprcal evaluatons usng real and synthetc datasets support our clams. We next revew CC and the ICA and ICA C algorthms. Then we ntroduce ICA MC. Fnally, we present our expermental evaluaton and dscuss future research ssues. Collectve Classfcaton Assume we are gven a graph G = (V,E,X,Y,C), where V s a set of nodes, E s a set of (possbly drected) edges, each x X s an attrbute vector for node v V, each Y Y s a label varable for v, and C s the set of possble labels. We are also gven a set of known label values Y for nodes V V, so that Y = {y v V }. Fnally, assume that we are gven a tranng graph G Tr, whch s defned smlarly to G except that every node n G Tr s a known node. Then the task s to nfer Y U =Y Y, whch are the values of Y for the nodes n G whose labels are unknown. For each node v, let y be the true label and ŷ be the predcted label.

2 For example, consder the task of predctng whether a web page belongs to a professor or a student. Conventonal supervsed learnng approaches gnore the lnks and classfy each page usng attrbutes derved from ts content (e.g., words present n the page). In contrast, a technque for relatonal classfcaton explctly uses the lnks to construct addtonal features for classfcaton (e.g., for each page, nclude as features the words from hyperlnked pages). These relatonal features can ncrease classfcaton accuracy, though not always (Chakrabart et al. 998). Alternatvely, even greater (and usually more relable) ncreases can occur when the class labels of the lnked pages are used to derve relevant relatonal features (Jensen et al. 2004). However, usng features based on these labels s challengng because some or all of these labels are ntally unknown. Thus, ther labels must frst be predcted (wthout usng relatonal features) and then re-predcted n some manner (usng all features). Ths process of ontly nferrng the labels of nterrelated nodes s known as collectve classfcaton (CC). We next descrbe two exstng collectve nference algorthms, ICA and ICA C, and then ntroduce ICA MC. Each algorthm reles on a gven node classfer (M AR ) that predcts a node s label usng both attrbutes and relatons. ICA: Inference usng all predcted labels Fgure shows pseudocode for ICA, ICA C, and ICA MC (dependng on AlgType). Step s a bootstrap step that predcts the class label ŷ for each node n V U usng only attrbutes (conf records the confdence of ths predcton, but ICA does not use t). ICA then terates (step 2). Durng each teraton, t selects all avalable label predctons (step 3), computes the relatonal features values based on them (step 4), and then re-predcts the class label of each node usng both attrbutes and relatonal features (step 5). Step 6 s gnored for ICA. After teratng, step 7 returns the fnal set of predcted class labels and ther confdence values. ICA C : Inference usng some predcted labels In steps 3-4 of Fgure, ICA assumes that the predcted node labels are all equally lkely to be correct. When AlgType s nstead ICA C, the nference becomes more cautous by only consderng more confdent predctons. Specfcally, step 3 commts nto Y only the most confdent m of the currently predcted labels; other labels are consdered mssng and are gnored. Step 4 computes the relatonal features usng only the commtted labels, and step 5 performs classfcaton usng ths nformaton. Step 3 gradually ncreases the fracton of predcted labels that are commtted per teraton. Node label assgnments commtted n an teraton h are not necessarly commtted agan n future teratons (and may n fact change). ICA C requres a confdence measure (conf n Fgure ) to rank the current label predctons. As wth pror work (Nevlle and Jensen 2000, McDowell et al. 2007), we set conf to be the posteror probablty of the most lkely class ICA_classfy(V,E,X,Y,M AR,M A,M M,n,AlgType,) = // V=nodes; E=edges; X=attr. vectors; Y =labels of known nodes // M AR =node classfer (uses attrs. & relatons); M A = classfer // that uses attrs. only; M M =meta-classfer (predcts correctness) // n=# ters; AlgType=ICA, ICA C, or ICA MC ; =est. class dstr. for each node v V U do // Bootstrap (ŷ,conf ) M A (x ) 2 for h = 0 to n do 3 // Select node labels for computng relatonal feat. values f (AlgType = ICA) // Use all labels: nown or predcted Y Y {ŷ v V U } else // ICA C or ICA MC : Use known and m m V U * (h/n) // most confdent predcted labels Y Y {ŷ v V U rank(conf ) m} 4 for each node v V U do // Use labels selected above f calcrelatfeats(v,e,y) // to compute feat. values 5 for each node v V U do // Re-predct labels usng (ŷ,conf ) M AR (x,f ) 6 f (AlgType = ICA MC ) Z U {(ŷ,conf ) v V U } for each node v V U do // attrbutes and features // Compute meta-features; // use to re-estmate conf. // values for each node mf calcmetafeatures(,v,e,x,y,y,z U,M A,) conf M M (mf ) 7 // Return most lkely label (and conf. estmate) for each node return { (ŷ,conf ) v V U } Fgure : Pseudocode for ICA, ICA C, or ICA MC. Based on pror work (McDowell et al. 2009), we set n=0 teratons. for each node v. Ths s computed by the node classfer M AR based on the attrbutes and relatonal features of v. ICA C performs well on a varety of real and synthetc data, and attans hgher accuraces than ICA and smlar accuraces as more tme-consumng algorthms such as Gbbs samplng or LBP (McDowell et al. 2009). However, ICA C s ablty to select the best predcted labels depends entrely on the confdence value estmates from the node classfer. Accuracy may decrease f a msclassfed node s nonetheless assgned a hgh confdence value. Improvng ICA C wth Meta-Cauton To address ths potental problem wth ICA C, we created ICA MC. They are dentcal except that ICA MC uses a separate meta classfer to predct how lkely each predcton ŷ s to be correct. Below we descrbe ICA MC s use of ths meta-classfer, methods for generatng ts tranng data, and methods for constructng ts features. ICA MC : Inference usng predcted correct labels Fgure shows that ICA MC changes ICA C only n step 6. In partcular, after usng the node classfer to predct the label ŷ (and assocated confdence conf ) for every node, ICA MC computes the meta-feature values and then uses the metaclassfer M M to predct how lkely ŷ s to be correct. These predctons serve as the new confdence values that are then used n Step 3 of the next teraton to select the

3 commtted set Y. If the meta-classfer s confdence predctons more accurately dentfy those nodes whose labels are correctly predcted (compared to ICA C s smple confdence values), then accuracy should ncrease. Generatng meta-tranng data Learnng the meta-classfer requres constructng approprate meta-tranng data, whch we represent as a set of vectors. Fgure 2 shows the pseudocode for ths task, whose algorthm employs a holdout graph (a subset of the tranng set) wth nodes V, edges E, attrbutes X, and true labels Y. For each of T trals, step 3 randomly selects lp% of the nodes to be known; ths value s chosen to replcate the fracton of known labels that are present n the test set. It then executes ICA C on the graph, gven the known nodes (step 4). Ths yelds the set Z U, whch contans the label predctons and assocated confdence values for each node n V U. Usng these and the expected class dstrbuton (from the tranng set), t then generates a meta-tranng vector per node (steps 5-7). Ths vector ncludes eght meta-features (descrbed later) and a Boolean value that ndcates whether predcton ŷ s correct. Ths tranng data s later used to learn the meta-classfer that predcts the correctness of the ŷ estmates gven the values of the metafeatures. We set T=0 to conduct ten trals wth dfferent known nodes each tme. The goal s to reduce the bas that mght otherwse occur due to the partcular selecton of Y n step 3. We later compare ths wth the one-tral approach (T=). Generatng meta-features from meta-tranng data ICA MC needs useful meta-features to predct when the node classfer has correctly classfed a node. The constructed features are based on two key premses. Frst, we assume that the data exhbts relatonal autocorrelaton (correlaton of class labels among nterrelated nodes, Jensen et al., 2004) for use by the node classfer. Thus, each node s predcted label wll be nfluenced by the predctons of ts neghborng labels. Second, snce ICA MC (lke ICA C ) explots only some of the predcted labels durng each teraton, not all neghbor labels wll affect the predcton for v. We assume that the accuracy of predcton ŷ for teraton s affected only by the neghbors of v that were ncluded n the commtted set Y durng that same teraton. Let N refer to the set of such neghbors for v. Based on these two premses and addtonal ntutons descrbed below, we desgned eght features for ths ntal study of ICA MC. The frst three features are based on ones used by Blgc and Getoor (2008) for a related problem that s dscussed later. Future work should examne these choces and others n more detal. Suppose the CC algorthm predcts ŷ to be the label for node v, wth confdence conf. Then v s features are:. Local score: The CC algorthm s predctons should dffer from those of an attrbute-only classfer (e.g., M A n Fgure ), or there s no pont n executng CC. However, f M A and the node classfer M AR agree on a predcton, then t s more lkely to be correct. Ths generatemetavectors(v,e,x,y,m AR,M A,n,T,lp,) = // V=nodes, E=edges, X=attrbute vectors, Y=node labels // M AR = node classfer (uses attrs. & relats), M A = classfer // (attrs. only), n=# ICA C ters., T=# randomzed trals to use // lp=labeled proporton, = expected class dstrbuton MetaTranVecs 2 for = to T do 3 // Randomly select some nodes to be known Y randomselectsomenodes(v, Y, lp) // Randomze V U {v y Y-Y } // Nodes used for predcton 4 // Run ICA C to predct labels and compute confdences Z U ICA_classfy(V,E,X,Y,M AR,M A,,n,ICA C,) 5 for each v V U do // Calc. and store meta-feature vectors 6 mf calcmetafeatures(,v,e,x,y,y,z U,M A,) 7 MetaTranVecs MetaTranVecs mf 8 return MetaTranVecs // return all vectors of meta-features Fgure 2: Pseudocode to generate tranng vectors for the meta classfer used by ICA MC. heurstc s captured by usng, for each v, M A s confdence value for the ŷ that was predcted by M AR. nown nodes are assumed to be fully correct (score of ), though ths could be reduced to account for possble nose: P( Y yˆ x ) v V lf v V 2. Relatonal score: If a node s surrounded by nodes whose predctons are more lkely (e.g., have hgh lf scores), then ts predcton s also more lkely: rf lf N 3. Global score: Let Pror(c) be the fracton of tranng nodes wth class label c, and Posteror(c) be the fracton of test set labels predcted as c by the CC algorthm. If Posteror(c) s much hgher than Pror(c), then many nodes wth predcted label c may be ncorrect. Thus, the global score measures whether class y s over or underrepresented n the posteror dstrbuton: Pror( yˆ ) Posteror( yˆ ) gf 2 4. Node confdence: If the node classfer s confdent n some predcton ŷ (hgh posteror probablty), then ths suggests that ŷ s more lkely to be correct: cf conf v N v V v V If only ths feature s used, ICA MC devolves to ICA C. 5. Neghbor confdence: As wth the relatonal score, more confdent neghbor predctons suggest that a node s predcton s more lkely to be correct: nf cf N v N

4 6. Neghbor agreement: If most of node v s neghbors have the same predcted label, ths may ndcate that ŷ s more lkely to be correct. Let count (N ) and count 2 (N ) ndcate the count of the two most frequent label predctons n N. If the former value s large and the latter s small, then neghbor agreement s hgh: naf count ( N ) count 2( N ) N 7. nown neghbors: Havng many known neghbors ncreases the chances that a node s predcton s correct: knf N V 8. nown vcnty: A node s predcton may also be nfluenced by known nodes that are lnked to t by one or more ntervenng nodes. We use a smple measure that favors drect known neghbors, then counts (wth reduced weght) any known nodes reached va one addtonal node v: kvf N V 2 v v' ( N N Each of these eght features may not be useful for every dataset. However, ICA MC needs only some of the features to be useful the meta-classfer (we use logstc regresson) wll learn approprate parameters for each feature based on ther predctve accuracy on the metatranng data. Also, features that provde no beneft are dscarded by the feature search process descrbed later. Evaluaton Hypotheses. By default, ICA MC uses feature search and ten randomzed tranng data trals. Ths ICA MC attans hgher accuraces than ICA C (Hypothess #), ICA MC wthout such trals (#2), ICA MC wthout feature search (#3), and ICA MC wth ust the three features used by Blgc and Getoor (#4). Data Sets. We used the followng data sets (see Table ):. Cora (see Sen et al. 2008): A collecton of machne learnng papers categorzed nto seven classes. 2. CteSeer (see Sen et al. 2008): A collecton of research papers drawn from the CteSeer collecton. 3. WebB (see Nevlle and Jensen 2007): A collecton of web pages from four computer scence departments. 4. Synthetc: We generate synthetc data usng Sen et al. s (2008) graph generator. Smlar to ther defaults, we use a degree of homophly of 0.7 and a lnk densty of 0.4. Table : Data sets summary Characterstcs Cora CteSeer WebB Syn. Total nodes n.a. Avg. # nodes per test set Avg. lnks per node Class labels Non-rel. features aval Non-rel. features used Relatonal features used Folds ) Feature Representaton. Our node representaton ncludes relatonal features and non-relatonal attrbutes, as descrbed below. Non-relatonal (content) attrbutes: The real datasets are all textual. We use a bag-of-words representaton for the textual content of each node, where the feature correspondng to a word s assgned true f t occurs n the node and false otherwse. Our verson of the WebB dataset has 00 words avalable. For Cora and CteSeer, we used nformaton gan to select the 00 hghest-scorng words, based on McDowell et al. (2007), whch reported that usng more dd not mprove performance. Our focus s on the case where relatvely few attrbutes are avalable (or the attrbutes are not very predctve) as may occur n large real-world networks (c.f., Macskassy and Provost 2007, Gallagher et al. 2008). Thus, for most of our experments we randomly select 0 of the 00 avalable words to use as attrbutes. We also brefly dscuss results when usng 00 attrbutes. For the synthetc data, ten bnary attrbutes are generated usng the technque descrbed by McDowell et al. (2009). Ths model has a parameter ap (attrbute predctveness) that ranges from 0.0 to.0; t ndcates how strongly predctve the attrbutes are of the class label. We evaluate ap usng the values {0.2, 0.4, 0.6}. Relatonal features: Each relatonal feature value s a multset. For nstance, a possble feature value s {3 A, 2 B, mssng}, whch ndcates that a node lnks to 3 other nodes whose predcted label s A, 2 nodes whose predcton s B, and node labeled mssng. Durng nference, each label n the multset (excludng mssng labels) s separately used to update the probablty that a node has label c. Ths s the ndependent value approach that was ntroduced by Nevlle et al. (2003), used by Nevlle and Jensen (2007), and shown to be superor to count or proporton features by McDowell et al. (2009). See Nevlle et al. (2003) for more detals. For Cora and CteSeer, we compute a multset feature usng only ncomng lnks, and a separate such feature usng only outgong lnks. For WebB, we also compute one such feature usng co-ctaton lnks (a co-ctaton lnk exsts between nodes and f some node k lnks to both of them). For the synthetc data, the lnks are undrected, so there s a sngle relatonal feature. Classfers. For the node classfer, we used a naïve Bayes classfer. McDowell et al. (2009) reported that, usng multset features, t attaned hgher accuraces than dd alternatves such as logstc regresson. For the metaclassfer, we used logstc regresson, as dd Blgc and Getoor (2008). Future work should consder other choces. Test Procedure. We conducted an n-fold cross-valdaton study for each tested algorthm. For WebB, we treated each of the four schools as a separate fold. For Cora and CteSeer, we created fve dsont test sets by usng smlarty-drven snowball samplng (McDowell et al. 2009). Ths s smlar to the approach of Sen et al. (2008).

5 For all 3 datasets we tested on one graph, traned on two others, and used the remanng two (one for WebB) as a holdout set for learnng the meta-classfer and performng the meta-feature search. For the synthetc data, we performed 25 separate trals. For each tral we generated three dsont graphs: one test set, one tranng set, and one holdout set. We randomly selected lp=0% of each test set to form V (nodes wth known labels). Ths s a sparsely labeled task, whch s common n real data (Gallagher et al. 2008). To search for whch of the eght meta-features to use wth ICA MC, we use the smple, greedy Backwards Sequental Elmnaton (BSE) algorthm (ttler, 986). It evaluates accuracy on the holdout set wth ICA MC, then recursvely elmnates any meta-feature whose removal ncreases accuracy. To ncrease robustness, accuracy s averaged over ten executons of ICA MC, each tme usng a dfferent set of ntal known labels (as done for T=0 n Fgure 2). The fnal set of meta-features s used for testng. Tested Algorthms. We tested ICA, ICA C, and ICA MC. In addton, to assess the utlty of ICA MC s desgn decsons, we also tested three of ts ablated varants:. tral nstead of 0 : ths uses only one randomzed tral to collect meta-tranng data (.e., T= n Fgure 2) and only one evaluaton tral for the meta-feature search. 2. No meta-feature search : Ths skps search and uses all eght meta-features that were prevously descrbed. 3. Only Blgc meta-feats : Ths uses ust features #, #2, and #3 the set used by Blgc and Getoor (2008). Performance Measure. We compared all the algorthms on ther average classfcaton error rate on the test sets. Analyss. We performed ndependent analyses for each predcton task and ont analyses by poolng the observatons, ether for all the real data sets or for all the synthetc data condtons shown. Our analyss uses onetaled pared t-tests accepted at the 95% confdence level. Results. Table 2 dsplays the classfcaton error rates averaged over all the folds for each algorthm. For each (data set, algorthm) par, the best result s shown n bold. Result : ICA MC sgnfcantly outperforms ICA C and ICA when attrbute predctveness s low: Comparng ICA MC wth ICA C, we fnd that ICA MC reduces classfcaton error by % for the real data, and.9-6.9% for the synthetc data. Ths mprovement s sgnfcant n every case (p <.03 for the real data and p <.045 for the synthetc data). In addton, the pooled analyses found sgnfcant gans for both the real and synthetc data. Therefore, we accept Hypothess #. For the synthetc data, the gans clearly decrease as attrbute predctveness (ap) ncreases. Ths s consstent wth the results of McDowell et al. (2009), who report that the cautous use of relatonal nformaton s more mportant for CC algorthms when ap and/or the number of attrbutes s small. Snce ICA MC s even more cautous than ICA C, ICA MC has larger gans over ICA C when ap s small (the same relatve trend exsts between ICA C and the non- Table 2: Average % classfcaton error rate Real datasets Synthetc data Core Algorthms Web Cora CS B ap=.2 ap=.4 ap=.6 ICA ICA C ICA MC Gan* Varants of ICA MC tral nstead of No meta-feat. search Only Blgc meta-feats ndcates sgnfcantly worse behavor than ICA MC. * ndcates gan from meta-cauton (ICA C ICA MC) cautous ICA). Nonetheless, ICA MC contnues to provde a small gan even when ap s hgh a gan of 0.9% when ap=0.8 (results not shown). For the real data, ICA MC provdes gans for all three datasets, where the largest gan s wth WebB. WebB has more complex and numerous lnkng patterns (Macskassy and Provost 2007). For ths reason, ICA MC s careful selecton of whch neghborng labels to use for predcton may be especally mportant wth WebB. We repeated these experments wth real data usng 2, 5, or 20 attrbutes (nstead of 0) and found smlar results. In every case pooled analyses found a sgnfcant gan for ICA MC over ICA C (average gans rangng from %), wth the largest gans occurrng wth WebB. As wth the synthetc data, these gans dmnsh when the attrbutes are more predctve. For nstance, when 00 attrbutes are used the gans of ICA MC remaned but were small (0.2-.0%) and statstcally nsgnfcant. These results suggest that ICA MC s especally helpful when the attrbutes alone are not very predctve, and at least does no harm otherwse. Result 2: ICA MC wth randomzed trals and meta-feature search outperforms smpler varants: The bottom of Table 2 shows results wth the varants of ICA MC that do not use multple randomzed trals or do less or no meta-feature search. ICA MC outperforms the tral nstead of 0 and Only Blgc meta-feats varants, often sgnfcantly, and pooled analyses fnd that ICA MC outperforms both, for the real and for the synthetc data. Thus, we accept Hypotheses #2 and #4. ICA MC also sgnfcantly outperforms the varant that uses all eght meta-features ( No meta-feat. search ) for the real data, but not for the synthetc data (perhaps because smpler, undrected lnkng patterns were used n the synthetc data). Thus, we reect Hypothess #3. Despte the reecton of one hypothess, ICA MC always outperformed all three varants (or lagged by at most 0.2%) and sgnfcantly outperformed all three varants on the real datasets. Some of the varants that smplfy ICA MC s search process sometmes performed notably worse than even ICA C. Together, these results suggest that the complete ICA MC, wth randomzed trals and feature search, s the most robust performer.

6 Dscusson ICA MC ncreased accuracy compared to ICA and ICA C. However, why does ICA MC s meta-classfer more effectvely dentfy relable predctons than does ICA C s node classfer? Frst, the meta-classfer s task s smpler: choosng between two values (correct or ncorrect) vs. between all possble class labels. Second, the metaclassfer can use addtonal nformaton, such as the number of known labels, whch has no obvous utlty for predctng a partcular label, but does help estmate the correctness of the resultant predcton. Fnally, usng two dfferent classfers helps to reduce the bas due to usng the Naïve Bayes node classfer alone. Meta-feature search often sgnfcantly ncreased ICA MC s accuracy. However, s the same set of features almost always chosen? On average, the global score and node confdence features were selected most often, and known neghbor least often. Ths vared substantally, however, wth some features selected 90% of the tme for one dataset and never for another. These results, combned wth the results from Table 2, suggest that search s essental to make ICA MC robust across dfferent data, even f the default set of meta-features s further refned. We are not aware of any other work that uses a metaclassfer to mprove the operaton of a CC nference algorthm, although Blgc and Getoor (2008) dd use a smlar predctor to dentfy the least lkely CC label predctons (n order to purchase the correct labels for them). In contrast, we seek the most lkely predctons (to favor them for nference). They consdered three features for ths dfferent task, whch our search algorthm selected for ICA MC 62%, 67%, and 9% of the tme, respectvely. Thus, ther features are also useful for our task, although the results of the prevous secton show that usng only those features leads to very poor performance for ICA MC. Compared to ICA C, ICA MC requres addtonal computaton: to execute ICA C when collectng metatranng data, to execute ICA MC for feature selecton, and to tran the meta-classfer for each combnaton of metafeatures that are consdered. However, n many real-world graphs each node lnks to at most k other nodes, n whch case each of these steps s lnear n the number of nodes. In addton, once the meta-classfer s learned, ICA MC requres lttle addtonal tme for nference compared to ICA C (.e., t needs only one addtonal executon of the meta-classfer per teraton). Concluson We demonstrated that Meta-Cautous ICA (ICA MC ) sgnfcantly outperforms ICA C for many tasks. Moreover, we showed that aspects of ICA MC n partcular, ts use of multple randomzed tranng data trals and ts use of search for selectng meta-features were essental to achevng performance that was robust across a range of datasets. Snce ICA C has already been shown to be a very effectve CC algorthm, these results suggest that ICA MC should be serously consdered for CC applcatons, partcularly when attrbutes alone do not yeld hgh predctve accuracy. Further work s needed to confrm our results usng other datasets, meta-features, and classfers, and to consder how meta-cauton mght be extended to other CC algorthms. In addton, we ntend to consder technques for further reducng the tme complexty of ICA MC compared to ICA C. Acknowledgements We thank the Naval Research Laboratory for supportng ths work, and Bran Gallagher, Gta Sukthankar, and the anonymous revewers for ther helpful comments. References Blgc, M. and Getoor, L. (2008). Effectve label acquston for collectve classfcaton. Proceedngs of the Fourteenth ACM SIGDD Internatonal Conference on nowledge Dscovery and Data Mnng (pp. 43 5). Las Vegas, NV: ACM. Chakrabart, S., Dom, B., and Indyk, P. (998). Enhanced hypertext categorzaton usng hyperlnks. Proceedngs of the ACM SIGMOD Internatonal Conference on Management of Data (pp ). Seattle, WA: ACM. Gallagher, B., Tong, H., Elass-Rad, T., and Faloutsos, C. (2008). Usng ghost edges for classfcaton n sparsely labeled networks. Proceedngs of the Fourteenth ACM SIGDD Internatonal Conference on nowledge Dscovery and Data Mnng (pp ). Las Vegas, NV: ACM. Jensen, D., Nevlle, J., and Gallagher, B. (2004). Why collectve nference mproves relatonal classfcaton. Proceedngs of the Tenth ACM SIGDD Internatonal Conference on nowledge Dscovery and Data Mnng (pp ). Seattle, WA: ACM. ttler, J. (986). Feature selecton and extracton. In T.Y. Young &. Fu (Eds.), Handbook of Pattern Recognton and Image Processng. San Dego, CA: Academc Press. Macskassy, S. and Provost, F. (2007). Classfcaton n network data: a toolkt and a unvarate case study. Journal of Machne Learnng Research, 8, McDowell, L., Gupta,. M., and Aha, D.W. (2007). Cautous nference n collectve classfcaton. In Proceedngs of the Twenty-Second AAAI Conference on Artfcal Intellgence (pp ). Vancouver, Canada: AAAI. McDowell, L., Gupta,. M., and Aha, D.W. (2009). Cautous collectve classfcaton. Journal of Machne Learnng Research, 0, Nevlle, J., and Jensen, D. (2000). Iteratve classfcaton n relatonal data. In L. Getoor and D. Jensen (Eds.) Learnng Statstcal Models from Relatonal Data: Papers from the AAAI Workshop (Techncal Report WS-00-06). Austn, TX: AAAI. Nevlle, J. and Jensen, D. (2007). Relatonal dependency networks. Journal of Machne Learnng Research, 8, Nevlle, J., Jensen, D., and Gallagher, B. (2003). Smple estmators for relatonal bayesan classfers. Proceedngs of the Thrd IEEE Internatonal Conference on Data Mnng (pp ). Melbourne, FL: IEEE. Sen, P., Namata, G., Blgc, M., Getoor, L., Gallagher, B., and Elass-Rad, T. (2008). Collectve classfcaton n network data. AI Magazne, 29(3), Taskar, B., Abbeel, P., and oller, D. (2002). Dscrmnatve probablstc models for relatonal data. Proceedngs of the Eghteenth Conference on Uncertanty n Artfcal Intellgence (pp ). Edmonton, (BC) Canada: Morgan aufmann.

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

Data Mining: Model Evaluation

Data Mining: Model Evaluation Data Mnng: Model Evaluaton Aprl 16, 2013 1 Issues: Evaluatng Classfcaton Methods Accurac classfer accurac: predctng class label predctor accurac: guessng value of predcted attrbutes Speed tme to construct

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

arxiv: v1 [cs.lg] 19 Mar 2017

arxiv: v1 [cs.lg] 19 Mar 2017 Recurrent Collectve Classfcaton Shuangfe Fan Dept of Electrcal and Computer Engneerng Vrgna Tech Bert Huang Dept of Computer Scence Vrgna Tech arxv:170306514v1 cslg 19 Mar 2017 Abstract We propose a new

More information

Multi-Label Collective Classification

Multi-Label Collective Classification Mult-Label Collectve Classfcaton Xangnan Kong Xaoxao Sh Phlp S. u Abstract Collectve classfcaton n relatonal data has become an mportant and actve research topc n the last decade, where class labels for

More information

Node Classification in Dynamic Social Networks

Node Classification in Dynamic Social Networks Node Classfcaton n Dynamc Socal Networks Yulong Pe, Janpeng Zhang, George H.L. Fletcher, and Mykola Pechenzky Department of Mathematcs and Computer Scence Endhoven Unversty of Technology, Endhoven, the

More information

Implementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status

Implementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status Internatonal Journal of Appled Busness and Informaton Systems ISSN: 2597-8993 Vol 1, No 2, September 2017, pp. 6-12 6 Implementaton Naïve Bayes Algorthm for Student Classfcaton Based on Graduaton Status

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Classifying Acoustic Transient Signals Using Artificial Intelligence

Classifying Acoustic Transient Signals Using Artificial Intelligence Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Intelligent Information Acquisition for Improved Clustering

Intelligent Information Acquisition for Improved Clustering Intellgent Informaton Acquston for Improved Clusterng Duy Vu Unversty of Texas at Austn duyvu@cs.utexas.edu Mkhal Blenko Mcrosoft Research mblenko@mcrosoft.com Prem Melvlle IBM T.J. Watson Research Center

More information

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors Onlne Detecton and Classfcaton of Movng Objects Usng Progressvely Improvng Detectors Omar Javed Saad Al Mubarak Shah Computer Vson Lab School of Computer Scence Unversty of Central Florda Orlando, FL 32816

More information

Investigating the Performance of Naïve- Bayes Classifiers and K- Nearest Neighbor Classifiers

Investigating the Performance of Naïve- Bayes Classifiers and K- Nearest Neighbor Classifiers Journal of Convergence Informaton Technology Volume 5, Number 2, Aprl 2010 Investgatng the Performance of Naïve- Bayes Classfers and K- Nearest Neghbor Classfers Mohammed J. Islam *, Q. M. Jonathan Wu,

More information

Application of Maximum Entropy Markov Models on the Protein Secondary Structure Predictions

Application of Maximum Entropy Markov Models on the Protein Secondary Structure Predictions Applcaton of Maxmum Entropy Markov Models on the Proten Secondary Structure Predctons Yohan Km Department of Chemstry and Bochemstry Unversty of Calforna, San Dego La Jolla, CA 92093 ykm@ucsd.edu Abstract

More information

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010 Smulaton: Solvng Dynamc Models ABE 5646 Week Chapter 2, Sprng 200 Week Descrpton Readng Materal Mar 5- Mar 9 Evaluatng [Crop] Models Comparng a model wth data - Graphcal, errors - Measures of agreement

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

An algorithm for correcting mislabeled data

An algorithm for correcting mislabeled data Intellgent Data Analyss 5 (2001) 491 2 491 IOS Press An algorthm for correctng mslabeled data Xnchuan Zeng and Tony R. Martnez Computer Scence Department, Brgham Young Unversty, Provo, UT 842, USA E-mal:

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Bayesian Approach for Fatigue Life Prediction from Field Inspection

Bayesian Approach for Fatigue Life Prediction from Field Inspection Bayesan Approach for Fatgue Lfe Predcton from Feld Inspecton Dawn An, and Jooho Cho School of Aerospace & Mechancal Engneerng, Korea Aerospace Unversty skal@nate.com, jhcho@kau.ac.kr Nam H. Km, and Srram

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Under-Sampling Approaches for Improving Prediction of the Minority Class in an Imbalanced Dataset

Under-Sampling Approaches for Improving Prediction of the Minority Class in an Imbalanced Dataset Under-Samplng Approaches for Improvng Predcton of the Mnorty Class n an Imbalanced Dataset Show-Jane Yen and Yue-Sh Lee Department of Computer Scence and Informaton Engneerng, Mng Chuan Unversty 5 The-Mng

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article. A selective ensemble classification method on microarray data

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article. A selective ensemble classification method on microarray data Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(6):2860-2866 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 A selectve ensemble classfcaton method on mcroarray

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

CAN COMPUTERS LEARN FASTER? Seyda Ertekin Computer Science & Engineering The Pennsylvania State University

CAN COMPUTERS LEARN FASTER? Seyda Ertekin Computer Science & Engineering The Pennsylvania State University CAN COMPUTERS LEARN FASTER? Seyda Ertekn Computer Scence & Engneerng The Pennsylvana State Unversty sertekn@cse.psu.edu ABSTRACT Ever snce computers were nvented, manknd wondered whether they mght be made

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

A Statistical Model Selection Strategy Applied to Neural Networks

A Statistical Model Selection Strategy Applied to Neural Networks A Statstcal Model Selecton Strategy Appled to Neural Networks Joaquín Pzarro Elsa Guerrero Pedro L. Galndo joaqun.pzarro@uca.es elsa.guerrero@uca.es pedro.galndo@uca.es Dpto Lenguajes y Sstemas Informátcos

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Fast and Scalable Training of Semi-Supervised CRFs with Application to Activity Recognition

Fast and Scalable Training of Semi-Supervised CRFs with Application to Activity Recognition Fast and Scalable Tranng of Sem-Supervsed CRFs wth Applcaton to Actvty Recognton Maryam Mahdavan Computer Scence Department Unversty of Brtsh Columba Vancouver, BC, Canada Tanzeem Choudhury Intel Research

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

Learning Non-Linearly Separable Boolean Functions With Linear Threshold Unit Trees and Madaline-Style Networks

Learning Non-Linearly Separable Boolean Functions With Linear Threshold Unit Trees and Madaline-Style Networks In AAAI-93: Proceedngs of the 11th Natonal Conference on Artfcal Intellgence, 33-1. Menlo Park, CA: AAAI Press. Learnng Non-Lnearly Separable Boolean Functons Wth Lnear Threshold Unt Trees and Madalne-Style

More information

A Robust Method for Estimating the Fundamental Matrix

A Robust Method for Estimating the Fundamental Matrix Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.

More information

A Lazy Ensemble Learning Method to Classification

A Lazy Ensemble Learning Method to Classification IJCSI Internatonal Journal of Computer Scence Issues, Vol. 7, Issue 5, September 2010 ISSN (Onlne): 1694-0814 344 A Lazy Ensemble Learnng Method to Classfcaton Haleh Homayoun 1, Sattar Hashem 2 and Al

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

A Multivariate Analysis of Static Code Attributes for Defect Prediction

A Multivariate Analysis of Static Code Attributes for Defect Prediction Research Paper) A Multvarate Analyss of Statc Code Attrbutes for Defect Predcton Burak Turhan, Ayşe Bener Department of Computer Engneerng, Bogazc Unversty 3434, Bebek, Istanbul, Turkey {turhanb, bener}@boun.edu.tr

More information

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines (IJCSIS) Internatonal Journal of Computer Scence and Informaton Securty, Herarchcal Web Page Classfcaton Based on a Topc Model and Neghborng Pages Integraton Wongkot Srura Phayung Meesad Choochart Haruechayasak

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Learning-Based Top-N Selection Query Evaluation over Relational Databases

Learning-Based Top-N Selection Query Evaluation over Relational Databases Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

CSCI 5417 Information Retrieval Systems Jim Martin!

CSCI 5417 Information Retrieval Systems Jim Martin! CSCI 5417 Informaton Retreval Systems Jm Martn! Lecture 11 9/29/2011 Today 9/29 Classfcaton Naïve Bayes classfcaton Ungram LM 1 Where we are... Bascs of ad hoc retreval Indexng Term weghtng/scorng Cosne

More information

A classification scheme for applications with ambiguous data

A classification scheme for applications with ambiguous data A classfcaton scheme for applcatons wth ambguous data Thomas P. Trappenberg Centre for Cogntve Neuroscence Department of Psychology Unversty of Oxford Oxford OX1 3UD, England Thomas.Trappenberg@psy.ox.ac.uk

More information

USING GRAPHING SKILLS

USING GRAPHING SKILLS Name: BOLOGY: Date: _ Class: USNG GRAPHNG SKLLS NTRODUCTON: Recorded data can be plotted on a graph. A graph s a pctoral representaton of nformaton recorded n a data table. t s used to show a relatonshp

More information

Deep Classification in Large-scale Text Hierarchies

Deep Classification in Large-scale Text Hierarchies Deep Classfcaton n Large-scale Text Herarches Gu-Rong Xue Dkan Xng Qang Yang 2 Yong Yu Dept. of Computer Scence and Engneerng Shangha Jao-Tong Unversty {grxue, dkxng, yyu}@apex.sjtu.edu.cn 2 Hong Kong

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Error Detection and Impact-Sensitive Instance Ranking in Noisy Datasets

Error Detection and Impact-Sensitive Instance Ranking in Noisy Datasets Error Detecton and Impact-Senstve Instance Ranng n osy Datasets Xngquan Zhu, Xndong Wu, and Yng Yang Department of Computer Scence, Unversty of Vermont, Burlngton VT 05405, USA {xqzhu, xwu, yyang}@cs.uvm.edu

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

A User Selection Method in Advertising System

A User Selection Method in Advertising System Int. J. Communcatons, etwork and System Scences, 2010, 3, 54-58 do:10.4236/jcns.2010.31007 Publshed Onlne January 2010 (http://www.scrp.org/journal/jcns/). A User Selecton Method n Advertsng System Shy

More information

Query Clustering Using a Hybrid Query Similarity Measure

Query Clustering Using a Hybrid Query Similarity Measure Query clusterng usng a hybrd query smlarty measure Fu. L., Goh, D.H., & Foo, S. (2004). WSEAS Transacton on Computers, 3(3), 700-705. Query Clusterng Usng a Hybrd Query Smlarty Measure Ln Fu, Don Hoe-Lan

More information

CMPS 10 Introduction to Computer Science Lecture Notes

CMPS 10 Introduction to Computer Science Lecture Notes CPS 0 Introducton to Computer Scence Lecture Notes Chapter : Algorthm Desgn How should we present algorthms? Natural languages lke Englsh, Spansh, or French whch are rch n nterpretaton and meanng are not

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK

FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK FINDING IMPORTANT NODES IN SOCIAL NETWORKS BASED ON MODIFIED PAGERANK L-qng Qu, Yong-quan Lang 2, Jng-Chen 3, 2 College of Informaton Scence and Technology, Shandong Unversty of Scence and Technology,

More information

NetCycle: Collective Evolution Inference in Heterogeneous Information Networks

NetCycle: Collective Evolution Inference in Heterogeneous Information Networks : Collectve Evoluton Inference n Heterogeneous Informaton Networks ABSTRACT Yzhou Zhang Fudan Unversty Shangha, Chna yzhouzhang4@fudan.edu.cn Xangnan Kong Worcester Polytechnc Insttute Worcester, MA, USA

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

A Hill-climbing Landmarker Generation Algorithm Based on Efficiency and Correlativity Criteria

A Hill-climbing Landmarker Generation Algorithm Based on Efficiency and Correlativity Criteria A Hll-clmbng Landmarker Generaton Algorthm Based on Effcency and Correlatvty Crtera Daren Ler, Irena Koprnska, and Sanjay Chawla School of Informaton Technologes, Unversty of Sydney Madsen Buldng F09,

More information

Fusion Performance Model for Distributed Tracking and Classification

Fusion Performance Model for Distributed Tracking and Classification Fuson Performance Model for Dstrbuted rackng and Classfcaton K.C. Chang and Yng Song Dept. of SEOR, School of I&E George Mason Unversty FAIRFAX, VA kchang@gmu.edu Martn Lggns Verdan Systems Dvson, Inc.

More information

Feature Selection as an Improving Step for Decision Tree Construction

Feature Selection as an Improving Step for Decision Tree Construction 2009 Internatonal Conference on Machne Learnng and Computng IPCSIT vol.3 (2011) (2011) IACSIT Press, Sngapore Feature Selecton as an Improvng Step for Decson Tree Constructon Mahd Esmael 1, Fazekas Gabor

More information

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics Ths module s part of the Memobust Handbook on Methodology of Modern Busness Statstcs 26 March 2014 Theme: Donor Imputaton Contents General secton... 3 1. Summary... 3 2. General descrpton... 3 2.1 Introducton

More information

Biological Sequence Mining Using Plausible Neural Network and its Application to Exon/intron Boundaries Prediction

Biological Sequence Mining Using Plausible Neural Network and its Application to Exon/intron Boundaries Prediction Bologcal Sequence Mnng Usng Plausble Neural Networ and ts Applcaton to Exon/ntron Boundares Predcton Kuochen L, Dar-en Chang, and Erc Roucha CECS, Unversty of Lousvlle, Lousvlle, KY 40292, USA Yuan Yan

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS

EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS P.G. Demdov Yaroslavl State Unversty Anatoly Ntn, Vladmr Khryashchev, Olga Stepanova, Igor Kostern EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS Yaroslavl, 2015 Eye

More information

Associative Based Classification Algorithm For Diabetes Disease Prediction

Associative Based Classification Algorithm For Diabetes Disease Prediction Internatonal Journal of Engneerng Trends and Technology (IJETT) Volume-41 Number-3 - November 016 Assocatve Based Classfcaton Algorthm For Dabetes Dsease Predcton 1 N. Gnana Deepka, Y.surekha, 3 G.Laltha

More information

Announcements. Supervised Learning

Announcements. Supervised Learning Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples

More information

Incremental Learning with Support Vector Machines and Fuzzy Set Theory

Incremental Learning with Support Vector Machines and Fuzzy Set Theory The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and

More information

NetCycle: Collective Evolution Inference in Heterogeneous Information Networks

NetCycle: Collective Evolution Inference in Heterogeneous Information Networks : Collectve Evoluton Inference n Heterogeneous Informaton Networks Yzhou Zhang,2 Yun Xong,2 Xangnan Kong 3 Yangyong Zhu,2 Shangha Key Laboratory of Data Scence, Shangha, Chna 2 School of Computer Scence,

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

Learning to Classify Documents with Only a Small Positive Training Set

Learning to Classify Documents with Only a Small Positive Training Set Learnng to Classfy Documents wth Only a Small Postve Tranng Set Xao-L L 1, Bng Lu 2, and See-Kong Ng 1 1 Insttute for Infocomm Research, Heng Mu Keng Terrace, 119613, Sngapore 2 Department of Computer

More information

y and the total sum of

y and the total sum of Lnear regresson Testng for non-lnearty In analytcal chemstry, lnear regresson s commonly used n the constructon of calbraton functons requred for analytcal technques such as gas chromatography, atomc absorpton

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

Comparing High-Order Boolean Features

Comparing High-Order Boolean Features Brgham Young Unversty BYU cholarsarchve All Faculty Publcatons 2005-07-0 Comparng Hgh-Order Boolean Features Adam Drake adam_drake@yahoo.com Dan A. Ventura ventura@cs.byu.edu Follow ths and addtonal works

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

GA-Based Learning Algorithms to Identify Fuzzy Rules for Fuzzy Neural Networks

GA-Based Learning Algorithms to Identify Fuzzy Rules for Fuzzy Neural Networks Seventh Internatonal Conference on Intellgent Systems Desgn and Applcatons GA-Based Learnng Algorthms to Identfy Fuzzy Rules for Fuzzy Neural Networks K Almejall, K Dahal, Member IEEE, and A Hossan, Member

More information

SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE

SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE Dorna Purcaru Faculty of Automaton, Computers and Electroncs Unersty of Craoa 13 Al. I. Cuza Street, Craoa RO-1100 ROMANIA E-mal: dpurcaru@electroncs.uc.ro

More information

Self-tuning Histograms: Building Histograms Without Looking at Data

Self-tuning Histograms: Building Histograms Without Looking at Data Self-tunng Hstograms: Buldng Hstograms Wthout Lookng at Data Ashraf Aboulnaga Computer Scences Department Unversty of Wsconsn - Madson ashraf@cs.wsc.edu Surajt Chaudhur Mcrosoft Research surajtc@mcrosoft.com

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research

More information

TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS. Muradaliyev A.Z.

TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS. Muradaliyev A.Z. TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS Muradalyev AZ Azerbajan Scentfc-Research and Desgn-Prospectng Insttute of Energetc AZ1012, Ave HZardab-94 E-mal:aydn_murad@yahoocom Importance of

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Contrary to Popular Belief Incremental Discretization can be Sound, Computationally Efficient and Extremely Useful for Streaming Data

Contrary to Popular Belief Incremental Discretization can be Sound, Computationally Efficient and Extremely Useful for Streaming Data Contrary to Popular Belef Incremental Dscretzaton can be Sound, Computatonally Effcent and Extremely Useful for Streamng Data Geoffrey I. Webb Faculty of Informaton Technology, Monash Unversty, Vctora,

More information

An enhanced representation of time series which allows fast and accurate classification, clustering and relevance feedback

An enhanced representation of time series which allows fast and accurate classification, clustering and relevance feedback An enhanced representaton of tme seres whch allows fast and accurate classfcaton, clusterng and relevance feedback Eamonn J. Keogh and Mchael J. Pazzan Department of Informaton and Computer Scence Unversty

More information

A Semi-parametric Regression Model to Estimate Variability of NO 2

A Semi-parametric Regression Model to Estimate Variability of NO 2 Envronment and Polluton; Vol. 2, No. 1; 2013 ISSN 1927-0909 E-ISSN 1927-0917 Publshed by Canadan Center of Scence and Educaton A Sem-parametrc Regresson Model to Estmate Varablty of NO 2 Meczysław Szyszkowcz

More information