A Multi-class SVM Classifier Utilizing Binary Decision Tree

Size: px
Start display at page:

Download "A Multi-class SVM Classifier Utilizing Binary Decision Tree"

Transcription

1 Informatica 33 (009) A Multi-class Classifier Utilizing Binary Decision Tree Gjorgji Mazarov, Dejan Gjorgjevikj an Ivan Chorbev Department of Computer Science an Engineering Faculty of Electrical Engineering an Information Technology Karpos b.b., 1000 Skopje, Maceonia mazarovg@feit.ukim.eu.mk Keywors: Support Vector Machine, multi-class classification, clustering, binary ecision tree architecture Receive: July 7, 008 In this paper a novel architecture of Support Vector Machine classifiers utilizing binary ecision tree (-BDT) for solving multiclass problems is presente. The hierarchy of binary ecision subtasks using s is esigne with a clustering algorithm. For consistency between the clustering moel an, the clustering moel utilizes istance measures at the kernel space, rather than at the input space. The propose base Binary Decision Tree architecture takes avantage of both the efficient computation of the ecision tree architecture an the high classification accuracy of s. The - BDT architecture was esigne to provie superior multi-class classification performance. Its performance was measure on samples from MNIST, Penigit, Optigit an Statlog atabases of hanwritten igits an letters. The results of the experiments inicate that while maintaining comparable or offering better accuracy with other base approaches, ensembles of trees (Bagging an Ranom Forest) an neural network, the training phase of -BDT is faster. During recognition phase, ue to its logarithmic complexity, -BDT is much faster than the wiely use multi-class methos like one-against-one an one-against-all, for multiclass problems. Furthermore, the experiments showe that the propose metho becomes more favourable as the number of classes in the recognition problem increases. Povzetek: Prestavljena je metoa granje binarnih reves z uporabo za večrazrene probleme. 1 Introuction The recent results in pattern recognition have shown that support vector machine () classifiers often have superior recognition rates in comparison to other classification methos. However, the was originally evelope for binary ecision problems, an its extension to multi-class problems is not straightforwar. How to effectively exten it for solving multiclass classification problem is still an on-going research issue. The popular methos for applying s to multiclass classification problems usually ecompose the multi-class problems into several two-class problems that can be aresse irectly using several s. For the reaers convenience, we introuce the briefly in section. A brief introuction to several wiely use multi-class classification methos that utilize binary s is given in section 3. The Kernelbase clustering introuce to convert the multi-class problem into -base binary ecision-tree architecture is explaine in section 4. In section 5, we iscuss relate works an compare -BDT with other multi-class methos via theoretical analysis an empirical estimation. The experimental results in section are presente to compare the performance of the propose -BDT with traitional multi-class approaches base on, ensemble of ecision trees an neural network. Section 7 gives a conclusion of the paper. Support vector machines for pattern recognition The support vector machine is originally a binary classification metho evelope by Vapnik an colleagues at Bell laboratories [1][], with further algorithm improvements by others [3]. For a binary problem, we have training ata points: {x i, y i }, i=1,...,l, y i {-1, 1}, x i R. Suppose we have some hyperplane which separates the positive from the negative examples (a separating hyperplane ). The points x which lie on the hyperplane satisfy w x + b = 0, where w is normal to the hyperplane, b / w is the perpenicular istance from the hyperplane to the origin, an w is the Eucliean norm of w. Let + ( - ) be the shortest istance from the separating hyperplane to the closest positive (negative) example. Define the margin of a separating hyperplane to be For the linearly separable case, the support vector algorithm simply looks for the separating hyperplane with largest margin. This can be formulate as follows: suppose that all the training ata satisfy the following constraints:

2 34 Informatica 33 (009) G. Mazarov et al. xi w b 1 for y i 1, ( 1 ) xi w b 1 for y i 1, ( ) These can be combine into one set of inequalities: yi x w b 1 0 i i, ( 3 ) Now consier the points for which the equality in Eq. (1) hols (requiring that there exists such a point) is equivalent to choosing a scale for w an b. These points lie on the hyperplane H 1 : x i w + b = 1 with normal w an perpenicular istance from the origin 1-b / w. Similarly, the points for which the equality in Eq. () hols lie on the hyperplane H : x i w + b = -1, with normal again w an perpenicular istance from the origin -1-b / w. Hence + = - = 1/ w an the margin is simply / w. constraints also form a convex set (any linear constraint efines a convex set, an a set of N simultaneous linear constraints efines the intersection of N convex sets, which is also a convex set). This means that we can equivalently solve the following ual problem: maximize L P, subject to the constraints that the graient of L P with respect to w an b vanish, an subject also to the constraints that the α i 0 (let s call that particular set of constraints C ). This particular ual formulation of the problem is calle the Wolfe ual [4]. It has the property that the maximum of L P, subject to constraints C, occurs at the same values of the w, b an α, as the minimum of L P, subject to constraints C 1. Requiring that the graient of L P with respect to w an b vanish gives the conitions: w i y i xi, ( 5 ) i i y i 0. ( ) i Since these are equality constraints in the ual formulation, we can substitute them into Eq. (4) to give L D i l 1 i i j yi y jxi x j, ( 7 ) i, j origin margin Figure 1 Linear separating hyperplanes for the separable case. The support vectors are circle. Note that H 1 an H are parallel (they have the same normal) an that no training points fall between them. Thus we can fin the pair of hyperplanes which gives the maximum margin by minimizing w, subject to constraints (3). Thus we expect the solution for a typical two imensional case to have the form shown on Fig. 1. We introuce nonnegative Lagrange multipliers α i, i = 1,..., l, one for each of the inequality constraints (3). Recall that the rule is that for constraints of the form c i 0, the constraint equations are multiplie by nonnegative Lagrange multipliers an subtracte from the objective function, to form the Lagrangian. For equality constraints, the Lagrange multipliers are unconstraine. This gives Lagrangian: L p 1 l l w i yi xi w b i, ( 4 ) i1 i1 We must now minimize L p with respect to w, b, an maximize with respect to all α i at the same time, all subject to the constraints α i 0 (let s call this particular set of constraints C 1 ). Now this is a convex quaratic programming problem, since the objective function is itself convex, an those points which satisfy the Note that we have now given the Lagrangian ifferent labels (P for primal, D for ual) to emphasize that the two formulations are ifferent: L P an L D arise from the same objective function but with ifferent constraints; an the solution is foun by minimizing L P or by maximizing L D. Note also that if we formulate the problem with b = 0, which amounts to requiring that all hyperplanes contain the origin, the constraint () oes not appear. This is a mil restriction for high imensional spaces, since it amounts to reucing the number of egrees of freeom by one. Support vector training (for the separable, linear case) therefore amounts to maximizing L D with respect to the α i, subject to constraints () an positivity of the α i, with solution given by (5). Notice that there is a Lagrange multiplier α i for every training point. In the solution, those points for which α i > 0 are calle support vectors, an lie on one of the hyperplanes H 1, H. All other training points have α i = 0 an lie either on H 1 or H (such that the equality in Eq. (3) hols), or on that sie of H 1 or H such that the strict inequality in Eq. (3) hols. For these machines, the support vectors are the critical elements of the training set. They lie closest to the ecision bounary; if all other training points were remove (or move aroun, but so as not to cross H 1 or H ), an training was repeate, the same separating hyperplane woul be foun. The above algorithm for separable ata, when applie to non-separable ata, will fin no feasible solution: this will be evience by the objective function (i.e. the ual Lagrangian) growing arbitrarily large. So how can we exten these ieas to hanle non-separable ata? We woul like to relax the constraints (1) an (), but only

3 A MULTI-CLASS CLASSIFIER... Informatica 33 (009) when necessary, that is, we woul like to introuce a further cost (i.e. an increase in the primal objective function) for oing so. This can be one by introucing positive slack variables e i ; i = 1,..., l, in the constraints, which then become: xi w b 1 e i for y i 1, ( 8 ) xi w b 1 e i for y i 1, ( 9 ) e i 0i. ( 10 ) Thus, for an error to occur, the corresponing e i must excee unity, so Σ i e i is an upper boun on the number of training errors. Hence a natural way to assign an extra cost for errors is to change the objective function to be minimize from w / to w / + C(Σ i e i ), where C is a parameter to be chosen by the user, a larger C corresponing to assigning a higher penalty to errors. How can the above methos be generalize to the case where the ecision function (f(x) whose sign represents the class assigne to ata point x) is not a linear function of the ata? First notice that the only way in which the ata appears in the training problem, is in the form of ot proucts, x i x j. Now suppose we first mappe the ata (Figure ) to some other (possibly even infinite imensional) Eucliean space H, using a mapping which we will call Ф: : R H, ( 11 ) Then of course the training algorithm woul only epen on the ata through ot proucts in H, i.e. on functions of the form Ф(x i ) Ф(x j ). Now if there were a kernel function K such that K(x i, x j ) = Ф(x i ) Ф(x j ), we woul only nee to use K in the training algorithm, an woul never nee to explicitly even know what Ф is. The kernel function has to satisfy Mercer s conition [1].One example for this function is Gaussian: xi x j Kx i, x j exp, ( 1 ) In this particular example, H is infinite imensional, so it woul not be very easy to work with Ф explicitly. However, if one replaces x i x j by K(x i, x j ) everywhere in the training algorithm, the algorithm will happily prouce a support vector machine which lives in an infinite imensional space, an furthermore o so in roughly the same amount of time it woul take to train on the un-mappe ata. All the consierations of the previous sections hol, since we are still oing a linear separation, but in a ifferent space. But how can we use this machine? After all, we nee w, an that will live in H. But in test phase an is use by computing ot proucts of a given test point x with w, or more specifically by computing the sign of Ns N s f ( x) i yis i x b i yi Ks i, x b ( 13 ) i1 i1 where the s i are the support vectors. So again we can avoi computing Ф(x) explicitly an use the K(s i, x) = Ф(s i ) Ф(x) instea. Figure General principle of : projection of ata in an optimal imensional space. 3 An overview of wiely use multiclass classification methos Although s were originally esigne as binary classifiers, approaches that aress a multi-class problem as a single all-together optimization problem exist [5], but are computationally much more expensive than solving several binary problems. A variety of techniques for ecomposition of the multi-class problem into several binary problems using Support Vector Machines as binary classifiers have been propose, an several wiely use are given in this section. 3.1 One-against-all (OvA) For the N-class problems (N>), N two-class classifiers are constructe []. The i th is traine while labeling the samples in the i th class as positive examples an all the rest as negative examples. In the recognition phase, a test example is presente to all N s an is labelle accoring to the maximum output among the N classifiers. The isavantage of this metho is its training complexity, as the number of training samples is large. Each of the N classifiers is traine using all available samples. 3. One-against-one (OvO) This algorithm constructs N(N-1)/ two-class classifiers, using all the binary pair-wise combinations of the N classes. Each classifier is traine using the samples of the

4 3 Informatica 33 (009) G. Mazarov et al. first class as positive examples an the samples of the secon class as negative examples. To combine these classifiers, the Max Wins algorithm is aopte. It fins the resultant class by choosing the class vote by the majority of the classifiers [7]. The number of samples use for training of each one of the OvO classifiers is smaller, since only samples from two of all N classes are taken in consieration. The lower number of samples causes smaller nonlinearity, resulting in shorter training times. The isavantage of this metho is that every test sample has to be presente to large number of classifiers N(N-1)/. This results in slower testing, especially when the number of the classes in the problem is big [8]. 3.3 Directe acyclic graph (DAG) Introuce by Platt [1] the DAG algorithm for training an N(N-1)/ classifiers is the same as in oneagainst-one. In the recognition phase, the algorithm epens on a roote binary irecte acyclic graph to make a ecision [9]. DAG creates a moel for each pair of classes. When one such moel, which is able to separate class c 1 from class c, classifies a certain test example into class c 1, it oes not really vote for class c 1, rather it votes against class c, because the example must lie on the other sie of the separating hyperplane than most of the class c samples. Therefore, from that point onwars the algorithm ignores all the moels involving the class c. This means that after each classification with one of the binary moels, one more class can be thrown out as a possible caniate, an after only N-1 steps just one caniate class remains, which therefore becomes the preiction for the current test example. This results in significantly faster testing, while achieving similar recognition rate as One-against-one. 3.4 Binary tree of (BTS) This metho uses multiple s arrange in a binary tree structure [10]. A in each noe of the tree is traine using two of the classes. The algorithm then employs probabilistic outputs to measure the similarity between the remaining samples an the two classes use for training. All samples in the noe are assigne to the two subnoes erive from the previously selecte classes by similarity. This step repeats at every noe until each noe contains only samples from one class. The main problem that shoul be consiere seriously here is training time, because asie training, one has to test all samples in every noe to fin out which classes shoul be assigne to which subnoe while builing the tree. This may ecrease the training performance consierably for huge training atasets. 4 Support vector machines utilizing a binary ecision tree In this paper we propose a binary ecision tree architecture that uses s for making the binary ecisions in the noes. The propose classifier architecture -BDT (Support Vector Machines utilizing Binary Decision Tree), takes avantage of both the efficient computation of the tree architecture an the high classification accuracy of s. Utilizing this architecture, N-1 s neee to be traine for an N class problem, but only at most log N s are require to be consulte to classify a sample. This can lea to a ramatic improvement in recognition spee when aressing problems with big number of classes. An example of -BDT that solves a 7 - class pattern recognition problem utilizing a binary tree, in which each noe makes binary ecision using a is shown on Figure 3. The hierarchy of binary ecision subtasks shoul be carefully esigne before the training of each classifier. The recognition of each sample starts at the root of the tree. At each noe of the binary tree a ecision is being mae about the assignment of the input pattern into one of the two possible groups represente by transferring the pattern to the left or to the right sub-tree. Each of these groups may contain multiple classes. This is repeate recursivly ownwar the tree until the sample reaches a leaf noe that represents the class it has been assigne to. There exist many ways to ivie N classes into two groups, an it is critical to have proper grouping for the goo performance of -BDT. For consistency between the clustering moel an the way calculates the ecision hyperplane, the clustering moel utilizes istance measures at the kernel space, rather than at the input space. Because of this, all training samples are mappe into the kernel space with the same kernel function that is to be use in the training phase., ,,3,4,5,,7,3,4,7 1,5, 4,7 Figure 3: Illustration of -BDT. 1 The -BDT metho that we propose is base on recursively iviing the classes in two isjoint groups in every noe of the ecision tree an training a that will ecie in which of the groups the incoming unknown sample shoul be assigne. The groups are etermine by a clustering algorithm accoring to their class membership. 1,5 5

5 A MULTI-CLASS CLASSIFIER... Informatica 33 (009) Let s take a set of samples x 1, x,..., x M each one labele by y i {c 1, c,..., c N } where N is the number of classes. -BDT metho starts with iviing the classes in two isjoint groups g 1 an g. This is performe by calculating N gravity centres for the N ifferent classes. Then, the two classes that have the biggest Eucliean istance from each other are assigne to each of the two clustering groups. After this, the class with the smallest Eucliean istance from one of the clustering groups is foun an assigne to the corresponing group. The gravity center of this group is then recalculate to represent the aition of the samples of the new class to the group. The process continues by fining the next unassigne class that is closest to either of the clustering groups, assigning it to the corresponing group an upating the group s gravity center, until all classes are assigne to one of the two possible groups. This efines a grouping of all the classes in two isjoint groups of classes. This grouping is then use to train a classifier in the root noe of the ecision tree, using the samples of the first group as positive examples an the samples of the secon group as negative examples. The classes from the first clustering group are being assigne to the first (left) subtree, while the classes of the secon clustering group are being assigne to the (right) secon subtree. The process continues recursively (iviing each of the groups into two subgroups applying the proceure explaine above), until there is only one class per group which efines a leaf in the ecision tree Figure 4: -BDT ivisions of the seven classes. For example, Figure 4 illustrates grouping of 7 classes, while Figure 3 shows the corresponing ecision tree of s. After calculating the gravity centers for all classes, the classes c an c 5 are foun to be the furthest apart from each other, consiering their Eucliean istance an are assigne to group g 1 an g accoringly. The closest to group g 1 is class c 3, so it is assigne to the group g 1, followe by recalculation of the g 1 s gravity center. In the next step, class c 1 is the closest to group g, so it is assigne to that group an the group s gravity center is recalculate. In the following iteration, class c 7 is assigne to g 1 an class c is assigne to g, folowe by recalculating of group s gravity centers. Finally class c 4 is assigne to g 1. This completes the first roun of grouping that efines the classes that will be transferre to the left an the right subtree of the root noe. The classifier in the root is traine by consiering samples from the classes {c, c 3, c 4, c 7 } as positive examples an samples from the classes {c 1, c 5, c } as negative examples. The grouping proceure is repeate inepenently for the classes of the left an the right subtree of the root, which results in grouping c 7 an c 4 in g 1,1 an c an c 3 in g 1, in the left noe of the tree an c 1 an c 5 in g,1 an c in g, in the right noe of the tree. The concept is repeate for each associate to a noe in the taxonomy. This will result in training only N-1 s for solving an N-class problem. 5 Relate work an iscussion Various multi-class classification algorithms can be compare by their preictive accuracy an their training an testing times. The training time T for a binary is estimate empirically by a power law [13] stating that T αm, where M is the number of training samples an is a proportionality constant. The parameter is a constant, which epens of the atasets an it is typically in the range [1, ]. Accoring to this law, the estimate training time for OvA is T OvA NM, ( 11 ) where N is the number of classes in the problem. Without loss of generality, let's assume that each of the N classes has the same number of training samples. Thus, each binary of OvO approach only requires M/N samples. Therefore, the training time for OvO is: N TOvO N 1 M N N M, ( 1 ) The training time for DAG is same as OvO. As for BTS an -BDT, the training time is summe over all the noes in the log N levels. In the i th level, there are i-1 noes an each noe uses M/N for BTS an M/ i-1 for -BDT training samples. Hence, the total training time for BTS is: logn i1 M TBTS i1 N, ( 13 ) log N M i1 1 N M N i1 an for -BDT is:

6 38 Informatica 33 (009) G. Mazarov et al. log N i1 M T BDT M, ( 14 ) i1 i1 It must be note that T -BDT in our algorithm oes not inclue the time to buil the hierarchy structure of the N classes, since it consumes insignificant time compare to the quaratic optimization time that ominates the total training time. On the other han, in the process of builing the tree, BTS requires testing of each traine with all the training samples in orer to etermine the next step, therefore significantly increasing the total training time. Accoring to the empirical estimation above, it is evient that the training spee of -BDT is comparable with OvA, OvO, DAG an BTS. In the testing phase, DAG performs faster than OvO an OvA, since it requires only N-1 binary evaluations. -BDT is even faster than DAG because the epth of the -BDT ecision tree is log N in the worst case, which is superior to N-1, especially when N>>. While testing, the inner prouct of the sample s feature vector an all the support vectors of the moel are calculate for each sample. The total number of support vectors in the traine moel irectly contributes to the major part of the evaluation time, which was also confirme by the experiments. A multistage (M) for multi-class problem has been propose by Liu et al. [11]. They use Support Vector Clustering (SVC) [1] to ivie the training ata into two parts that are use to train a binary. For each partition, the same proceure is recursively repeate until the binary gives an exact label of class. An unsolve problem in M is how to control the SVC to ivie the training ataset into exact two parts. However, this proceure is painful an unfeasible, especially for large atasets. The training set from one class coul belong to both clusters, resulting in ecrease preictive accuracy. There are ifferent approaches for solving multi-class problems which are not base on. Some of them are presente in the following iscussion. However, the experimental results clearly show that their classification accuracy is significantly smaller than the base methos. Ensemble techniques have receive consierable attention within the recent machine learning research [1][17][18][19]. The basic goal is to train a iverse set of classifiers for a single learning problem an to vote or average their preictions. The approach is simple as well as powerful, an the obtaine accuracy gains often have soli theoretical founations [0][0][1]. Averaging the preictions of these classifiers helps to reuce the variance an often increases the reliability of the preictions. There are several techniques for obtaining a iverse set of classifiers. The most common technique is to use subsampling to iversify the training sets as in Bagging [1] an Boosting [0]. Other techniques inclue the use of ifferent feature subsets for every classifier in the ensemble [3], to exploit the ranomness of the base algorithms [4], possibly by artificially ranomizing their behavior [5], or to use multiple representations of the omain objects. Finally, classifier iversity can be ensure by moifying the output labels, i.e., by transforming the learning tasks into a collection of relate learning tasks that use the same input examples, but ifferent assignments of the class labels. Error-correcting output coes are the most prominent example for this type of ensemble methos []. Error-correcting output coes are a popular an powerful class binarization technique. The basic iea is to transform an N-class problem into n binary problems (n > N), where each binary problem uses a subset of the classes as the positive class an the remaining classes as a negative class. As a consequence, each original class is encoe as an n-imensional binary vector, one imension for each preiction of a binary problem (+1 for positive an 1 for negative). The resulting matrix of the form { 1, +1} N n is calle the coing matrix. New examples are classifie by etermining the row in the matrix that is closest to the binary vector obtaine by submitting the example to the n classifiers. If the binary problems are chosen in a way that maximizes the istance between the class vectors, the reliability of the classification can be significantly increase. Errorcorrecting output coes can also be easily parallelize, but each subtask requires the total training set. Similar to binarization, some approaches suggest mapping the original multiple classes into three clsses. A relate technique where multi-class problems are mappe to 3-class problems is propose by Angulo an Catal a []. Like with pairwise classification, they propose generating one training set for each pair of classes. They label the two class values with target values +1 an 1, an aitionally, samples of all other classes are labele to a thir class, with a target value of 0. This iea leas to increase size of the training set compare to the binary classification. The mapping into three classes was also use by Kalousis an Theoharis [7] for preicting the most suitable learning algorithm(s) for a given ataset. They traine a nearest-neighbor learner to preict the better algorithm of each pair of learning algorithms. Each of these pairwise problems ha three classes: one for each algorithm an a thir class name tie, where both algorithms ha similar performances. Johannes Fürnkranz has investigate the use of roun robin binarization (or pair-wise classification) [8] as a technique for hanling multi-class problems with separate-an-conquer rule learning algorithms (aka covering algorithms). In particular, roun robin binarization helps Ripper [9] outperform C5.0 on multiclass problems, whereas C5.0 outperforms the original version of Ripper on the same problems. Experimental results In this section, we present the results of our experiments with several multi-class problems. The performance was measure on the problem of recognition of hanwritten igits an letters.

7 A MULTI-CLASS CLASSIFIER... Informatica 33 (009) Here, we compare the results of the propose - BDT metho with the following methos: 1) one-against-all (OvA); ) one-against-one (OvO); 3) DAG; 4) BTS; 5) Bagging ) Ranom Forests 7) Multilayer Perceptron (MLP, neural network) The training an testing of the s base methos (OvO, OvA, DAG, BTS an -BDT) was performe using a custom evelope application that uses the Torch library [14]. For solving the partial binary classification problems, we use s with Gaussian kernel. In these methos, we ha to optimize the values of the kernel parameter σ an penalty C. For parameter optimization we use experimental results. The achieve parameter values for the given atasets are given in Table 1. Table 1. The optimize values for σ an C for the use atasets. MNIST Penigit Optigit Statlog σ C We also evelope an application that uses the same (Torch) library for the neural network classification. One hien layer with 5 units was use by the neural network. The number of hien units was etermine experimentally. The classifications base on ensembles of ecision trees [30] (Bagging an Ranom Forest) was performe by Clus, a popular ecision tree learner base on the principles state by Blockeel et al. [31]. There were 100 moels in the ensembles. The pruning metho that we use was C4.5. The number of selecte features in the Ranom Forest metho was log M, where M is the number of features in the ataset. The most important criterion in evaluating the performance of a classifier is usually its recognition rate, but very often the training an testing time of the classifier are equally important. In our experiments, four ifferent multi-class classification problems were aresse by each of the eight previously mentione methos. The training an testing time an the recognition performance were recore for every metho. The first problem was recognition of isolate hanwritten igits (10 classes) from the MNIST atabase. The MNIST atabase [15] contains grayscale images of isolate hanwritten igits. From each igit image, after performing a slant correction, 40 features were extracte. The features are consiste of 10 horizontal, 8 vertical an iagonal projections [5]. The MNIST atabase contains training samples, an testing samples. The secon an the thir problem are 10 class problems from the UCI Repository [33] of machine learning atabases: Optigit an Penigit. Penigit has 1 features, 7494 training samples, an 3498 testing samples. Optigit has 4 features, 383 training samples, an 1797 testing samples. The fourth problem was recognition of isolate hanwritten letters a -class problem from the Statlog collection [34]. Statlog-letter contains training samples, an testing samples, where each sample is represente by 1 features. The classifiers were traine using all available training samples of the set an were evaluate by recognizing all the test samples from the corresponing set. All tests were performe on a personal computer with an Intel CoreDuo processor at 1.8GHz with the Winows XP operating system. Tables through 4 show the results of the experiments using 8 ifferent approaches (5 approaches base on, two base on ensembles of ecision trees an one neural network) on each of the 4 ata sets. The first column of each table escribes the classification metho. Table gives the preiction error rate of each metho applie on each of the atasets. Table 3 an table 4 shows the testing an training time of each algorithm, for the atasets, measure in secons, respectively. The results in the tables show that base methos outperform the other approaches, in terms of classification accuracy. In terms of spee, base methos are faster, with ifferent ratios for ifferent atasets. In overall, the base algorithms were significantly better compare to the non base methos. The results in table show that for all atasets, the one-against-all (OvA) metho achieve the lowest error rate. For the MNIST, Penigit an Optigit atasets, the other base methos (OvO, DAG, BTS an our metho - -BDT) achieve higher, but similar error rates. For the recognition of hanwritten letters from the Statlog atabase, the OvO an DAG methos achieve very similar error rates that were about 1.5% higher than the OvA metho. The BTS metho showe the lowest error rate of all methos using oneagainst-one s. Our -BDT metho achieve better recognition rate than all the methos using oneagainst-one s, incluing BTS. Of the non base methos, the Ranom Forest metho achieve the best recognition accuracy for all atasets. The preiction performance of the MLP metho was comparable to the Ranom Forest metho for the 10-class problems, but noticeably worse for the -class problem. The MLP metho is the fastest one in terms of training an testing time, which is evient in Table 3 an Table 4. The classification methos base on ensembles of trees were the slowest in the training an the testing phase, especially the Bagging metho. Overall, the Ranom Forest metho was more accurate than the other non base methos, while the MLP metho was the fastest. The results in Table 3 show that the DAG metho achieve the fastest testing time of all the base methos for the MNIST ataset. For the other atasets, the testing time of DAG is comparable

8 40 Informatica 33 (009) G. Mazarov et al. with BTS an -BDT methos an their testing time is noticeably better than the one-against-all (OvA) an one-against-one (OvO) methos. The -BDT metho was faster in the recognition phase for the Penigit ataset an slightly slower than DAG metho for the Statlog ataset. Table. The preiction error rate (%) of each metho for every ataset Classifier MNIST Penigit Optigit Statlog OvA OvO DAG ВТЅ BDT R. Forest Bagging MLP Table 3. Testing time of each metho for every ataset measure in secons Classifier MNIST Penigit Optigit Statlog OvA OvO DAG ВТЅ BDT R. Forest Bagging MLP Table 4. Training time of each metho for every ataset measure in secons Classifier MNIST Penigit Optigit Statlog OvA OvO DAG ВТЅ BDT R. Forest Bagging MLP In terms of training spees, it is evient in Table 4 that among the base methos, -BDT is the fastest one in the training phase. For the three 10-class problems the time neee to train the 10 classifiers for the OvA approach took about 4 times longer than training the 45 classifiers for the OvO an DAG methos. Due to the huge number of training samples in the MNIST ataset (0000), -BDT s training time was longer compare to other one-against-one methos. The huge number of training samples increases the nonlinearity of the hyperplane in the, resulting in an increse number of support vectors an increase training time. Also, the elay exists only in the first level of the tree, where the entire training ataset is use for training. In the lower levels, the training time of ivie subsets is not as significant as the first level s elay. In the other 10 class problems, our metho achieve the shortest training time. For the Statlog ataset, the time neee for training of the one-against-all s was almost 7 times longer than the time for training the 35 one-against-one s. The BTS metho is the slowest one in the training phase of the methos using one-against-one s. It must be note that as the number of classes in the ataset increases, the avantage of -BDT becomes more evient. The -BDT metho was the fastest while training, achieving better recognition rate than the methos using one-against-one s. It was only slightly slower in recognition than DAG. 7 Conclusion A novel architecture of Support Vector Machine classifiers utilizing binary ecision tree (-BDT) for solving multiclass problems was presente. The - BDT architecture was esigne to provie superior multi-class classification performance, utilizing a ecision tree architecture that requires much less computation for eciing a class for an unknown sample. A clustering algorithm that utilizes istance measures at the kernel space is use to convert the multi-class problem into binary ecision tree, in which the binary ecisions are mae by the s. The results of the experiments show that the spee of training an testing are improve, while keeping comparable or offering better recognition rates than the other multi-class methos. The experiments showe that this metho becomes more favourable as the number of classes in the recognition problem increases. References [1] V. Vapnik. The Nature of Statistical Learning Theory, n E. Springer, New York, [] C. J. C. Burges. A tutorial on support vector machine for pattern recognition. Data Min. Knowl. Disc. (1998) 11. [3] T. Joachims. Making large scale learning practical. in B. Scholkopf, C. Bruges an A. Smola (es). Avances in kernel methos-support vector learning, MIT Press, Cambrige, MA, [4] R. Fletcher. Practical Methos of Optimization. n E. John Wiley & Sons. Chichester (1987). [5] J. Weston, C. Watkins. Multi-class support vector machines. Proceeings of ESANN99, M. Verleysen, E., Brussels, Belgium, 1999.

9 A MULTI-CLASS CLASSIFIER... Informatica 33 (009) [] V. Vapnik. Statistical Learning Theory. Wiley, New York, [7] J. H. Frieman. Another approach to polychotomous classification. Technical report. Department of Statistics, Stanfor University, [8] P. Xu, A. K. Chan. Support vector machine for multi-class signal classification with unbalance samples. Proceeings of the International Joint Conference on Neural Networks 003. Portlan, pp , 003. [9] Platt, N. Cristianini, J. Shawe-Taylor. Large margin DAG s for multiclass classification. Avances in Neural Information Processing System. Vol. 1, pp , 000. [10] B. Fei, J. Liu. Binary Tree of : A New Fast Multiclass Training an Classification Algorithm. IEEE Transaction on neural networks, Vol. 17, No. 3, May 00. [11] X. Liu, H. Xing, X. Wang. A multistage support vector machine. n International Conference on Machine Learning an Cybernetics, pages , 003. [1] A. Ben-Hur, D. Horn, H. Siegelmann, V. Vapnik. Support vector clustering. Journal of Machine Learning Research, vol. :15-137, 001. [13] J. Platt. Fast training of support vector machines using sequential minimal optimization. In Avances in Kernel Methos - Support Vector Learning. Pages , Cambrige, MA, MIT Press. [14] R. Collobert, S. Bengio, J. Mariéthoz. Torch: a moular machine learning software library. Technical Report IDIAP-RR 0-4, IDIAP, 00. [15], MNIST, MiniNIST, USA [1] T. G. Dietterich. Machine learning research: Four current irections. AI Magazine, 18(4): 97 13, Winter [17] G. Dietterich. Ensemble methos in machine learning. In J. Kittler an F. Roli (es.) First International Workshop on Multiple Classifier Systems, pp Springer-Verlag, 000a. [18] D. Opitz an R. Maclin. Popular ensemble methos: An empirical stuy. Journal of Artificial Intelligence Research, 11:19 198, [19] E. Bauer an R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting, an variants. Machine Learning, 3:105 19, [0] Y. Freun an R. E. Schapire. A ecision-theoretic generalization of on-line learning an an application to boosting. Journal of Computer an System Sciences, 55(1): , [1] L. Breiman. Bagging preictors. Machine Learning, 4():13 140, 199. [] T. G. Dietterich an G. Bakiri. Solving multiclass learning problems via error-correcting output coes. Journal of Artificial Intelligence Research, :3 8, [3] S. D. Bay. Nearest neighbor classification from multiple feature subsets. Intelligent Data Analysis, 3(3):191 09, [4] J. F. Kolen an J. B. Pollack. Back propagation is sensitive to initial conitions. In Avances in Neural Information Processing Systems 3 (NIPS- 90), pp Morgan Kaufmann, [5] T. G. Dietterich. An experimental comparison of three methos for constructing ensembles of ecision trees: Bagging, boosting, an ranomization. Machine Learning, 40(): , 000b. [] C. Angulo an A. Catal`a. K-SVCR. A multi-class support vector machine. In R. L opez e M antaras an E. Plaza (es.) Proceeings of the 11th European Conference on Machine Learning (ECML-000), pp Springer-Verlag, 000. [7] A. Kalousis an T. Theoharis. Noemon: Design, implementation an performance results of an intelligent assistant for classifier selection. Intelligent Data Analysis, 3(5): , [8] Johannes Fürnkranz, Roun robin classification, The Journal of Machine Learning Research,, p , 3/1/00 [9] W. W. Cohen. Fast effective rule inuction. In A. Prieitis an S. Russell (es.) Proceeings of the 1th International Conference on Machine Learning (ML-95), pp , Lake Tahoe, CA, Morgan Kaufmann. [30] D. Kocev, C. Vens, J. Struyf an S. Dˇzeroski. Ensembles of multi-objective ecision trees. Proceeings of the 18th European Conference on Machine Learning (pp. 4 31) (007). Springer. [31] H. Blockeel, J. Struyf. Efficient Algorithms for Decision Tree Cross-valiation. Journal of Machine Learning Research 3:1-50, 00. [3] D. Gorgevik, D. Cakmakov. An Efficient Three- Stage Classifier for Hanwritten Digit Recognition. Proceeings of 17th Int. Conference on Pattern Recognition, ICPR004. Vol. 4, pp , IEEE Computer Society, Cambrige, UK, 3- August 004. [33] C. Blake, E. Keogh an C. Merz. UCI Repository of Machine Learning Databases, (1998). Statlog Data Set, [Online] [34] Statlog Data Set, atasets/letter+recognition [Online]

10 4 Informatica 33 (009) G. Mazarov et al.

Classifying Facial Expression with Radial Basis Function Networks, using Gradient Descent and K-means

Classifying Facial Expression with Radial Basis Function Networks, using Gradient Descent and K-means Classifying Facial Expression with Raial Basis Function Networks, using Graient Descent an K-means Neil Allrin Department of Computer Science University of California, San Diego La Jolla, CA 9237 nallrin@cs.ucs.eu

More information

Image Segmentation using K-means clustering and Thresholding

Image Segmentation using K-means clustering and Thresholding Image Segmentation using Kmeans clustering an Thresholing Preeti Panwar 1, Girhar Gopal 2, Rakesh Kumar 3 1M.Tech Stuent, Department of Computer Science & Applications, Kurukshetra University, Kurukshetra,

More information

Particle Swarm Optimization Based on Smoothing Approach for Solving a Class of Bi-Level Multiobjective Programming Problem

Particle Swarm Optimization Based on Smoothing Approach for Solving a Class of Bi-Level Multiobjective Programming Problem BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 17, No 3 Sofia 017 Print ISSN: 1311-970; Online ISSN: 1314-4081 DOI: 10.1515/cait-017-0030 Particle Swarm Optimization Base

More information

Skyline Community Search in Multi-valued Networks

Skyline Community Search in Multi-valued Networks Syline Community Search in Multi-value Networs Rong-Hua Li Beijing Institute of Technology Beijing, China lironghuascut@gmail.com Jeffrey Xu Yu Chinese University of Hong Kong Hong Kong, China yu@se.cuh.eu.h

More information

1 Surprises in high dimensions

1 Surprises in high dimensions 1 Surprises in high imensions Our intuition about space is base on two an three imensions an can often be misleaing in high imensions. It is instructive to analyze the shape an properties of some basic

More information

Non-homogeneous Generalization in Privacy Preserving Data Publishing

Non-homogeneous Generalization in Privacy Preserving Data Publishing Non-homogeneous Generalization in Privacy Preserving Data Publishing W. K. Wong, Nios Mamoulis an Davi W. Cheung Department of Computer Science, The University of Hong Kong Pofulam Roa, Hong Kong {wwong2,nios,cheung}@cs.hu.h

More information

6 Gradient Descent. 6.1 Functions

6 Gradient Descent. 6.1 Functions 6 Graient Descent In this topic we will iscuss optimizing over general functions f. Typically the function is efine f : R! R; that is its omain is multi-imensional (in this case -imensional) an output

More information

Transient analysis of wave propagation in 3D soil by using the scaled boundary finite element method

Transient analysis of wave propagation in 3D soil by using the scaled boundary finite element method Southern Cross University epublications@scu 23r Australasian Conference on the Mechanics of Structures an Materials 214 Transient analysis of wave propagation in 3D soil by using the scale bounary finite

More information

Learning Polynomial Functions. by Feature Construction

Learning Polynomial Functions. by Feature Construction I Proceeings of the Eighth International Workshop on Machine Learning Chicago, Illinois, June 27-29 1991 Learning Polynomial Functions by Feature Construction Richar S. Sutton GTE Laboratories Incorporate

More information

Kinematic Analysis of a Family of 3R Manipulators

Kinematic Analysis of a Family of 3R Manipulators Kinematic Analysis of a Family of R Manipulators Maher Baili, Philippe Wenger an Damien Chablat Institut e Recherche en Communications et Cybernétique e Nantes, UMR C.N.R.S. 6597 1, rue e la Noë, BP 92101,

More information

Loop Scheduling and Partitions for Hiding Memory Latencies

Loop Scheduling and Partitions for Hiding Memory Latencies Loop Scheuling an Partitions for Hiing Memory Latencies Fei Chen Ewin Hsing-Mean Sha Dept. of Computer Science an Engineering University of Notre Dame Notre Dame, IN 46556 Email: fchen,esha @cse.n.eu Tel:

More information

A Classification of 3R Orthogonal Manipulators by the Topology of their Workspace

A Classification of 3R Orthogonal Manipulators by the Topology of their Workspace A Classification of R Orthogonal Manipulators by the Topology of their Workspace Maher aili, Philippe Wenger an Damien Chablat Institut e Recherche en Communications et Cybernétique e Nantes, UMR C.N.R.S.

More information

Generalized Edge Coloring for Channel Assignment in Wireless Networks

Generalized Edge Coloring for Channel Assignment in Wireless Networks TR-IIS-05-021 Generalize Ege Coloring for Channel Assignment in Wireless Networks Chun-Chen Hsu, Pangfeng Liu, Da-Wei Wang, Jan-Jan Wu December 2005 Technical Report No. TR-IIS-05-021 http://www.iis.sinica.eu.tw/lib/techreport/tr2005/tr05.html

More information

Image compression predicated on recurrent iterated function systems

Image compression predicated on recurrent iterated function systems 2n International Conference on Mathematics & Statistics 16-19 June, 2008, Athens, Greece Image compression preicate on recurrent iterate function systems Chol-Hui Yun *, Metzler W. a an Barski M. a * Faculty

More information

Coupling the User Interfaces of a Multiuser Program

Coupling the User Interfaces of a Multiuser Program Coupling the User Interfaces of a Multiuser Program PRASUN DEWAN University of North Carolina at Chapel Hill RAJIV CHOUDHARY Intel Corporation We have evelope a new moel for coupling the user-interfaces

More information

Fast Fractal Image Compression using PSO Based Optimization Techniques

Fast Fractal Image Compression using PSO Based Optimization Techniques Fast Fractal Compression using PSO Base Optimization Techniques A.Krishnamoorthy Visiting faculty Department Of ECE University College of Engineering panruti rishpci89@gmail.com S.Buvaneswari Visiting

More information

Random Clustering for Multiple Sampling Units to Speed Up Run-time Sample Generation

Random Clustering for Multiple Sampling Units to Speed Up Run-time Sample Generation DEIM Forum 2018 I4-4 Abstract Ranom Clustering for Multiple Sampling Units to Spee Up Run-time Sample Generation uzuru OKAJIMA an Koichi MARUAMA NEC Solution Innovators, Lt. 1-18-7 Shinkiba, Koto-ku, Tokyo,

More information

Online Appendix to: Generalizing Database Forensics

Online Appendix to: Generalizing Database Forensics Online Appenix to: Generalizing Database Forensics KYRIACOS E. PAVLOU an RICHARD T. SNODGRASS, University of Arizona This appenix presents a step-by-step iscussion of the forensic analysis protocol that

More information

Generalized Edge Coloring for Channel Assignment in Wireless Networks

Generalized Edge Coloring for Channel Assignment in Wireless Networks Generalize Ege Coloring for Channel Assignment in Wireless Networks Chun-Chen Hsu Institute of Information Science Acaemia Sinica Taipei, Taiwan Da-wei Wang Jan-Jan Wu Institute of Information Science

More information

Solution Representation for Job Shop Scheduling Problems in Ant Colony Optimisation

Solution Representation for Job Shop Scheduling Problems in Ant Colony Optimisation Solution Representation for Job Shop Scheuling Problems in Ant Colony Optimisation James Montgomery, Carole Faya 2, an Sana Petrovic 2 Faculty of Information & Communication Technologies, Swinburne University

More information

filtering LETTER An Improved Neighbor Selection Algorithm in Collaborative Taek-Hun KIM a), Student Member and Sung-Bong YANG b), Nonmember

filtering LETTER An Improved Neighbor Selection Algorithm in Collaborative Taek-Hun KIM a), Student Member and Sung-Bong YANG b), Nonmember 107 IEICE TRANS INF & SYST, VOLE88 D, NO5 MAY 005 LETTER An Improve Neighbor Selection Algorithm in Collaborative Filtering Taek-Hun KIM a), Stuent Member an Sung-Bong YANG b), Nonmember SUMMARY Nowaays,

More information

a 1 (x ) a 1 (x ) a 1 (x ) ) a 2 3 (x Input Variable x

a 1 (x ) a 1 (x ) a 1 (x ) ) a 2 3 (x Input Variable x Support Vector Learning for Fuzzy Rule-Base Classification Systems Yixin Chen, Stuent Member, IEEE, James Z. Wang, Member, IEEE Abstract To esign a fuzzy rule-base classification system (fuzzy classifier)

More information

A Neural Network Model Based on Graph Matching and Annealing :Application to Hand-Written Digits Recognition

A Neural Network Model Based on Graph Matching and Annealing :Application to Hand-Written Digits Recognition ITERATIOAL JOURAL OF MATHEMATICS AD COMPUTERS I SIMULATIO A eural etwork Moel Base on Graph Matching an Annealing :Application to Han-Written Digits Recognition Kyunghee Lee Abstract We present a neural

More information

Intensive Hypercube Communication: Prearranged Communication in Link-Bound Machines 1 2

Intensive Hypercube Communication: Prearranged Communication in Link-Bound Machines 1 2 This paper appears in J. of Parallel an Distribute Computing 10 (1990), pp. 167 181. Intensive Hypercube Communication: Prearrange Communication in Link-Boun Machines 1 2 Quentin F. Stout an Bruce Wagar

More information

Rule extraction from support vector machines

Rule extraction from support vector machines Rule extraction from support vector machines Haydemar Núñez 1,3 Cecilio Angulo 1,2 Andreu Català 1,2 1 Dept. of Systems Engineering, Polytechnical University of Catalonia Avda. Victor Balaguer s/n E-08800

More information

Table-based division by small integer constants

Table-based division by small integer constants Table-base ivision by small integer constants Florent e Dinechin, Laurent-Stéphane Diier LIP, Université e Lyon (ENS-Lyon/CNRS/INRIA/UCBL) 46, allée Italie, 69364 Lyon Ceex 07 Florent.e.Dinechin@ens-lyon.fr

More information

SURVIVABLE IP OVER WDM: GUARANTEEEING MINIMUM NETWORK BANDWIDTH

SURVIVABLE IP OVER WDM: GUARANTEEEING MINIMUM NETWORK BANDWIDTH SURVIVABLE IP OVER WDM: GUARANTEEEING MINIMUM NETWORK BANDWIDTH Galen H Sasaki Dept Elec Engg, U Hawaii 2540 Dole Street Honolul HI 96822 USA Ching-Fong Su Fuitsu Laboratories of America 595 Lawrence Expressway

More information

Study of Network Optimization Method Based on ACL

Study of Network Optimization Method Based on ACL Available online at www.scienceirect.com Proceia Engineering 5 (20) 3959 3963 Avance in Control Engineering an Information Science Stuy of Network Optimization Metho Base on ACL Liu Zhian * Department

More information

NEW METHOD FOR FINDING A REFERENCE POINT IN FINGERPRINT IMAGES WITH THE USE OF THE IPAN99 ALGORITHM 1. INTRODUCTION 2.

NEW METHOD FOR FINDING A REFERENCE POINT IN FINGERPRINT IMAGES WITH THE USE OF THE IPAN99 ALGORITHM 1. INTRODUCTION 2. JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 13/009, ISSN 164-6037 Krzysztof WRÓBEL, Rafał DOROZ * fingerprint, reference point, IPAN99 NEW METHOD FOR FINDING A REFERENCE POINT IN FINGERPRINT IMAGES

More information

Computer Organization

Computer Organization Computer Organization Douglas Comer Computer Science Department Purue University 250 N. University Street West Lafayette, IN 47907-2066 http://www.cs.purue.eu/people/comer Copyright 2006. All rights reserve.

More information

Classical Mechanics Examples (Lagrange Multipliers)

Classical Mechanics Examples (Lagrange Multipliers) Classical Mechanics Examples (Lagrange Multipliers) Dipan Kumar Ghosh Physics Department, Inian Institute of Technology Bombay Powai, Mumbai 400076 September 3, 015 1 Introuction We have seen that the

More information

Cluster Center Initialization Method for K-means Algorithm Over Data Sets with Two Clusters

Cluster Center Initialization Method for K-means Algorithm Over Data Sets with Two Clusters Available online at www.scienceirect.com Proceia Engineering 4 (011 ) 34 38 011 International Conference on Avances in Engineering Cluster Center Initialization Metho for K-means Algorithm Over Data Sets

More information

Indexing the Edges A simple and yet efficient approach to high-dimensional indexing

Indexing the Edges A simple and yet efficient approach to high-dimensional indexing Inexing the Eges A simple an yet efficient approach to high-imensional inexing Beng Chin Ooi Kian-Lee Tan Cui Yu Stephane Bressan Department of Computer Science National University of Singapore 3 Science

More information

New Version of Davies-Bouldin Index for Clustering Validation Based on Cylindrical Distance

New Version of Davies-Bouldin Index for Clustering Validation Based on Cylindrical Distance New Version of Davies-Boulin Inex for lustering Valiation Base on ylinrical Distance Juan arlos Roas Thomas Faculta e Informática Universia omplutense e Mari Mari, España correoroas@gmail.com Abstract

More information

Feature Extraction and Rule Classification Algorithm of Digital Mammography based on Rough Set Theory

Feature Extraction and Rule Classification Algorithm of Digital Mammography based on Rough Set Theory Feature Extraction an Rule Classification Algorithm of Digital Mammography base on Rough Set Theory Aboul Ella Hassanien Jafar M. H. Ali. Kuwait University, Faculty of Aministrative Science, Quantitative

More information

Almost Disjunct Codes in Large Scale Multihop Wireless Network Media Access Control

Almost Disjunct Codes in Large Scale Multihop Wireless Network Media Access Control Almost Disjunct Coes in Large Scale Multihop Wireless Network Meia Access Control D. Charles Engelhart Anan Sivasubramaniam Penn. State University University Park PA 682 engelhar,anan @cse.psu.eu Abstract

More information

THE APPLICATION OF ARTICLE k-th SHORTEST TIME PATH ALGORITHM

THE APPLICATION OF ARTICLE k-th SHORTEST TIME PATH ALGORITHM International Journal of Physics an Mathematical Sciences ISSN: 2277-2111 (Online) 2016 Vol. 6 (1) January-March, pp. 24-6/Mao an Shi. THE APPLICATION OF ARTICLE k-th SHORTEST TIME PATH ALGORITHM Hua Mao

More information

A fast embedded selection approach for color texture classification using degraded LBP

A fast embedded selection approach for color texture classification using degraded LBP A fast embee selection approach for color texture classification using egrae A. Porebski, N. Vanenbroucke an D. Hama Laboratoire LISIC - EA 4491 - Université u Littoral Côte Opale - 50, rue Ferinan Buisson

More information

Handling missing values in kernel methods with application to microbiology data

Handling missing values in kernel methods with application to microbiology data an Machine Learning. Bruges (Belgium), 24-26 April 2013, i6oc.com publ., ISBN 978-2-87419-081-0. Available from http://www.i6oc.com/en/livre/?gcoi=28001100131010. Hanling missing values in kernel methos

More information

Blind Data Classification using Hyper-Dimensional Convex Polytopes

Blind Data Classification using Hyper-Dimensional Convex Polytopes Blin Data Classification using Hyper-Dimensional Convex Polytopes Brent T. McBrie an Gilbert L. Peterson Department of Electrical an Computer Engineering Air Force Institute of Technology 9 Hobson Way

More information

Lab work #8. Congestion control

Lab work #8. Congestion control TEORÍA DE REDES DE TELECOMUNICACIONES Grao en Ingeniería Telemática Grao en Ingeniería en Sistemas e Telecomunicación Curso 2015-2016 Lab work #8. Congestion control (1 session) Author: Pablo Pavón Mariño

More information

Learning convex bodies is hard

Learning convex bodies is hard Learning convex boies is har Navin Goyal Microsoft Research Inia navingo@microsoftcom Luis Raemacher Georgia Tech lraemac@ccgatecheu Abstract We show that learning a convex boy in R, given ranom samples

More information

Learning Subproblem Complexities in Distributed Branch and Bound

Learning Subproblem Complexities in Distributed Branch and Bound Learning Subproblem Complexities in Distribute Branch an Boun Lars Otten Department of Computer Science University of California, Irvine lotten@ics.uci.eu Rina Dechter Department of Computer Science University

More information

Rough Set Approach for Classification of Breast Cancer Mammogram Images

Rough Set Approach for Classification of Breast Cancer Mammogram Images Rough Set Approach for Classification of Breast Cancer Mammogram Images Aboul Ella Hassanien Jafar M. H. Ali. Kuwait University, Faculty of Aministrative Science, Quantitative Methos an Information Systems

More information

Divide-and-Conquer Algorithms

Divide-and-Conquer Algorithms Supplment to A Practical Guie to Data Structures an Algorithms Using Java Divie-an-Conquer Algorithms Sally A Golman an Kenneth J Golman Hanout Divie-an-conquer algorithms use the following three phases:

More information

A Plane Tracker for AEC-automation Applications

A Plane Tracker for AEC-automation Applications A Plane Tracker for AEC-automation Applications Chen Feng *, an Vineet R. Kamat Department of Civil an Environmental Engineering, University of Michigan, Ann Arbor, USA * Corresponing author (cforrest@umich.eu)

More information

Pairwise alignment using shortest path algorithms, Gunnar Klau, November 29, 2005, 11:

Pairwise alignment using shortest path algorithms, Gunnar Klau, November 29, 2005, 11: airwise alignment using shortest path algorithms, Gunnar Klau, November 9,, : 3 3 airwise alignment using shortest path algorithms e will iscuss: it graph Dijkstra s algorithm algorithm (GDU) 3. References

More information

New Geometric Interpretation and Analytic Solution for Quadrilateral Reconstruction

New Geometric Interpretation and Analytic Solution for Quadrilateral Reconstruction New Geometric Interpretation an Analytic Solution for uarilateral Reconstruction Joo-Haeng Lee Convergence Technology Research Lab ETRI Daejeon, 305 777, KOREA Abstract A new geometric framework, calle

More information

Table of Contents. Recognition of Facial Gestures... 1 Attila Fazekas

Table of Contents. Recognition of Facial Gestures... 1 Attila Fazekas Table of Contents Recognition of Facial Gestures...................................... 1 Attila Fazekas II Recognition of Facial Gestures Attila Fazekas University of Debrecen, Institute of Informatics

More information

On the Role of Multiply Sectioned Bayesian Networks to Cooperative Multiagent Systems

On the Role of Multiply Sectioned Bayesian Networks to Cooperative Multiagent Systems On the Role of Multiply Sectione Bayesian Networks to Cooperative Multiagent Systems Y. Xiang University of Guelph, Canaa, yxiang@cis.uoguelph.ca V. Lesser University of Massachusetts at Amherst, USA,

More information

APPLYING GENETIC ALGORITHM IN QUERY IMPROVEMENT PROBLEM. Abdelmgeid A. Aly

APPLYING GENETIC ALGORITHM IN QUERY IMPROVEMENT PROBLEM. Abdelmgeid A. Aly International Journal "Information Technologies an Knowlege" Vol. / 2007 309 [Project MINERVAEUROPE] Project MINERVAEUROPE: Ministerial Network for Valorising Activities in igitalisation -

More information

Nearest Neighbor Search using Additive Binary Tree

Nearest Neighbor Search using Additive Binary Tree Nearest Neighbor Search using Aitive Binary Tree Sung-Hyuk Cha an Sargur N. Srihari Center of Excellence for Document Analysis an Recognition State University of New York at Buffalo, U. S. A. E-mail: fscha,sriharig@cear.buffalo.eu

More information

An Adaptive Routing Algorithm for Communication Networks using Back Pressure Technique

An Adaptive Routing Algorithm for Communication Networks using Back Pressure Technique International OPEN ACCESS Journal Of Moern Engineering Research (IJMER) An Aaptive Routing Algorithm for Communication Networks using Back Pressure Technique Khasimpeera Mohamme 1, K. Kalpana 2 1 M. Tech

More information

Shift-map Image Registration

Shift-map Image Registration Shift-map Image Registration Svärm, Linus; Stranmark, Petter Unpublishe: 2010-01-01 Link to publication Citation for publishe version (APA): Svärm, L., & Stranmark, P. (2010). Shift-map Image Registration.

More information

CS 106 Winter 2016 Craig S. Kaplan. Module 01 Processing Recap. Topics

CS 106 Winter 2016 Craig S. Kaplan. Module 01 Processing Recap. Topics CS 106 Winter 2016 Craig S. Kaplan Moule 01 Processing Recap Topics The basic parts of speech in a Processing program Scope Review of syntax for classes an objects Reaings Your CS 105 notes Learning Processing,

More information

Frequent Pattern Mining. Frequent Item Set Mining. Overview. Frequent Item Set Mining: Motivation. Frequent Pattern Mining comprises

Frequent Pattern Mining. Frequent Item Set Mining. Overview. Frequent Item Set Mining: Motivation. Frequent Pattern Mining comprises verview Frequent Pattern Mining comprises Frequent Pattern Mining hristian Borgelt School of omputer Science University of Konstanz Universitätsstraße, Konstanz, Germany christian.borgelt@uni-konstanz.e

More information

Shift-map Image Registration

Shift-map Image Registration Shift-map Image Registration Linus Svärm Petter Stranmark Centre for Mathematical Sciences, Lun University {linus,petter}@maths.lth.se Abstract Shift-map image processing is a new framework base on energy

More information

Bagging and Boosting Algorithms for Support Vector Machine Classifiers

Bagging and Boosting Algorithms for Support Vector Machine Classifiers Bagging and Boosting Algorithms for Support Vector Machine Classifiers Noritaka SHIGEI and Hiromi MIYAJIMA Dept. of Electrical and Electronics Engineering, Kagoshima University 1-21-40, Korimoto, Kagoshima

More information

Short-term prediction of photovoltaic power based on GWPA - BP neural network model

Short-term prediction of photovoltaic power based on GWPA - BP neural network model Short-term preiction of photovoltaic power base on GWPA - BP neural networ moel Jian Di an Shanshan Meng School of orth China Electric Power University, Baoing. China Abstract In recent years, ue to China's

More information

Optimal Oblivious Path Selection on the Mesh

Optimal Oblivious Path Selection on the Mesh Optimal Oblivious Path Selection on the Mesh Costas Busch Malik Magon-Ismail Jing Xi Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 280, USA {buschc,magon,xij2}@cs.rpi.eu Abstract

More information

Threshold Based Data Aggregation Algorithm To Detect Rainfall Induced Landslides

Threshold Based Data Aggregation Algorithm To Detect Rainfall Induced Landslides Threshol Base Data Aggregation Algorithm To Detect Rainfall Inuce Lanslies Maneesha V. Ramesh P. V. Ushakumari Department of Computer Science Department of Mathematics Amrita School of Engineering Amrita

More information

A Convex Clustering-based Regularizer for Image Segmentation

A Convex Clustering-based Regularizer for Image Segmentation Vision, Moeling, an Visualization (2015) D. Bommes, T. Ritschel an T. Schultz (Es.) A Convex Clustering-base Regularizer for Image Segmentation Benjamin Hell (TU Braunschweig), Marcus Magnor (TU Braunschweig)

More information

On Effectively Determining the Downlink-to-uplink Sub-frame Width Ratio for Mobile WiMAX Networks Using Spline Extrapolation

On Effectively Determining the Downlink-to-uplink Sub-frame Width Ratio for Mobile WiMAX Networks Using Spline Extrapolation On Effectively Determining the Downlink-to-uplink Sub-frame With Ratio for Mobile WiMAX Networks Using Spline Extrapolation Panagiotis Sarigianniis, Member, IEEE, Member Malamati Louta, Member, IEEE, Member

More information

Using Vector and Raster-Based Techniques in Categorical Map Generalization

Using Vector and Raster-Based Techniques in Categorical Map Generalization Thir ICA Workshop on Progress in Automate Map Generalization, Ottawa, 12-14 August 1999 1 Using Vector an Raster-Base Techniques in Categorical Map Generalization Beat Peter an Robert Weibel Department

More information

Chapter 5 Proposed models for reconstituting/ adapting three stereoscopes

Chapter 5 Proposed models for reconstituting/ adapting three stereoscopes Chapter 5 Propose moels for reconstituting/ aapting three stereoscopes - 89 - 5. Propose moels for reconstituting/aapting three stereoscopes This chapter offers three contributions in the Stereoscopy area,

More information

d 3 d 4 d d d d d d d d d d d 1 d d d d d d

d 3 d 4 d d d d d d d d d d d 1 d d d d d d Proceeings of the IASTED International Conference Software Engineering an Applications (SEA') October 6-, 1, Scottsale, Arizona, USA AN OBJECT-ORIENTED APPROACH FOR MANAGING A NETWORK OF DATABASES Shu-Ching

More information

Exploring Context with Deep Structured models for Semantic Segmentation

Exploring Context with Deep Structured models for Semantic Segmentation 1 Exploring Context with Deep Structure moels for Semantic Segmentation Guosheng Lin, Chunhua Shen, Anton van en Hengel, Ian Rei between an image patch an a large backgroun image region. Explicitly moeling

More information

Technical Report TR Navigation Around an Unknown Obstacle for Autonomous Surface Vehicles Using a Forward-Facing Sonar

Technical Report TR Navigation Around an Unknown Obstacle for Autonomous Surface Vehicles Using a Forward-Facing Sonar Technical Report Department of Computer Science an Engineering niversity of Minnesota 4-192 Keller Hall 2 nion Street SE Minneapolis, MN 55455-159 SA TR 15-5 Navigation Aroun an nknown Obstacle for Autonomous

More information

Research Article Inviscid Uniform Shear Flow past a Smooth Concave Body

Research Article Inviscid Uniform Shear Flow past a Smooth Concave Body International Engineering Mathematics Volume 04, Article ID 46593, 7 pages http://x.oi.org/0.55/04/46593 Research Article Invisci Uniform Shear Flow past a Smooth Concave Boy Abullah Mura Department of

More information

Multilevel Linear Dimensionality Reduction using Hypergraphs for Data Analysis

Multilevel Linear Dimensionality Reduction using Hypergraphs for Data Analysis Multilevel Linear Dimensionality Reuction using Hypergraphs for Data Analysis Haw-ren Fang Department of Computer Science an Engineering University of Minnesota; Minneapolis, MN 55455 hrfang@csumneu ABSTRACT

More information

Navigation Around an Unknown Obstacle for Autonomous Surface Vehicles Using a Forward-Facing Sonar

Navigation Around an Unknown Obstacle for Autonomous Surface Vehicles Using a Forward-Facing Sonar Navigation Aroun an nknown Obstacle for Autonomous Surface Vehicles sing a Forwar-Facing Sonar Patrick A. Plonski, Joshua Vaner Hook, Cheng Peng, Narges Noori, Volkan Isler Abstract A robotic boat is moving

More information

Algorithm for Intermodal Optimal Multidestination Tour with Dynamic Travel Times

Algorithm for Intermodal Optimal Multidestination Tour with Dynamic Travel Times Algorithm for Intermoal Optimal Multiestination Tour with Dynamic Travel Times Neema Nassir, Alireza Khani, Mark Hickman, an Hyunsoo Noh This paper presents an efficient algorithm that fins the intermoal

More information

Overlap Interval Partition Join

Overlap Interval Partition Join Overlap Interval Partition Join Anton Dignös Department of Computer Science University of Zürich, Switzerlan aignoes@ifi.uzh.ch Michael H. Böhlen Department of Computer Science University of Zürich, Switzerlan

More information

Additional Divide and Conquer Algorithms. Skipping from chapter 4: Quicksort Binary Search Binary Tree Traversal Matrix Multiplication

Additional Divide and Conquer Algorithms. Skipping from chapter 4: Quicksort Binary Search Binary Tree Traversal Matrix Multiplication Aitional Divie an Conquer Algorithms Skipping from chapter 4: Quicksort Binary Search Binary Tree Traversal Matrix Multiplication Divie an Conquer Closest Pair Let s revisit the closest pair problem. Last

More information

Dense Disparity Estimation in Ego-motion Reduced Search Space

Dense Disparity Estimation in Ego-motion Reduced Search Space Dense Disparity Estimation in Ego-motion Reuce Search Space Luka Fućek, Ivan Marković, Igor Cvišić, Ivan Petrović University of Zagreb, Faculty of Electrical Engineering an Computing, Croatia (e-mail:

More information

Support Vector Machines

Support Vector Machines Support Vector Machines About the Name... A Support Vector A training sample used to define classification boundaries in SVMs located near class boundaries Support Vector Machines Binary classifiers whose

More information

Exploring Context with Deep Structured models for Semantic Segmentation

Exploring Context with Deep Structured models for Semantic Segmentation APPEARING IN IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, APRIL 2017. 1 Exploring Context with Deep Structure moels for Semantic Segmentation Guosheng Lin, Chunhua Shen, Anton van en

More information

Queueing Model and Optimization of Packet Dropping in Real-Time Wireless Sensor Networks

Queueing Model and Optimization of Packet Dropping in Real-Time Wireless Sensor Networks Queueing Moel an Optimization of Packet Dropping in Real-Time Wireless Sensor Networks Marc Aoun, Antonios Argyriou, Philips Research, Einhoven, 66AE, The Netherlans Department of Computer an Communication

More information

Animated Surface Pasting

Animated Surface Pasting Animate Surface Pasting Clara Tsang an Stephen Mann Computing Science Department University of Waterloo 200 University Ave W. Waterloo, Ontario Canaa N2L 3G1 e-mail: clftsang@cgl.uwaterloo.ca, smann@cgl.uwaterloo.ca

More information

Support Vector Machines for Face Recognition

Support Vector Machines for Face Recognition Chapter 8 Support Vector Machines for Face Recognition 8.1 Introduction In chapter 7 we have investigated the credibility of different parameters introduced in the present work, viz., SSPD and ALR Feature

More information

Backpressure-based Packet-by-Packet Adaptive Routing in Communication Networks

Backpressure-based Packet-by-Packet Adaptive Routing in Communication Networks 1 Backpressure-base Packet-by-Packet Aaptive Routing in Communication Networks Eleftheria Athanasopoulou, Loc Bui, Tianxiong Ji, R. Srikant, an Alexaner Stolyar Abstract Backpressure-base aaptive routing

More information

MANJUSHA K.*, ANAND KUMAR M., SOMAN K. P.

MANJUSHA K.*, ANAND KUMAR M., SOMAN K. P. Journal of Engineering Science an echnology Vol. 13, No. 1 (2018) 141-157 School of Engineering, aylor s University IMPLEMENAION OF REJECION SRAEGIES INSIDE MALAYALAM CHARACER RECOGNIION SYSEM BASED ON

More information

Considering bounds for approximation of 2 M to 3 N

Considering bounds for approximation of 2 M to 3 N Consiering bouns for approximation of to (version. Abstract: Estimating bouns of best approximations of to is iscusse. In the first part I evelop a powerseries, which shoul give practicable limits for

More information

Design of Policy-Aware Differentially Private Algorithms

Design of Policy-Aware Differentially Private Algorithms Design of Policy-Aware Differentially Private Algorithms Samuel Haney Due University Durham, NC, USA shaney@cs.ue.eu Ashwin Machanavajjhala Due University Durham, NC, USA ashwin@cs.ue.eu Bolin Ding Microsoft

More information

Fuzzy Clustering in Parallel Universes

Fuzzy Clustering in Parallel Universes Fuzzy Clustering in Parallel Universes Bern Wisweel an Michael R. Berthol ALTANA-Chair for Bioinformatics an Information Mining Department of Computer an Information Science, University of Konstanz 78457

More information

arxiv: v2 [cs.lg] 22 Jan 2019

arxiv: v2 [cs.lg] 22 Jan 2019 Spatial Variational Auto-Encoing via Matrix-Variate Normal Distributions Zhengyang Wang Hao Yuan Shuiwang Ji arxiv:1705.06821v2 [cs.lg] 22 Jan 2019 Abstract The key iea of variational auto-encoers (VAEs)

More information

Implementation and Evaluation of NAS Parallel CG Benchmark on GPU Cluster with Proprietary Interconnect TCA

Implementation and Evaluation of NAS Parallel CG Benchmark on GPU Cluster with Proprietary Interconnect TCA Implementation an Evaluation of AS Parallel CG Benchmark on GPU Cluster with Proprietary Interconnect TCA Kazuya Matsumoto 1, orihisa Fujita 2, Toshihiro Hanawa 3, an Taisuke Boku 1,2 1 Center for Computational

More information

THE BAYESIAN RECEIVER OPERATING CHARACTERISTIC CURVE AN EFFECTIVE APPROACH TO EVALUATE THE IDS PERFORMANCE

THE BAYESIAN RECEIVER OPERATING CHARACTERISTIC CURVE AN EFFECTIVE APPROACH TO EVALUATE THE IDS PERFORMANCE БСУ Международна конференция - 2 THE BAYESIAN RECEIVER OPERATING CHARACTERISTIC CURVE AN EFFECTIVE APPROACH TO EVALUATE THE IDS PERFORMANCE Evgeniya Nikolova, Veselina Jecheva Burgas Free University Abstract:

More information

A PSO Optimized Layered Approach for Parametric Clustering on Weather Dataset

A PSO Optimized Layered Approach for Parametric Clustering on Weather Dataset Vol.3, Issue.1, Jan-Feb. 013 pp-504-508 ISSN: 49-6645 A PSO Optimize Layere Approach for Parametric Clustering on Weather Dataset Shikha Verma, 1 Kiran Jyoti 1 Stuent, Guru Nanak Dev Engineering College

More information

A Cost Model For Nearest Neighbor Search. High-Dimensional Data Space

A Cost Model For Nearest Neighbor Search. High-Dimensional Data Space A Cost Moel For Nearest Neighbor Search in High-Dimensional Data Space Stefan Berchtol University of Munich Germany berchtol@informatikuni-muenchene Daniel A Keim University of Munich Germany keim@informatikuni-muenchene

More information

Estimating Velocity Fields on a Freeway from Low Resolution Video

Estimating Velocity Fields on a Freeway from Low Resolution Video Estimating Velocity Fiels on a Freeway from Low Resolution Vieo Young Cho Department of Statistics University of California, Berkeley Berkeley, CA 94720-3860 Email: young@stat.berkeley.eu John Rice Department

More information

Multi-camera tracking algorithm study based on information fusion

Multi-camera tracking algorithm study based on information fusion International Conference on Avance Electronic Science an Technolog (AEST 016) Multi-camera tracking algorithm stu base on information fusion a Guoqiang Wang, Shangfu Li an Xue Wen School of Electronic

More information

Offloading Cellular Traffic through Opportunistic Communications: Analysis and Optimization

Offloading Cellular Traffic through Opportunistic Communications: Analysis and Optimization 1 Offloaing Cellular Traffic through Opportunistic Communications: Analysis an Optimization Vincenzo Sciancalepore, Domenico Giustiniano, Albert Banchs, Anreea Picu arxiv:1405.3548v1 [cs.ni] 14 May 24

More information

Optimization of cable-stayed bridges with box-girder decks

Optimization of cable-stayed bridges with box-girder decks Avances in Engineering Software 31 (2000) 417 423 www.elsevier.com/locate/avengsoft Optimization of cable-staye briges with box-girer ecks L.M.C. Simões*, J.H.J.O. Negrão Department of Civil Engineering,

More information

6.823 Computer System Architecture. Problem Set #3 Spring 2002

6.823 Computer System Architecture. Problem Set #3 Spring 2002 6.823 Computer System Architecture Problem Set #3 Spring 2002 Stuents are strongly encourage to collaborate in groups of up to three people. A group shoul han in only one copy of the solution to the problem

More information

Efficient Pairwise Classification

Efficient Pairwise Classification Efficient Pairwise Classification Sang-Hyeun Park and Johannes Fürnkranz TU Darmstadt, Knowledge Engineering Group, D-64289 Darmstadt, Germany Abstract. Pairwise classification is a class binarization

More information

2-connected graphs with small 2-connected dominating sets

2-connected graphs with small 2-connected dominating sets 2-connecte graphs with small 2-connecte ominating sets Yair Caro, Raphael Yuster 1 Department of Mathematics, University of Haifa at Oranim, Tivon 36006, Israel Abstract Let G be a 2-connecte graph. A

More information

0607 CAMBRIDGE INTERNATIONAL MATHEMATICS

0607 CAMBRIDGE INTERNATIONAL MATHEMATICS CAMBRIDGE INTERNATIONAL EXAMINATIONS International General Certificate of Seconary Eucation MARK SCHEME for the May/June 03 series 0607 CAMBRIDGE INTERNATIONAL MATHEMATICS 0607/4 Paper 4 (Extene), maximum

More information

A Spectral Clustering Approach to Optimally Combining Numerical Vectors with a Modular Network

A Spectral Clustering Approach to Optimally Combining Numerical Vectors with a Modular Network A Spectral Clustering Approach to Optimally Combining Numerical Vectors with a Moular Networ Motoi Shiga Bioinformatics Center Kyoto University Goasho Uji 6-, Japan shiga@uicryotouacjp Ichigau Taigawa

More information

Chapter 9 Memory Management

Chapter 9 Memory Management Contents 1. Introuction 2. Computer-System Structures 3. Operating-System Structures 4. Processes 5. Threas 6. CPU Scheuling 7. Process Synchronization 8. Dealocks 9. Memory Management 10.Virtual Memory

More information