Layer-Specific Adaptive Learning Rates for Deep Networks

Size: px
Start display at page:

Download "Layer-Specific Adaptive Learning Rates for Deep Networks"

Transcription

1 Layer-Specific Adaptive Learning Rates for Deep Networks arxiv: v1 [cs.cv] 15 Oct 2015 Bharat Singh, Soham De, Yangmuzi Zhang, Thomas Godstein, and Gavin Tayor Department of Computer Science Department of Eectrica & Computer Engineering University of Maryand, Coege Park, MD Department of Computer Science, US Nava Academy, Annapois, MD Emai: {bharat, sohamde, Abstract The increasing compexity of deep earning architectures is resuting in training time requiring weeks or even months. This sow training is due in part to vanishing gradients, in which the gradients used by back-propagation are extremey arge for weights connecting deep ayers (ayers near the output ayer), and extremey sma for shaow ayers (near the input ayer); this resuts in sow earning in the shaow ayers. Additionay, it has aso been shown that in highy non-convex probems, such as deep neura networks, there is a proiferation of high-error ow curvature sadde points, which sows down earning dramaticay [1]. In this paper, we attempt to overcome the two above probems by proposing an optimization method for training deep neura networks which uses earning rates which are both specific to each ayer in the network and adaptive to the curvature of the function, increasing the earning rate at ow curvature points. This enabes us to speed up earning in the shaow ayers of the network and quicky escape high-error ow curvature sadde points. We test our method on standard image cassification datasets such as MNIST, CIFAR10 and ImageNet, and demonstrate that our method increases accuracy as we as reduces the required training time over standard agorithms. I. INTRODUCTION Deep neura networks have been extremey successfu over the past few years, achieving state of the art performance on a arge number of tasks such as image cassification [2], face recognition [3], sentiment anaysis [4], speech recognition [5], etc. One can spot a genera trend in these papers: resuts tend to get better as the amount of training data increases, aong with an increase in the compexity of the deep network architecture. However, increasingy compex deep networks can take weeks or months to train, even with high-performance hardware. Thus, there is a need for more efficient methods for training deep networks. Deep neura networks earn high-eve features by performing a sequence of non-inear transformations. Let our training data set A be composed of n data points a 1, a 2,..., a n R m and corresponding abes B = {b i } n i=1. Let us consider a 3-ayer network with activation function f. Let X 1 and X 2 denote the weights on each ayer that we are trying to earn, i.e., X 1 denotes the weights between nodes of the first ayer and the second ayer, and X 2 denotes the weights between nodes of the second ayer and the third ayer. The earning probem for this specific exampe can be formuated as the foowing optimization probem: minimize X 1,X 2 f(f(a X 1 ) X 2 ) B 2 2 (1) The activation function f can be any non-inear mapping, and is traditionay a sigmoid or tanh function. Recenty, rectified inear (ReLu) units (f(z) = max{0, z}) have become popuar because they tend to be easy to train and yied superior resuts for some probems [6]. The non-convex objective (1) is usuay minimized using iterative methods (such as back-propagation) with the hope of converging to a good oca minima. Most iterative schemes generate additive updates to a set of parameters x (in our case, the weight matrices) of the form x (k+1) = x (k) + (2) where is some appropriatey chosen update. Notice we use sighty different notation here from standard optimization iterature in that we incorporate the step size or earning rate t (k) within. This is done to hep us describe other optimization agorithms easiy in the foowing sections. Thus, denotes the update in the parameters, and comprises of a search direction and a step size or earning rate t (k), which contros how arge of a step to take in that direction. Most common update rues are variants of gradient descent, where the search direction is given by the negative gradient g (k) : = t (k) g (k) = t (k) f(x (k) ) (3) Since the size of the training data for these deep networks is usuay of the order of miions or biions of data points, exact computation of the gradient is not feasibe. Rather, the gradient is often estimated using a singe data point or a sma batch of data points. This is the basis for stochastic gradient descent (SGD) [7], which is the most widey used method for training deep nets. SGD requires manuay seecting an initia earning rate, and then designing an update rue for the earning rate which decreases it over time (for exampe, exponentia decay with time). The performance of SGD, however, is very sensitive to this choice of update, eading to adaptive methods that automaticay adjust the earning rate as the system earns [8], [9]. When these descent methods are used to train deep networks, additiona probems are introduced. As the number of ayers in a network increases, the gradients that are propagated back to the initia ayers get very sma. This dramaticay sows down the rate of earning in the initia ayers, and sows down convergence of the whoe network [10].

2 Recenty, it has aso been shown that for high-dimensiona non-convex probems, such as deep networks, the existence of oca minima which have high error reative to the goba minima is exponentiay sma in the number of dimensions. Instead, in these probems, there is an exponentiay arge number of high error sadde points with ow curvature [1], [11], [12]. Gradient descent methods, in genera, move away from sadde points by foowing the directions of negative curvature. However, due to the ow curvature of sma negative eigenvaues, the steps taken become very sma, thus sowing down earning consideraby. In this paper, we propose a method that aeviates the probems mentioned above. The main contribution of our method is summarized beow: The earning rates are specific to each ayer in the network. This aows arger earning rates to compensate for the sma size of gradients in shaow ayers. The earning rates for each ayer tend to increase at ow curvature points. This enabes the method to quicky escape from high-error, ow-curvature sadde points, which occur in abundance in deep network. It is appicabe to most existing stochastic gradient optimization methods which use a goba earning rate. It requires very itte extra computation over standard stochastic gradient methods, and requires no extra storage of previous gradients required as in AdaGrad [9]. In Section II, we review some popuar gradient methods that have been successfu for deep networks. In Section III, we describe our optimization agorithm. Finay, in Section IV we compare our method to standard optimization agorithms on datasets ike MNIST, CIFAR10 and ImageNet. II. RELATED WORK Stochastic Gradient Descent (SGD) sti remains one of the most widey used methods for arge-scae machine earning, argey due to its ease in impementation. In SGD, the updates for the parameters are defined by equations (2) and (3), and the earning rate is decreased over time as iterates approach a oca optimum. A standard earning rate update is given by t (k) = t (0) /(1 + γk) p (4) where the initia earning rate t (0), γ and p are hyperparameters chosen by the user. Many modifications to the basic gradient descent agorithm have been proposed. A popuar method in the convex optimization iterature is Newton s method, which uses the Hessian of the objective function f(x) to determine the step size: nt = 2 f(x (k) ) 1 g (k) (5) Unfortunatey, as the number of parameters increases, even to moderate size, computing the Hessian becomes very computationay expensive. Thus, there have been many modifications proposed which either try to improve the use of first-order information or try to approximate the Hessian of the objective function. In this paper, we focus on modifications to first-order methods. The cassica momentum method [13] is a technique that increases the earning rate for parameters for which the gradient consistenty points in the same direction, whie decreasing the earning rate for parameters for which the gradient is changing fast. Thus, the update equation keeps track of previous updates to the parameters with an exponentia decay: = µ x (k 1) tg (k) (6) where µ [0, 1] is caed the momentum coefficient, and t > 0 is the goba earning rate. Nesterov s Acceerated Gradient (NAG) [14], a first order method, has a better convergence rate than gradient descent in certain situations. This method predicts the gradient for the next iteration and changes the earning rate for the current iteration based on the predicted gradient. Thus, if the gradient is higher for the next step, it woud increase the earning rate for the current iteration and if it is ow, it woud sow down. Recenty, [15] showed that this method can be thought of as a momentum method with the update equation as foows: = µ x (k 1) t f(x (k 1) + µ x (k 1) ) (7) Through a carefuy designed random initiaization and using a particuar type of sowy increasing schedue for µ, this method can reach high eves of performance when used on deep networks [15]. Rather than using a singe earning rate over a parameters, recent work has shown that using a earning rate specific to each parameter can be a much more successfu approach. A method that has gained popuarity is AdaGrad [9], which uses the foowing update rue: t = k g (k) i=1 (g(i) ) 2 The denominator is the 2 norm of a the gradients of the previous iterations. This scaes the goba earning rate t, which is shared by a the parameters, to give a parameterspecific earning rate. One disadvantage of AdaGrad is that it accumuates the gradients over a previous iterations, the sum of which continues to grow throughout training. This (aong with weight decay) shrinks the earning rate on each parameter unti each is infinitesimay sma, imiting the number of iterations of usefu training. A method which buids on AdaGrad and attempts to address some of the above-mentioned disadvantages is AdaDeta [8]. AdaDeta accumuates the gradients in the previous time steps using an exponentiay decaying average of the squared gradients. This prevents the denominator from becoming infinitesimay sma, and ensures that the parameters continue to be updated even after a arge number of iterations. It aso repaces the goba earning rate t with an exponentiay decaying average of the squares of the parameter updates x over the previous iterations. This method has been shown to perform reativey we when used to train deep networks, and is much ess sensitive to the choice of hyper-parameters. However, it does not perform as we as other methods ike SGD and AdaGrad in terms of accuracy [8]. (8)

3 III. OUR APPROACH Because of the vanishing gradients phenomenon, shaow network ayers tend to have much smaer gradients than deep ayers sometimes differing by orders of magnitude from one ayer to the next [10]. In most previous work in optimization for deep networks, methods either keep a goba earning rate that is shared over a parameters, or use an adaptive earning rate specific to each parameter. Our method expoits the foowing observation: parameters in the same ayer have gradients of simiar magnitudes, and can thus efficienty share a common earning rate. Layer-specific earning rates can be used to acceerate ayers with smaer gradients. Another advantage of this approach is that by avoiding the computation of arge numbers of parameter-specific earning rates, our method remains computationay efficient. Finay, as mentioned in Section I, to avoid sowing down earning at high-error ow curvature sadde points, we aso want our method to take arge steps at ow curvature points. Let t (k) be the earning rate at the k-th iteration for any standard optimization method. In case of SGD, this woud be given by equation 4, whie for AdaGrad it woud just be the goba earning rate t as in equation 8. We propose to modify t (k) as foows: t (k) = t (k) (1 + og(1 + 1/( g (k) 2 ))) (9) Here t (k) denotes the new earning rate for the parameters in the -th ayer at the k-th iteration and g (k) denotes a vector of the gradients of the parameters in the -th ayer at the k-th iteration. Thus, we see that we use ony the gradients in the same ayer to determine the earning rate for that ayer. It is aso important to note that we do not use any gradients from previous iterations, and thus save on storage. From equation 9, we see that when the gradients in a ayer are very arge, the equation just reduces to using the norma earning rate t (k). However, when the gradients are very sma, we are more ikey to be near a ow curvature point. Thus, the equation scaes up the earning rate to ensure that the initia ayers of the network earn faster, and that we escape higherror ow curvature sadde points quicky. We can use this ayer-specific earning rate on top of SGD. Using equation 3, the update in that case, woud be: = t (k) g (k) (10) = t (k) (1 + og(1 + 1/( g (k) 2 )))g (k) (11) where denotes the update in the parameters of the -th ayer at the k-th iteration. Simiary, we can modify AdaGrad s update equation (8) to use our modified earning rates. = t (k) k i=1 (g(i) (k) g ) 2 (12) Note that, unike AdaGrad which uses a distinct earning rate for each parameter, we use a different earning rate for each ayer, which is shared by a weights in that ayer. Additionay, AdaGrad modifies the earning rate based on the entire history of gradients observed for that weight whie we update a ayer s earning rate based ony on gradients observed for a weights in a specific ayer in the current iteration. Thus, our scheme avoids both storing gradient information from previous iterations and computing earning rates for each parameter; it is therefore ess computationay and memory intensive when compared to AdaGrad. The proposed ayer specific earning rates aso works we on arge scae datasets ike ImageNet (when appied over SGD), where AdaGrad fais to converge to a good soution. The proposed method can be used with any existing optimization technique which uses a goba earning rate, provides a ayer-specific earning rate, and escapes sadde points quicky, a without sacrificing computation or memory usage. As we show in Section IV, using our adaptive earning rates on top of existing optimization techniques amost aways improves performance on standard datasets. The proposed method can be used with any existing optimization technique which uses a goba earning rate. This heps in getting a ayer-specific earning rate, as we as, heps in escaping sadde points quicker, with very itte computationa overhead. As we show in Section IV, using our adaptive earning rates on top of existing optimization techniques amost aways improves performance on standard datasets. A. Dataset IV. EXPERIMENTAL RESULTS We present image cassification resuts on three standard datasets: MNIST, CIFAR10 and ImageNet (ILSVRC 2012 dataset, part of the ImageNet chaenge). MNIST contains 60,000 handwritten digit images for training and 10,000 handwritten digit images for testing. CIFAR10 contains has 10 casses with 6,000 images in each cass. ImageNet contains 1.2 miion coor images from 1000 different casses. B. Experimenta Detais We use Caffe [16] to impement our method. Caffe provides optimization methods for Stochastic Gradient Descent (SGD), Nesterov s Acceerated Gradient (NAG) and AdaGrad. For a fair comparison between state-of-the-art methods, we add our adaptive ayer-specific earning rate method on top of each of these optimization methods. In our experiments, we demonstrate the effectiveness of our agorithm on convoutiona neura networks on 3 datasets. On CIFAR10, we use the same goba earning rate as provided in Caffe. Since our method aways increases the ayer-specific earning rate (with respect to other optimization methods) based on the goba earning rate, we start with a sighty smaer earning rate of to make the earning ess aggressive for the ImageNet experiment. SGD was initiaized with the earning rate used in [2] for experiments done on ImageNet. 1) MNIST: We use the same architecture as LeNet for our experiments on MNIST. We present the resuts of using our proposed ayer-specific earning rates on top of stochastic gradient descent, Nesterov s acceerated gradient method and AdaGrad on the MNIST dataset. Since a methods converge very quicky on this dataset, we present the accuracy and oss ony for the first 2,000 iterations. Tabe I shows the

4 Iteration SGD Ours-SGD Nesterov Ours-NAG AdaGrad Ours-AdaGrad ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± 0.08 TABLE I: Mean error rate on MNIST after different iterations for stochastic gradient descent, Nesterov s acceerated gradient and AdaGrad with their ayer specific adaptive versions are shown in the tabe. Each method was run 10 times and their mean and standard deviation is reported Accuracy k Step Down 60k Step Down SGD 50k Step Down SGD 60k Step Down Nesterov Adagrad 0.72 (a) Stochastic Gradient Descent 0.73 (b) Nesterov s Acceerated Gradient (c) AdaGrad Fig. 1: On CIFAR data set: pots showing accuracies (Figures 1a-1c) comparing SGD, NAG and AdaGrad, each with our adaptive ayer-wise earning rates. For the SGD pot, we show resuts both when we step down the earning rate at 50,000 iterations as we as 60,000 iterations. mean accuracy and standard deviation when each method was run 10 times. We observe that our proposed ayer-specific earning rate is consistenty better than Nesterov s acceerated gradient, stochastic gradient descent and AdaGrad. In a the experiments, the proposed method aso attains the maximum accuracy of 99.2% just ike stochastic gradient descent, Nesterov s acceerated gradient and AdaGrad. 2) CIFAR10: On CIFAR10 we use a convoutiona neura network with 2 ayers of 32 feature maps from 5 5 convoution kernes, each foowed by 3 3 max pooing ayers. After this we have another convoution ayer with 64 feature maps from a 5 5 convoution kerne foowed by a 3 3 max pooing ayer. Finay, we have a fuy connected ayer with 10 hidden nodes and a soft-max ogistic regression ayer. After each convoution ayer a ReLu non-inearity is appied. This is the same architecture as specified in Caffe. For the first 60,000 iterations the earning rate was and it was dropped by a factor of 10 at 60,000 and 65,000 iterations. On this dataset, we again observe that fina error and oss of our method is consistenty ower than SGD, NAG and AdaGrad (Tabe II). After step down, our adaptive method reaches a ower accuracy than both SGD and NAG. Note that just using our optimization method (without changing the network architecture) we can get an improvement of 0.32% over the mean accuracy for SGD. Even if we step down the earning rate at 50,000 iterations (taking iterations in tota), we obtain an accuracy of 82.08%, which is better than SGD after 70,000 iterations, significanty cutting down on required training time Fig. 1. Since our method converges much faster when used with SGD, it is possibe to perform the step down on the earning rate even earier, potentiay reducing training time even further. Athough Adagrad does not perform very we on CIFAR10 with defaut parameters, we observe an improvement of 1.3% over the mean fina accuracy, with again a significant speed-up in training time. 3) ImageNet: We use an impementation of AexNet [2] in Caffe, a deep convoutiona neura network architecture, for comparing our method with other optimization agorithms. AexNet consists of 5 convoution ayers foowed by 3 fuy connected ayers. More detais regarding the architecture can be found in the paper [2]. Since AexNet is a deep neura network with significant compexity, it is suitabe to appy our method on this network architecture. Fig 2 shows the resuts of using our method over SGD. We observe that our method obtains significanty greater accuracy and ower oss after 100,000 and 200,000 iterations. Further, we are aso abe to reach the maximum accuracy of 57.5% on the vaidation set after 295,000 iterations which is achieved by SGD ony after 345,000 iterations, resuting in a reduction of 15% in training time. Given that such a arge mode takes more than a week to train propery, this is a significant reduction. Our oss is aso consistenty ower than SGD across a iterations. In the existing mode, we perform a step down by a factor of 10 after every 100,000 iterations. In order to anayze how our method performs when we reduce the number of training iterations, we vary the number of training iterations at a specific earning rate before performing a step down. Tabe III shows the fina accuracy after 350,000 iterations of SGD and our method. Athough the fina accuracy drops sighty as we decrease the number of iterations after which we perform the step down in the earning rate, it is ceary evident that our method achieves better accuracy than SGD. Note that we report top-1 cass accuracy. Since we use the Caffe impementation of the AexNet architecture and do not use any data augmentation techniques, our resuts are sighty ower than those reported in [2]. V. CONCLUSIONS In this paper we propose a genera method for training deep neura networks using ayer-specific adaptive earning rates,

5 Iteration SGD Ours-SGD Nesterov Ours-NAG AdaGrad Ours-AdaGrad ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± 0.39 TABLE II: Mean accuracy on CIFAR10 after different iterations for SGD, NAG and AdaGrad with their ayer specific adaptive versions are shown in the tabe. The mean and standard deviation over 5 runs is reported SGD the University of Maryand supercomputing resources ( 2.6 REFERENCES Loss Accuracy Number Of Iterations x (a) Loss with Stochastic Gradient Descent SGD Number Of Iterations x 10 5 (b) Accuracy with Stochastic Gradient Descent Fig. 2: On ImageNet data set: pot comparing stochastic gradient descent with our adaptive ayer-wise earning rates. We can see a consistent improvement in accuracy and oss over the reguar SGD method across a iterations. Iterations SGD 70, % 55.84% 80, % 56.57% 90, % 57.13% TABLE III: Comparison of stochastic gradient descent and Our Method with step-down at different iterations on ImageNet which can be used on top of any optimization method with a goba earning rate. The method uses gradients from each ayer to compute an adaptive earning rate for each ayer. It aims to speed up convergence when the parameters are in a ow curvature sadde point region. Layer-specific earning rates aso enabe the method to prevent sow earning in initia ayers of the deep network, usuay caused by very sma gradient vaues. ACKNOWLEDGMENT The authors acknowedge ONR Grant numbers N WX01341 and N , as we as [1] R. Pascanu, Y. N. Dauphin, S. Gangui, and Y. Bengio, On the sadde point probem for non-convex optimization, arxiv preprint arxiv: , [2] A. Krizhevsky, I. Sutskever, and G. E. Hinton, Imagenet cassification with deep convoutiona neura networks, in Advances in neura information processing systems, 2012, pp [3] Y. Taigman, M. Yang, M. Ranzato, and L. Wof, Deepface: Cosing the gap to human-eve performance in face verification, in Computer Vision and Pattern Recognition (CVPR), 2014 IEEE Conference on. IEEE, 2014, pp [4] R. Socher, A. Pereygin, J. Y. Wu, J. Chuang, C. D. Manning, A. Y. Ng, and C. Potts, Recursive deep modes for semantic compositionaity over a sentiment treebank, in Proceedings of the Conference on Empirica Methods in Natura Language Processing (EMNLP). Citeseer, 2013, pp [5] G. Hinton, L. Deng, D. Yu, G. E. Dah, A.-r. Mohamed, N. Jaity, A. Senior, V. Vanhoucke, P. Nguyen, T. N. Sainath et a., Deep neura networks for acoustic modeing in speech recognition: The shared views of four research groups, Signa Processing Magazine, IEEE, vo. 29, no. 6, pp , [6] X. Gorot, A. Bordes, and Y. Bengio, Deep sparse rectifier networks, in Proceedings of the 14th Internationa Conference on Artificia Inteigence and Statistics. JMLR W&CP Voume, vo. 15, 2011, pp [7] H. Robbins, S. Monro et a., A stochastic approximation method, The Annas of Mathematica Statistics, vo. 22, no. 3, pp , [8] M. D. Zeier, Adadeta: An adaptive earning rate method, arxiv preprint arxiv: , [9] J. Duchi, E. Hazan, and Y. Singer, Adaptive subgradient methods for onine earning and stochastic optimization, The Journa of Machine Learning Research, vo. 12, pp , [10] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neura computation, vo. 9, no. 8, pp , [11] A. J. Bray and D. S. Dean, Statistics of critica points of gaussian fieds on arge-dimensiona spaces, Physica review etters, vo. 98, no. 15, p , [12] Y. V. Fyodorov and I. Wiiams, Repica symmetry breaking condition exposed by random matrix cacuation of andscape compexity, Journa of Statistica Physics, vo. 129, no. 5-6, pp , [13] B. T. Poyak, Some methods of speeding up the convergence of iteration methods, USSR Computationa Mathematics and Mathematica Physics, vo. 4, no. 5, pp. 1 17, [14] Y. Nesterov, A method of soving a convex programming probem with convergence rate o (1/k2), in Soviet Mathematics Dokady, vo. 27, no. 2, 1983, pp [15] I. Sutskever, J. Martens, G. Dah, and G. Hinton, On the importance of initiaization and momentum in deep earning, in Proceedings of the 30th Internationa Conference on Machine Learning (ICML-13), 2013, pp [16] Y. Jia, E. Shehamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darre, Caffe: Convoutiona architecture for fast feature embedding, arxiv preprint arxiv: , 2014.

Nearest Neighbor Learning

Nearest Neighbor Learning Nearest Neighbor Learning Cassify based on oca simiarity Ranges from simpe nearest neighbor to case-based and anaogica reasoning Use oca information near the current query instance to decide the cassification

More information

Learning to Learn Second-Order Back-Propagation for CNNs Using LSTMs

Learning to Learn Second-Order Back-Propagation for CNNs Using LSTMs Learning to Learn Second-Order Bac-Propagation for CNNs Using LSTMs Anirban Roy SRI Internationa Meno Par, USA anirban.roy@sri.com Sinisa Todorovic Oregon State University Corvais, USA sinisa@eecs.oregonstate.edu

More information

Automatic Grouping for Social Networks CS229 Project Report

Automatic Grouping for Social Networks CS229 Project Report Automatic Grouping for Socia Networks CS229 Project Report Xiaoying Tian Ya Le Yangru Fang Abstract Socia networking sites aow users to manuay categorize their friends, but it is aborious to construct

More information

Language Identification for Texts Written in Transliteration

Language Identification for Texts Written in Transliteration Language Identification for Texts Written in Transiteration Andrey Chepovskiy, Sergey Gusev, Margarita Kurbatova Higher Schoo of Economics, Data Anaysis and Artificia Inteigence Department, Pokrovskiy

More information

Mobile App Recommendation: Maximize the Total App Downloads

Mobile App Recommendation: Maximize the Total App Downloads Mobie App Recommendation: Maximize the Tota App Downoads Zhuohua Chen Schoo of Economics and Management Tsinghua University chenzhh3.12@sem.tsinghua.edu.cn Yinghui (Catherine) Yang Graduate Schoo of Management

More information

The Classification of Stored Grain Pests based on Convolutional Neural Network

The Classification of Stored Grain Pests based on Convolutional Neural Network 2017 2nd Internationa Conference on Mechatronics and Information Technoogy (ICMIT 2017) The Cassification of Stored Grain Pests based on Convoutiona Neura Network Dexian Zhang1, Wenun Zhao*, 1 1 Schoo

More information

JOINT IMAGE REGISTRATION AND EXAMPLE-BASED SUPER-RESOLUTION ALGORITHM

JOINT IMAGE REGISTRATION AND EXAMPLE-BASED SUPER-RESOLUTION ALGORITHM JOINT IMAGE REGISTRATION AND AMPLE-BASED SUPER-RESOLUTION ALGORITHM Hyo-Song Kim, Jeyong Shin, and Rae-Hong Park Department of Eectronic Engineering, Schoo of Engineering, Sogang University 35 Baekbeom-ro,

More information

Sensitivity Analysis of Hopfield Neural Network in Classifying Natural RGB Color Space

Sensitivity Analysis of Hopfield Neural Network in Classifying Natural RGB Color Space Sensitivity Anaysis of Hopfied Neura Network in Cassifying Natura RGB Coor Space Department of Computer Science University of Sharjah UAE rsammouda@sharjah.ac.ae Abstract: - This paper presents a study

More information

Distance Weighted Discrimination and Second Order Cone Programming

Distance Weighted Discrimination and Second Order Cone Programming Distance Weighted Discrimination and Second Order Cone Programming Hanwen Huang, Xiaosun Lu, Yufeng Liu, J. S. Marron, Perry Haaand Apri 3, 2012 1 Introduction This vignette demonstrates the utiity and

More information

NestedNet: Learning Nested Sparse Structures in Deep Neural Networks

NestedNet: Learning Nested Sparse Structures in Deep Neural Networks NestedNet: Learning Nested Sparse Structures in Deep Neura Networks Eunwoo Kim Chanho Ahn Songhwai Oh Department of ECE and ASRI, Seou Nationa University, South Korea {kewoo15, mychahn, songhwai}@snu.ac.kr

More information

A Design Method for Optimal Truss Structures with Certain Redundancy Based on Combinatorial Rigidity Theory

A Design Method for Optimal Truss Structures with Certain Redundancy Based on Combinatorial Rigidity Theory 0 th Word Congress on Structura and Mutidiscipinary Optimization May 9 -, 03, Orando, Forida, USA A Design Method for Optima Truss Structures with Certain Redundancy Based on Combinatoria Rigidity Theory

More information

Hiding secrete data in compressed images using histogram analysis

Hiding secrete data in compressed images using histogram analysis University of Woongong Research Onine University of Woongong in Dubai - Papers University of Woongong in Dubai 2 iding secrete data in compressed images using histogram anaysis Farhad Keissarian University

More information

A Memory Grouping Method for Sharing Memory BIST Logic

A Memory Grouping Method for Sharing Memory BIST Logic A Memory Grouping Method for Sharing Memory BIST Logic Masahide Miyazai, Tomoazu Yoneda, and Hideo Fuiwara Graduate Schoo of Information Science, Nara Institute of Science and Technoogy (NAIST), 8916-5

More information

ACTIVE LEARNING ON WEIGHTED GRAPHS USING ADAPTIVE AND NON-ADAPTIVE APPROACHES. Eyal En Gad, Akshay Gadde, A. Salman Avestimehr and Antonio Ortega

ACTIVE LEARNING ON WEIGHTED GRAPHS USING ADAPTIVE AND NON-ADAPTIVE APPROACHES. Eyal En Gad, Akshay Gadde, A. Salman Avestimehr and Antonio Ortega ACTIVE LEARNING ON WEIGHTED GRAPHS USING ADAPTIVE AND NON-ADAPTIVE APPROACHES Eya En Gad, Akshay Gadde, A. Saman Avestimehr and Antonio Ortega Department of Eectrica Engineering University of Southern

More information

Automatic Hidden Web Database Classification

Automatic Hidden Web Database Classification Automatic idden Web atabase Cassification Zhiguo Gong, Jingbai Zhang, and Qian Liu Facuty of Science and Technoogy niversity of Macau Macao, PRC {fstzgg,ma46597,ma46620}@umac.mo Abstract. In this paper,

More information

Research of Classification based on Deep Neural Network

Research of  Classification based on Deep Neural Network 2018 Internationa Conference on Sensor Network and Computer Engineering (ICSNCE 2018) Research of Emai Cassification based on Deep Neura Network Wang Yawen Schoo of Computer Science and Engineering Xi

More information

GPU Implementation of Parallel SVM as Applied to Intrusion Detection System

GPU Implementation of Parallel SVM as Applied to Intrusion Detection System GPU Impementation of Parae SVM as Appied to Intrusion Detection System Sudarshan Hiray Research Schoar, Department of Computer Engineering, Vishwakarma Institute of Technoogy, Pune, India sdhiray7@gmai.com

More information

Neural Networks. Aarti Singh. Machine Learning Nov 3, Slides Courtesy: Tom Mitchell

Neural Networks. Aarti Singh. Machine Learning Nov 3, Slides Courtesy: Tom Mitchell Neura Networks Aarti Singh Machine Learning 10-601 Nov 3, 2011 Sides Courtesy: Tom Mitche 1 Logis0c Regression Assumes the foowing func1ona form for P(Y X): Logis1c func1on appied to a inear func1on of

More information

Optimization and Application of Support Vector Machine Based on SVM Algorithm Parameters

Optimization and Application of Support Vector Machine Based on SVM Algorithm Parameters Optimization and Appication of Support Vector Machine Based on SVM Agorithm Parameters YAN Hui-feng 1, WANG Wei-feng 1, LIU Jie 2 1 ChongQing University of Posts and Teecom 400065, China 2 Schoo Of Civi

More information

A Fast Block Matching Algorithm Based on the Winner-Update Strategy

A Fast Block Matching Algorithm Based on the Winner-Update Strategy In Proceedings of the Fourth Asian Conference on Computer Vision, Taipei, Taiwan, Jan. 000, Voume, pages 977 98 A Fast Bock Matching Agorithm Based on the Winner-Update Strategy Yong-Sheng Chenyz Yi-Ping

More information

Fast Methods for Kernel-based Text Analysis

Fast Methods for Kernel-based Text Analysis Proceedings of the 41st Annua Meeting of the Association for Computationa Linguistics, Juy 2003, pp. 24-31. Fast Methods for Kerne-based Text Anaysis Taku Kudo and Yuji Matsumoto Graduate Schoo of Information

More information

Research on UAV Fixed Area Inspection based on Image Reconstruction

Research on UAV Fixed Area Inspection based on Image Reconstruction Research on UAV Fixed Area Inspection based on Image Reconstruction Kun Cao a, Fei Wu b Schoo of Eectronic and Eectrica Engineering, Shanghai University of Engineering Science, Abstract Shanghai 20600,

More information

Neural Network Enhancement of the Los Alamos Force Deployment Estimator

Neural Network Enhancement of the Los Alamos Force Deployment Estimator Missouri University of Science and Technoogy Schoars' Mine Eectrica and Computer Engineering Facuty Research & Creative Works Eectrica and Computer Engineering 1-1-1994 Neura Network Enhancement of the

More information

Neural Networks. Aarti Singh & Barnabas Poczos. Machine Learning / Apr 24, Slides Courtesy: Tom Mitchell

Neural Networks. Aarti Singh & Barnabas Poczos. Machine Learning / Apr 24, Slides Courtesy: Tom Mitchell Neura Networks Aarti Singh & Barnabas Poczos Machine Learning 10-701/15-781 Apr 24, 2014 Sides Courtesy: Tom Mitche 1 Logis0c Regression Assumes the foowing func1ona form for P(Y X): Logis1c func1on appied

More information

Solving Large Double Digestion Problems for DNA Restriction Mapping by Using Branch-and-Bound Integer Linear Programming

Solving Large Double Digestion Problems for DNA Restriction Mapping by Using Branch-and-Bound Integer Linear Programming The First Internationa Symposium on Optimization and Systems Bioogy (OSB 07) Beijing, China, August 8 10, 2007 Copyright 2007 ORSC & APORC pp. 267 279 Soving Large Doube Digestion Probems for DNA Restriction

More information

Chapter Multidimensional Direct Search Method

Chapter Multidimensional Direct Search Method Chapter 09.03 Mutidimensiona Direct Search Method After reading this chapter, you shoud be abe to:. Understand the fundamentas of the mutidimensiona direct search methods. Understand how the coordinate

More information

Further Optimization of the Decoding Method for Shortened Binary Cyclic Fire Code

Further Optimization of the Decoding Method for Shortened Binary Cyclic Fire Code Further Optimization of the Decoding Method for Shortened Binary Cycic Fire Code Ch. Nanda Kishore Heosoft (India) Private Limited 8-2-703, Road No-12 Banjara His, Hyderabad, INDIA Phone: +91-040-3378222

More information

On Trivial Solution and High Correlation Problems in Deep Supervised Hashing

On Trivial Solution and High Correlation Problems in Deep Supervised Hashing On Trivia Soution and High Correation Probems in Deep Supervised Hashing Yuchen Guo, Xin Zhao, Guiguang Ding, Jungong Han Schoo of Software, Tsinghua University, Beijing 84, China Schoo of Computing and

More information

Binarized support vector machines

Binarized support vector machines Universidad Caros III de Madrid Repositorio instituciona e-archivo Departamento de Estadística http://e-archivo.uc3m.es DES - Working Papers. Statistics and Econometrics. WS 2007-11 Binarized support vector

More information

Image Segmentation Using Semi-Supervised k-means

Image Segmentation Using Semi-Supervised k-means I J C T A, 9(34) 2016, pp. 595-601 Internationa Science Press Image Segmentation Using Semi-Supervised k-means Reza Monsefi * and Saeed Zahedi * ABSTRACT Extracting the region of interest is a very chaenging

More information

CNN and RNN Based Neural Networks for Action Recognition

CNN and RNN Based Neural Networks for Action Recognition Journa of Physics: Conference Series PAPER OPEN ACCESS CNN and RNN Based Neura Networks for Action Recognition To cite this artice: Chen Zhao et a 2018 J. Phys.: Conf. Ser. 1087 062013 View the artice

More information

University of Illinois at Urbana-Champaign, Urbana, IL 61801, /11/$ IEEE 162

University of Illinois at Urbana-Champaign, Urbana, IL 61801, /11/$ IEEE 162 oward Efficient Spatia Variation Decomposition via Sparse Regression Wangyang Zhang, Karthik Baakrishnan, Xin Li, Duane Boning and Rob Rutenbar 3 Carnegie Meon University, Pittsburgh, PA 53, wangyan@ece.cmu.edu,

More information

Multi-level Shape Recognition based on Wavelet-Transform. Modulus Maxima

Multi-level Shape Recognition based on Wavelet-Transform. Modulus Maxima uti-eve Shape Recognition based on Waveet-Transform oduus axima Faouzi Aaya Cheikh, Azhar Quddus and oncef Gabbouj Tampere University of Technoogy (TUT), Signa Processing aboratory, P.O. Box 553, FIN-33101

More information

Topology-aware Key Management Schemes for Wireless Multicast

Topology-aware Key Management Schemes for Wireless Multicast Topoogy-aware Key Management Schemes for Wireess Muticast Yan Sun, Wade Trappe,andK.J.RayLiu Department of Eectrica and Computer Engineering, University of Maryand, Coege Park Emai: ysun, kjriu@gue.umd.edu

More information

Transformation Invariance in Pattern Recognition: Tangent Distance and Propagation

Transformation Invariance in Pattern Recognition: Tangent Distance and Propagation Transformation Invariance in Pattern Recognition: Tangent Distance and Propagation Patrice Y. Simard, 1 Yann A. Le Cun, 2 John S. Denker, 2 Bernard Victorri 3 1 Microsoft Research, 1 Microsoft Way, Redmond,

More information

Lecture outline Graphics and Interaction Scan Converting Polygons and Lines. Inside or outside a polygon? Scan conversion.

Lecture outline Graphics and Interaction Scan Converting Polygons and Lines. Inside or outside a polygon? Scan conversion. Lecture outine 433-324 Graphics and Interaction Scan Converting Poygons and Lines Department of Computer Science and Software Engineering The Introduction Scan conversion Scan-ine agorithm Edge coherence

More information

A Fast-Convergence Decoding Method and Memory-Efficient VLSI Decoder Architecture for Irregular LDPC Codes in the IEEE 802.

A Fast-Convergence Decoding Method and Memory-Efficient VLSI Decoder Architecture for Irregular LDPC Codes in the IEEE 802. A Fast-Convergence Decoding Method and Memory-Efficient VLSI Decoder Architecture for Irreguar LDPC Codes in the IEEE 82.16e Standards Yeong-Luh Ueng and Chung-Chao Cheng Dept. of Eectrica Engineering,

More information

WHILE estimating the depth of a scene from a single image

WHILE estimating the depth of a scene from a single image JOURNAL OF L A T E X CLASS FILES, VOL. 4, NO. 8, AUGUST 05 Monocuar Depth Estimation using Muti-Scae Continuous CRFs as Sequentia Deep Networks Dan Xu, Student Member, IEEE, Eisa Ricci, Member, IEEE, Wani

More information

Response Surface Model Updating for Nonlinear Structures

Response Surface Model Updating for Nonlinear Structures Response Surface Mode Updating for Noninear Structures Gonaz Shahidi a, Shamim Pakzad b a PhD Student, Department of Civi and Environmenta Engineering, Lehigh University, ATLSS Engineering Research Center,

More information

A New Supervised Clustering Algorithm Based on Min-Max Modular Network with Gaussian-Zero-Crossing Functions

A New Supervised Clustering Algorithm Based on Min-Max Modular Network with Gaussian-Zero-Crossing Functions 2006 Internationa Joint Conference on Neura Networks Sheraton Vancouver Wa Centre Hote, Vancouver, BC, Canada Juy 16-21, 2006 A New Supervised Custering Agorithm Based on Min-Max Moduar Network with Gaussian-Zero-Crossing

More information

A Comparison of a Second-Order versus a Fourth- Order Laplacian Operator in the Multigrid Algorithm

A Comparison of a Second-Order versus a Fourth- Order Laplacian Operator in the Multigrid Algorithm A Comparison of a Second-Order versus a Fourth- Order Lapacian Operator in the Mutigrid Agorithm Kaushik Datta (kdatta@cs.berkeey.edu Math Project May 9, 003 Abstract In this paper, the mutigrid agorithm

More information

On-Chip CNN Accelerator for Image Super-Resolution

On-Chip CNN Accelerator for Image Super-Resolution On-Chip CNN Acceerator for Image Super-Resoution Jung-Woo Chang and Suk-Ju Kang Dept. of Eectronic Engineering, Sogang University, Seou, South Korea {zwzang91, sjkang}@sogang.ac.kr ABSTRACT To impement

More information

Deep Hashing Network for Efficient Similarity Retrieval

Deep Hashing Network for Efficient Similarity Retrieval Proceedings of the Thirtieth AAAI Conference on Artificia Inteigence (AAAI-16) Deep Hashing Network for Efficient Simiarity Retrieva Han Zhu, Mingsheng Long, Jianmin Wang and Yue Cao Schoo of Software,

More information

CLOUD RADIO ACCESS NETWORK WITH OPTIMIZED BASE-STATION CACHING

CLOUD RADIO ACCESS NETWORK WITH OPTIMIZED BASE-STATION CACHING CLOUD RADIO ACCESS NETWORK WITH OPTIMIZED BASE-STATION CACHING Binbin Dai and Wei Yu Ya-Feng Liu Department of Eectrica and Computer Engineering University of Toronto, Toronto ON, Canada M5S 3G4 Emais:

More information

An Introduction to Design Patterns

An Introduction to Design Patterns An Introduction to Design Patterns 1 Definitions A pattern is a recurring soution to a standard probem, in a context. Christopher Aexander, a professor of architecture Why woud what a prof of architecture

More information

Multi-Scale Continuous CRFs as Sequential Deep Networks for Monocular Depth Estimation

Multi-Scale Continuous CRFs as Sequential Deep Networks for Monocular Depth Estimation Muti-Scae Continuous CRFs as Sequentia Deep Networks for Monocuar Depth Estimation Dan Xu 1, Eisa Ricci 4,5, Wani Ouyang 2,3, Xiaogang Wang 2, Nicu Sebe 1 1 University of Trento, 2 The Chinese University

More information

An Adaptive Two-Copy Delayed SR-ARQ for Satellite Channels with Shadowing

An Adaptive Two-Copy Delayed SR-ARQ for Satellite Channels with Shadowing An Adaptive Two-Copy Deayed SR-ARQ for Sateite Channes with Shadowing Jing Zhu, Sumit Roy zhuj@ee.washington.edu Department of Eectrica Engineering, University of Washington Abstract- The paper focuses

More information

Semi-Supervised Learning with Sparse Distributed Representations

Semi-Supervised Learning with Sparse Distributed Representations Semi-Supervised Learning with Sparse Distributed Representations David Zieger dzieger@stanford.edu CS 229 Fina Project 1 Introduction For many machine earning appications, abeed data may be very difficut

More information

Improvement of Nearest-Neighbor Classifiers via Support Vector Machines

Improvement of Nearest-Neighbor Classifiers via Support Vector Machines From: FLAIRS-01 Proceedings. Copyright 2001, AAAI (www.aaai.org). A rights reserved. Improvement of Nearest-Neighbor Cassifiers via Support Vector Machines Marc Sebban and Richard Nock TRIVIA-Department

More information

InnerSpec: Technical Report

InnerSpec: Technical Report InnerSpec: Technica Report Fabrizio Guerrini, Aessandro Gnutti, Riccardo Leonardi Department of Information Engineering, University of Brescia Via Branze 38, 25123 Brescia, Itay {fabrizio.guerrini, a.gnutti006,

More information

COMPRESSIVE sensing (CS), which aims at recovering

COMPRESSIVE sensing (CS), which aims at recovering D-Net: Deep Learning pproach for Compressive Sensing RI Yan Yang, Jian Sun, Huibin Li, and Zongben u ariv:705.06869v [cs.cv] 9 ay 07 bstract Compressive sensing (CS) is an effective approach for fast agnetic

More information

A Discriminative Global Training Algorithm for Statistical MT

A Discriminative Global Training Algorithm for Statistical MT Discriminative Goba Training gorithm for Statistica MT Christoph Timann IBM T.J. Watson Research Center Yorktown Heights, N.Y. 10598 cti@us.ibm.com Tong Zhang Yahoo! Research New York Cit, N.Y. 10011 tzhang@ahoo-inc.com

More information

Deep Quantization Network for Efficient Image Retrieval

Deep Quantization Network for Efficient Image Retrieval Proceedings of the Thirtieth AAAI Conference on Artificia Inteigence (AAAI-16) Deep Quantiation Network for Efficient Image Retrieva Yue Cao, Mingsheng Long, Jianmin Wang, Han Zhu and Qingfu Wen Schoo

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Schoo of Computer Science Probabiistic Graphica Modes Distributed Agorithms for ML David (Wei) Dai Lecture 21, Apri 5, 2017 Eric Xing @ CMU, 2005-2017 1 Massive Data 1B+ USERS 30+ PETABYTES 32 miion pages

More information

Deep Fisher Networks for Large-Scale Image Classification

Deep Fisher Networks for Large-Scale Image Classification Deep Fisher Networs for Large-Scae Image Cassification Karen Simonyan Andrea Vedadi Andrew Zisserman Visua Geometry Group, University of Oxford {aren,vedadi,az}@robots.ox.ac.u Abstract As massivey parae

More information

AN EVOLUTIONARY APPROACH TO OPTIMIZATION OF A LAYOUT CHART

AN EVOLUTIONARY APPROACH TO OPTIMIZATION OF A LAYOUT CHART 13 AN EVOLUTIONARY APPROACH TO OPTIMIZATION OF A LAYOUT CHART Eva Vona University of Ostrava, 30th dubna st. 22, Ostrava, Czech Repubic e-mai: Eva.Vona@osu.cz Abstract: This artice presents the use of

More information

THE PERCENTAGE OCCUPANCY HIT OR MISS TRANSFORM

THE PERCENTAGE OCCUPANCY HIT OR MISS TRANSFORM 17th European Signa Processing Conference (EUSIPCO 2009) Gasgow, Scotand, August 24-28, 2009 THE PERCENTAGE OCCUPANCY HIT OR MISS TRANSFORM P. Murray 1, S. Marsha 1, and E.Buinger 2 1 Dept. of Eectronic

More information

Resource Optimization to Provision a Virtual Private Network Using the Hose Model

Resource Optimization to Provision a Virtual Private Network Using the Hose Model Resource Optimization to Provision a Virtua Private Network Using the Hose Mode Monia Ghobadi, Sudhakar Ganti, Ghoamai C. Shoja University of Victoria, Victoria C, Canada V8W 3P6 e-mai: {monia, sganti,

More information

Fault detection and classification by unsupervised feature extraction and dimensionality reduction

Fault detection and classification by unsupervised feature extraction and dimensionality reduction Compex Inte. Syst. (2015) 1:25 33 DOI 10.1007/s40747-015-0004-2 ORIGINAL ARTICLE Faut detection and cassification by unsupervised feature extraction and dimensionaity reduction Praveen Chopra 1,2 Sandeep

More information

PRUNING CONVOLUTIONAL NEURAL NETWORKS

PRUNING CONVOLUTIONAL NEURAL NETWORKS Pubished as a conference paper at ICLR 2017 PRUNING CONVOLUTIONAL NEURAL NETWORKS FOR RESOURCE EFFICIENT INFERENCE Pavo Mochanov, Stephen Tyree, Tero Karras, Timo Aia, Jan Kautz NVIDIA {pmochanov, styree,

More information

CORRELATION filters (CFs) are a useful tool for a variety

CORRELATION filters (CFs) are a useful tool for a variety Zero-Aiasing Correation Fiters for Object Recognition Joseph A. Fernandez, Student Member, IEEE, Vishnu Naresh Boddeti, Member, IEEE, Andres Rodriguez, Member, IEEE, B. V. K. Vijaya Kumar, Feow, IEEE arxiv:4.36v

More information

Special Edition Using Microsoft Excel Selecting and Naming Cells and Ranges

Special Edition Using Microsoft Excel Selecting and Naming Cells and Ranges Specia Edition Using Microsoft Exce 2000 - Lesson 3 - Seecting and Naming Ces and.. Page 1 of 8 [Figures are not incuded in this sampe chapter] Specia Edition Using Microsoft Exce 2000-3 - Seecting and

More information

Portable Compiler Optimisation Across Embedded Programs and Microarchitectures using Machine Learning

Portable Compiler Optimisation Across Embedded Programs and Microarchitectures using Machine Learning Portabe Compier Optimisation Across Embedded Programs and Microarchitectures using Machine Learning Christophe Dubach, Timothy M. Jones, Edwin V. Bonia Members of HiPEAC Schoo of Informatics University

More information

Stereo Matching with Energy Minimizing Snake Grid for 3D Face Modeling

Stereo Matching with Energy Minimizing Snake Grid for 3D Face Modeling Stereo Matching with Energy Minimizing Snake Grid for 3D Face Modeing Shafik Huq 1, Besma Abidi 1, Ardeshir Goshtasby 2, and Mongi Abidi 1 1 Imaging, Robotics, and Inteigent System (IRIS) Laboratory, Department

More information

Incremental Discovery of Object Parts in Video Sequences

Incremental Discovery of Object Parts in Video Sequences Incrementa Discovery of Object Parts in Video Sequences Stéphane Drouin, Patrick Hébert and Marc Parizeau Computer Vision and Systems Laboratory, Department of Eectrica and Computer Engineering Lava University,

More information

Lecture Notes for Chapter 4 Part III. Introduction to Data Mining

Lecture Notes for Chapter 4 Part III. Introduction to Data Mining Data Mining Cassification: Basic Concepts, Decision Trees, and Mode Evauation Lecture Notes for Chapter 4 Part III Introduction to Data Mining by Tan, Steinbach, Kumar Adapted by Qiang Yang (2010) Tan,Steinbach,

More information

Backing-up Fuzzy Control of a Truck-trailer Equipped with a Kingpin Sliding Mechanism

Backing-up Fuzzy Control of a Truck-trailer Equipped with a Kingpin Sliding Mechanism Backing-up Fuzzy Contro of a Truck-traier Equipped with a Kingpin Siding Mechanism G. Siamantas and S. Manesis Eectrica & Computer Engineering Dept., University of Patras, Patras, Greece gsiama@upatras.gr;stam.manesis@ece.upatras.gr

More information

An Alternative Approach for Solving Bi-Level Programming Problems

An Alternative Approach for Solving Bi-Level Programming Problems American Journa of Operations Research, 07, 7, 9-7 http://www.scirp.org/ourna/aor ISSN Onine: 60-889 ISSN rint: 60-880 An Aternative Approach for Soving Bi-Leve rogramming robems Rashmi Bira, Viay K. Agarwa,

More information

Real-Time Feature Descriptor Matching via a Multi-Resolution Exhaustive Search Method

Real-Time Feature Descriptor Matching via a Multi-Resolution Exhaustive Search Method 297 Rea-Time Feature escriptor Matching via a Muti-Resoution Ehaustive Search Method Chi-Yi Tsai, An-Hung Tsao, and Chuan-Wei Wang epartment of Eectrica Engineering, Tamang University, New Taipei City,

More information

Interpreting Individual Classifications of Hierarchical Networks

Interpreting Individual Classifications of Hierarchical Networks Interpreting Individua Cassifications of Hierarchica Networks Wi Landecker, Michae D. Thomure, Luís M. A. Bettencourt, Meanie Mitche, Garrett T. Kenyon, and Steven P. Brumby Department of Computer Science

More information

Learning Dynamic Guidance for Depth Image Enhancement

Learning Dynamic Guidance for Depth Image Enhancement Learning Dynamic Guidance for Depth Image Enhancement Shuhang Gu 1, Wangmeng Zuo 2, Shi Guo 2, Yunjin Chen 3, Chongyu Chen 4,1, Lei Zhang 1, 1 The Hong Kong Poytechnic University, 2 Harbin Institute of

More information

Application of Intelligence Based Genetic Algorithm for Job Sequencing Problem on Parallel Mixed-Model Assembly Line

Application of Intelligence Based Genetic Algorithm for Job Sequencing Problem on Parallel Mixed-Model Assembly Line American J. of Engineering and Appied Sciences 3 (): 5-24, 200 ISSN 94-7020 200 Science Pubications Appication of Inteigence Based Genetic Agorithm for Job Sequencing Probem on Parae Mixed-Mode Assemby

More information

Joint disparity and motion eld estimation in. stereoscopic image sequences. Ioannis Patras, Nikos Alvertos and Georgios Tziritas y.

Joint disparity and motion eld estimation in. stereoscopic image sequences. Ioannis Patras, Nikos Alvertos and Georgios Tziritas y. FORTH-ICS / TR-157 December 1995 Joint disparity and motion ed estimation in stereoscopic image sequences Ioannis Patras, Nikos Avertos and Georgios Tziritas y Abstract This work aims at determining four

More information

Optimized Base-Station Cache Allocation for Cloud Radio Access Network with Multicast Backhaul

Optimized Base-Station Cache Allocation for Cloud Radio Access Network with Multicast Backhaul Optimized Base-Station Cache Aocation for Coud Radio Access Network with Muticast Backhau Binbin Dai, Student Member, IEEE, Ya-Feng Liu, Member, IEEE, and Wei Yu, Feow, IEEE arxiv:804.0730v [cs.it] 28

More information

MACHINE learning techniques can, automatically,

MACHINE learning techniques can, automatically, Proceedings of Internationa Joint Conference on Neura Networks, Daas, Texas, USA, August 4-9, 203 High Leve Data Cassification Based on Network Entropy Fiipe Aves Neto and Liang Zhao Abstract Traditiona

More information

A Novel Linear-Polynomial Kernel to Construct Support Vector Machines for Speech Recognition

A Novel Linear-Polynomial Kernel to Construct Support Vector Machines for Speech Recognition Journa of Computer Science 7 (7): 99-996, 20 ISSN 549-3636 20 Science Pubications A Nove Linear-Poynomia Kerne to Construct Support Vector Machines for Speech Recognition Bawant A. Sonkambe and 2 D.D.

More information

WATERMARKING GIS DATA FOR DIGITAL MAP COPYRIGHT PROTECTION

WATERMARKING GIS DATA FOR DIGITAL MAP COPYRIGHT PROTECTION WATERMARKING GIS DATA FOR DIGITAL MAP COPYRIGHT PROTECTION Shen Tao Chinese Academy of Surveying and Mapping, Beijing 100039, China shentao@casm.ac.cn Xu Dehe Institute of resources and environment, North

More information

file://j:\macmillancomputerpublishing\chapters\in073.html 3/22/01

file://j:\macmillancomputerpublishing\chapters\in073.html 3/22/01 Page 1 of 15 Chapter 9 Chapter 9: Deveoping the Logica Data Mode The information requirements and business rues provide the information to produce the entities, attributes, and reationships in ogica mode.

More information

Extended Node-Arc Formulation for the K-Edge-Disjoint Hop-Constrained Network Design Problem

Extended Node-Arc Formulation for the K-Edge-Disjoint Hop-Constrained Network Design Problem Extended Node-Arc Formuation for the K-Edge-Disjoint Hop-Constrained Network Design Probem Quentin Botton Université cathoique de Louvain, Louvain Schoo of Management, (Begique) botton@poms.uc.ac.be Bernard

More information

Quality of Service Evaluations of Multicast Streaming Protocols *

Quality of Service Evaluations of Multicast Streaming Protocols * Quaity of Service Evauations of Muticast Streaming Protocos Haonan Tan Derek L. Eager Mary. Vernon Hongfei Guo omputer Sciences Department University of Wisconsin-Madison, USA {haonan, vernon, guo}@cs.wisc.edu

More information

Model-driven Collaboration and Information Integration for Enhancing Video Semantic Concept Detection

Model-driven Collaboration and Information Integration for Enhancing Video Semantic Concept Detection Mode-driven Coaboration and Information Integration for Enhancing Video Semantic Concept Detection Tao Meng, Mei-Ling Shyu Department of Eectrica and Computer Engineering University of Miami Cora Gabes,

More information

Minimizing Resource Cost for Camera Stream Scheduling in Video Data Center

Minimizing Resource Cost for Camera Stream Scheduling in Video Data Center Gao YH, Ma HD, Liu W. Minimizing resource cost for camera stream scheduing in video data center. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 32(3): 555 570 May 2017. DOI 10.1007/s11390-017-1743-x Minimizing

More information

FACE RECOGNITION WITH HARMONIC DE-LIGHTING. s: {lyqing, sgshan, wgao}jdl.ac.cn

FACE RECOGNITION WITH HARMONIC DE-LIGHTING.  s: {lyqing, sgshan, wgao}jdl.ac.cn FACE RECOGNITION WITH HARMONIC DE-LIGHTING Laiyun Qing 1,, Shiguang Shan, Wen Gao 1, 1 Graduate Schoo, CAS, Beijing, China, 100080 ICT-ISVISION Joint R&D Laboratory for Face Recognition, CAS, Beijing,

More information

AUTOMATIC IMAGE RETARGETING USING SALIENCY BASED MESH PARAMETERIZATION

AUTOMATIC IMAGE RETARGETING USING SALIENCY BASED MESH PARAMETERIZATION S.Sai Kumar et a. / (IJCSIT Internationa Journa of Computer Science and Information Technoogies, Vo. 1 (4, 010, 73-79 AUTOMATIC IMAGE RETARGETING USING SALIENCY BASED MESH PARAMETERIZATION 1 S.Sai Kumar,

More information

arxiv: v3 [cs.ds] 6 Apr 2017

arxiv: v3 [cs.ds] 6 Apr 2017 Ke Li Jitendra Maik University of Caifornia, Berkeey, CA 9470, United States KELI@EECSBERKELEYEDU MALIK@EECSBERKELEYEDU arxiv:1510044v3 csds] 6 Apr 017 Abstract Existing methods for retrieving k-nearest

More information

Forgot to compute the new centroids (-1); error in centroid computations (-1); incorrect clustering results (-2 points); more than 2 errors: 0 points.

Forgot to compute the new centroids (-1); error in centroid computations (-1); incorrect clustering results (-2 points); more than 2 errors: 0 points. Probem 1 a. K means is ony capabe of discovering shapes that are convex poygons [1] Cannot discover X shape because X is not convex. [1] DBSCAN can discover X shape. [1] b. K-means is prototype based and

More information

arxiv: v1 [cs.cv] 29 Jul 2018

arxiv: v1 [cs.cv] 29 Jul 2018 Joint Representation and Truncated Inference Learning for Correation Fiter based Tracking Yingjie Yao 1[0000 000 3533 1569], Xiaohe Wu 1[0000 0001 6884 911], Lei Zhang [0000 000 444 494], Shiguang Shan

More information

Load Balancing by MPLS in Differentiated Services Networks

Load Balancing by MPLS in Differentiated Services Networks Load Baancing by MPLS in Differentiated Services Networks Riikka Susitaiva, Jorma Virtamo, and Samui Aato Networking Laboratory, Hesinki University of Technoogy P.O.Box 3000, FIN-02015 HUT, Finand {riikka.susitaiva,

More information

Deep Learning with Tensorflow AlexNet

Deep Learning with Tensorflow   AlexNet Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification

More information

Interpreting Individual Classifications of Hierarchical Networks

Interpreting Individual Classifications of Hierarchical Networks Portand State University PDXSchoar Computer Science Facuty Pubications and Presentations Computer Science 2013 Interpreting Individua Cassifications of Hierarchica Networks Wi Landecker Portand State University,

More information

Planar Graphs of Bounded Degree have Constant Queue Number

Planar Graphs of Bounded Degree have Constant Queue Number Panar Graphs of Bounded Degree have Constant Queue Number Michae A. Bekos, Henry Förster, Martin Gronemann, Tamara Mchedidze 3 Fabrizio Montecchiani 4, Chrysanthi Raftopouou 5, Torsten Ueckerdt 3 Institute

More information

Efficient method to design RF pulses for parallel excitation MRI using gridding and conjugate gradient

Efficient method to design RF pulses for parallel excitation MRI using gridding and conjugate gradient Origina rtice Efficient method to design RF puses for parae excitation MRI using gridding and conjugate gradient Shuo Feng, Jim Ji Department of Eectrica & Computer Engineering, Texas & M University, Texas,

More information

Hour 3: The Network Access Layer Page 1 of 10. Discuss how TCP/IP s Network Access layer relates to the OSI networking model

Hour 3: The Network Access Layer Page 1 of 10. Discuss how TCP/IP s Network Access layer relates to the OSI networking model Hour 3: The Network Access Layer Page 1 of 10 Hour 3: The Network Access Layer At the base of the TCP/IP protoco stack is the Network Access ayer, the coection of services and specifications that provide

More information

Outline. Parallel Numerical Algorithms. Forward Substitution. Triangular Matrices. Solving Triangular Systems. Back Substitution. Parallel Algorithm

Outline. Parallel Numerical Algorithms. Forward Substitution. Triangular Matrices. Solving Triangular Systems. Back Substitution. Parallel Algorithm Outine Parae Numerica Agorithms Chapter 8 Prof. Michae T. Heath Department of Computer Science University of Iinois at Urbana-Champaign CS 554 / CSE 512 1 2 3 4 Trianguar Matrices Michae T. Heath Parae

More information

Proceedings of the International Conference on Systolic Arrays, San Diego, California, U.S.A., May 25-27, 1988 AN EFFICIENT ASYNCHRONOUS MULTIPLIER!

Proceedings of the International Conference on Systolic Arrays, San Diego, California, U.S.A., May 25-27, 1988 AN EFFICIENT ASYNCHRONOUS MULTIPLIER! [1,2] have, in theory, revoutionized cryptography. Unfortunatey, athough offer many advantages over conventiona and authentication), such cock synchronization in this appication due to the arge operand

More information

Quality Assessment using Tone Mapping Algorithm

Quality Assessment using Tone Mapping Algorithm Quaity Assessment using Tone Mapping Agorithm Nandiki.pushpa atha, Kuriti.Rajendra Prasad Research Schoar, Assistant Professor, Vignan s institute of engineering for women, Visakhapatnam, Andhra Pradesh,

More information

On Upper Bounds for Assortment Optimization under the Mixture of Multinomial Logit Models

On Upper Bounds for Assortment Optimization under the Mixture of Multinomial Logit Models On Upper Bounds for Assortment Optimization under the Mixture of Mutinomia Logit Modes Sumit Kunnumka September 30, 2014 Abstract The assortment optimization probem under the mixture of mutinomia ogit

More information

Testing Whether a Set of Code Words Satisfies a Given Set of Constraints *

Testing Whether a Set of Code Words Satisfies a Given Set of Constraints * JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 6, 333-346 (010) Testing Whether a Set of Code Words Satisfies a Given Set of Constraints * HSIN-WEN WEI, WAN-CHEN LU, PEI-CHI HUANG, WEI-KUAN SHIH AND MING-YANG

More information

UNCORRECTED PROOF ARTICLE IN PRESS. , Scott Schoenfeld b. SMM 4402 No. of Pages 5, DTD = July 2003 Disk used

UNCORRECTED PROOF ARTICLE IN PRESS. , Scott Schoenfeld b. SMM 4402 No. of Pages 5, DTD = July 2003 Disk used SMM 2 No. of Pages 5, DTD =.3. 2 Juy 23 Disk used 2 Evoution of crysta orientation distribution coefficients 3 during pastic deformation D.S. Li a, H. Garmestani a, *, Scott Schoenfed b 5 a Schoo of Materias

More information

QoS-Aware Data Transmission and Wireless Energy Transfer: Performance Modeling and Optimization

QoS-Aware Data Transmission and Wireless Energy Transfer: Performance Modeling and Optimization QoS-Aware Data Transmission and Wireess Energy Transfer: Performance Modeing and Optimization Dusit Niyato, Ping Wang, Yeow Wai Leong, and Tan Hwee Pink Schoo of Computer Engineering, Nanyang Technoogica

More information