Homework 1. An Introduction to Neural Networks

Size: px
Start display at page:

Download "Homework 1. An Introduction to Neural Networks"

Transcription

1 Hoework An Introduction to Neural Networks -785: Introduction to Deep Learning Spring 09 OUT: January 4, 09 DUE: February 6, 09, :59 PM Start Here Collaboration policy: You are expected to coply with the University Policy on Acadeic Integrity and Plagiaris. You are allowed to talk with / work with other students on hoework assignents You can share ideas but not code, you ust subit your own code. All subitted code will be copared against all code subitted this seester and in previous seesters using MOSS. Overview: Part : All of the probles in Part will be graded on Autolab. You can download the starter code hw.py fro Autolab as well. This assignent has 00 points total, including 90 that are autograded. Part : This section of the hoework is an open ended copetition hosted on Kaggle.co, a popular service for hosting predictive odeling and data analytics copetitions. All of the details for copleting Part can be found on the copetition page. Subission: Part : The copressed handout folder hosted on Autolab contains a single python file, hw.py. This contains a teplate for solving all of the probles and includes soe useful helper functions. Make sure that all class and function definitions originally specified as well as class attributes and ethods in hw.py are fully ipleented to the specification provided in this writeup and in the docstrings or coents in hw.py. Your subission ust be titled handin.tar gzip forat and it is inially required to contain a directory called hw, which contains the hw.py file ipleenting the classes, functions, and ethods specified in the writeup. After copleting Training Statistics hw will also contain iages. Other source files that are iported by hw.py are allowed. Please do not iport any other external libraries other than Nupy, as extra packages that do not exist in the autograder iage will cause a subission failure. Part : See the the copetition page for details. Local autograder: To ease your life, we provide in the handout for part a local autograder that you can use on your own code. It roughly has the sae behavior as the one on Autolab, except that it will copare your outputs and attributes with prestored results on prestored data while autolab directly copares you to a solution. In theory, passing one eans that you pass the other, but we recoend you also subit partial solutions on autolab fro tie to tie while copleting the hoework. To run the local autograder : ove your hw folder in local autograder/ and run runner.py fro the local autograder folder.

2 Neural Networks For part of the hoework you will write your own ipleentation of the backpropagation algorith for training your own neural network, as well as a few other features such as activations and batch-noralization. You are required to do this assignent in the Python Python version 3 prograing language. Do not use any autodiff toolboxes PyTorch, TensorFlow, Keras, etc - you are only peritted and recoended to vectorize your coputation using the Nupy library. In the file hw.py we provide you with classes whose ethods you have to fill. The autograder tests will copare the outputs of your ethods and attributes of your classes with a reference solution. Therefore we enforce for a large part the design of your code ; however, you still have a lot of freedo in your ipleentation. We recoend that you look through all of the probles before attepting the first proble. However we do recoend you coplete the probles in order as the difficulty increases, and questions often rely on the copletion of previous questions. Activations A nuber of activation classes including Sigoid, Tanh and ReLU ust be ipleented. Each of these coe in the for of a class which has both a forward and derivative ethods that ust be ipleented. The Identity activation has been ipleented for you as an exaple. All of these classes derive fro Activation which acts as an abstract base class / interface. You only need to ipleent the two ethods for each class. No helper functions or additional classes should be necessary. Keep your code as concise as possible and leverage Nupy as uch as possible. e indful that you are working with atrices which have a given shape as well as a set of values.. Softax Cross Entropy [5 points] For this hoework, we will be using the softax cross entropy loss detailed in the appendix of this writeup. Use the LogSuExp trick to ensure nuerical stability. Ipleent the ethods specified in the class definition SoftaxCrossEntropy. This class inherits the base Criterion class. Again no helper ethods should be needed and you should stride to keep your code as siple as possible. Use e-8 for eps... SoftaxCrossEntropy.forward Ipleent the softax cross entropy operation on a batch of output vectors. Hint: Add a class attribute to keep track of interediate values necessary for the backward coputation. Input shapes: x: batch size, 0 y: batch size, 0 Output Shape: out: batch size,

3 .. SoftaxCrossEntropy.backward Perfor the backward pass of softax cross entropy operation using interediate values saved in the forward pass. Output shapes: out: batch size, 0 3 Multi-Layer Perceptron MLP In this section of the hoework, you will be ipleenting a Multi-Layer Perceptron with an API siilar to popular Autoatic Differentiation Libraries like PyTorch, which you will be allowed and encouraged to use in the second part of the hoework. Provided in hw.py is a teplate the MLP class. Go through the functions of the given MLP class thoroughly and ake sure you understand what each function in the class does so that you can create a generic ipleentation that supports an arbitrary nuber of layers, types of activations and network sizes. This section is purely descriptive and for reference as you tackle the next probles. Nothing to ipleent... yet. The paraeters for the MLP class are: input size: The size of each individual data exaple. output size: The nuber of outputs. hiddens: A list with the nuber of units in each hidden layer. activations: A list of Activation objects for each layer. weight init fn: A function applied to each weight atrix before training. bias init fn: A function applied to each bias vector before training. criterion: A Criterion object to copute the loss and its derivative. lr: The learning rate. oentu: Moentu scale Should be 0.0 until copleting 3.6. nu bn layers: Nuber of atchnor layers start fro upstrea Should be 0 until copleting 3.5. The attributes of the MLP class The list of weight The list of weight atrix A list of bias A list of bias vector layers: A list of atchnor objects. Should be None until copleting 3.5. The ethods of the MLP class are: forward: Forward pass. Accepts a ini-batch of data and return a batch of output activations. backward: ackward pass. Accepts ground truth labels and coputes gradients for all paraeters. Hint: Use state stored in activations during forward pass to siplify your code. 3

4 zero grads: Set all gradient ters to 0. step: Apply gradients coputed in backward to the paraeters. train Already ipleented: Set the ode of the network to train. eval Already ipleented: Set the ode of the network to evaluation. Note: MLP ethods train and eval will be useful in 3.5. Note: Pay attention to the data structures being passed into the constructor and the class attributes specified initially. Saple constructor call: MLP784, 0, [64, 64, 3], [Sigoid, Sigoid, Sigoid, Identity], weight_init_fn, bias_init_fn, SoftaxCrossEntropy, 0.008, oentu=0.9, nu_bn_layers=0 3. Weight Initialization [0 points] Good initialization has been shown to ake a great difference in the training process. Ipleent two basic paraeter initialization functions rando noral weight init: Accepts two ints specifying the nuber of inputs and units. Returns an appropriately shaped ndarray with each entry initialized with an independent saple fro the standard noral distribution. zeros bias init: Accepts the nuber of bias units and returns an appropriately shaped ndarray with each entry initialized to zero. 3. Linear Classifier [9 points] Ipleent a linear classifier MLP with no hidden layers using the MLP class. We suggest you read through the entire assignent before you start ipleenting this part. For this proble, you will have to fill in the parts of the MLP class and pass the following paraeters: input size: The size of each individual data exaple. output size: The nuber of outputs. hiddens: An epty list because there are no hidden layers. activations: A single Identity activation. weight init fn: Function nae of a weight initializer function you will write for local testing and will be passed to you by the autograder. bias init fn: Function nae of a bias initializer function you will write for local testing and will be passed to you by the autograder. criterion: SoftaxCrossEntropy object. Consider that a linear transforation with an identity activation is siply a linear odel. You will have to ipleent the forward and backward function of the MLP class so that it can at least work as a linear classifier. All paraeters should be aintained as class attributes originally specified in the original handout code e.g. self.w, self.b There won t be any autograder tests that check for weight initializer ipleentations, but you will need to ipleent a rando noral initializer and a zeros initializer to test your ipleentation for

5 The step function also needs to ipleented, as it will be invoked after every backward pass to update the paraeters of the network the gradient descent algorith. After all parts are filled, train the odel for 00 epochs with a batch size of 00 and try visualizing the training curves using the utilities provided in test.py. 3.3 Non-Linearities [ points] Linear odels are cool, but we are here to build neural networks. Ipleent all of the non-linearities specified in hw.py. A saple ipleentation of the Identity activation is provided for reference. The output of the activation should be stored in the self.state variable of the class. The self.state variable should be further used for calculating the derivative during the backward pass. These are the following list of classes for which you would have to ipleent the forward and derivative function. The autograder will test these ipleentations individually.. Sigoid [4 points]. Hyperbolic Tangent Tanh [4 points] 3. Rectified Linear unit ReLU [4 points] 3.4 Hidden Layers [40 points] Update the MLP class that previously just supported a single fully connected layer a linear odel such that it can now support an arbitrary nuber hidden layers, each with an arbitrary nuber of units. Specifically, the hiddens arguent of the MLP class will no longer be assued to be an epty list and can be of arbitrary length arbitrary nuber of layers and contain arbitrary positive integers arbitrary nuber of units in each layer. The ipleentation should apply the Activation objects, passed as activations, to their respective layers. For exaple, activations[0] should be applied to the activity of the first hidden layer. While at risk of being pedantic, here is a clarification of the expected arguents to be passed to the MLP constructor once it can support arbitrary nubers of layers with an arbitrary assortent of activation functions. input size: The size of each individual data exaple. output size: The nuber of outputs. hiddens: A list of layer sizes nuber of units per layer. activations: A list of Activation objects to be applied after each linear transforation respectively. weight init fn: Function nae of a weight initializer function you will write for local testing and will be passed to you by the autograder. bias init fn: Function nae of a bias initializer function you will write for local testing and will be passed to you by the autograder. criterion: SoftaxCrossEntropy object. 5

6 3.5 atch Noralization atchnor [5 points] In this proble you will ipleent the atch Noralization technique fro Ioffe and Szegedy [05]. Ipleent the atchnor class and ake sure that all provided class attributes are set correctly. Apply the atchnor transforation of the first nu bn layers specified as an arguent to MLP. For the sake of the autograder tests, you can assue that batch nor will be applied to networks with sigoid non-linearities. 3.6 Moentu [0 points] Modify the step function present in the MLP class to include oentu in your gradient descent. The oentu value will be passed as a paraeter. Your function should perfor epoch nuber of epochs and return the resulting weights. Instead of calling the step function after copleting your backward pass, you would have to call the oentu function to update the paraeters. Also, reeber to invoke the zero grad function after each batch. Note: Please ensure that you shuffle the training set after each epoch by using np.rando.shuffle and generate a list of indices and perforing a gather operation on the data using these indices. 3.7 Training Statistics [0 points] In this proble you will be passed the MNIST data set provided as a 3-tuple of {train set, val set, test set}, each in the sae forat described in, and a series of network and training paraeters. training losses: A Nupy ndarray containing the average training loss for each epoch. training errors: A Nupy ndarray containing the classification error on the training set at each epoch Hint: You should not be evaluating perforance on the training set after each epoch, but copute an approxiation of the training classification error for each training batch. validation losses: A Nupy ndarray containing the average validation loss loss for each epoch. validation errors: A Nupy ndarray containing the classification error on the validation set after each epoch. Fill the function get training stats that, given a network, a dataset, a nuber of epochs and a batch size, trains the network and returns those quantities. Then execute the provided test file, that will run get training stats on MNIST for a given network and plot the previous arrays into files. Hand-in those iage files along with your code on autolab. Note: You will not be autograded on that part. Instead we will grade you anually by looking at your code and plots. We will check that your code and the statistics you obtain ake sense. Contrary to the other questions, we do not require an exact atch with the solution, only that you run your forward, backward and step functions in the right order, and use the partitions of the dataset correctly train and val at least, with shuffling of the training set between epochs. Note: We provide exaples of the plots that you should obtain. If you get siilar or better values at the end of the training loss around 0.3, error around 0., then you are very likely to get all the points. If you are a bit above but that the errors and losses still drop significantly during training, it s possible that you get all the points as well. References Sergey Ioffe and Christian Szegedy. atch noralization: Accelerating deep network training by reducing internal covariate shift. CoRR, abs/ , 05. URL 6

7 Appendix A Softax Cross Entropy Let s define softax, a useful function that offers a sooth and differentiable version of the ax function with a nice probabilistic interpretation. We denote the softax function for a logit x j of K logits outputs as σx j. e xj σx j = K Notice that the softax function properly noralizes the k logits, so we can interpret each σx j as a probability and the largest logit x j will have the greatest ass in the distribution. σx j = K j= e xj = For two distributions P and Q, we define cross-entropy over discrete events X as j= CE = P x log Qx = P x log Qx x X x X Cross-entropy coes fro inforation theory, where it is defined as the expected inforation quantified as log q of soe subjective distribution Q over an objective distribution P. It quantifies how uch inforation we our odel Q receive when we observe true outcoes fro Q it tells us how far our odel is fro the true distribution P. We iniize Cross-Entropy when P = Q. The value of cross-entropy in this case is known siply as the entropy. H = P x log P x = P x log P x x X x X This is the irreducible inforation we receive when we observe the outcoe of a rando process. Consider a coin toss. Even if we know the ernoulli paraeter p the probability of heads ahead of tie, we will never have absolute certainty about the outcoe until we actually observe the toss. The greater p is, the ore certain we are about the outcoe and the less inforation we expect to receive upon observation. We can noralize our network output using softax and then use Cross-Entropy as an objective function. Our softax outputs represent Q, our subjective distribution. We will denote each softax output as ŷ j and represent the true distribution P with output labels y j = when the label is for each output. We let y j = when the label is j and y j = 0 otherwise. The result is a degenerate distribution that will ai to estiate P when averaged over the training set. Let s foralize this objective function and take the derivative. 7

8 Lŷ, y = CEŷ, y = y i log ŷ i = = y i log σx i y i log ŷ, y = = = = e xi K y i log e xi K y i log e xi log e x k k= K y i log e xi + y i log k= K y i x i + y i log k= e x k e x k The next step is a little bit ore subtle, but recall there is only a single true label for each exaple and therefore only a single y i is equal to ; all others are 0. Therefore we can iagine expanding the su over i in the second ter and only one ter of this su will be and all the others will be zero. ŷ, y = K y i x i + log k= e x k Now we take the partial derivative and reeber that the derivative of a su is the su of the derivative of its ters and that any ter without x j can be discarded. ŷ, y = y j x j + = y j + = y j + = y j + K log k= K e xj e xj K e x k k= e x k That second ter looks very failiar, huh? 8

9 ŷ, y e xj = y j + x K j = y j + σx j = σx j y j = ŷ j y j After all that work we end up with a very siple and elegant expression for the derivative of softax with cross-entropy divergence with respect to its input. What this is telling us is that when y j =, the gradient is negative and thus the opposite direction of the gradient is positive: it is telling us to increase the probability ass of that specific output through the softax. atch Noralization atch Noralization coonly referred to as atchnor is a wildly successful and siple technique for accelerating training and learning better neural network representations. The general otivation of atchnor is the non-stationarity of unit activity during training that requires downstrea units to adapt to a non-stationary input distribution. This co-adaptation proble, which the paper authors refer to as internal covariate shift, significantly slows learning. Just as it is coon to whiten training data standardize and de-correlate the covariates, we can invoke the abstraction of a neural network as a hierarchical set of feature filters and consider the activity of each layer to be the covariates for the subsequent layer and consider whitening the activity over all training exaples after each update. Whitening layer activity across the training data for each paraeter update is coputationally infeasible, so instead we ake soe large assuptions that end up working well anyway. The ain assuption we ake is that the activity of a given unit is independent of the activity of all other units in a given layer. That is, for a layer l with units, individual unit activities consider each a rando variable x = {x k,..., x d } are independent of each other {x... x k... x d }. Under this independence assuption, the covariates are not correlated and therefore we only need to noralize the individual unit activities. Since it is not practical in neural network optiization to perfor updates with a full-batch gradient, we typically use an approxiation of the true full-batch gradient over a subset of the training data. We ake the sae approxiation in our noralization by approxiating the ean and variance of the unit activities over this sae subset of the training data. Given this setup, consider u to be the ean and σ the variance of a unit s activity over a subset of the training data, which we will hence refer to as a batch of data. For a training set X with n exaples, we partition it into n/ batches of size. For an arbitrary unit k, we copute the batch statistics u and σ and noralize as follows: u k σ k ˆx i x i u σ + ɛ x k i x k i u k 3 A significant issue posed by siply noralizing individual unit activity across batches is that it liits the set of possible network representations. A way around this is to introduce a set of learnable paraeters for 9

10 each unit that ensure the atchnor transforation can be learned to perfor an identity transforation. To do so, these per-unit learnable paraeters γ k and β k, rescale and reshift the noralized unit activity. Thus the output of the atchnor transforation for a data exaple, y i is given as follows, y i γˆx i + β 4 We can now derive the analytic partial derivatives of the atchnor transforation. Let L be the training loss over the batch and y the derivative of the loss with respect to the output of the atchnor i transforation for a single data exaple x i. = = γ 5 β = γ = σ Solve for ˆxi = = = β = γ = = σ σ 6 ˆx i 7 [ ] x i µ σ + ɛ x i µ σ + ɛ 3 0 = [ ] x i µ σ + ɛ = σ + ɛ + xi µ [ σ + ɛ = σ + ɛ + xi µ = σ + ɛ x i µ = σ + ɛ x i µ σ + ɛ 3 ] x i u + ɛ 3 x i u + ɛ x i µ x i µ 5 6 Now sub this expression for ˆxi into = = σ + ɛ x i µ σ + ɛ 3 x i µ 7 0

11 Notice that part of the expression in the second ter is just σ = = Now for the grand finale, let s solve for x i σ + ɛ + σ σ + ɛ σ x i µ x i µ x i = x i = [ ˆxi x i + σ σ + ˆx ] i x i x i 8 9 = + σ x i σ + 0 x i x i = = σ x i + + x i σ [ x i µ σ + ɛ x i = [ σ + ɛ ] + σ σ = [ ] σ + ɛ + σ = [ ] σ + ɛ + σ = [ ] σ + ɛ + σ = [ ] σ + ɛ + σ x i ] + σ + x i x i σ + 3 x i x i x j µ + 4 x i µ j= x i [ ] x i µ + 5 x i [ ] x i µ + x j 6 x i j= [ ] x i µ + [ ] 7 In suary, we have derived the following quantities required in the forward and backward coputation for a atchnor applied to a single unit: 8 Forward u = σ = ˆx i = x i u σ + ɛ x i 9 x i u 30 3

12 ackward = = γ 3 β = γ = σ = β = γ = σ 33 ˆx i 34 35

TensorFlow and Keras-based Convolutional Neural Network in CAT Image Recognition Ang LI 1,*, Yi-xiang LI 2 and Xue-hui LI 3

TensorFlow and Keras-based Convolutional Neural Network in CAT Image Recognition Ang LI 1,*, Yi-xiang LI 2 and Xue-hui LI 3 2017 2nd International Conference on Coputational Modeling, Siulation and Applied Matheatics (CMSAM 2017) ISBN: 978-1-60595-499-8 TensorFlow and Keras-based Convolutional Neural Network in CAT Iage Recognition

More information

Structural Balance in Networks. An Optimizational Approach. Andrej Mrvar. Faculty of Social Sciences. University of Ljubljana. Kardeljeva pl.

Structural Balance in Networks. An Optimizational Approach. Andrej Mrvar. Faculty of Social Sciences. University of Ljubljana. Kardeljeva pl. Structural Balance in Networks An Optiizational Approach Andrej Mrvar Faculty of Social Sciences University of Ljubljana Kardeljeva pl. 5 61109 Ljubljana March 23 1994 Contents 1 Balanced and clusterable

More information

EE 364B Convex Optimization An ADMM Solution to the Sparse Coding Problem. Sonia Bhaskar, Will Zou Final Project Spring 2011

EE 364B Convex Optimization An ADMM Solution to the Sparse Coding Problem. Sonia Bhaskar, Will Zou Final Project Spring 2011 EE 364B Convex Optiization An ADMM Solution to the Sparse Coding Proble Sonia Bhaskar, Will Zou Final Project Spring 20 I. INTRODUCTION For our project, we apply the ethod of the alternating direction

More information

Predicting x86 Program Runtime for Intel Processor

Predicting x86 Program Runtime for Intel Processor Predicting x86 Progra Runtie for Intel Processor Behra Mistree, Haidreza Haki Javadi, Oid Mashayekhi Stanford University bistree,hrhaki,oid@stanford.edu 1 Introduction Progras can be equivalent: given

More information

Identifying Converging Pairs of Nodes on a Budget

Identifying Converging Pairs of Nodes on a Budget Identifying Converging Pairs of Nodes on a Budget Konstantina Lazaridou Departent of Inforatics Aristotle University, Thessaloniki, Greece konlaznik@csd.auth.gr Evaggelia Pitoura Coputer Science and Engineering

More information

A Novel Fast Constructive Algorithm for Neural Classifier

A Novel Fast Constructive Algorithm for Neural Classifier A Novel Fast Constructive Algorith for Neural Classifier Xudong Jiang Centre for Signal Processing, School of Electrical and Electronic Engineering Nanyang Technological University Nanyang Avenue, Singapore

More information

Computer Aided Drafting, Design and Manufacturing Volume 26, Number 2, June 2016, Page 13

Computer Aided Drafting, Design and Manufacturing Volume 26, Number 2, June 2016, Page 13 Coputer Aided Drafting, Design and Manufacturing Volue 26, uber 2, June 2016, Page 13 CADDM 3D reconstruction of coplex curved objects fro line drawings Sun Yanling, Dong Lijun Institute of Mechanical

More information

Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression

Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression Sha M. Kakade Microsoft Research and Wharton, U Penn skakade@icrosoft.co Varun Kanade SEAS, Harvard University

More information

Oblivious Routing for Fat-Tree Based System Area Networks with Uncertain Traffic Demands

Oblivious Routing for Fat-Tree Based System Area Networks with Uncertain Traffic Demands Oblivious Routing for Fat-Tree Based Syste Area Networks with Uncertain Traffic Deands Xin Yuan Wickus Nienaber Zhenhai Duan Departent of Coputer Science Florida State University Tallahassee, FL 3306 {xyuan,nienaber,duan}@cs.fsu.edu

More information

A simplified approach to merging partial plane images

A simplified approach to merging partial plane images A siplified approach to erging partial plane iages Mária Kruláková 1 This paper introduces a ethod of iage recognition based on the gradual generating and analysis of data structure consisting of the 2D

More information

Clustering. Cluster Analysis of Microarray Data. Microarray Data for Clustering. Data for Clustering

Clustering. Cluster Analysis of Microarray Data. Microarray Data for Clustering. Data for Clustering Clustering Cluster Analysis of Microarray Data 4/3/009 Copyright 009 Dan Nettleton Group obects that are siilar to one another together in a cluster. Separate obects that are dissiilar fro each other into

More information

Different criteria of dynamic routing

Different criteria of dynamic routing Procedia Coputer Science Volue 66, 2015, Pages 166 173 YSC 2015. 4th International Young Scientists Conference on Coputational Science Different criteria of dynaic routing Kurochkin 1*, Grinberg 1 1 Kharkevich

More information

Spectral Clustering with Two Views

Spectral Clustering with Two Views Spectral Clustering with Two Views Virginia R. de Sa Departent of Cognitive Science, 55 University of California, San Diego 95 Gilan Dr. La Jolla, CA 993-55 desa@ucsd.edu Abstract In this paper we develop

More information

Solving the Damage Localization Problem in Structural Health Monitoring Using Techniques in Pattern Classification

Solving the Damage Localization Problem in Structural Health Monitoring Using Techniques in Pattern Classification Solving the Daage Localization Proble in Structural Health Monitoring Using Techniques in Pattern Classification CS 9 Final Project Due Dec. 4, 007 Hae Young Noh, Allen Cheung, Daxia Ge Introduction Structural

More information

On the Computation and Application of Prototype Point Patterns

On the Computation and Application of Prototype Point Patterns On the Coputation and Application of Prototype Point Patterns Katherine E. Tranbarger Freier 1 and Frederic Paik Schoenberg 2 Abstract This work addresses coputational probles related to the ipleentation

More information

NON-RIGID OBJECT TRACKING: A PREDICTIVE VECTORIAL MODEL APPROACH

NON-RIGID OBJECT TRACKING: A PREDICTIVE VECTORIAL MODEL APPROACH NON-RIGID OBJECT TRACKING: A PREDICTIVE VECTORIAL MODEL APPROACH V. Atienza; J.M. Valiente and G. Andreu Departaento de Ingeniería de Sisteas, Coputadores y Autoática Universidad Politécnica de Valencia.

More information

Image Filter Using with Gaussian Curvature and Total Variation Model

Image Filter Using with Gaussian Curvature and Total Variation Model IJECT Vo l. 7, Is s u e 3, Ju l y - Se p t 016 ISSN : 30-7109 (Online) ISSN : 30-9543 (Print) Iage Using with Gaussian Curvature and Total Variation Model 1 Deepak Kuar Gour, Sanjay Kuar Shara 1, Dept.

More information

A Beam Search Method to Solve the Problem of Assignment Cells to Switches in a Cellular Mobile Network

A Beam Search Method to Solve the Problem of Assignment Cells to Switches in a Cellular Mobile Network A Bea Search Method to Solve the Proble of Assignent Cells to Switches in a Cellular Mobile Networ Cassilda Maria Ribeiro Faculdade de Engenharia de Guaratinguetá - DMA UNESP - São Paulo State University

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor Willia Hoff Dept of Electrical Engineering &Coputer Science http://inside.ines.edu/~whoff/ 1 Caera Calibration 2 Caera Calibration Needed for ost achine vision and photograetry tasks (object

More information

CS 361 Meeting 8 9/24/18

CS 361 Meeting 8 9/24/18 CS 36 Meeting 8 9/4/8 Announceents. Hoework 3 due Friday. Review. The closure properties of regular languages provide a way to describe regular languages by building the out of sipler regular languages

More information

1 Extended Boolean Model

1 Extended Boolean Model 1 EXTENDED BOOLEAN MODEL It has been well-known that the Boolean odel is too inflexible, requiring skilful use of Boolean operators to obtain good results. On the other hand, the vector space odel is flexible

More information

Evaluation of a multi-frame blind deconvolution algorithm using Cramér-Rao bounds

Evaluation of a multi-frame blind deconvolution algorithm using Cramér-Rao bounds Evaluation of a ulti-frae blind deconvolution algorith using Craér-Rao bounds Charles C. Beckner, Jr. Air Force Research Laboratory, 3550 Aberdeen Ave SE, Kirtland AFB, New Mexico, USA 87117-5776 Charles

More information

ELEN : Project Progress Report. Second Order Delay Computation for RC networks with Non-Tree Topology

ELEN : Project Progress Report. Second Order Delay Computation for RC networks with Non-Tree Topology ELEN 689-603: Project Progress eport Second Order Delay Coputation for C networks wi Non-Tree Topology ajeshwary Tayade 000 07 573 //003 PDF created wi pdffactory trial version www.pdffactory.co : Introduction

More information

Geo-activity Recommendations by using Improved Feature Combination

Geo-activity Recommendations by using Improved Feature Combination Geo-activity Recoendations by using Iproved Feature Cobination Masoud Sattari Middle East Technical University Ankara, Turkey e76326@ceng.etu.edu.tr Murat Manguoglu Middle East Technical University Ankara,

More information

IMAGE MOSAICKING FOR ESTIMATING THE MOTION OF AN UNDERWATER VEHICLE. Rafael García, Xevi Cufí and Lluís Pacheco

IMAGE MOSAICKING FOR ESTIMATING THE MOTION OF AN UNDERWATER VEHICLE. Rafael García, Xevi Cufí and Lluís Pacheco IMAGE MOSAICKING FOR ESTIMATING THE MOTION OF AN UNDERWATER VEHICLE Rafael García, Xevi Cufí and Lluís Pacheco Coputer Vision and Robotics Group Institute of Inforatics and Applications, University of

More information

OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS

OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS Key words SOA, optial, coplex service, coposition, Quality of Service Piotr RYGIELSKI*, Paweł ŚWIĄTEK* OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS One of the ost iportant tasks in service oriented

More information

Leveraging Relevance Cues for Improved Spoken Document Retrieval

Leveraging Relevance Cues for Improved Spoken Document Retrieval Leveraging Relevance Cues for Iproved Spoken Docuent Retrieval Pei-Ning Chen 1, Kuan-Yu Chen 2 and Berlin Chen 1 National Taiwan Noral University, Taiwan 1 Institute of Inforation Science, Acadeia Sinica,

More information

Detection of Outliers and Reduction of their Undesirable Effects for Improving the Accuracy of K-means Clustering Algorithm

Detection of Outliers and Reduction of their Undesirable Effects for Improving the Accuracy of K-means Clustering Algorithm Detection of Outliers and Reduction of their Undesirable Effects for Iproving the Accuracy of K-eans Clustering Algorith Bahan Askari Departent of Coputer Science and Research Branch, Islaic Azad University,

More information

Compiling an Honest but Curious Protocol

Compiling an Honest but Curious Protocol 6.876/18.46: Advanced Cryptography May 7, 003 Lecture 1: Copiling an Honest but Curious Protocol Scribed by: Jonathan Derryberry 1 Review In previous lectures, the notion of secure ultiparty coputing was

More information

Mapping Data in Peer-to-Peer Systems: Semantics and Algorithmic Issues

Mapping Data in Peer-to-Peer Systems: Semantics and Algorithmic Issues Mapping Data in Peer-to-Peer Systes: Seantics and Algorithic Issues Anastasios Keentsietsidis Marcelo Arenas Renée J. Miller Departent of Coputer Science University of Toronto {tasos,arenas,iller}@cs.toronto.edu

More information

Supplementary Section. A. Algorithm. B. Datasets. C. More on Labeling Functions

Supplementary Section. A. Algorithm. B. Datasets. C. More on Labeling Functions Suppleentary Section In this section, we include ore details about our algorith, labeling functions, datasets as well as additional results and coparisons, which could not be included in the ain paper

More information

Brian Noguchi CS 229 (Fall 05) Project Final Writeup A Hierarchical Application of ICA-based Feature Extraction to Image Classification Brian Noguchi

Brian Noguchi CS 229 (Fall 05) Project Final Writeup A Hierarchical Application of ICA-based Feature Extraction to Image Classification Brian Noguchi A Hierarchical Application of ICA-based Feature Etraction to Iage Classification Introduction Iage classification poses one of the greatest challenges in the achine vision and achine learning counities.

More information

COMPUTER GENERATED HOLOGRAMS Optical Sciences 627 W.J. Dallas (Monday, August 23, 2004, 12:38 PM) PART III: CHAPTER ONE DIFFUSERS FOR CGH S

COMPUTER GENERATED HOLOGRAMS Optical Sciences 627 W.J. Dallas (Monday, August 23, 2004, 12:38 PM) PART III: CHAPTER ONE DIFFUSERS FOR CGH S COPUTER GEERATED HOLOGRAS Optical Sciences 67 W.J. Dallas (onday, August 3, 004, 1:38 P) PART III: CHAPTER OE DIFFUSERS FOR CGH S Part III: Chapter One Page 1 of 8 Introduction Hologras for display purposes

More information

6.1 Topological relations between two simple geometric objects

6.1 Topological relations between two simple geometric objects Chapter 5 proposed a spatial odel to represent the spatial extent of objects in urban areas. The purpose of the odel, as was clarified in Chapter 3, is ultifunctional, i.e. it has to be capable of supplying

More information

PROBABILISTIC LOCALIZATION AND MAPPING OF MOBILE ROBOTS IN INDOOR ENVIRONMENTS WITH A SINGLE LASER RANGE FINDER

PROBABILISTIC LOCALIZATION AND MAPPING OF MOBILE ROBOTS IN INDOOR ENVIRONMENTS WITH A SINGLE LASER RANGE FINDER nd International Congress of Mechanical Engineering (COBEM 3) Noveber 3-7, 3, Ribeirão Preto, SP, Brazil Copyright 3 by ABCM PROBABILISTIC LOCALIZATION AND MAPPING OF MOBILE ROBOTS IN INDOOR ENVIRONMENTS

More information

Summary. Reconstruction of data from non-uniformly spaced samples

Summary. Reconstruction of data from non-uniformly spaced samples Is there always extra bandwidth in non-unifor spatial sapling? Ralf Ferber* and Massiiliano Vassallo, WesternGeco London Technology Center; Jon-Fredrik Hopperstad and Ali Özbek, Schluberger Cabridge Research

More information

QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS

QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS Guofei Jiang and George Cybenko Institute for Security Technology Studies and Thayer School of Engineering Dartouth College, Hanover NH 03755

More information

MiPPS: A Generative Model for Multi-Manifold Clustering

MiPPS: A Generative Model for Multi-Manifold Clustering Manifold Learning and its Applications: Papers fro the AAAI Fall Syposiu (FS-9-) MiPPS: A Generative Model for Multi-Manifold Clustering Oluwasani Koyejo and Joydeep Ghosh Electrical and Coputer Engineering

More information

Reconstruction of Time Series using Optimal Ordering of ICA Components

Reconstruction of Time Series using Optimal Ordering of ICA Components Reconstruction of Tie Series using Optial Ordering of ICA Coponents Ar Goneid and Abear Kael Departent of Coputer Science & Engineering, The Aerican University in Cairo, Cairo, Egypt e-ail: goneid@aucegypt.edu

More information

An Efficient Approach for Content Delivery in Overlay Networks

An Efficient Approach for Content Delivery in Overlay Networks An Efficient Approach for Content Delivery in Overlay Networks Mohaad Malli, Chadi Barakat, Walid Dabbous Projet Planète, INRIA-Sophia Antipolis, France E-ail:{alli, cbarakat, dabbous}@sophia.inria.fr

More information

Investigation of The Time-Offset-Based QoS Support with Optical Burst Switching in WDM Networks

Investigation of The Time-Offset-Based QoS Support with Optical Burst Switching in WDM Networks Investigation of The Tie-Offset-Based QoS Support with Optical Burst Switching in WDM Networks Pingyi Fan, Chongxi Feng,Yichao Wang, Ning Ge State Key Laboratory on Microwave and Digital Counications,

More information

MAPPING THE DATA FLOW MODEL OF COMPUTATION INTO AN ENHANCED VON NEUMANN PROCESSOR * Peter M. Maurer

MAPPING THE DATA FLOW MODEL OF COMPUTATION INTO AN ENHANCED VON NEUMANN PROCESSOR * Peter M. Maurer MAPPING THE DATA FLOW MODEL OF COMPUTATION INTO AN ENHANCED VON NEUMANN PROCESSOR * Peter M. Maurer Departent of Coputer Science and Engineering University of South Florida Tapa, FL 33620 Abstract -- The

More information

Trees. Linear vs. Branching CSE 143. Branching Structures in CS. What s in a Node? A Tree. [Chapter 10]

Trees. Linear vs. Branching CSE 143. Branching Structures in CS. What s in a Node? A Tree. [Chapter 10] CSE 143 Trees [Chapter 10] Linear vs. Branching Our data structures so far are linear Have a beginning and an end Everything falls in order between the ends Arrays, lined lists, queues, stacs, priority

More information

I ve Seen Enough : Incrementally Improving Visualizations to Support Rapid Decision Making

I ve Seen Enough : Incrementally Improving Visualizations to Support Rapid Decision Making I ve Seen Enough : Increentally Iproving Visualizations to Support Rapid Decision Making Sajjadur Rahan Marya Aliakbarpour 2 Ha Kyung Kong Eric Blais 3 Karrie Karahalios,4 Aditya Paraeswaran Ronitt Rubinfield

More information

Super-Resolution on Moving Objects using a Polygon-Based Object Description

Super-Resolution on Moving Objects using a Polygon-Based Object Description Super-Resolution on Moving Objects using a Polgon-Based Object Description A. van Eekeren K. Schutte O.R. Oudegeest L.J. van Vliet TNO Defence, Securit and Safet, P.O. Box 96864, 2509 JG, The Hague, The

More information

Medical Biophysics 302E/335G/ st1-07 page 1

Medical Biophysics 302E/335G/ st1-07 page 1 Medical Biophysics 302E/335G/500 20070109 st1-07 page 1 STEREOLOGICAL METHODS - CONCEPTS Upon copletion of this lesson, the student should be able to: -define the ter stereology -distinguish between quantitative

More information

A Trajectory Splitting Model for Efficient Spatio-Temporal Indexing

A Trajectory Splitting Model for Efficient Spatio-Temporal Indexing A Trajectory Splitting Model for Efficient Spatio-Teporal Indexing Slobodan Rasetic Jörg Sander Jaes Elding Mario A. Nasciento Departent of Coputing Science University of Alberta Edonton, Alberta, Canada

More information

ELEVATION SURFACE INTERPOLATION OF POINT DATA USING DIFFERENT TECHNIQUES A GIS APPROACH

ELEVATION SURFACE INTERPOLATION OF POINT DATA USING DIFFERENT TECHNIQUES A GIS APPROACH ELEVATION SURFACE INTERPOLATION OF POINT DATA USING DIFFERENT TECHNIQUES A GIS APPROACH Kulapraote Prathuchai Geoinforatics Center, Asian Institute of Technology, 58 Moo9, Klong Luang, Pathuthani, Thailand.

More information

A Measurement-Based Model for Parallel Real-Time Tasks

A Measurement-Based Model for Parallel Real-Time Tasks A Measureent-Based Model for Parallel Real-Tie Tasks Kunal Agrawal 1 Washington University in St. Louis St. Louis, MO, USA kunal@wustl.edu https://orcid.org/0000-0001-5882-6647 Sanjoy Baruah 2 Washington

More information

Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA February 9, 2017 1 / 24 OUTLINE 1 Introduction Keras: Deep Learning library for Theano and TensorFlow 2 Installing Keras Installation

More information

Rule Extraction using Artificial Neural Networks

Rule Extraction using Artificial Neural Networks Rule Extraction using Artificial Neural Networks S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract Artificial neural networks have been successfully applied to a variety of business application probles involving

More information

Smarter Balanced Assessment Consortium Claims, Targets, and Standard Alignment for Math

Smarter Balanced Assessment Consortium Claims, Targets, and Standard Alignment for Math Sarter Balanced Assessent Consortiu s, s, Stard Alignent for Math The Sarter Balanced Assessent Consortiu (SBAC) has created a hierarchy coprised of clais targets that together can be used to ake stateents

More information

Relief shape inheritance and graphical editor for the landscape design

Relief shape inheritance and graphical editor for the landscape design Relief shape inheritance and graphical editor for the landscape design Egor A. Yusov Vadi E. Turlapov Nizhny Novgorod State University after N. I. Lobachevsky Nizhny Novgorod Russia yusov_egor@ail.ru vadi.turlapov@cs.vk.unn.ru

More information

Verdict Accuracy of Quick Reduct Algorithm using Clustering, Classification Techniques for Gene Expression Data

Verdict Accuracy of Quick Reduct Algorithm using Clustering, Classification Techniques for Gene Expression Data Verdict Accuracy of Quick Reduct Algorith using Clustering, Classification Techniques for Gene Expression Data T.Chandrasekhar 1, K.Thangavel 2 and E.N.Sathishkuar 3 1 Departent of Coputer Science, eriyar

More information

A Hybrid Network Architecture for File Transfers

A Hybrid Network Architecture for File Transfers JOURNAL OF IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, VOL. 6, NO., JANUARY 9 A Hybrid Network Architecture for File Transfers Xiuduan Fang, Meber, IEEE, Malathi Veeraraghavan, Senior Meber,

More information

Geometry. The Method of the Center of Mass (mass points): Solving problems using the Law of Lever (mass points). Menelaus theorem. Pappus theorem.

Geometry. The Method of the Center of Mass (mass points): Solving problems using the Law of Lever (mass points). Menelaus theorem. Pappus theorem. Noveber 13, 2016 Geoetry. The Method of the enter of Mass (ass points): Solving probles using the Law of Lever (ass points). Menelaus theore. Pappus theore. M d Theore (Law of Lever). Masses (weights)

More information

GAN-OPC: Mask Optimization with Lithography-guided Generative Adversarial Nets

GAN-OPC: Mask Optimization with Lithography-guided Generative Adversarial Nets GAN-OPC: Mask Optiization with Lithography-guided Generative Adversarial Nets Haoyu Yang hyyang@cse.cuhk.edu.hk Shuhe Li shli@cse.cuhk.edu.hk Yuzhe Ma yza@cse.cuhk.edu.hk ABSTRACT Bei Yu byu@cse.cuhk.edu.hk

More information

Effects of Desingularization and Collocation-Point Shift on Steady Waves with Forward Speed

Effects of Desingularization and Collocation-Point Shift on Steady Waves with Forward Speed Effects of Desingularization and Collocation-Point Shift on Steady Waves with Forward Speed Yonghwan Ki* & Dick K.P. Yue** Massachusetts Institute of Technology, Departent of Ocean Engineering, Cabridge,

More information

AN APPROACH ON BIMODAL BIOMETRIC SYSTEMS

AN APPROACH ON BIMODAL BIOMETRIC SYSTEMS AN APPROACH ON BIODAL BIOETRIC SYSTES Eugen LUPU, Siina EERICH Technical University of Cluj-Napoca, 26-28 Baritiu str. Cluj-Napoca phone: +40-264-40-266; fax: +40-264-592-055; e-ail: Eugen.Lupu @co.utcluj.ro

More information

Efficient Estimation of Inclusion Coefficient using HyperLogLog Sketches

Efficient Estimation of Inclusion Coefficient using HyperLogLog Sketches Efficient Estiation of Inclusion Coefficient using HyperLogLog Sketches Azade Nazi, Bolin Ding, Vivek Narasayya, Surajit Chaudhuri Microsoft Research {aznazi, bolind, viveknar, surajitc}@icrosoft.co ABSTRACT

More information

LookNN: Neural Network with No Multiplication

LookNN: Neural Network with No Multiplication LookNN: Neural Network with No Multiplication Mohaad Saragh Razlighi, Mohsen Iani, Farinaz Koushanfar and Tajana Rosing ECE Departent, CSE Departent, UC San Diego, La Jolla, CA 993, USA {saragh, oiani,

More information

cm3520 cm3525 Security Function

cm3520 cm3525 Security Function wwwiagisticsco c3520 c3525 Security Function Contents Contents 1 Security 11 Introduction 1-2 12 Tradearks and Registered Tradearks 1-2 13 Copliance with the ISO15408 Standard 1-2 14 Operating Precautions1-2

More information

Automatic Graph Drawing Algorithms

Automatic Graph Drawing Algorithms Autoatic Graph Drawing Algoriths Susan Si sisuz@turing.utoronto.ca Deceber 7, 996. Ebeddings of graphs have been of interest to theoreticians for soe tie, in particular those of planar graphs and graphs

More information

Implementation of fast motion estimation algorithms and comparison with full search method in H.264

Implementation of fast motion estimation algorithms and comparison with full search method in H.264 IJCSNS International Journal of Coputer Science and Network Security, VOL.8 No.3, March 2008 139 Ipleentation of fast otion estiation algoriths and coparison with full search ethod in H.264 A.Ahadi, M.M.Azadfar

More information

Video summarization based on shot boundary detection with penalized contrasts

Video summarization based on shot boundary detection with penalized contrasts Video suarization based on shot boundary detection with penalized contrasts Paschalina Medentzidou and Constantine Kotropoulos Departent of Inforatics Aristotle University of Thessaloniki Thessaloniki

More information

Heterogeneous Radial Basis Function Networks

Heterogeneous Radial Basis Function Networks Proceedings of the International Conference on Neural Networks (ICNN ), vol. 2, pp. 23-2, June. Heterogeneous Radial Basis Function Networks D. Randall Wilson, Tony R. Martinez e-ail: randy@axon.cs.byu.edu,

More information

Modal Masses Estimation in OMA by a Consecutive Mass Change Method.

Modal Masses Estimation in OMA by a Consecutive Mass Change Method. Modal Masses Estiation in OMA by a Consecutive Mass Change Method. F. Pelayo University of Oviedo, Departent of Construction and Manufacturing Engineering, Gijón, Spain M. López-Aenlle University of Oviedo,

More information

Cassia County School District #151. Expected Performance Assessment Students will: Instructional Strategies. Performance Standards

Cassia County School District #151. Expected Performance Assessment Students will: Instructional Strategies. Performance Standards Unit 1 Congruence, Proof, and Constructions Doain: Congruence (CO) Essential Question: How do properties of congruence help define and prove geoetric relationships? Matheatical Practices: 1. Make sense

More information

2013 IEEE Conference on Computer Vision and Pattern Recognition. Compressed Hashing

2013 IEEE Conference on Computer Vision and Pattern Recognition. Compressed Hashing 203 IEEE Conference on Coputer Vision and Pattern Recognition Copressed Hashing Yue Lin Rong Jin Deng Cai Shuicheng Yan Xuelong Li State Key Lab of CAD&CG, College of Coputer Science, Zhejiang University,

More information

Optimized stereo reconstruction of free-form space curves based on a nonuniform rational B-spline model

Optimized stereo reconstruction of free-form space curves based on a nonuniform rational B-spline model 1746 J. Opt. Soc. A. A/ Vol. 22, No. 9/ Septeber 2005 Y. J. Xiao and Y. F. Li Optiized stereo reconstruction of free-for space curves based on a nonunifor rational B-spline odel Yi Jun Xiao Departent of

More information

QoS and Sensible Routing Decisions

QoS and Sensible Routing Decisions QoS and Sensible Routing Decisions Erol Gelenbe Dept. of Electrical & Electronic Engineering Iperial College London SW7 2BT e.gelenbe@iperial.ac.uk Abstract Network Quality of Service (QoS) criteria of

More information

Resolution. Super-Resolution Imaging. Problem

Resolution. Super-Resolution Imaging. Problem Resolution Super-Resolution Iaging Resolution: Sallest easurable detail in a visual presentation Subhasis Chaudhuri Departent of Electrical Engineering Indian institute of Technology Bobay Powai, Mubai-400

More information

Efficient Constraint Evaluation Algorithms for Hierarchical Next-Best-View Planning

Efficient Constraint Evaluation Algorithms for Hierarchical Next-Best-View Planning Efficient Constraint Evaluation Algoriths for Hierarchical Next-Best-View Planning Ko-Li Low National University of Singapore lowl@cop.nus.edu.sg Anselo Lastra University of North Carolina at Chapel Hill

More information

Derivation of an Analytical Model for Evaluating the Performance of a Multi- Queue Nodes Network Router

Derivation of an Analytical Model for Evaluating the Performance of a Multi- Queue Nodes Network Router Derivation of an Analytical Model for Evaluating the Perforance of a Multi- Queue Nodes Network Router 1 Hussein Al-Bahadili, 1 Jafar Ababneh, and 2 Fadi Thabtah 1 Coputer Inforation Systes Faculty of

More information

Smarter Balanced Assessment Consortium Claims, Targets, and Standard Alignment for Math

Smarter Balanced Assessment Consortium Claims, Targets, and Standard Alignment for Math Sarter Balanced Assessent Consortiu s, s, Stard Alignent for Math The Sarter Balanced Assessent Consortiu (SBAC) has created a hierarchy coprised of clais targets that together can be used to ake stateents

More information

M a c intosh Cyr i l lic Lang u age Ki t. Installation and User s Manual Manuel d installation et d u t i l i s a t i o n

M a c intosh Cyr i l lic Lang u age Ki t. Installation and User s Manual Manuel d installation et d u t i l i s a t i o n apple M a c intosh Cyr i l lic Lang u age Ki t Installation and User s Manual Manuel d installation et d u t i l i s a t i o n K Apple Coputer, Inc. This anual and the software described in it are copyrighted

More information

The optimization design of microphone array layout for wideband noise sources

The optimization design of microphone array layout for wideband noise sources PROCEEDINGS of the 22 nd International Congress on Acoustics Acoustic Array Systes: Paper ICA2016-903 The optiization design of icrophone array layout for wideband noise sources Pengxiao Teng (a), Jun

More information

The Henryk Niewodniczański INSTITUTE OF NUCLEAR PHYSICS Polish Academy of Sciences ul. Radzikowskiego 152, Kraków

The Henryk Niewodniczański INSTITUTE OF NUCLEAR PHYSICS Polish Academy of Sciences ul. Radzikowskiego 152, Kraków The Henryk Niewodniczański INSTITUT OF NUCLAR PHYSICS Polish Acadey of Sciences ul. Radzikowskiego 152, 31-342 Kraków www.ifj.edu.pl/reports/2005.htl Kraków, October 2005 Report No. 1968/D The theroluinescence

More information

A Broadband Spectrum Sensing Algorithm in TDCS Based on ICoSaMP Reconstruction

A Broadband Spectrum Sensing Algorithm in TDCS Based on ICoSaMP Reconstruction MATEC Web of Conferences 73, 03073(08) https://doi.org/0.05/atecconf/087303073 SMIMA 08 A roadband Spectru Sensing Algorith in TDCS ased on I Reconstruction Liu Yang, Ren Qinghua, Xu ingzheng and Li Xiazhao

More information

Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization

Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization 1 Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization Jonathan van de Belt, Haed Ahadi, and Linda E. Doyle The Centre for Future Networks and Counications - CONNECT,

More information

Fair Resource Allocation for Heterogeneous Tasks

Fair Resource Allocation for Heterogeneous Tasks Fair Resource Allocation for Heterogeneous Tasks Koyel Mukherjee, Partha utta, Gurulingesh Raravi, Thangaraj Balasubraania, Koustuv asgupta, Atul Singh Xerox Research Center India, Bangalore, India 560105

More information

Feature Based Registration for Panoramic Image Generation

Feature Based Registration for Panoramic Image Generation IJCSI International Journal of Coputer Science Issues, Vol. 10, Issue 6, No, Noveber 013 www.ijcsi.org 13 Feature Based Registration for Panoraic Iage Generation Kawther Abbas Sallal 1, Abdul-Mone Saleh

More information

A New Generic Model for Vision Based Tracking in Robotics Systems

A New Generic Model for Vision Based Tracking in Robotics Systems A New Generic Model for Vision Based Tracking in Robotics Systes Yanfei Liu, Ada Hoover, Ian Walker, Ben Judy, Mathew Joseph and Charly Heranson lectrical and Coputer ngineering Departent Cleson University

More information

The Internal Conflict of a Belief Function

The Internal Conflict of a Belief Function The Internal Conflict of a Belief Function Johan Schubert Abstract In this paper we define and derive an internal conflict of a belief function We decopose the belief function in question into a set of

More information

The Boundary Between Privacy and Utility in Data Publishing

The Boundary Between Privacy and Utility in Data Publishing The Boundary Between Privacy and Utility in Data Publishing Vibhor Rastogi Dan Suciu Sungho Hong ABSTRACT We consider the privacy proble in data publishing: given a database instance containing sensitive

More information

Feature Selection to Relate Words and Images

Feature Selection to Relate Words and Images The Open Inforation Systes Journal, 2009, 3, 9-13 9 Feature Selection to Relate Words and Iages Wei-Chao Lin 1 and Chih-Fong Tsai*,2 Open Access 1 Departent of Coputing, Engineering and Technology, University

More information

Preprocessing of fmri data (basic)

Preprocessing of fmri data (basic) Preprocessing of fmri data (basic) Practical session SPM Course 2016, Zurich Andreea Diaconescu, Maya Schneebeli, Jakob Heinzle, Lars Kasper, and Jakob Sieerkus Translational Neuroodeling Unit (TNU) Institute

More information

MIDA: AN IDA SEARCH WITH DYNAMIC CONTROL

MIDA: AN IDA SEARCH WITH DYNAMIC CONTROL April 1991 UILU-ENG-91-2216 CRHC-91-9 Center fo r Reliable and High-Perforance Coputing MIDA: AN IDA SEARCH WITH DYNAMIC CONTROL Benjain W. Wah Coordinated Science Laboratory College of Engineering UNIVERSITY

More information

A High-Speed VLSI Fuzzy Inference Processor for Trapezoid-Shaped Membership Functions *

A High-Speed VLSI Fuzzy Inference Processor for Trapezoid-Shaped Membership Functions * JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 21, 607-626 (2005) A High-Speed VLSI Fuzzy Inference Processor for Trapezoid-Shaped Mebership Functions * SHIH-HSU HUANG AND JIAN-YUAN LAI + Departent of

More information

HIGH PERFORMANCE PRE-SEGMENTATION ALGORITHM FOR SONAR IMAGES

HIGH PERFORMANCE PRE-SEGMENTATION ALGORITHM FOR SONAR IMAGES HIGH PERFORMANCE PRE-SEGMENTATION ALGORITHM FOR SONAR IMAGES Benjain Lehann*, Konstantinos Siantidis*, Dieter Kraus** *ATLAS ELEKTRONIK GbH Sebaldsbrücker Heerstraße 235 D-28309 Breen, GERMANY Eail: benjain.lehann@atlas-elektronik.co

More information

Impact of Network Delay Variations on Multicast Sessions with TCP-like Congestion Control

Impact of Network Delay Variations on Multicast Sessions with TCP-like Congestion Control Ipact of Network Delay Variations on Multicast Sessions with TCP-like Congestion Control Augustin Chaintreau y François Baccelli z Christophe Diot yy y Ecole Norale Supérieure 45 rue d Ul 75005 Paris FRANCE

More information

KS3 Maths Progress Delta 2-year Scheme of Work

KS3 Maths Progress Delta 2-year Scheme of Work KS3 Maths Progress Delta 2-year Schee of Work See the spreadsheet for the detail. Essentially this is an express route through our higher-ability KS3 Maths Progress student books Delta 1, 2 and 3 as follows:

More information

Image Processing for fmri John Ashburner. Wellcome Trust Centre for Neuroimaging, 12 Queen Square, London, UK.

Image Processing for fmri John Ashburner. Wellcome Trust Centre for Neuroimaging, 12 Queen Square, London, UK. Iage Processing for fmri John Ashburner Wellcoe Trust Centre for Neuroiaging, 12 Queen Square, London, UK. Contents * Preliinaries * Rigid-Body and Affine Transforations * Optiisation and Objective Functions

More information

Sports Video Classification from Multimodal Information Using Deep Neural Networks

Sports Video Classification from Multimodal Information Using Deep Neural Networks Ho Should Intelligence Be Abstracted in AI Research... AAAI echnical Report FS-3-0 Sports Video Classification fro Multiodal Inforation Using Deep Neural Netors Devendra Singh Sachan, Uesh ean Ait Sethi

More information

Design and Implementation of Business Logic Layer Object-Oriented Design versus Relational Design

Design and Implementation of Business Logic Layer Object-Oriented Design versus Relational Design Design and Ipleentation of Business Logic Layer Object-Oriented Design versus Relational Design Ali Alharthy Faculty of Engineering and IT University of Technology, Sydney Sydney, Australia Eail: Ali.a.alharthy@student.uts.edu.au

More information

COMPARISON OF ANT COLONY OPTIMIZATION & PARTICLE SWARM OPTIMIZATION IN GRID ENVIRONMENT

COMPARISON OF ANT COLONY OPTIMIZATION & PARTICLE SWARM OPTIMIZATION IN GRID ENVIRONMENT COMPARISON OF ANT COLONY OPTIMIZATION & PARTICLE SWARM OPTIMIZATION IN GRID ENVIRONMENT B.BOOBA 1 Dr. T.V.GOPAL 2 1 Research Scholar, Assist. Professor(Sl.Gr),Departent of Coputer Applications, Easwari

More information

POSITION-PATCH BASED FACE HALLUCINATION VIA LOCALITY-CONSTRAINED REPRESENTATION. Junjun Jiang, Ruimin Hu, Zhen Han, Tao Lu, and Kebin Huang

POSITION-PATCH BASED FACE HALLUCINATION VIA LOCALITY-CONSTRAINED REPRESENTATION. Junjun Jiang, Ruimin Hu, Zhen Han, Tao Lu, and Kebin Huang IEEE International Conference on ultiedia and Expo POSITION-PATCH BASED FACE HALLUCINATION VIA LOCALITY-CONSTRAINED REPRESENTATION Junjun Jiang, Ruiin Hu, Zhen Han, Tao Lu, and Kebin Huang National Engineering

More information

Mirror Localization for a Catadioptric Imaging System by Projecting Parallel Lights

Mirror Localization for a Catadioptric Imaging System by Projecting Parallel Lights 2007 IEEE International Conference on Robotics and Autoation Roa, Italy, 10-14 April 2007 FrC2.5 Mirror Localization for a Catadioptric Iaging Syste by Projecting Parallel Lights Ryusuke Sagawa, Nobuya

More information

Real-Time Detection of Invisible Spreaders

Real-Time Detection of Invisible Spreaders Real-Tie Detection of Invisible Spreaders MyungKeun Yoon Shigang Chen Departent of Coputer & Inforation Science & Engineering University of Florida, Gainesville, FL 3, USA {yoon, sgchen}@cise.ufl.edu Abstract

More information

LOSSLESS COMPRESSION OF BAYER MASK IMAGES USING AN OPTIMAL VECTOR PREDICTION TECHNIQUE

LOSSLESS COMPRESSION OF BAYER MASK IMAGES USING AN OPTIMAL VECTOR PREDICTION TECHNIQUE 1th European Signal Processing Conference (EUSIPCO 2006), Florence, Italy, Septeber -8, 2006, copyright by EUASIP LOSSLESS COMPESSION OF AYE MASK IMAES USIN AN OPTIMAL VECTO PEDICTION TECHNIQUE Stefano

More information