Robot behaviour learning using Topological Gaussian Adaptive Resonance Hidden Markov Model

Size: px
Start display at page:

Download "Robot behaviour learning using Topological Gaussian Adaptive Resonance Hidden Markov Model"

Transcription

1 DOI /s x ORIGINAL ARTICLE Robot behaviour learning using Topological Gaussian Adaptive Resonance Hidden Markov Model Farhan Dawood 1 Chu Kiong Loo 2 Received: 14 January 2015 / Accepted: 9 August 2015 The Natural Computing Applications Forum 2015 Abstract Behaviour learning by robots is an emerging research field. The robot behaviour learning in realistic environment requires the robot to be able to learn incrementally in order to adapt itself to the dynamic environment. The incremental learning system should be stable and acquire new information without corrupting previous knowledge. Probabilistic approach has become the prevalent epitome in robotic learning systems. In probabilistic learning models, such as Hidden Markov Model, estimating the optimal number of states is difficult which limits the models capabilities by limiting the number of motion patterns to be learned a priori. In this paper, we presented an approach for learning behaviour patterns through continuous observation. We have proposed a novel architecture for learning the spatio-temporal sequences using Topological Gaussian Adaptive Resonance Hidden Markov Model. The proposed model dynamically generates the graph-based structure for the observed patterns through a novel topological mapping architecture. The structure (number of states) of the probabilistic model is not fixed and is adjusted based on the topological map of the acquired motion features. The topological map consists of nodes connected with edges, where each node represents the encoded motion features to adaptively generalize the observed behaviour patterns. The model combines the self-organizing, self-stabilizing & Chu Kiong Loo ckloo.um@gmail.com; ckloo.um@um.edu.my Farhan Dawood farhan.dawood@gmail.com 1 2 Advanced Robotics Lab, University of Malaya, Lembah Pantai, Kuala Lumpur, Malaysia Department of Artificial Intelligence, University of Malaya, Lembah Pantai, Kuala Lumpur, Malaysia properties and time series processing features to learn the spatio-temporal sequences. To demonstrate the properties of the proposed algorithm, we have performed a set of experiments through simulation on DARwIn-OP humanoid robot. Keywords Behaviour learning Adaptive resonance theory Gaussian distribution Hidden Markov Model Incremental learning Topological map 1 Introduction Due to the recent developments in robotics, the robots are able to move and act in human-centred environment, taking part in our daily lives. This has introduced the need for building robotic systems equipped with behaviour learning capabilities. Motivated by the fact that when a robot is in an environment, some novel events may be encountered by the robot which were not learned previously. This will make the robots ineffective for such dynamic environments. Therefore, the robot must be capable of updating these changes so that it can adapt itself efficaciously to novel situations encountered during learning. A natural way of representing these behaviours by the robot is through the use of probabilistic models. Hidden Markov Models (HMM) [1], originally developed for speech recognition and synthesis, have been widely used for human motion recognition and generation. HMM takes sequential data as input and generates the probabilities. The use of HMM is ubiquitous in time series signal processing. The ability of HMMs to generalize human actions has led to the development of several methods for HMM-based movement recognition and generation [2, 3]. However, most of these methods operate offline where the model structure is static and determined beforehand. In offline or

2 batch learning method, the motion patterns are processed sequentially. Such interaction might seem artificial, since the patterns are pre-defined by the designer and cannot accommodate the new properties of various tasks. Therefore, the stochastic model should allow new information to be incorporated incrementally. One major dilemma associated with incremental learning and memorizing the observed behavioural actions into its repertoire is the stability and plasticity dilemma [4], i.e. how can a system retain old memories but also learn new ones. For life-long learning in dynamic environment, the system must be able to learn to adapt to the changing situations by observing new patterns and keep the previous knowledge in a stable manner. Another issue associated with the well-established learning algorithms based on stochastic models is in estimating the structure of the model. Setting up appropriate number of states for the model is difficult because the number of motion patterns in the data is unknown. If there is less number of states than the observed trajectory components, the model cannot explain patterns considerably. On contrary, if the model states are large in number, the system will require too much training data and excessive computational requirement. Thus, the major question is to find a suitable model structure that would best approximate the given environment without loss of information. In the literature, the structure of HMM, i.e. the number of HMM states, is either to be chosen manually or to be estimated using other methods such as Bayesian information criterion (BIC) [5] or Akaike information criterion (AIC) [6]. This limits the number of motion patterns to be learned a priori and results in a trade-off between model fitness and number of parameters. In this paper, we are proposing a novel probabilistic model for encoding spatio-temporal motion patterns to develop a self-stabilizing and self-scaling architecture. The major contribution of our work is in developing a novel incremental learning architecture using an adaptive structure to solve the problems of: (1) How to select the structure for the probabilistic model, i.e. estimating the number of states to efficiently encode the sensory data without restricting the learning capabilities of the robot? (2) How the model can retain previously learned data and also acquire new observed knowledge in a self-organizing and self-adaptive manner? (3) How the robot can interact with the environment incrementally so that when new pattern is observed the model should incorporate it? For this purpose, we have proposed a novel Topological Gaussian Adaptive Resonance Hidden Markov Model (TGAR-HMM). In contrast to conventional HMM, the developed probabilistic learning algorithm is based on incrementally learning spatio-temporal behavioural sequences by developing the graph-based structure of the behaviour patterns in the form of topological map using Topological Gaussian Adaptive Resonance Map (TGARM). The topological model incrementally updates the number of states and parameters required by the probabilistic model to encode the observed motion elements. This compactly describes the environment as a collection of nodes linked by edges. TGARM creates a stable representation while incorporating new knowledge. After organization of the data, the temporal sequence is learned through the TGAR-HMM, where each node represents a state of the probabilistic model while an edge between two nodes represents the transition between these states. This state-space distribution allows transitions between neighbouring states only. The structure of the model and its parameters representing the probabilities are updated with the acquisition of new observation. Figure 1 shows the graphical representation of the proposed architecture. In order to evaluate the performance of our proposed architecture, we have tested our algorithm on two different types of datasets and analysed its performance. These tests were simulated on DARwIn-OP humanoid robot. The actions involved during testing range from simple behavioural gestures to complex action of kicking the ball. The rest of the paper is organized as follows: The next section reviews the related work. Later, in Sect. 3, we will discuss the proposed novel approach for motion patterns learning through TGAR-HMM. The experimental results and analysis of the data are described in Sect. 4. Finally, in Sect. 5, we will conclude our work with some discussion on the proposed model. 2 Related work Adaptive resonance theory (ART) models the cognitive and neural theory of how brain independently learns to categorize and recognize the events in the dynamic environment [7, 8]. The prominent features of this family of neural networks led the engineers to design various models such as ART-1 [9] (for binary input patterns), ART-2 [10] (for analog and binary input patterns) and Fuzzy ART [11] (combination of fuzzy logic and ART). However, learning complex spatio-temporal sequences through ART networks is still under development. To learn the temporal orders of sequential events, Bradski et al. [12] presented a working memory model called Sustained Temporal Order REcurrent (STORE) model. This architecture makes use of the ART network to recognize the temporal event sequences. Growing neural gas (GNG) is well known as unsupervised learning method. Furao & Hasegawa [13, 14] suggest several enhancements to the unsupervised variant of GNG and particularly focusing on the life-long learning of nonstationary data for solving the issues like the topology learning of images and clustering. They applied a two-

3 Fig. 1 Overview of the TGAR-HMM architecture. The observed behaviour sequence is first arranged through topological map. This topological map is then used to update the state structure for estimating the optimal number of states and the transition probabilities among these states layered network, where the first layer is used to generate a topology structure of the input data and the second layer is used to determine the number of clusters. Unfortunately, such neural networks do not work with spatio-temporal data such as speech or motion patterns. The high requirement for node insertion of the SOINN and use of the utility parameter restrains topographical formation and organization of the input data to the SOINN network. Botzheim et al. [15] presented a structured learning approach by combining GNG and Spiking Neural Network (SNN), suitable for spatio-temporal learning. The approach contains two stages, a topology learning phase and a spatiotemporal learning phase. The growing neural gas (GNG) is applied in the first phase for information extraction, and the spiking neural network (SNN) is applied in the second stage to recognize the gesture. In the second step, spiking neural network is applied to understand the spatiotemporal information and estimate the activity. Seyhan et al. [16] developed a behaviour learning model for simple and complex actions in robot using HMM and a correlation-based ART (CobART) network. CobART is a ART 2 type of network. The motion primitives acquired from the CobART are modelled through HMM to represent a relation between these motion primitives. The model generates different categories for the same behaviours but with slight variation between them, thus providing a correlation between the motion patterns. A similar model was presented by [17] through hierarchically integrating CobART networks. The model learns the spatio-temporal sequences; however, the structure of the behaviour HMM proposed by them is fixed and cannot grow incrementally. Kulic et al. [6] proposed an approach for online learning of motion primitives while observing and interaction with human collaborator. The method is based on online segmentation of the observed motion patterns for generating motion primitives, and incremental learning algorithm is used to provide the relationship between motion primitives for motion generation process. A hierarchical tree structure is incrementally formed representing the motion learned by the robot. Each node in the tree represents a motion primitive. The observed behaviours are encoded in the HMM and then clustered based on intra-modal distance measured using the relative log likelihood. In addition to the tree structure, a directed graph model is built representing the sequential relationship between the motion primitives which is then used for motion generation [18, 19]. A similar approach has been used by Kulic et al. [20] using Factorial HMM. FHMM is a generalization of HMM where

4 multiple HMM chains interact to generate single output. Each dynamic chain models the observed motion pattern with its own transition and output probability parameters. At each time step, the output generated by each chain is summed to generate the optimized observed output. A symbol space is constructed to discriminate between behaviour patterns. In all these methods, the number of HMM models encoding the motion patterns is not fixed; however, the number of states in each HMM model is pre-defined. Calinon et al. [21] presented two incremental learning approaches: direct and generative method. These methods are used to update the models parameters when novel skills are demonstrated. The probabilistic model is first learned using joint angles trajectories observed using motion sensors and is progressively refined using kinaesthetic process. For direct update method, the idea is to adapt the EM algorithm by separating the parts dedicated to the data already used to train the model and the one dedicated to the newly available data. The model is first created with data points and updated iteratively during EM step, until convergence to the set of parameters. When new data are available, regression is used to generate stochastically new data by considering the current GMR model. Cho et al. [22] describe an incremental learning approach for learning of behaviours by teaching a humanoid robot kinaesthetically by manually moving robots joints. The motion patterns are encoded in GMM described in a latent space with reduced dimensionality computed using PCA. When a new behaviour is taught, the latent space is modified and GMM is updated. It is assumed that complete data are not available, and therefore, an approximation of BIC is utilized to estimate the GMM parameters. The generalized motion patterns are generated using Gaussian Mixture Regression (GMR). New motions are generated through the fusion of two previously learned motion patterns using GMR. Similarly, new motion can also be generated combining motions spatially or from different time intervals [23]. Vasquez et al. [24] have developed an incremental model for learning and predicting through growing HMM. The model updates the structure of HMM through the use of topological mapping algorithm called Instantaneous Topological Mapping (ITM). The model was designed for vehicle motion learning and prediction. Similar to our approach, the growing HMM architecture uses the topological mapping structure to define the HMM structure. In an earlier version of the algorithm, they have employed growing neural gas (GNG). [25] developed a topological structure for unsupervised learning using Fuzzy ART. Using TopoART algorithm, a stable representation of the data is created. The model was only used for clustering and is not suitable for spatio-temporal learning. Fung et al. [26] developed an ART network-based taskindependent robot behaviour learning mechanism. They called their model the Behaviour Learning Operating Modular (BLOM) Architecture. In this architecture, the vigilance parameter of ART network is updated through game-theoretic adaptation strategy. The perceptual sensory inputs and actions are categorized separately through the ART network. The association between these two category types is established through the associative memories. The BLOM architecture was tested on simple path planning experiments; however, this architecture is not suitable for learning spatio-temporal behavioural sequences. [27] proposed an adaptive neural architecture for robot behaviour learning which dynamically grows when required (GWR). The GWR model clusters the sensory-motor data through the radial basis function (RBF) network. 3 Topological Gaussian Adaptive Resonance Hidden Markov Model In this section, we will explain the proposed incremental spatio-temporal sequence learning algorithm. The motion features (joint angle values) acquired from robot sensors are first organized through topological map using TGARM algorithm. This topological structure is then used to update and estimate the structure of probabilistic model to encode acquired features. Structurally, TGAR-HMM is similar to the standard HMMs, besides the fact that the transition structure and the number of states are not constant but vary as more input observation sequences are processed. In addition to that, the learning algorithm is able to incrementally update the model. In our proposed architecture, an HMM can be considered as graph whose nodes represent states attainable by the object and whose edges represent transitions between these states. The system is assumed to be at a particular state and to develop stochastically at discrete time steps by following the graph edges according to a transition probability. TGAR- HMM is described as a time evolving HMM with continuous observation variables, where the number of HMM states, structure and probability parameters are updated every time that a new observation sequence is available. The main idea for developing the proposed probabilistic model is that the structure of the model should consider the spatial structure of the state-space discretization, where the transition among discrete states are only permitted if the corresponding regions are neighbours. Hence, structure learning essentially consists of estimating the suitable space discretization from the observed data and identifying neighbouring regions. We have addressed this problem by proposing a topological map using the Topological Gaussian Adaptive Resonance Map (TGARM) algorithm. For parameter learning, we have utilized the incremental expectation maximization (EM) approach to handle the changing number of states and with continuous observation.

5 3.1 Motion primitive modelling topological Gaussian adaptive resonance map In this phase, we extract useful motion sequences which are represented as the nodes linked with each other through edges. Through topological map, the continuous environment is represented as a state-space distribution organized as a graph where each node is connected through the edges. The function of the topological map is to develop a discrete structure of the continuous environment. The input to the learning algorithm consists of a series of discrete observation (i.e. joint angle values from sensor reading) describing the motion features. In addition, the observations are arranged in sequences O 1:T ¼½O 1 ;...; O T :¼ O t such that every sequence described the trajectory of action. We adopt the competitive Hebbian rule proposed by Martinez [28] in topology-preserving networks to build connections between neural nodes. The competitive Hebbian rule can be described as: for each input signal, connect the two closest nodes (measured by distance) by an edge. It has been proved that each edge of the generated graph belongs to the Delaunay triangulation corresponding to the given set of reference vectors, and the graph is optimally topology-preserving in a very general sense (Fig. 2). We designed a topological map of the observed data through a novel algorithm called TGARM. The TGARM model is based on the Gaussian mixture model of the input space where each Gaussian component represents a category node. TGARM has following properties: The structure grows incrementally by incorporating new knowledge without defiling the previously learned data and adaptively responds to the information acquired from the environment. This avoids catastrophic forgetting, i.e. retain previously learned knowledge effectively. Each node in the topological map is defined as a Gaussian distribution, with mean and covariance. The parameters of each node in the topological map are updated for each input sequence observed rather than acquiring an entire data set. TGARM has more resistance to noise and take into consideration the statistical distribution of the input pattern. Figure 3 shows the structure of topological map model. The motion features, such as joint angle values, are provided as input to the architecture. Each input neuron is connected to the output neurons through the bottom-up weights; conversely, each neuron in the output layer is connected to the input layer through the top-down weights. The bottom-up weights or the activation function provides likelihood that an input pattern is a probable candidate for being a node or a category, whereas the top-down weights or the matching function provides a confidence measure about the selected candidate to be added as a node in the Fig. 2 Example of space distribution through topological mapping showing nodes connected through edges Fig. 3 Architecture for creation of topological map network. This confidence measure is controlled by the vigilance parameter, q. Otherwise, the feature space is searched for new candidate. The output layer creates a topological structure of the input data. Each node weight is defined by a vector l j and matrix R j representing its mean and covariance, respectively. Another associated parameter with each node is the node count or learning rate n j representing the number of nodes or the number of input patterns learned by TGARM. The network is initialized with two parameters: the baseline vigilance parameter q which takes the values within interval (0, 1) and the initial covariance matrix, c. The variables that define the contents of the node are summarized in Table 1. At the beginning of learning, the mean value is initialized with the input parameter and the covariance matrix is initialized with some suitable value (through experiments). As time

6 Table 1 Topological Gaussian Adaptive Resonance Map (TGARM) parameters Parameters x J l j R j q n j Meaning Winning node Mean value weight parameter Covariance matrix weight parameter initialized by initial covariance matrix c Vigilance parameter initialized by baseline vigilance parameter q with values (0, 1) Node count or learning rate progresses, the nodes are added into the network and their associated weight values are updated. The learning algorithm grows its neural structure starting with the first node. The motion features, represented by joint angle values, are encoded as a Gaussian node in the structure, and two nodes are connected with the edges. This allows the flow of information among neighbouring nodes. This algorithm allows the observer to incrementally learn and update the structure of the model based on the observed motion patterns. The algorithm for topological mapping is summarized in Algorithm 1. Algorithm 1 Algorithm for Topological Gaussian Adaptive Resonance Structure Require: : Observation Vector O t Covariance Matrix Σ Baseline Vigilance Parameter ρ Ensure: : Nodes N Edges E Input the observation vector O t. if Thereisnonodeinthenetworkthen Add O t in the network as new node. N N i {O t } ; n i =0 Update the weights of the node N i (n, μ, Σ) using Eq. (4) (6) else Determine the winner node ω J using (1) and (2) Determine the vigilance criterion for the winner node ω J using Eq. (3) if calcvig< ρ then Add as a new node N N {O t } Update the weights of the ω J (winner node) N (n J,μ J,Σ J )usingeq.(4) (6) Add edge between the previous winner and current winner nodes E E {(prevw inner, ω J )} else if calcvig ρ then Reset the winner node and find a new winner from observation vector. Update the weights of previous winner node. Obtain the new observation vector O t. If the learning is not completed, go to Step 6 to process the next observation. end if end if end if During learning a winning node w J is selected from an input pattern based on the highest probability. Since each node is represented by Gaussian components defined by the mean values, l j and the covariance matrix R j, therefore, the conditional density of O t given the winning node j or the bottom-up input activation value for a node is calculated as: 1 pðo t jjþ ¼ ð2pþ M=2 jr j j 1=2 exp 1 2 ðo i l j Þ T R 1 j ðo i l j Þ ð1þ where M is the dimensionality of the input motion patterns. For each input pattern, the activation value is calculated using (1) and the neuron with highest activation value is selected which determines the node with the highest probability (2). J ¼ arg max pðo t jjþ j ð2þ But the node represented by its weights determined by mean and covariance is only allowed to be updated if the vigilance criterion or matching between the given input and the selected winner node is fulfilled. A node w J passes the vigilance criterion if its match function value exceeds the vigilance parameter value q, that is, if: exp 1 2 ðo t l j Þ T R 1 j ðo t l j Þ q ð3þ The vigilance is a measure of similarity between the input and the node s mean relative to its standard deviation. If the winning node fails to pass the vigilance test (3), the current winner node is disqualified and its activation value is reset. Then, the observation pattern is searched for the new winning best-matching neuron. If no satisfactory neuron is found, a new neuron representing the input pattern with n J ¼ 0 is integrated satisfying the vigilance criterion. When the winning neuron, satisfying the vigilance condition representing the input pattern, is selected, its parameters, i.e. count, mean and variance, are updated using (4) (6). n J ¼ n J þ 1 l J ¼ 1 1 n J R J ¼ l J þ 1 1 R J þ n J 1 n J 1 n J O t ðo t l J ÞðO t l J Þ T ð4þ ð5þ ð6þ This algorithm allows the observer to incrementally learn and update the structure of the model based on the observed motion patterns. When the resonating neuron is determined, a lateral connection or an edge is established

7 between the current and the previous winner node. This mechanism will provide a stable architecture for providing a link between the previously learned knowledge and also integrate newly observed data in order to map temporal correlation between them. The performance of the TGARM depends on two parameters: the vigilance parameter q, and the initial covariance matrix c. The vigilance parameter directly influences the formation of new nodes when novel information is detected. For higher values of the vigilance parameter, the system becomes more sensitive to the changes in the input and the network becomes complicated. On the other hand, for the lower values of the vigilance parameter the system becomes less sensitive and faster. Therefore, the decision about the vigilance parameter value greatly influences the convergence and recognition properties of the system. Furthermore, the generalization performance of the network is greatly affected by the selection of values for initial covariance matrix. 3.2 Incremental motion learning After updating the structure of the model, motion patterns are learned through the probabilistic module. This is done through Hidden Markov Model (HMM). HMM is a doubly stochastic model composed of states which are not directly observed. HMM explicitly includes time, resulting in efficient learning of temporal sequences and compensate for uncertainties. Each state in the HMM emits an observation as output which infers the most likely dynamical system. Each state is connected by transitions between the states and generates an output pattern. Figure 4 shows an example of left-to-right HMM where each motion feature is encoded in the HMM state fs 1 ; s 2 ;...; s N g and allows the observation symbol, b i ðþ, to be emitted from each state. The probability of moving from one state to another is given by transition probability a ij. In order to select the appropriate structure of the HMM or selecting the optimum numbers of HMM states, TGARM algorithm is employed. After updating the model structure, the remaining parameters of HMM, such as transition probability and prior probabilities, are updated using expectation maximization (EM) algorithm [1, 29]. In our model, the transition between different states is characterized by the edges connecting neighbouring states, and thus, the system can only make transition between the neighbouring states. Since behaviour patterns are ordered sequences composed of different atomic actions or motion primitives, each motion element is encoded as a Gaussian. Therefore, we used left-to-right HMM model structure for representing observed motion patterns to allow the data to flow in a sequential order in forward direction of time. In left-to-right HMM, the self-transition loop is also allowed. An HMM is characterized by the following parameters: State prior probability p i ¼ Pðs 0 ¼ iþ. This represents the prior probability for the corresponding state. State transition probability matrix a ij ¼ Pðs tþ1 ¼ jjs t ¼ iþ. This represents the probability of transition from state i to state j. Observation probability distribution B ¼ PðO t js t ¼ iþ. This represents the probability distribution of observation vector from state i. This distribution is represented by Gaussian function denoted by the parameters NðO t jm i ; C i Þ, where m i and C i are mean vector and the covariance matrix for the i-th state in HMM. These HMM parameters are denoted as k ¼fp; A; Bg ¼ fp; A; m; Cg. Each hidden state in HMM encodes and abstracts an observed motion pattern where a sequence of these motion pattern is estimated using the transition between these hidden states. Table 2 summarizes the contents of the state of HMM. Table 2 Hidden Markov Model Parameters Fig. 4 Motion sequences encoded in Hidden Markov Model (HMM) Parameters p i a ij m i C i O t Meaning State prior probability every node added in the map is initialized with state prior value State transition probability updated through the edges connecting two nodes in TGARM. Mean value for Gaussian in HMM updated through topological map for each node ðm i ¼ l i Þ Covariance matrix for each node in HMM updated through topological map for each node ðc i ¼ R i Þ Observation probability distribution represented through Gaussian distribution with parameters NðO t jm i ; C i Þ

8 3.2.1 Updating structure and parameter of HMM After updating the topological map, the structure of HMM is also updated based on the added nodes and edges. The structure of HMM is updated whenever a new behaviour is observed. Corresponding to every node added in the topological map, a state in the HMM is also added. Each added state is initialized with the prior probability p i ¼ p 0 and self-transition probability a i;i ¼ a 0, where i represents the new node. Similarly, for addition of every new node and the new edges (i, j) connecting these nodes, the transition probabilities are also initialized with state transition probability value a i;j ¼ a 0. After updating the HMM structure, the parameters of HMM are also updated. The mean and the covariance values related to each Gaussian observation are updated during the structure (topological map) updating process discussed in previous section. The same values are used by the HMM. However, the remaining parameters such as transition probability and state prior probabilities need to be re-estimated. These parameters are updated using the expectation maximization algorithm. Traditionally, Baum Welch algorithm [1] which is a type of EM algorithm is used for learning the initial state probability distribution and the sate transition model. Re-estimate the transition probability using Eq. (7). P T 1 t¼1 a ij ¼ a tðiþa ij b j ðo tþ1 Þb tþ1 ðjþ P T 1 t¼1 a ð7þ tðiþb t ðiþ p i ¼ a 1ðiÞb 1 ðiþ ð8þ PðOjs i Þ In Eqs. (7) and (8), the a i and b i represent the forward and backward variables [1]. Table 3 summarizes the computation of these variables. PðOjS i Þ in (8) determines the joint observation probability. In order to update the parameters incrementally for new observed data, an incremental learning rule is applied as follows: Table 3 Recursive Computation of Forward and Backward variables Algorithm Forward variable Backward variable Computation a i ð1þ ¼p i b i ðo 1 Þ a j ðt þ 1Þ ¼ PN a i ðtþa ij b j ðo tþ1 Þ i¼1 b i ðtþ ¼1 b i ðiþ ¼ PN a ij b j ðo tþ1 Þb j ðt þ 1Þ j¼1 a ij ¼ a ij þðn p 1Þa ij N p ð9þ p i ¼ p i þðn p 1Þp i ð10þ N p where N p is the number input patterns that have been observed until the current time. 4 Experimental results In this section, we will discuss the experimental setup and datasets used along with the analysis of the proposed work on these datasets. The proposed algorithm was tested on open humanoid platform DARwIn-OP [30] developed by Robotis Co. Ltd. For simulation purposes, we have used the Webots simulator. DARwIn-OP has 20 degrees of freedom with 3-axis accelerometer (Fig. 5). The proposed approach for robot behaviour learning was tested on a dataset containing a series of different action sequences obtained through robot sensors. The dataset contains joint angle values for multiple observations. These joint angle values are used as input to the learning algorithm. In the experiment, the performance of incremental learning algorithm is computed in terms of reproducing the joint angle values. For learning the number of hidden states for each behaviour was estimated using TGARM. The experiments are conducted to test the efficacy of the proposed model for learning and reproducing the motion patterns. Motion sequences are presented one at a time to the algorithm incrementally, simulating online, sequential acquisition. The algorithm was tested for two types of datasets. The first experiment consists of different kinds of behavioural gestures. The dataset consists of variety of different actions involving upper part of the body. These actions include standing up, raising and lowering left and right arm by 180 one at a time, raising and lowering both arms by 180 simultaneously, raising and lowering left and right arm by 90 one at a time, raising and lowering left and right arm by 90 one at a time. All these gestures are required to achieve different kinds of goal position representing each category of behaviour. The observed behavioural trajectory is acquired from different motion sensors and mapped onto different categories used by the model. The first experiment was conducted on the behaviour gestures generated by the robot. The behavioural actions are generated using the movement of joints in the Webots. The action features consist of rotational angle values of respective joints measured in radians. Figure 6 shows the visualization of actions performed. Each image in the figure shows different frames extracted from the action sequences.

9 Neural Comput & Applic Fig. 5 Kinematic information of open humanoid platform DARwIn-OP Fig. 6 Different samples of behavioural actions performed by the robot during experimentation

10 Fig. 7 Different frames from the walk approach kick experiment performed on DARwIn-op humanoid robot in Webots The second experiment consists of complex behavioural patterns of finding the ball, walking towards the ball and then kicking the ball. During these experiments, the robot first stands up and searches for the ball using image processing techniques. The discussions of these processing algorithms are out of the scope of this paper and are considered to be already available for experimentation as they have no direct effect on the experimentation. After finding the ball, the robot starts walking towards the ball and kicks the ball either with left or with right foot. The joint angle values for respective joints (involved in action) are recorded. The experiment was conducted for different position and distance of the ball with respect to the robot. For example, placing the ball behind or in front of the robot involves turning of robot in the direction of the ball. Figure 7 shows the random frames extracted from the sequence of dataset-2. As discussed earlier, the performance of TGARM greatly depends on the selection of values for the vigilance parameter and initial covariance matrix. In current experiments, the vigilance parameter values are selected in an ad hoc fashion by trial and error choosing the optimal value for the parameters. The experiments were performed for different values of vigilance parameter, selecting the value which efficiently generalizes the observed patterns. Similarly, the initial covariance matrix determines the isotropic spread in feature space of a new node s distribution. For large c, the learning will be slow with fewer nodes, while for smaller values of c, the training will be faster with more number of nodes. The initial covariance matrix is selected in the same manner as the vigilance parameter. We evaluated the performance of the system using the error between he demonstrated and generalized motion to determine the appropriate adapting learned motion. The error metric provides a measure for the evaluation of the generalization capability of the proposed model. The mean error is used as a metric to evaluate the sustainability of the generalized motion with respect to the demonstrated motion. We utilize heuristics to find the suitable parameters for the TGAR-HMM. We have applied different values for the vigilance and the initial covariance matrix to find the suitable candidate values for these parameters. For behaviour dataset, a value of 0.95 is observed to give a good generalization performance and a compromise between the number of nodes and the mean square error. The reason for selecting a value of 0.95 for the vigilance parameter is to generate the motion pattern as close as possible to the original pattern. This results in selection of optimal number of states or nodes during learning. Similarly, a value of 0.3 is chosen for the initial covariance matrix. Table 4 summarizes the selected parameters. Figure 8a shows the effect of selecting various values for vigilance parameter on the mean square error and the number of nodes. As the value of vigilance parameter is increased, the number of nodes also increased. Inversely, as the value of vigilance parameter is decreased, the value of MSE decreases. This shows that the higher value of vigilance parameter efficiently generalizes the learned behaviour patterns. Figure 8b shows the effect of vigilance parameter on the compression ratio (CR) (11). The compression ratio is defined through the ratio between the original motion samples and the number of nodes generated by the learning algorithm. As the value of vigilance parameter is increased, compression ratio is reduced by Table 4 Selected parameters and mean square error Algorithm Parameters MSE GHMM q ¼ 0:95, c ¼ 0: TGAR-HMM c ¼ 0:3, s ¼ 0:

11 (a) (b) Fig. 8 a Plot of MSE and number of nodes for different values of vigilance parameter. b Effect of vigilance parameter on compression ratio Fig. 9 Computation times adding number of nodes and minimizes the mean square error to better generalize the observed behaviour patterns. No: of samples CR ¼ ð11þ No. of nodes Figure 9 plots the processing time taken by learning algorithm with respect to the number of trajectories. The model size is represented as the number of nodes in the TGAR-HMM structure is also given as a reference. As may be anticipated, the learning time appears to be linearly dependent on the size of model. An interesting observation is that the time per observation is lower than 1ms. Thus, algorithm is well adapted for online applications in the environment where it is likely to observe fewer trajectories per second. We evaluated the performance of the system using error between the demonstrated and generalized motion to determine the appropriate adapting learned motion. The mean error is used as a metric to evaluate the sustainability of the generalized motion with respect to the demonstrated motion. This error metric provides a measure for the evaluation of generalization capability of proposed learning model. The motions generated from the RBA and LBA for the behaviour dataset are summarized in Fig. 10. Here

12 (a) (b) Fig. 10 Plot of original and generated motion patterns for different parts of the robots joints during behavioural gestures experiment of a RBA, b LBA 4.1 Performance evaluation Fig. 11 Generalization result for walk approach kick motion pattern. The figure displays the joint angle pattern for right hip yaw joint movement during experiment we have only shown the original and generalized motion patterns acquired from the left right shoulder and left right elbow of DARwIn-OP robot. These experiments show that the proposed algorithm can learn the behaviours and generate the same behaviours efficiently. Using the same heuristics applied for the behavioural gestures data, we select the vigilance parameter value for the find-walk-kick ball experiment. A value of 0.95 is selected as a suitable candidate. Figure 11 shows the generalization result for the second experiment involving the complex behaviour of walking, approaching and kicking the ball. Due to the large dimensionality of the dataset, for visualization, we only show the result of right hip joint angle value since these play the major role in walking motion. In order to assess the effectiveness of the proposed architecture, we have performed a comparison of our approach against the most similar approach, called Growing Hidden Markov Model (GHMM) proposed by Vasquez et al. [24]. Comparing heterogeneous techniques, even if they solve a common problem, is a difficult task. Often their theoretical bases are too different, making it difficult to evaluate them fairly. Here we propose a measure, despite its simplicity, still provide useful indicator about the parsimony of the models produced by the different approaches. Vasquez et al. [24] have developed an incremental model for learning and predicting through growing HMM. The model updates the structure of HMM through the use of topological mapping algorithm called instantaneous topological mapping (ITM). In ITM, the insertion of nodes is controlled by the Mahalanobis distance between the nodes and the insertion threshold parameter s. The model was designed for vehicle motion learning and prediction. Similar to our approach, the growing HMM architecture uses the topological mapping structure to define the HMM structure. The structure of the transition matrix is the fundamental complexity factor for inference on HMMs. Therefore, we measure the size of the model by the number of nodes in the transition graph. The parameters we have chosen for every algorithm are shown in Table 4. Owing to the difficulty in choosing adequate parameters to compare the approaches, we have started by making an educated guess and then refining them by trial and error. The growth on the model size with respect to the number of trajectories on the

13 (a) (b) Fig. 12 Model size (number of nodes) comparison between TGAR-HMM and GHMM for two different datasets. a Behaviour dataset, b walk approach kick dataset learning dataset is displayed in Fig. 12. As can be seen in the figure, the size of the TGAR-HMM models is small with respect to the discussed approach. The proposed approach is related to the growing HMM (GHMM) model. However, unlike GHMM, in the proposed algorithm we used Gaussian distribution for adding and updating new nodes. Furthermore, instead of using fixed covariance matrix for modelling the observation probabilities, the TGAR-HMM model uses the covariance matrix which is updated based on the input patterns. 5 Conclusion In this paper, we have developed a novel architecture for incremental and online learning for continuous flow of motion patterns based on novel architecture called Topological Gaussian Adaptive Resonance Hidden Markov Model (TGAR-HMM). The observed motion patterns are encoded through TGAR-HMM. The structure of the model is updated incrementally through a novel topological map. Based on this topological map, the model aggregates the information as observed and organize the acquired information in an efficient growing and self-organizing manner. The dynamic architecture grows incrementally to adapt to the new data from the environment. The fundamental characteristics of the network are the stable learning eliminating the stability and plasticity dilemma and fast convergence. As the new sequence of data is learned, the network converges to create a stable topological structure of the input trajectory. Based on this feature, the proposed architecture does not suffer ruinous forgetting and also retains previously learned data and accommodates the new information. The TGAR-HMM model efficiently learns and encodes the spatio-temporal patterns and computes the probability that observation sequences could be generated. Secondly, the novel HMM architecture adaptively selects the models structure based on the observed data and is not pre-defined based on some prior knowledge. The proposed model has been tested for different kinds of behaviour on an open humanoid platform DARwIn-OP through simulation. The model can encode and generalize simple type of behaviours (such as behavioural gestures) as well as the complex behaviour patterns consisting of combination of different simple behaviours. The aim of the proposed algorithm is not to exactly trace back the original trajectory sequence as this will result in large number of nodes or HMM states. This results in selection of optimal number of states or nodes from the measured data. This selection is controlled by the vigilance parameter. However, the selection of vigilance is done manually using trial and error method, but the current experiments show that once the suitable value of vigilance parameter is selected for a articular dataset, it can be efficiently applied to different kinds of motion sequences. However, an improved method for selection or adaptively modifying the value of vigilance parameter will efficiently provide a better generalization performance. Acknowledgments This research is supported by High Impact Research Grant UM.C/625/1/HIR/MoE/FCSIT/10 from the Ministry of Education Malaysia. References 1. Rabiner L (1989) A tutorial on hidden Markov models and selected applications in speech recognition. Proc IEEE 77(2): Inamura T, Toshima I, Tanie H, Nakamura Y (2004) Embodied symbol emergence based on mimesis theory. Int J Robot Res 23(4 5):

14 3. Calinon S, Billard A (2004) Stochastic gesture production and recognition model for a humanoid robot. In: Proceedings of the IEEE/RSJ international conference on intelligent robots and systems, 2004 (IROS 2004), pp Grossberg S (1987) Competitive learning: from interactive activation to adaptive resonance. Cogn Sci 11(1): Billard AG, Calinon S, Guenter F (2006) Discriminative and adaptive imitation in uni-manual and bi-manual tasks. Robot Auton Syst 54(5): Kulic D, Takano W, Nakamura Y (2007) Incremental on-line hierarchical clustering of whole body motion patterns. In: The 16th robot and human interactive communication 2007, RO MAN Grossberg S (2013) Adaptive resonance theory: how a brain learns to consciously attend, learn, and recognize a changing world. Neural Netw 37: Jain L, Seera M, Lim C, Balasubramaniam P (2014) A review of online learning in supervised neural networks. Neural Comput Appl 25(3 4): Carpenter GA, Grossberg S (1987) A massively parallel architecture for a self-organizing neural pattern recognition machine. Comput Vision Graph Image Process 37(1): Carpenter GA, Grossberg S (1987) Art 2: self-organization of stable category recognition codes for analog input patterns. Appl Opt 26(23): Carpenter G, Grossberg S, Rosen D (1991) Art 2 A: an adaptive resonance algorithm for rapid category learning and recognition. In: International joint conference on neural networks, IJCNN 91, pp Bradski G, Carpenter G, Grossberg S (1994) Store working memory networks for storage and recall of arbitrary temporal sequences. Biol Cybern 71(6): Furao S, Hasegawa O (2006) An incremental network for on-line unsupervised classification and topology learning. Neural Netw 19(1): Furao S, Hasegawa O (2008) A fast nearest neighbor classifier based on self-organizing incremental neural network. Neural Netw 21(10): Botzheim J, Kubota N (2012) Growing neural gas for information extraction in gesture recognition and reproduction of robot partners. In: IEEE 2012 international symposium on micronanomechatronics and human science (MHS), pp Seyhan SS, Alpaslan FN, Yava M (2013) Simple and complex behavior learning using behavior hidden markov model and cobart. Neurocomputing 103: Yavas M, Alpaslan FN (2012) Hierarchical behavior categorization using correlation based adaptive resonance theory. Neurocomputing 77(1): Kulic D, Lee D, Ott C, Nakamura Y (2008) Incremental learning of full body motion primitives for humanoid robots. In: IEEE RAS 8th international conference on humanoid robots, humanoids 2008, pp Kulic D, Ott C, Lee D, Ishikawa J, Nakamura Y (2012) Incremental learning of full body motion primitives and their sequencing through human motion observation. Int J Robot Res 31: Kulic D, Takano W, Nakamura Y (2011) Towards lifelong learning and organization of whole body motion patterns. In: Kaneko M, Nakamura Y (eds) Robotics research, springer tracts in advanced robotics, vol 66. Springer, Berlin, pp Calinon S, Billard A (2007) Incremental learning of gestures by imitation in a humanoid robot. In: Proceedings of the ACM/IEEE international conference on Human-robot interaction, ACM, New York, NY, USA, HRI 07, pp Cho S, Jo S (2011) Kinesthetic learning of behaviors in a humanoid robot. In: th international conference on control, automation and systems (ICCAS), pp Cho S, Jo S (2012) Incremental motion learning through kinesthetic teachings and new motion production from learned motions by a humanoid robot. Int J Control Autom Syst 10(1): Vasquez D, Fraichard T, Laugier C (2009) Incremental learning of statistical motion patterns with growing hidden markov models. IEEE Trans Intell Transp Syst 10(3): Tscherepanow M, Kortkamp M, Kammer M (2011) A hierarchical art network for the stable incremental learning of topological structures and associations from noisy data. Neural Netw 24(8): Wk Fung, Liu YH (2003) Adaptive categorization of art networks in robot behavior learning using game-theoretic formulation. Neural Netw 16(10): Jun L, Duckett T (2003) Robot behavior learning with a dynamically adaptive RBF network: experiments in offline and online learning. In: Proceedings of the second international conference on computational intelligence, robotics and autonomous system, Singapore 28. Martinetz T, Schulten K (1994) Topology representing networks. Neural Netw 7(3): Neal RM, Hinton GE (1999) A view of the EM algorithm that justifies incremental, sparse, and other variants. In: Jordan MI (ed) Learning in graphical models. MIT Press, Cambridge, pp Haa I, Tamuraa Y, Asamaa H (2013) Development of open platform humanoid robot darwin-op. Adv Robot 27(3):

Representability of Human Motions by Factorial Hidden Markov Models

Representability of Human Motions by Factorial Hidden Markov Models Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems San Diego, CA, USA, Oct 29 - Nov 2, 2007 WeC10.1 Representability of Human Motions by Factorial Hidden Markov

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

Computational Intelligence Meets the NetFlix Prize

Computational Intelligence Meets the NetFlix Prize Computational Intelligence Meets the NetFlix Prize Ryan J. Meuth, Paul Robinette, Donald C. Wunsch II Abstract The NetFlix Prize is a research contest that will award $1 Million to the first group to improve

More information

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models Gleidson Pegoretti da Silva, Masaki Nakagawa Department of Computer and Information Sciences Tokyo University

More information

Kinesthetic Teaching via Fast Marching Square

Kinesthetic Teaching via Fast Marching Square Kinesthetic Teaching via Fast Marching Square Javier V. Gómez, David Álvarez, Santiago Garrido and Luis Moreno Abstract This paper presents a novel robotic learning technique based on Fast Marching Square

More information

Controlling the spread of dynamic self-organising maps

Controlling the spread of dynamic self-organising maps Neural Comput & Applic (2004) 13: 168 174 DOI 10.1007/s00521-004-0419-y ORIGINAL ARTICLE L. D. Alahakoon Controlling the spread of dynamic self-organising maps Received: 7 April 2004 / Accepted: 20 April

More information

Incremental learning of full body motion primitives and their sequencing through human motion observation

Incremental learning of full body motion primitives and their sequencing through human motion observation Incremental learning of full body motion primitives and their sequencing through human motion observation The International Journal of Robotics Research 0(0) 1 16 The Author(s) 2011 Reprints and permission:

More information

Modeling time series with hidden Markov models

Modeling time series with hidden Markov models Modeling time series with hidden Markov models Advanced Machine learning 2017 Nadia Figueroa, Jose Medina and Aude Billard Time series data Barometric pressure Temperature Data Humidity Time What s going

More information

Incremental learning of motion primitives for full body motions

Incremental learning of motion primitives for full body motions Incremental learning of motion primitives for full body motions p. 1/43 Incremental learning of motion primitives for full body motions Dana Kulić Department of Electrical and Computer Engineering, University

More information

Incremental learning of full body motions via adaptive Factorial Hidden Markov Models

Incremental learning of full body motions via adaptive Factorial Hidden Markov Models Incremental learning of full body motions via adaptive Factorial Hidden Markov Models Dana Kulić, Wataru Takano and Yoshihiko Nakamura Department of Mechano-Informatics, University of Tokyo 7-3- Hongo,

More information

Incremental on-line hierarchical clustering of whole body motion patterns

Incremental on-line hierarchical clustering of whole body motion patterns Incremental on-line hierarchical clustering of whole body motion patterns Dana Kulić, Wataru Takano and Yoshihiko Nakamura Department of Mechano-Informatics, Graduate School of Information Science and

More information

Unsupervised Learning

Unsupervised Learning Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised

More information

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2 A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation Kwanyong Lee 1 and Hyeyoung Park 2 1. Department of Computer Science, Korea National Open

More information

An Introduction to Pattern Recognition

An Introduction to Pattern Recognition An Introduction to Pattern Recognition Speaker : Wei lun Chao Advisor : Prof. Jian-jiun Ding DISP Lab Graduate Institute of Communication Engineering 1 Abstract Not a new research field Wide range included

More information

Short Survey on Static Hand Gesture Recognition

Short Survey on Static Hand Gesture Recognition Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of

More information

Function approximation using RBF network. 10 basis functions and 25 data points.

Function approximation using RBF network. 10 basis functions and 25 data points. 1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Incremental Learning of Full Body Motion Primitives for Humanoid Robots

Incremental Learning of Full Body Motion Primitives for Humanoid Robots 2008 8 th IEEE-RAS International Conference on Humanoid Robots December 1 ~ 3, 2008 / Daejeon, Korea Incremental Learning of Full Body Motion Primitives for Humanoid Robots Dana Kulić, Dongheui Lee, Christian

More information

Hidden Markov Model for Sequential Data

Hidden Markov Model for Sequential Data Hidden Markov Model for Sequential Data Dr.-Ing. Michelle Karg mekarg@uwaterloo.ca Electrical and Computer Engineering Cheriton School of Computer Science Sequential Data Measurement of time series: Example:

More information

A Neural Network for Real-Time Signal Processing

A Neural Network for Real-Time Signal Processing 248 MalkofT A Neural Network for Real-Time Signal Processing Donald B. Malkoff General Electric / Advanced Technology Laboratories Moorestown Corporate Center Building 145-2, Route 38 Moorestown, NJ 08057

More information

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping 1 Masaki Ogino, 1 Shigeo Matsuyama, 1 Jun ichiro Ooga, and 1, Minoru Asada 1 Dept. of Adaptive Machine Systems, HANDAI

More information

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation Chan-Su Lee and Ahmed Elgammal Rutgers University, Piscataway, NJ, USA {chansu, elgammal}@cs.rutgers.edu Abstract. We

More information

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging 1 CS 9 Final Project Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging Feiyu Chen Department of Electrical Engineering ABSTRACT Subject motion is a significant

More information

Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition

Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition André Maurer, Micha Hersch and Aude G. Billard Ecole Polytechnique Fédérale de Lausanne (EPFL) Swiss Federal Institute

More information

Combining Automated On-line Segmentation and Incremental Clustering for Whole Body Motions

Combining Automated On-line Segmentation and Incremental Clustering for Whole Body Motions 2008 IEEE International Conference on Robotics and Automation Pasadena, CA, USA, May 9-23, 2008 Combining Automated On-line Segmentation and Incremental Clustering for Whole Body Motions Dana Kulić, Wataru

More information

I How does the formulation (5) serve the purpose of the composite parameterization

I How does the formulation (5) serve the purpose of the composite parameterization Supplemental Material to Identifying Alzheimer s Disease-Related Brain Regions from Multi-Modality Neuroimaging Data using Sparse Composite Linear Discrimination Analysis I How does the formulation (5)

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting

More information

Skill. Robot/ Controller

Skill. Robot/ Controller Skill Acquisition from Human Demonstration Using a Hidden Markov Model G. E. Hovland, P. Sikka and B. J. McCarragher Department of Engineering Faculty of Engineering and Information Technology The Australian

More information

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Evaluation

More information

New wavelet based ART network for texture classification

New wavelet based ART network for texture classification University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 1996 New wavelet based ART network for texture classification Jiazhao

More information

Robot Learning by Demonstration

Robot Learning by Demonstration School of Computer Science Internship Project Report Robot Learning by Demonstration Author: Kaushik Subramanian Rutgers University Supervisor: Dr. Gerhard Lakemeyer Head of Knowledge Based Systems Group

More information

Human Skill Transfer System via Novint Falcon

Human Skill Transfer System via Novint Falcon Human Skill Transfer System via Novint Falcon Tarinee Tonggoed and Siam Charoenseang Abstract This paper presents a skill transfer system of hand movement via Novint Falcon. In the research, expert can

More information

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern

More information

An Adaptable Neural-Network Model for Recursive Nonlinear Traffic Prediction and Modeling of MPEG Video Sources

An Adaptable Neural-Network Model for Recursive Nonlinear Traffic Prediction and Modeling of MPEG Video Sources 150 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL 14, NO 1, JANUARY 2003 An Adaptable Neural-Network Model for Recursive Nonlinear Traffic Prediction and Modeling of MPEG Video Sources Anastasios D Doulamis,

More information

Image Classification Using Wavelet Coefficients in Low-pass Bands

Image Classification Using Wavelet Coefficients in Low-pass Bands Proceedings of International Joint Conference on Neural Networks, Orlando, Florida, USA, August -7, 007 Image Classification Using Wavelet Coefficients in Low-pass Bands Weibao Zou, Member, IEEE, and Yan

More information

Graph-based High Level Motion Segmentation using Normalized Cuts

Graph-based High Level Motion Segmentation using Normalized Cuts Graph-based High Level Motion Segmentation using Normalized Cuts Sungju Yun, Anjin Park and Keechul Jung Abstract Motion capture devices have been utilized in producing several contents, such as movies

More information

Representation of 2D objects with a topology preserving network

Representation of 2D objects with a topology preserving network Representation of 2D objects with a topology preserving network Francisco Flórez, Juan Manuel García, José García, Antonio Hernández, Departamento de Tecnología Informática y Computación. Universidad de

More information

Person Authentication from Video of Faces: A Behavioral and Physiological Approach Using Pseudo Hierarchical Hidden Markov Models

Person Authentication from Video of Faces: A Behavioral and Physiological Approach Using Pseudo Hierarchical Hidden Markov Models Person Authentication from Video of Faces: A Behavioral and Physiological Approach Using Pseudo Hierarchical Hidden Markov Models Manuele Bicego 1, Enrico Grosso 1, and Massimo Tistarelli 2 1 DEIR - University

More information

Topological Correlation

Topological Correlation Topological Correlation K.A.J. Doherty, R.G. Adams and and N. Davey University of Hertfordshire, Department of Computer Science College Lane, Hatfield, Hertfordshire, UK Abstract. Quantifying the success

More information

Comparison of supervised self-organizing maps using Euclidian or Mahalanobis distance in classification context

Comparison of supervised self-organizing maps using Euclidian or Mahalanobis distance in classification context 6 th. International Work Conference on Artificial and Natural Neural Networks (IWANN2001), Granada, June 13-15 2001 Comparison of supervised self-organizing maps using Euclidian or Mahalanobis distance

More information

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra Pattern Recall Analysis of the Hopfield Neural Network with a Genetic Algorithm Susmita Mohapatra Department of Computer Science, Utkal University, India Abstract: This paper is focused on the implementation

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,

More information

NOVEL HYBRID GENETIC ALGORITHM WITH HMM BASED IRIS RECOGNITION

NOVEL HYBRID GENETIC ALGORITHM WITH HMM BASED IRIS RECOGNITION NOVEL HYBRID GENETIC ALGORITHM WITH HMM BASED IRIS RECOGNITION * Prof. Dr. Ban Ahmed Mitras ** Ammar Saad Abdul-Jabbar * Dept. of Operation Research & Intelligent Techniques ** Dept. of Mathematics. College

More information

Movement Learning and Control for Robots in Interaction

Movement Learning and Control for Robots in Interaction Movement Learning and Control for Robots in Interaction Dr.-Ing. Michael Gienger Honda Research Institute Europe Carl-Legien-Strasse 30 63073 Offenbach am Main Seminar Talk Machine Learning Lab, University

More information

Hidden Markov Models. Slides adapted from Joyce Ho, David Sontag, Geoffrey Hinton, Eric Xing, and Nicholas Ruozzi

Hidden Markov Models. Slides adapted from Joyce Ho, David Sontag, Geoffrey Hinton, Eric Xing, and Nicholas Ruozzi Hidden Markov Models Slides adapted from Joyce Ho, David Sontag, Geoffrey Hinton, Eric Xing, and Nicholas Ruozzi Sequential Data Time-series: Stock market, weather, speech, video Ordered: Text, genes Sequential

More information

Efficient Object Extraction Using Fuzzy Cardinality Based Thresholding and Hopfield Network

Efficient Object Extraction Using Fuzzy Cardinality Based Thresholding and Hopfield Network Efficient Object Extraction Using Fuzzy Cardinality Based Thresholding and Hopfield Network S. Bhattacharyya U. Maulik S. Bandyopadhyay Dept. of Information Technology Dept. of Comp. Sc. and Tech. Machine

More information

9.1. K-means Clustering

9.1. K-means Clustering 424 9. MIXTURE MODELS AND EM Section 9.2 Section 9.3 Section 9.4 view of mixture distributions in which the discrete latent variables can be interpreted as defining assignments of data points to specific

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised

More information

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Int. J. Advance Soft Compu. Appl, Vol. 9, No. 1, March 2017 ISSN 2074-8523 The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Loc Tran 1 and Linh Tran

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational

More information

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction Stefan Müller, Gerhard Rigoll, Andreas Kosmala and Denis Mazurenok Department of Computer Science, Faculty of

More information

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,

More information

CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs

CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs Felix Wang fywang2 John Wieting wieting2 Introduction We implement a texture classification algorithm using 2-D Noncausal Hidden

More information

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology Learning the Three Factors of a Non-overlapping Multi-camera Network Topology Xiaotang Chen, Kaiqi Huang, and Tieniu Tan National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Human Motion Database with a Binary Tree and Node Transition Graphs

Human Motion Database with a Binary Tree and Node Transition Graphs Human Motion Database with a Binary Tree and Node Transition Graphs Katsu Yamane Disney Research, Pittsburgh kyamane@disneyresearch.com Yoshifumi Yamaguchi Dept. of Mechano-Informatics University of Tokyo

More information

Neural Network Weight Selection Using Genetic Algorithms

Neural Network Weight Selection Using Genetic Algorithms Neural Network Weight Selection Using Genetic Algorithms David Montana presented by: Carl Fink, Hongyi Chen, Jack Cheng, Xinglong Li, Bruce Lin, Chongjie Zhang April 12, 2005 1 Neural Networks Neural networks

More information

Online labelling strategies for growing neural gas

Online labelling strategies for growing neural gas Online labelling strategies for growing neural gas Oliver Beyer and Philipp Cimiano Semantic Computing Group, CITEC, Bielefeld University, obeyer@cit-ec.uni-bielefeld.de http://www.sc.cit-ec.uni-bielefeld.de

More information

Classification of Printed Chinese Characters by Using Neural Network

Classification of Printed Chinese Characters by Using Neural Network Classification of Printed Chinese Characters by Using Neural Network ATTAULLAH KHAWAJA Ph.D. Student, Department of Electronics engineering, Beijing Institute of Technology, 100081 Beijing, P.R.CHINA ABDUL

More information

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed

More information

Based on Raymond J. Mooney s slides

Based on Raymond J. Mooney s slides Instance Based Learning Based on Raymond J. Mooney s slides University of Texas at Austin 1 Example 2 Instance-Based Learning Unlike other learning algorithms, does not involve construction of an explicit

More information

Self-Organizing Maps for cyclic and unbounded graphs

Self-Organizing Maps for cyclic and unbounded graphs Self-Organizing Maps for cyclic and unbounded graphs M. Hagenbuchner 1, A. Sperduti 2, A.C. Tsoi 3 1- University of Wollongong, Wollongong, Australia. 2- University of Padova, Padova, Italy. 3- Hong Kong

More information

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Neural Networks CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Biological and artificial neural networks Feed-forward neural networks Single layer

More information

Online Learning for Object Recognition with a Hierarchical Visual Cortex Model

Online Learning for Object Recognition with a Hierarchical Visual Cortex Model Online Learning for Object Recognition with a Hierarchical Visual Cortex Model Stephan Kirstein, Heiko Wersing, and Edgar Körner Honda Research Institute Europe GmbH Carl Legien Str. 30 63073 Offenbach

More information

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation IJCSNS International Journal of Computer Science and Network Security, VOL.13 No.11, November 2013 1 Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial

More information

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves

More information

Machine Learning. Unsupervised Learning. Manfred Huber

Machine Learning. Unsupervised Learning. Manfred Huber Machine Learning Unsupervised Learning Manfred Huber 2015 1 Unsupervised Learning In supervised learning the training data provides desired target output for learning In unsupervised learning the training

More information

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES 6.1 INTRODUCTION The exploration of applications of ANN for image classification has yielded satisfactory results. But, the scope for improving

More information

Comparison of Recommender System Algorithms focusing on the New-Item and User-Bias Problem

Comparison of Recommender System Algorithms focusing on the New-Item and User-Bias Problem Comparison of Recommender System Algorithms focusing on the New-Item and User-Bias Problem Stefan Hauger 1, Karen H. L. Tso 2, and Lars Schmidt-Thieme 2 1 Department of Computer Science, University of

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

Assignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018

Assignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018 Assignment 2 Unsupervised & Probabilistic Learning Maneesh Sahani Due: Monday Nov 5, 2018 Note: Assignments are due at 11:00 AM (the start of lecture) on the date above. he usual College late assignments

More information

Spatio-Temporal Stereo Disparity Integration

Spatio-Temporal Stereo Disparity Integration Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz

More information

Automatic Linguistic Indexing of Pictures by a Statistical Modeling Approach

Automatic Linguistic Indexing of Pictures by a Statistical Modeling Approach Automatic Linguistic Indexing of Pictures by a Statistical Modeling Approach Abstract Automatic linguistic indexing of pictures is an important but highly challenging problem for researchers in content-based

More information

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition Linear Discriminant Analysis in Ottoman Alphabet Character Recognition ZEYNEB KURT, H. IREM TURKMEN, M. ELIF KARSLIGIL Department of Computer Engineering, Yildiz Technical University, 34349 Besiktas /

More information

How to Use the SOINN Software: User s Guide (Version 1.0)

How to Use the SOINN Software: User s Guide (Version 1.0) How to Use the SOINN Software: User s Guide (Version 1.0) Kazuhiro Yamasaki 1, Naoya Makibuchi 1, Furao Shen 2, and Osamu Hasegawa 1 1 Department of Computational Intelligence and Systems Science, Tokyo

More information

Tracking of Human Body using Multiple Predictors

Tracking of Human Body using Multiple Predictors Tracking of Human Body using Multiple Predictors Rui M Jesus 1, Arnaldo J Abrantes 1, and Jorge S Marques 2 1 Instituto Superior de Engenharia de Lisboa, Postfach 351-218317001, Rua Conselheiro Emído Navarro,

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

COMBINING NEURAL NETWORKS FOR SKIN DETECTION

COMBINING NEURAL NETWORKS FOR SKIN DETECTION COMBINING NEURAL NETWORKS FOR SKIN DETECTION Chelsia Amy Doukim 1, Jamal Ahmad Dargham 1, Ali Chekima 1 and Sigeru Omatu 2 1 School of Engineering and Information Technology, Universiti Malaysia Sabah,

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

Motion Interpretation and Synthesis by ICA

Motion Interpretation and Synthesis by ICA Motion Interpretation and Synthesis by ICA Renqiang Min Department of Computer Science, University of Toronto, 1 King s College Road, Toronto, ON M5S3G4, Canada Abstract. It is known that high-dimensional

More information

Some questions of consensus building using co-association

Some questions of consensus building using co-association Some questions of consensus building using co-association VITALIY TAYANOV Polish-Japanese High School of Computer Technics Aleja Legionow, 4190, Bytom POLAND vtayanov@yahoo.com Abstract: In this paper

More information

Feature Extraction in Wireless Personal and Local Area Networks

Feature Extraction in Wireless Personal and Local Area Networks Feature Extraction in Wireless Personal and Local Area Networks 29. October 2003, Singapore Institut für Praktische Informatik Johannes Kepler Universität Linz, Austria rene@soft.uni-linz.ac.at < 1 > Content

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

Background Subtraction in Video using Bayesian Learning with Motion Information Suman K. Mitra DA-IICT, Gandhinagar

Background Subtraction in Video using Bayesian Learning with Motion Information Suman K. Mitra DA-IICT, Gandhinagar Background Subtraction in Video using Bayesian Learning with Motion Information Suman K. Mitra DA-IICT, Gandhinagar suman_mitra@daiict.ac.in 1 Bayesian Learning Given a model and some observations, the

More information

Modeling pigeon behaviour using a Conditional Restricted Boltzmann Machine

Modeling pigeon behaviour using a Conditional Restricted Boltzmann Machine Modeling pigeon behaviour using a Conditional Restricted Boltzmann Machine Matthew D. Zeiler 1,GrahamW.Taylor 1, Nikolaus F. Troje 2 and Geoffrey E. Hinton 1 1- University of Toronto - Dept. of Computer

More information

3D Human Motion Analysis and Manifolds

3D Human Motion Analysis and Manifolds D E P A R T M E N T O F C O M P U T E R S C I E N C E U N I V E R S I T Y O F C O P E N H A G E N 3D Human Motion Analysis and Manifolds Kim Steenstrup Pedersen DIKU Image group and E-Science center Motivation

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

Analyzing and Segmenting Finger Gestures in Meaningful Phases

Analyzing and Segmenting Finger Gestures in Meaningful Phases 2014 11th International Conference on Computer Graphics, Imaging and Visualization Analyzing and Segmenting Finger Gestures in Meaningful Phases Christos Mousas Paul Newbury Dept. of Informatics University

More information

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant

More information

Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification

Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification Tomohiro Tanno, Kazumasa Horie, Jun Izawa, and Masahiko Morita University

More information

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks Ritika Luthra Research Scholar Chandigarh University Gulshan Goyal Associate Professor Chandigarh University ABSTRACT Image Skeletonization

More information

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Costas Panagiotakis and Anastasios Doulamis Abstract In this paper, an unsupervised, automatic video human members(human

More information

Neural Networks (Overview) Prof. Richard Zanibbi

Neural Networks (Overview) Prof. Richard Zanibbi Neural Networks (Overview) Prof. Richard Zanibbi Inspired by Biology Introduction But as used in pattern recognition research, have little relation with real neural systems (studied in neurology and neuroscience)

More information

HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation

HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation 009 10th International Conference on Document Analysis and Recognition HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation Yaregal Assabie and Josef Bigun School of Information Science,

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Introduction to Mobile Robotics Clustering Wolfram Burgard Cyrill Stachniss Giorgio Grisetti Maren Bennewitz Christian Plagemann Clustering (1) Common technique for statistical data analysis (machine learning,

More information

Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods

Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods Tobias Michels TU Darmstadt Aaron Hochländer TU Darmstadt Abstract In this paper we experiment with reinforcement

More information

Ambiguity Detection by Fusion and Conformity: A Spectral Clustering Approach

Ambiguity Detection by Fusion and Conformity: A Spectral Clustering Approach KIMAS 25 WALTHAM, MA, USA Ambiguity Detection by Fusion and Conformity: A Spectral Clustering Approach Fatih Porikli Mitsubishi Electric Research Laboratories Cambridge, MA, 239, USA fatih@merl.com Abstract

More information

Introduction to Trajectory Clustering. By YONGLI ZHANG

Introduction to Trajectory Clustering. By YONGLI ZHANG Introduction to Trajectory Clustering By YONGLI ZHANG Outline 1. Problem Definition 2. Clustering Methods for Trajectory data 3. Model-based Trajectory Clustering 4. Applications 5. Conclusions 1 Problem

More information