Optical Music Recognition using Hidden Markov Models
|
|
- Mavis Park
- 5 years ago
- Views:
Transcription
1 Optical Music Recognition using Hidden Markov Models Natalie Wilkinson April 25, Introduction Optical Music Recognition software (OMR) is software that has the ability to read sheet music. This type of software can be very useful as it can convert scanned music into a digital form that directly corresponds to music instead of having to save scanned music as an image. Some OMRs also have the ability to create a sound file from a scanned sheet of music, without requiring a trained musician to play the piece. Thus this type of software can be incredibly useful for students learning to read sheet music. Optical Character Recognition software (OCR) is a similar but much more commonplace software that converts images of text into a form that can be edited, such as a text file or a pdf. OCR software heavily influences OMR software, however there are different challenges for each type of software. Despite this, understanding OCR algorithms can be beneficial for implementing OMRs. There are many different ways OMRs can be implemented, and each implementation has its own strengths and weaknesses. The goal of this research was to investigate a particular mechanism that is used for the identification process of OMRs called hidden Markov models (HMM). HMMs are a mechanism that are trained by showing examples of items to be identified. Previous implementations of HMMs in OCRs for handwriting have been shown to be very successful [1]. Since sheet music appears to be of the same difficulty as handwritten text, it was assumed that HMMs in OMRs would be just as successful. However, previous implementations of OMRs using HMMs have been unsatisfactory compared to the other algorithms currently in use [12]. In this paper, we investigate whether an increase in the amount of training an HMM receives, could improve the accuracy of HMMs and hopefully provide a more useful mechanism for OMRs. 2 Related Work The research into OMRs started in the 1960s at MIT. They created a rudimentary OMR that could handle chords but not rests, clefs or time signatures. The next OMR that was created, in the 1970s, was based upon MIT s OMR but used a heuristic algorithm and could handle all basic symbols but could not parse chords. In the 1980s, a full OMR robot was built that could read sheet music placed in front of it and then play the corresponding piece on an organ [2]. The first attempt to handle handwritten music occurred in the early 1990s 1
2 End State S 0 S 1... S n S 0 p 00 p p 0n S 1 p 10 p p 1n S n p n0 p n1... p nn Observation E 0 E 1... E m S 0 p 00 p p 0m S 1 p 10 p p 1m S n p n0 p n1... p nm Start State State State S 0 S 1... S n [ p 0 p 1... p n ] Starting Transitional Conditional Figure 1: The matrices of a HMM [9]. In 2001, Bainbridge [2] summarized the challenges of Optical Music Recognition and provided a detailed history of OMR work up to that point. However, there are still many challenges in OMR leading to numerous open problems [9] [13]. As mentioned previously, OMRs were heavily influenced by OCRs. Thus it is worthwhile to mention the various mechanisms that are used in OCRs: pattern matching, nearest neighbors, and hidden Markov models [1]. For each of these mechanisms, the image storing the text has to first be segmented into the various individual characters. Then the identification mechanism can be applied on the image of a character. Pattern matching directly compares the image to others stored in a database, and works best with typewritten text of similar fonts. Nearest neighbors and hidden Markov models compare features computed from the image against features representing the characters and choose the best match. The performance of hidden Markov models tend to excel when used for handwritten text, compare to other mechanisms [1]. Hidden Markov models are a mechanism that creates models using a type of algorithm known as machine learning. This type of algorithm makes predictions about a data set, given prior knowledge of similar data sets. In a broad sense, machine learning can be compared to human learning, in that we make observations of our surroundings and learn from it, allowing us to make relatively accurate predictions about future events. In computer science, machine learning is typically implemented by first training a model on a set of data, then using the model(s) to correctly interpret similar data sets. It is in this way that hidden Markov Models are applied to OMRs. More specifically, hidden Markov models are probabilistic models used on systems with unobservable or hidden properties, which we refer to as states. In the system it is assumed that the hidden states can be estimated using probabilities learned from observations of the system. In the case of OMRs, an HMM attempts to model the sequence of states for a given musical symbol (whole note, quarter note, rest, etc.), based off of the features of the musical symbol (called observations) [10] [11]. The sequence of hidden states differs for each symbol, but the number of states will remain constant throughout. Digitally, an HMM is composed of three probability matrices, as shown in Figure 1. The first matrix holds the state-transition probabilities, which is of size n n, with n being the number of states. These probabilities indicate the likelihood of transitioning from one hidden state S i to another hidden state S j. The second matrix holds the conditional probabilities, 2
3 Symbol # of Symbols Neural Network Nearest Neighbour Support Vector Machines Hidden Markov Models Quarter Rest 63 85% 100% 100% 83% Treble Clef % 100% 100% 58% Flat % 100% 99% 96% Sharp 13 97% 100% 100% 99% Natural % 100% 100% 95% Figure 2: Comparison Data which is of size n m, with m being the number of possible observations. These probabilities indicate the likelihood of being in a hidden state S i, given a certain observation E j. The third matrix holds the initial-state probabilities, of size n, with probabilities indicating the likelihood of starting in a hidden state S i. The paper written by A. Rebelo et al. [12] is a comparative study of the performance of four of the main OMR algorithms on both handwritten and printed music symbols. Performance of the algorithms fluctuated based on which type of symbols were being tested, however their study found that the performance of HMMs for both handwritten and printed sheet music was lower than the other three algorithms. Their results are shown in Figure 2. This was a surprising result considering the algorithm s performance in OCRs on handwritten text, as stated by Arica and Yarman-Vural [1]. It was expected that HMMs would have comparable performance to the other algorithms for OMRs, since it outperforms them in OCRs on handwritten text. Due to this, A. Rebelo et al. listed hidden Markov models as an open problem to be investigated further. 3 Methods Most OMRs generally process sheet music in a series of stages. They first implement a pre-processing stage that prepares the sheet music for segmentation. Then they segment the image into the various musical symbols, and finally identify the musical symbols. Once that has been accomplished, they can then reconstruct the piece of music in a digital form. The whole process is fully described in the paper by Bainbridge and Bell [2], as well in the paper by Rebelo [13]. Our research began with an attempt to create a simplistic OMR in which to implement the HMM algorithm. As our hypothesis dealt with an increase in training data, we first needed to create the pre-processing and segmentation step of the OMR to allow us to generate a larger set of training data. Once this was achieved, we then created the identification step of the OMR. Our implementation of the simplistic OMR is described in the following sections. Figure 3 shows the general structure of the created software. Observation Extraction deals with the beginning of the identification step, creating observations for each image. Expectation Maximization is the algorithm used for creation of the HMMs, while Viterbi is the algorithm used for testing of the HMMs. The full identification step would only use Viterbi, once the 3
4 Figure 3: Basic OMR Structure HMMs were created. 3.1 Pre-Processing OMRs do not always need a pre-processing stage, however aligning and removal of staff lines eased the segmentation process by ensuring that most musical symbols would be surrounded by white space, and arranged in a line. Other OMRs have included the staff lines, which then become part of the model for each symbol. However, since the limited training data acquired from Rebelo et al. [12] did not include staff lines, we removed staff lines so that the training data we generated would be consistent with the acquired data. A piece of sheet music in Western music notation consists of multiple grand staves per page. These grand staves are denoted by curly braces at the beginning of sections of staff lines. Each grand stave is made up of two staves, each containing five staff lines. Notes and other symbols are placed among staff lines, where staff lines indicate the pitch of the note. The beginning of the stave contains the clef symbol as well as the time signature. This is illustrated in Figure 4. To ease the segmentation process of the OMR, we aligned the staff lines by musical staves. An alternate approach would require keeping track of the top and bottom of each grand stave, as well as the midpoint of the grand stave, and incorporating the values into the process for segmenting symbols. Thus it was simpler to find these values and then use them to split up the sheet music by staves. This was done by creating a horizontal histogram of the black pixels in an image as shown in Figure 4 and 5, and using the histogram to find the first instance with no black pixels in the horizontal direction above and below grand staves. Then the grand staves were split up by calculating the midpoint between the top and bottom of the staves. Once the staves were split, they were then aligned by staff line in the form show in Figure 6 Removal of the staff lines is not necessary, however as mentioned previously, this ensured that our segmented images would match our acquired data set. It also eases the segmentation process by ensuring that there is vertical white space between every individual symbol. The removal of staff lines was accomplished at the same time as aligning the grand staves. The peaks in the horizontal histogram corresponded to staff line locations. Thus during aligning, the lines which corresponded to staff lines were neglected, leading to an aligned piece of sheet music with staff lines removed, as shown in Figure 6. 4
5 Figure 4: Sample Sheet Music Figure 5: Sample staff line histogram Figure 6: Aligned sheet music with staff lines removed 5
6 Figure 7: Segmentation of Beam 3.2 Segmentation Although there were other techniques that could be used in place of segmentation, such as implementing a sliding window [10], we decided to use segmentation. This was done because segmentation simplified the feature extraction step, and created images for each musical symbol consistent with our acquired data. Creating an image for each musical symbol on a given piece of sheet music was beneficial since afterwards all images of a specific symbol could then be used to train an HMM for that symbol. The implementation of segmentation was done in multiple steps. First the program was given the pre-processed image, of the form shown in Figure 6. A vertical histogram of the number of black pixels was then created for the image, which was used to segment the image into sections of symbols by identifying white space. Musical symbols such as beams required multiple passes of the segmentation due to the beam interfering with the simple white space based segmentation. Using the initial vertical histogram of the aligned image, sections of symbols were segmented by splitting the image where the vertical histogram had a value of 0. In the pre-processing stage, the values n and d were found where n was the height of a single staff line and d was the height between two adjacent staff lines. For each symbol, we set w to be the width of a symbol and if w < 2d, then the symbol was classified as a single musical symbol. Otherwise, the symbol was classified as a beam and required additional segmentation [7]. This condition is valid because of the structure of Western music notation. No single symbol is wider than two staff line spaces To break up a beam into the individual symbols, an additional vertical histogram was created and is shown in Figure 7. This histogram was created by computing the maximum number of adjacent black pixels for that column, whose value was between 2n and 2n + d. This ensured that only beams and notes were counted in the histogram, excluding note stems. Beams were then removed from the histogram by average the histogram values and setting to 0 any column in the histogram whose value was less than the average. This process ensured that the beams themselves were not classified as an individual symbol, and allowed the musical symbols to be extracted out of the beam [5]. 6
7 3.3 Creation of Training Data Since our program could already segment an individual piece of sheet music into the subsequent symbols, the creation of the training data involved acquiring full pages of sheet music that could then be converted into the symbols required for the testing data. To do so, we simply had to run the pre-processing and segmentation stages of the program on the acquired sheet music images. Because our program removed the staff lines using a histogram method, all sheet music acquired for creation of the training data was required to be completely level, ensuring that a single staff line remained on the same row of pixels for the entire width of the image. Thus older pieces of scanned music that were tilted or fuzzy could not be used. All words that appeared in the music, such as verses, were removed using GIMP, a free open source alternative to photoshop, to ensure that only musical symbols would be segmented. Changes to the pre-processing stage of the OMR could straighten staff lines, remove blur, and potentially remove words, however this outside the scope of this research. Once the appropriate sheet music was acquired, the segmentation component of our OMR program was used to segment the sheet music into the individual musical symbols. Due to the nature of sheet music, every page of sheet music was segmented into hundreds of images of musical symbols. While each symbol of the same type could be similar to other symbols of the same type, due to their different positioning in the sheet music variation did occur. This was mostly due to the pre-processing stage which potentially removed sections of an image which overlapped with staff lines. Variation could also occur due to parts of the symbol image overlapping with other neighboring symbols. Because of the method in which segmentation created the images, these symbols were not sorted by type of musical symbol, which was needed for the training of the HMMs. Thus a manual sorting of all 18,000 images segmented from the 30 unique pages of sheet music was required. 3.4 Observation Extraction Before training of an HMM can begin, observation sequences had to be created for every musical symbol. These sequences contained a series of integer observations in the range of 0 to m, corresponding to the observations in the conditional probability matrix as shown in Figure 1. To do so, features had to first be extracted from the image, creating feature vectors, and then these vectors were combined and normalized to create the sequences of observations. Features were extracted from the image using a sliding window, with a width of 2 pixels [10]. For each window, six features were computed and normalized to a value between 0 and 1. These features were chosen based off the work by Pugin [10]. Feature 1 corresponds to 1 the number of distinct connected components of black pixels, and is computed using 1+n where n is the number of distinct connect black pixels. Features 2 and 3 calculate c x and c y, the gravity centers of x and y respectively. For each equation let w = width of window, h = height of window, and for each distinct black zone i, c i x, and c i y correspond to its x and y gravity center, and a i corresponds to its area. The computations for these features are shown in equations 1 and 2 respectively. 7
8 n i=1 c x = ci x a i A w n i=1 c y = ci y a i A h (1) (2) Feature 4 corresponds to the area of the largest black element, computed using a(n i) where S a(n i ) is the area of the largest black element, and S is the area of the window. Feature 5 corresponds to the area of the smallest white element, computed using a(n j) where a(n S j ) is the are of the smallest white element. The final feature corresponds to the total area of the black elements. Since Pugin s work kept staff lines in the images, to accurately determine the total area of the black elements they introduced a weighting mask. Since we removed staff lines from images, we simply computed this value using the equation A where A is the S total area of the black elements. After feature extraction for each window, the 6 features were converted to an integer in the range 0 to m. This was done by averaging the 6 features. Then, the minimum and maximum averaged value was computed across all the windows. This allowed us to create bins for every number in the range 0 to 6, and then place each averaged value in the appropriate bin, scaling them to the appropriate integer observation. 3.5 Expectation Maximization Creating and training a HMM requires the use of the Expectation Maximization (EM) Algorithm. First the type of HMM model, and the initial estimates of the HMM model must be chosen. Different types of models are described in detail in the paper by Chen [3]. Based off of Pugin s work [10], as well as work by Mohamed [8] in handwritten word recognition, we decided to use left-right HMMs. In these, transitions from hidden states can only occur from state i to state j where j >= i. Initial estimates were chosen to be uniform, as these have been shown to work well [11]. Once the model is chosen, the parameters of the HMMs have to be determined. The EM algorithm makes makes use of the forward-backward algorithm, to compute the expectation. We will begin with an explanation of the forward-backward algorithm Expectation In the forward-backward algorithm, the conditional probabilities are estimated given a sequence of observations. It does this in two passes. First, it computes the forward probabilities, which is the likelihood of being in a state given the first k observations. Second, it computes the backward probabilities, which is the likelihood of being in a state and seeing the remaining observations from the k th observation onward. Figure 8 illustrates the process of computing the forward probabilities for an HMM with 2 states. For each node S i, α it is the probabilities of being in S i having observed t observations. For each transition, p jit is computed from the probability of being in state i given observation t and from the probabilities of transitioning from S j to S i, where s stands for Start. The backward probabilities are computed the same 8
9 α 00 = p s00 S 0 p 001 = P (S 0 S 0 ) P (e 1 S 0 ) S 0 α 01 = α 00 p 001 +α 10 p p s00 = P (S 0 Start) P (e 0 S 0 ) Start p s10 = P (S 1 Start) P (e 0 S 1 ) p 101 = P (S 0 S 1 ) P (e 1 S 0 ) p 011 = P (S 1 S 0 ) P (e 1 S 1 ) α 10 = p s10 S 1 p 111 = P (S 1 S 1 ) P (e 1 S 1 ) S 1 α 11 = α 00 p α 10 p Figure 8: Forward-Backward Algorithm Diagram way as the forward probabilities, traveling through the observation sequence backwards. The algorithm for the forward and backward passes are explained in the paper by Eisner [4]. Once the forward and backward probabilities are computed, temporary values are computed using these probabilities to be used in the maximization step. These temporary variables are γ and ξ, and are computed using the following equations: γ it = α it βit N l=1 α lt β lt ξ ijt = α it a ij β j(t+1) b j(t+1) N l=1 α lt Where N is the number of states, T is the last observation, and a and b stand for the transitional and conditional matrices respectively Maximization Using the temporary variables γ and xi computed above for each sequence, the EM algorithm then iteratively re-estimates the three HMM matrices. This is done by running the forward-backward on each observation sequence, and then storing the temporary variables computed from each. Using these temporary variables, the three matrices are re-estimated using the following equations: π i = γ i (0) T 1 t=1 a ij = ξ ijt T 1 b i(et) = t=1 γ it T t=1 γ it bin(e t, t) T t=1 γ it 9
10 α 00 = p s00 p s00 = P (S 0 Start) P (e 0 S 0 ) Start S 0 p 101 = P (S 0 S 1 ) P (e 1 S 0 ) p s10 = P (S 1 Start) P (e 0 S 1 ) α 10 = p s10 S 1 p 001 = P (S 0 S 0 ) P (e 1 S 0 ) p 011 = P (S 1 S 0 ) P (e 1 S 1 ) p 111 = P (S 1 S 1 ) P (e 1 S 1 ) S 0 α 01 = max(α 00 p 001, α 10 p 101 )... S 1... α 11 = max(α 00 p 011, α 10 p 111 ) p 00T p 00T p 00T p 11T S 0 α 0T = max(α 0(T 1) p 00T, α 1(T 1) p 10T ) p 0(T +1) End p 1(T +1) S 1 α 1T = max(α 0(T 1) p 01T, α 1(T 1) p 11T ) α (T +1) = max( α 0T p 0(T +1), α 1T p 1(T +1) ) Figure 9: Viterbi Algorithm Diagram Where bin(e t, t) is 1 if e t = t and 0 if not, and π is the starting probability matrix. Each observation sequence produces its own re-estimation of the matrices. These re-estimations are then combined by averaging the values of each matrix. Then the full EM process, using the forward-backward algorithm and then re-estimating matrices, is repeated until the log-likelihood converges, resulting in one final set of matrices. The log-likelihood is computed by computing the sum of the values in the forward step of the forward-backward algorithm, computing the log of these, and then for each observation sequence, summing these logs. As the matrices are re-estimated, these values should decrease, indicating that our probabilities lead to smooth transitions. 3.6 Viterbi Once the HMMs have been created for each musical symbol, the Viterbi algorithm is used to determine how closely a symbol follows a given hidden Markov model. This is done in a similar method to the forward-backward algorithm. Figure 9 shows the viterbi trellis diagram for a hidden Markov model with 2 states, derived from the paper by Levy [6]. For each node S i, α i t is the probability of the most likely path through the trellis, ending at observation t in S i. This computed similarly to the forward-backwards algorithm, except instead of adding each transition, α it p jit, the maximum of the transitions is computed. The maximum value at the end state determines how closely a symbol follows the model [6]. Using the Viterbi algorithm we then tested our models. For each image, we ran the Viterbi algorithm with each of the models for each type of symbol. The model which produced the highest probability was returned, indicating that the image was likely that type of symbol. 10
11 Quarter Rest Natural Treble Clef Flat Sharp Symbol # of Symbols Correctly Identified Percentage Quarter Rest % Natural % Treble Clef % Flat % Sharp % Figure 10: Data 4 Results We successfully created a simplistic OMR with the ability to extract individual musical symbols out of an image of sheet music. These symbols can then, with the created software, be converted into a series of observations for use by the EM algorithm and the Viterbi Algorithm. The created OMR software is available at research/. The created software was successful in generating accurate HMMs for various musical symbols and testing the generated HMMs. HMMs were generated for Quarter Rests, Naturals, Treble Clefs, Sharps, and Flats, using the EM algorithm as detailed above. Other musical symbols were not considered for a variety of reasons. Notes were not included due to our segmentation program not handling chords. We were unsure if multiple notes versus single notes would be problematic, and thus did not consider them. The Bass Clef was not considered due to the nature of the clef, which has 3 distinct pieces. Since our segmentation program broke the symbol up, we did not include the symbol. The remaining common musical symbols were not included due to a limited number of the symbols, which would not have resulted in statistically significant results. The trained HMMs were then tested using the Viterbi algorithm, similar to Rebelo [12]. We ran the Viterbi algorithm in stages, once for each set of observation sequences of the same musical symbol, and then we counted the number of correct identifications. This was done to automate the counting process. The results are shown in Figure 10. From the data, it is clear that HMMs worked well for all of the trained symbols, and outperformed the HMM in the comparative study by Rebelo [12]. See Figure 2 for comparison. Thus we have shown that different implementations of HMMs can lead to higher performance. As we also had a greater number of training data, this could indicate that higher training data does improve performance, as hypothesized. However, since Rebelo also included deformations in their data set, this could be the reason for differences in performance. 11
12 5 Future Work The goal of the research was to accurately identify a symbol that was extracted from a piece of sheet music. For the symbols chosen, this was done successfully. However, there may still the potential for improvement in the models, to improve performance to the level of other algorithms. We have currently been using six states for the Markov models, based on our number of observations, as stated in the work by Pugin [10]. A change in the number of states may improve the performance of the models. Thus a goal for future work could be to run the program with a variety of different states and determine the number of states which leads to the highest performance of the HMMs. Changing the types of features computed may also result in an increased performance of the HMMs, however this seems unlikely as in the work by Pugin, they found that the six features used were optimal for their OMR [10]. However, an investigation into other features that could be implemented might be useful. This investigation is a bit more complicated than optimizing the number of states, as it requires reworking the feature extraction part of the program. However it is certainly very feasible, and if changing the number of states does not result in a better performance, this would be something to investigate. A secondary goal for future work would be to create a fully working OMR. Our OMR can currently only segment sheet music and then identify a given music symbol from the sheet music. A full OMR software not only identifies symbols but is able to correctly interpret an entire piece of music. Thus future work could entail adding the component that would recreate the piece of sheet music digitally, once the symbols have been identified. To do so, we will need to use a digital language that can accurately represent any given piece of sheet music, and then store a given piece of sheet music in this digital form. The most common digital language for music currently is musicxml. While there are already OMRs which can do this, incorporating this component would make the software more complete. We would also like the program to be able to segment a greater variety of sheet music as well. Currently it is assumed that the staff lines are completely horizontal with no deviation, and it is assumed that the given symbols are not blurred in any way, as mentioned previously. Thus, future work can include making the software more robust, in order to be able to handle scanned sheet music that may be tilted or blurred. Currently the number of symbols that the program can accurately identify is also limited. Thus, more hidden Markov models should be made to encompass all possible symbols that could be encountered. This would require further improvement in segmentation to break up chords into single notes, and to keep the Bass Clef connected. More data would also need to be acquired to produce the other models. Breaking up the chords will likely require implementation of a vertical histogram on the chords, to determine where individual notes are and break them up accordingly. This is outside the scope of this research. 6 Conclusion The software created for this research is currently functional but has room for improvement. The pre-processing stage successfully removes staff lines and aligns the piece by musical staves. The segmentation process accurately segments all symbols including beams, but 12
13 does not break up chords into the individual notes. However this is sufficient for the testing of HMMs. The feature extraction on each segmented symbol is also fully functional, resulting in integer observation sequences, as are the two algorithms required for this research: EM, and Viterbi. The evaluation of our program shows that HMMs work quite well for all of the symbols used, with accuracy above 95%. While there is more work that can be done to further improve the models, the data collected shows that it can be a viable algorithm for OMRs, and had better performance compared to prior work. Overall, we were successfully able to implement a working simplistic OMR and fully completed the identification process. References [1] N. Arica and F. T. Yarman-Vural. An overview of character recognition focused on off-line handwriting. IEEE Transactions on Systems, Man, and Cyber Part C, 31(2): , May [2] David Bainbridge and Tim Bell. The challenge of optical music recognition. Computers and the Humanities, 35(2):95 121, [3] Mou-Yen Chen, A. Kundu, and Jian Zhou. Off-line handwritten word recognition using a hidden markov model type stochastic network. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 16(5): , May [4] Jason Eisner. An interactive spreadsheet for teaching the forward-backward algorithm. In Dragomir Radev and Chris Brew, editors, Proceedings of the ACL Workshop on Effective Tools and Methodologies for Teaching NLP and CL, pages 10 18, Philadelphia, July [5] Susan Ella. George. Visual perception of music notation : on-line and off-line recognition / Susan Ella George. IRM Press Hershey, Pa, [6] Roger Levy. Lecture notes on linguistics/cse 256: Hidden markov model inference with the viterbi algorithm: a mini-example, Winter [7] S. Marinai and P. Nesi. Projection based segmentation of musical sheets. In Document Analysis and Recognition, ICDAR 99. Proceedings of the Fifth International Conference on, pages , Sep [8] M. Mohamed and P. Gader. Handwritten word recognition using segmentation-free hidden markov modeling and segmentation-based dynamic programming techniques. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 18(5): , May [9] Jiri Novotn and Jaroslav Pokorn. Introduction to optical music recognition: Overview and practical challenges. In DATESO,
14 [10] Laurent Pugin. Optical music recognition of early typographic prints using hidden markov models. In Proceedings of the 7th International Conference on Music Information Retrieval, Victoria (BC), Canada, October ismir.net/papers/ismir06152_paper.pdf. [11] Lawrence R. Rabiner. Readings in speech recognition. chapter A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition, pages Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, [12] A. Rebelo, G. Capela, and JaimeS. Cardoso. Optical recognition of music symbols. International Journal on Document Analysis and Recognition (IJDAR), 13(1):19 31, [13] Ana Rebelo, Ichiro Fujinaga, Filipe Paszkiewicz, Andre R. S. Marcal, Carlos Guedes, and Jaime S. Cardoso. Optical music recognition: state-of-the-art and open issues. International Journal of Multimedia Information Retrieval, 1(3): ,
A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models
A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models Gleidson Pegoretti da Silva, Masaki Nakagawa Department of Computer and Information Sciences Tokyo University
More informationWhich Line Is it Anyway : Clustering for Staff Line Removal in Optical Music Recognition. Sam Vinitsky
Which Line Is it Anyway : Clustering for Staff Line Removal in Optical Music Recognition Sam Vinitsky Optical Music Recognition: An Overview Optical Music Recognition: An Overview Sheet Music Primer Sheet
More informationAdvanced Image Processing, TNM034 Optical Music Recognition
Advanced Image Processing, TNM034 Optical Music Recognition Linköping University By: Jimmy Liikala, jimli570 Emanuel Winblad, emawi895 Toms Vulfs, tomvu491 Jenny Yu, jenyu080 1 Table of Contents Optical
More informationInvariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction
Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction Stefan Müller, Gerhard Rigoll, Andreas Kosmala and Denis Mazurenok Department of Computer Science, Faculty of
More informationCS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs
CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs Felix Wang fywang2 John Wieting wieting2 Introduction We implement a texture classification algorithm using 2-D Noncausal Hidden
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationOptical Character Recognition (OCR) for Printed Devnagari Script Using Artificial Neural Network
International Journal of Computer Science & Communication Vol. 1, No. 1, January-June 2010, pp. 91-95 Optical Character Recognition (OCR) for Printed Devnagari Script Using Artificial Neural Network Raghuraj
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationRobust line segmentation for handwritten documents
Robust line segmentation for handwritten documents Kamal Kuzhinjedathu, Harish Srinivasan and Sargur Srihari Center of Excellence for Document Analysis and Recognition (CEDAR) University at Buffalo, State
More informationOCR For Handwritten Marathi Script
International Journal of Scientific & Engineering Research Volume 3, Issue 8, August-2012 1 OCR For Handwritten Marathi Script Mrs.Vinaya. S. Tapkir 1, Mrs.Sushma.D.Shelke 2 1 Maharashtra Academy Of Engineering,
More informationShort Survey on Static Hand Gesture Recognition
Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of
More informationIdentifying Layout Classes for Mathematical Symbols Using Layout Context
Rochester Institute of Technology RIT Scholar Works Articles 2009 Identifying Layout Classes for Mathematical Symbols Using Layout Context Ling Ouyang Rochester Institute of Technology Richard Zanibbi
More informationOne Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition
One Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition Nafiz Arica Dept. of Computer Engineering, Middle East Technical University, Ankara,Turkey nafiz@ceng.metu.edu.
More informationFinal Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.
CS 188 Spring 2010 Introduction to Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators
More informationOrganizing and Summarizing Data
1 Organizing and Summarizing Data Key Definitions Frequency Distribution: This lists each category of data and how often they occur. : The percent of observations within the one of the categories. This
More informationABJAD: AN OFF-LINE ARABIC HANDWRITTEN RECOGNITION SYSTEM
ABJAD: AN OFF-LINE ARABIC HANDWRITTEN RECOGNITION SYSTEM RAMZI AHMED HARATY and HICHAM EL-ZABADANI Lebanese American University P.O. Box 13-5053 Chouran Beirut, Lebanon 1102 2801 Phone: 961 1 867621 ext.
More informationAssignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018
Assignment 2 Unsupervised & Probabilistic Learning Maneesh Sahani Due: Monday Nov 5, 2018 Note: Assignments are due at 11:00 AM (the start of lecture) on the date above. he usual College late assignments
More informationLecture 21 : A Hybrid: Deep Learning and Graphical Models
10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation
More informationDetection of Man-made Structures in Natural Images
Detection of Man-made Structures in Natural Images Tim Rees December 17, 2004 Abstract Object detection in images is a very active research topic in many disciplines. Probabilistic methods have been applied
More informationClustering Sequences with Hidden. Markov Models. Padhraic Smyth CA Abstract
Clustering Sequences with Hidden Markov Models Padhraic Smyth Information and Computer Science University of California, Irvine CA 92697-3425 smyth@ics.uci.edu Abstract This paper discusses a probabilistic
More informationCS 532c Probabilistic Graphical Models N-Best Hypotheses. December
CS 532c Probabilistic Graphical Models N-Best Hypotheses Zvonimir Rakamaric Chris Dabrowski December 18 2004 Contents 1 Introduction 3 2 Background Info 3 3 Brute Force Algorithm 4 3.1 Description.........................................
More informationEffect of Initial HMM Choices in Multiple Sequence Training for Gesture Recognition
Effect of Initial HMM Choices in Multiple Sequence Training for Gesture Recognition Nianjun Liu, Richard I.A. Davis, Brian C. Lovell and Peter J. Kootsookos Intelligent Real-Time Imaging and Sensing (IRIS)
More informationRobust PDF Table Locator
Robust PDF Table Locator December 17, 2016 1 Introduction Data scientists rely on an abundance of tabular data stored in easy-to-machine-read formats like.csv files. Unfortunately, most government records
More informationBMI/CS Lecture #22 - Stochastic Context Free Grammars for RNA Structure Modeling. Colin Dewey (adapted from slides by Mark Craven)
BMI/CS Lecture #22 - Stochastic Context Free Grammars for RNA Structure Modeling Colin Dewey (adapted from slides by Mark Craven) 2007.04.12 1 Modeling RNA with Stochastic Context Free Grammars consider
More informationConditional Random Fields and beyond D A N I E L K H A S H A B I C S U I U C,
Conditional Random Fields and beyond D A N I E L K H A S H A B I C S 5 4 6 U I U C, 2 0 1 3 Outline Modeling Inference Training Applications Outline Modeling Problem definition Discriminative vs. Generative
More informationModelStructureSelection&TrainingAlgorithmsfor an HMMGesture Recognition System
ModelStructureSelection&TrainingAlgorithmsfor an HMMGesture Recognition System Nianjun Liu, Brian C. Lovell, Peter J. Kootsookos, and Richard I.A. Davis Intelligent Real-Time Imaging and Sensing (IRIS)
More informationHandwritten Gurumukhi Character Recognition by using Recurrent Neural Network
139 Handwritten Gurumukhi Character Recognition by using Recurrent Neural Network Harmit Kaur 1, Simpel Rani 2 1 M. Tech. Research Scholar (Department of Computer Science & Engineering), Yadavindra College
More informationSimplifying OCR Neural Networks with Oracle Learning
SCIMA 2003 - International Workshop on Soft Computing Techniques in Instrumentation, Measurement and Related Applications Provo, Utah, USA, 17 May 2003 Simplifying OCR Neural Networks with Oracle Learning
More informationBiology 644: Bioinformatics
A statistical Markov model in which the system being modeled is assumed to be a Markov process with unobserved (hidden) states in the training data. First used in speech and handwriting recognition In
More informationThe Method of User s Identification Using the Fusion of Wavelet Transform and Hidden Markov Models
The Method of User s Identification Using the Fusion of Wavelet Transform and Hidden Markov Models Janusz Bobulski Czȩstochowa University of Technology, Institute of Computer and Information Sciences,
More informationANALYTIC WORD RECOGNITION WITHOUT SEGMENTATION BASED ON MARKOV RANDOM FIELDS
ANALYTIC WORD RECOGNITION WITHOUT SEGMENTATION BASED ON MARKOV RANDOM FIELDS CHRISTOPHE CHOISY AND ABDEL BELAID LORIA/CNRS Campus scientifique, BP 239, 54506 Vandoeuvre-les-Nancy cedex, France Christophe.Choisy@loria.fr,
More informationImage classification by a Two Dimensional Hidden Markov Model
Image classification by a Two Dimensional Hidden Markov Model Author: Jia Li, Amir Najmi and Robert M. Gray Presenter: Tzung-Hsien Ho Hidden Markov Chain Goal: To implement a novel classifier for image
More informationStaff Line Detection by Skewed Projection
Staff Line Detection by Skewed Projection Diego Nehab May 11, 2003 Abstract Most optical music recognition systems start image analysis by the detection of staff lines. This work explores simple techniques
More informationIntroduction to SLAM Part II. Paul Robertson
Introduction to SLAM Part II Paul Robertson Localization Review Tracking, Global Localization, Kidnapping Problem. Kalman Filter Quadratic Linear (unless EKF) SLAM Loop closing Scaling: Partition space
More informationOPTIMIZING A VIDEO PREPROCESSOR FOR OCR. MR IBM Systems Dev Rochester, elopment Division Minnesota
OPTIMIZING A VIDEO PREPROCESSOR FOR OCR MR IBM Systems Dev Rochester, elopment Division Minnesota Summary This paper describes how optimal video preprocessor performance can be achieved using a software
More informationLinear Discriminant Analysis in Ottoman Alphabet Character Recognition
Linear Discriminant Analysis in Ottoman Alphabet Character Recognition ZEYNEB KURT, H. IREM TURKMEN, M. ELIF KARSLIGIL Department of Computer Engineering, Yildiz Technical University, 34349 Besiktas /
More informationCharacter Recognition
Character Recognition 5.1 INTRODUCTION Recognition is one of the important steps in image processing. There are different methods such as Histogram method, Hough transformation, Neural computing approaches
More informationImage Mining: frameworks and techniques
Image Mining: frameworks and techniques Madhumathi.k 1, Dr.Antony Selvadoss Thanamani 2 M.Phil, Department of computer science, NGM College, Pollachi, Coimbatore, India 1 HOD Department of Computer Science,
More information27: Hybrid Graphical Models and Neural Networks
10-708: Probabilistic Graphical Models 10-708 Spring 2016 27: Hybrid Graphical Models and Neural Networks Lecturer: Matt Gormley Scribes: Jakob Bauer Otilia Stretcu Rohan Varma 1 Motivation We first look
More informationWriter Identification In Music Score Documents Without Staff-Line Removal
Writer Identification In Music Score Documents Without Staff-Line Removal Anirban Hati, Partha P. Roy and Umapada Pal Computer Vision and Pattern Recognition Unit Indian Statistical Institute Kolkata,
More informationA Generalized Method to Solve Text-Based CAPTCHAs
A Generalized Method to Solve Text-Based CAPTCHAs Jason Ma, Bilal Badaoui, Emile Chamoun December 11, 2009 1 Abstract We present work in progress on the automated solving of text-based CAPTCHAs. Our method
More informationAnalysis of Image and Video Using Color, Texture and Shape Features for Object Identification
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 16, Issue 6, Ver. VI (Nov Dec. 2014), PP 29-33 Analysis of Image and Video Using Color, Texture and Shape Features
More informationTopological Mapping. Discrete Bayes Filter
Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be
More informationCluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010
Cluster Analysis Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX 7575 April 008 April 010 Cluster Analysis, sometimes called data segmentation or customer segmentation,
More informationLECTURE 6 TEXT PROCESSING
SCIENTIFIC DATA COMPUTING 1 MTAT.08.042 LECTURE 6 TEXT PROCESSING Prepared by: Amnir Hadachi Institute of Computer Science, University of Tartu amnir.hadachi@ut.ee OUTLINE Aims Character Typology OCR systems
More informationCSI5387: Data Mining Project
CSI5387: Data Mining Project Terri Oda April 14, 2008 1 Introduction Web pages have become more like applications that documents. Not only do they provide dynamic content, they also allow users to play
More informationII. WORKING OF PROJECT
Handwritten character Recognition and detection using histogram technique Tanmay Bahadure, Pranay Wekhande, Manish Gaur, Shubham Raikwar, Yogendra Gupta ABSTRACT : Cursive handwriting recognition is a
More informationNeural Network Classifier for Isolated Character Recognition
Neural Network Classifier for Isolated Character Recognition 1 Ruby Mehta, 2 Ravneet Kaur 1 M.Tech (CSE), Guru Nanak Dev University, Amritsar (Punjab), India 2 M.Tech Scholar, Computer Science & Engineering
More informationHMM-Based Handwritten Amharic Word Recognition with Feature Concatenation
009 10th International Conference on Document Analysis and Recognition HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation Yaregal Assabie and Josef Bigun School of Information Science,
More informationTo earn the extra credit, one of the following has to hold true. Please circle and sign.
CS 188 Spring 2011 Introduction to Artificial Intelligence Practice Final Exam To earn the extra credit, one of the following has to hold true. Please circle and sign. A I spent 3 or more hours on the
More informationThe Detection of Faces in Color Images: EE368 Project Report
The Detection of Faces in Color Images: EE368 Project Report Angela Chau, Ezinne Oji, Jeff Walters Dept. of Electrical Engineering Stanford University Stanford, CA 9435 angichau,ezinne,jwalt@stanford.edu
More informationTHE preceding chapters were all devoted to the analysis of images and signals which
Chapter 5 Segmentation of Color, Texture, and Orientation Images THE preceding chapters were all devoted to the analysis of images and signals which take values in IR. It is often necessary, however, to
More informationProject Report for EE7700
Project Report for EE7700 Name: Jing Chen, Shaoming Chen Student ID: 89-507-3494, 89-295-9668 Face Tracking 1. Objective of the study Given a video, this semester project aims at implementing algorithms
More informationLearning C language Programming with executable flowchart language
Paper ID #8872 Learning C language Programming with executable flowchart language Prof. Cho Sehyeong, Myong Ji University 1992 Ph.D. in Computer Science, Pennsylvania State University 1992-1999 Electronics
More informationIntroduction to Graphical Models
Robert Collins CSE586 Introduction to Graphical Models Readings in Prince textbook: Chapters 10 and 11 but mainly only on directed graphs at this time Credits: Several slides are from: Review: Probability
More informationHidden Markov Model for Sequential Data
Hidden Markov Model for Sequential Data Dr.-Ing. Michelle Karg mekarg@uwaterloo.ca Electrical and Computer Engineering Cheriton School of Computer Science Sequential Data Measurement of time series: Example:
More informationHandwritten Character Recognition with Feedback Neural Network
Apash Roy et al / International Journal of Computer Science & Engineering Technology (IJCSET) Handwritten Character Recognition with Feedback Neural Network Apash Roy* 1, N R Manna* *Department of Computer
More informationIndian Multi-Script Full Pin-code String Recognition for Postal Automation
2009 10th International Conference on Document Analysis and Recognition Indian Multi-Script Full Pin-code String Recognition for Postal Automation U. Pal 1, R. K. Roy 1, K. Roy 2 and F. Kimura 3 1 Computer
More informationCLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS
CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of
More informationOptimization of HMM by the Tabu Search Algorithm
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 20, 949-957 (2004) Optimization of HMM by the Tabu Search Algorithm TSONG-YI CHEN, XIAO-DAN MEI *, JENG-SHYANG PAN AND SHENG-HE SUN * Department of Electronic
More informationTexture Sensitive Image Inpainting after Object Morphing
Texture Sensitive Image Inpainting after Object Morphing Yin Chieh Liu and Yi-Leh Wu Department of Computer Science and Information Engineering National Taiwan University of Science and Technology, Taiwan
More informationInformation Retrieval and Web Search Engines
Information Retrieval and Web Search Engines Lecture 7: Document Clustering December 4th, 2014 Wolf-Tilo Balke and José Pinto Institut für Informationssysteme Technische Universität Braunschweig The Cluster
More informationSUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract
More informationRegularization and Markov Random Fields (MRF) CS 664 Spring 2008
Regularization and Markov Random Fields (MRF) CS 664 Spring 2008 Regularization in Low Level Vision Low level vision problems concerned with estimating some quantity at each pixel Visual motion (u(x,y),v(x,y))
More informationFrom Gaze to Focus of Attention
From Gaze to Focus of Attention Rainer Stiefelhagen, Michael Finke, Jie Yang, Alex Waibel stiefel@ira.uka.de, finkem@cs.cmu.edu, yang+@cs.cmu.edu, ahw@cs.cmu.edu Interactive Systems Laboratories University
More informationAbstract. Problem Statement. Objective. Benefits
Abstract The purpose of this final year project is to create an Android mobile application that can automatically extract relevant information from pictures of receipts. Users can also load their own images
More informationNote Set 4: Finite Mixture Models and the EM Algorithm
Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for
More informationA Fuzzy ARTMAP Based Classification Technique of Natural Textures
A Fuzzy ARTMAP Based Classification Technique of Natural Textures Dimitrios Charalampidis Orlando, Florida 328 16 dcl9339@pegasus.cc.ucf.edu Michael Georgiopoulos michaelg @pegasus.cc.ucf.edu Takis Kasparis
More informationOverview. Frequency Distributions. Chapter 2 Summarizing & Graphing Data. Descriptive Statistics. Inferential Statistics. Frequency Distribution
Chapter 2 Summarizing & Graphing Data Slide 1 Overview Descriptive Statistics Slide 2 A) Overview B) Frequency Distributions C) Visualizing Data summarize or describe the important characteristics of a
More informationConstraints in Particle Swarm Optimization of Hidden Markov Models
Constraints in Particle Swarm Optimization of Hidden Markov Models Martin Macaš, Daniel Novák, and Lenka Lhotská Czech Technical University, Faculty of Electrical Engineering, Dep. of Cybernetics, Prague,
More informationNoteWorthy Composer Version 2.75a
NoteWorthy Composer Version 2.75a Tutorial Table of Contents Creating a New File... 3 Adding, Naming and Connecting Staves... 4 Saving the File... 5 Clef, Time Signature, and Key Signature Entry... 6 Note
More informationSpeech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri
Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute eugenew@cs.nyu.edu Slide Credit: Mehryar Mohri Speech Recognition Components Acoustic and pronunciation model:
More informationPixels. Orientation π. θ π/2 φ. x (i) A (i, j) height. (x, y) y(j)
4th International Conf. on Document Analysis and Recognition, pp.142-146, Ulm, Germany, August 18-20, 1997 Skew and Slant Correction for Document Images Using Gradient Direction Changming Sun Λ CSIRO Math.
More informationData Partitioning. Figure 1-31: Communication Topologies. Regular Partitions
Data In single-program multiple-data (SPMD) parallel programs, global data is partitioned, with a portion of the data assigned to each processing node. Issues relevant to choosing a partitioning strategy
More informationKeywords Connected Components, Text-Line Extraction, Trained Dataset.
Volume 4, Issue 11, November 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Language Independent
More informationMachine Problem 8 - Mean Field Inference on Boltzman Machine
CS498: Applied Machine Learning Spring 2018 Machine Problem 8 - Mean Field Inference on Boltzman Machine Professor David A. Forsyth Auto-graded assignment Introduction Mean-Field Approximation is a useful
More informationChapter 3. Speech segmentation. 3.1 Preprocessing
, as done in this dissertation, refers to the process of determining the boundaries between phonemes in the speech signal. No higher-level lexical information is used to accomplish this. This chapter presents
More informationCS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning
CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning Justin Chen Stanford University justinkchen@stanford.edu Abstract This paper focuses on experimenting with
More informationHandwritten Word Recognition using Conditional Random Fields
Handwritten Word Recognition using Conditional Random Fields Shravya Shetty Harish Srinivasan Sargur Srihari Center of Excellence for Document Analysis and Recognition (CEDAR) Department of Computer Science
More informationDecoding recaptcha. COSC 174 Term project Curtis R. Jones & Jacob Russell
Decoding recaptcha COSC 174 Term project Curtis R. Jones & Jacob Russell May 28, 2012 Curtis Jones & Jacob Russell Decoding recaptcha Page 1 Project Objective Our project is on solving recaptcha puzzles.
More informationHANDWRITTEN GURMUKHI CHARACTER RECOGNITION USING WAVELET TRANSFORMS
International Journal of Electronics, Communication & Instrumentation Engineering Research and Development (IJECIERD) ISSN 2249-684X Vol.2, Issue 3 Sep 2012 27-37 TJPRC Pvt. Ltd., HANDWRITTEN GURMUKHI
More informationFREEMAN CODE BASED ONLINE HANDWRITTEN CHARACTER RECOGNITION FOR MALAYALAM USING BACKPROPAGATION NEURAL NETWORKS
FREEMAN CODE BASED ONLINE HANDWRITTEN CHARACTER RECOGNITION FOR MALAYALAM USING BACKPROPAGATION NEURAL NETWORKS Amritha Sampath 1, Tripti C 2 and Govindaru V 3 1 Department of Computer Science and Engineering,
More informationUnsupervised Learning
Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover
More informationFrequency Distributions
Displaying Data Frequency Distributions After collecting data, the first task for a researcher is to organize and summarize the data so that it is possible to get a general overview of the results. Remember,
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationProbabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation
Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation Daniel Lowd January 14, 2004 1 Introduction Probabilistic models have shown increasing popularity
More informationText-Tracking Wearable Camera System for the Blind
2009 10th International Conference on Document Analysis and Recognition Text-Tracking Wearable Camera System for the Blind Hideaki Goto Cyberscience Center Tohoku University, Japan hgot @ isc.tohoku.ac.jp
More informationIndoor Object Recognition of 3D Kinect Dataset with RNNs
Indoor Object Recognition of 3D Kinect Dataset with RNNs Thiraphat Charoensripongsa, Yue Chen, Brian Cheng 1. Introduction Recent work at Stanford in the area of scene understanding has involved using
More informationTime series, HMMs, Kalman Filters
Classic HMM tutorial see class website: *L. R. Rabiner, "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition," Proc. of the IEEE, Vol.77, No.2, pp.257--286, 1989. Time series,
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationProduction of Video Images by Computer Controlled Cameras and Its Application to TV Conference System
Proc. of IEEE Conference on Computer Vision and Pattern Recognition, vol.2, II-131 II-137, Dec. 2001. Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System
More informationImage Coding with Active Appearance Models
Image Coding with Active Appearance Models Simon Baker, Iain Matthews, and Jeff Schneider CMU-RI-TR-03-13 The Robotics Institute Carnegie Mellon University Abstract Image coding is the task of representing
More informationProblem definition Image acquisition Image segmentation Connected component analysis. Machine vision systems - 1
Machine vision systems Problem definition Image acquisition Image segmentation Connected component analysis Machine vision systems - 1 Problem definition Design a vision system to see a flat world Page
More informationNOVATEUR PUBLICATIONS INTERNATIONAL JOURNAL OF INNOVATIONS IN ENGINEERING RESEARCH AND TECHNOLOGY [IJIERT] ISSN: VOLUME 2, ISSUE 1 JAN-2015
Offline Handwritten Signature Verification using Neural Network Pallavi V. Hatkar Department of Electronics Engineering, TKIET Warana, India Prof.B.T.Salokhe Department of Electronics Engineering, TKIET
More informationOptimization of vehicle licence plate segmentation and symbol recognition
Optimization of vehicle licence plate segmentation and symbol recognition R.P. van Heerden and E.C. Botha Department of Electrical, Electronic and Computer engineering University of Pretoria, South Africa
More informationTopics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning
Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound Lecture 12: Deep Reinforcement Learning Types of Learning Supervised training Learning from the teacher Training data includes
More informationCharacter Recognition Using Matlab s Neural Network Toolbox
Character Recognition Using Matlab s Neural Network Toolbox Kauleshwar Prasad, Devvrat C. Nigam, Ashmika Lakhotiya and Dheeren Umre B.I.T Durg, India Kauleshwarprasad2gmail.com, devnigam24@gmail.com,ashmika22@gmail.com,
More informationCIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]
CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.
More informationInformation Retrieval and Web Search Engines
Information Retrieval and Web Search Engines Lecture 7: Document Clustering May 25, 2011 Wolf-Tilo Balke and Joachim Selke Institut für Informationssysteme Technische Universität Braunschweig Homework
More informationResearch on Emotion Recognition for Facial Expression Images Based on Hidden Markov Model
e-issn: 2349-9745 p-issn: 2393-8161 Scientific Journal Impact Factor (SJIF): 1.711 International Journal of Modern Trends in Engineering and Research www.ijmter.com Research on Emotion Recognition for
More information