GPU Accelerated Model Combination for Robust Speech Recognition and Keyword Search
|
|
- Grace Flowers
- 5 years ago
- Views:
Transcription
1 GPU Accelerated Model Combination for Robust Speech Recognition and Keyword Search Wonkyum Lee Jungsuk Kim Ian Lane Electrical and Computer Engineering Carnegie Mellon University March 26, 1
2 Overview Introduc4on Acous4c Model Acous4c Model combina4on GPU Accelerated Model Combina4on Evalua4on Results Summary 2
3 Introduc4on 3
4 Introduc4on ASR (Automa4c Speech Recogni4on) Acoustic Model Lexicon Feature Extraction Decoder Word #1 Word String Word #2 Language Model 4
5 Introduc4on KWS (Keyword Search) Indexer Keyword Search Task Keyword: Speech- To- Text (by ASR) GTC Welcome to GTC two thousands frourteen Thousands of Hours of Indexed Audio Hit: GTC 5
6 Introduc4on Speech RecogniBon Speech to Text EvaluaBon Metric: Word Error Rate KWS Spot the Keyword in Audio EvaluaBon Metric: Actual Term Weighted Value - > Both Tasks require Robust ASR! 6
7 Acous4c Model 7
8 Acous4c Model Gaussian Mixture Model (GMM) 8
9 Acous4c Model GMM/HMM GMMs trained using the EM algorithm are able to self organize to fit a data set Hidden Markov Model models sequenbal paqerns Technical Advances over past 10 years AdaptaBon, DiscriminaBve Training, SGMM 9
10 Acous4c Model Deep Neural Network (DNN) George E. Dahl, Context-Dependent Pre-Trained Deep Neural Networks for Large-Vocabulary Speech Recognition 10
11 Acous4c Model DNN/HMM Called Hybrid DNN/HMM system Has good discriminabon Temporal aspects are deal with HMM, like lez- to- right HMM models Drawback is computabon is expensive! 11
12 Acous4c Model Combina4on How can we improve ASR with AcousBc Model Robust AcousBc Model More and More Data - > BeQer and BeQer Accuracy Robust Feature(BoQle- neck Feature, Noise Robust Feature) AcousBc Model CombinaBon 12
13 Acous4c Model Combina4on GMM1 DNN1 Log likelihood PaHern by Acous4c Model F e a t u r e GMM2 DNN2 Model Structure 13
14 Acous4c Model Combina4on Different Acous4c Models(model structure, features) have disbnct speech recognibon paqern. - > different performance in Speech RecogniBon and Keyword Search The goal is to find a way to combine different acous4c models for robust speech recogni4on and keyword search Considera4on Data type to be combined for AM combinabon WeighBng criterion Total system run Bme (Real- Bme factor) 14
15 Acous4c Model Combina4on Mul4- stream based AM combina4on Combine mulbple AMs at the AM score level WeighBng Criterion(ArithmeBc, Geometric, Harmonic) One- pass and One :me decoding Other combina4on Method Labce CombinaBon, Rover, Combmnz Intermediate/output level combinabon 15
16 Acous4c Model Combina4on Features o 1 o N AM 1 AM N s 1,1 s 1,2 s 1,3 s N,1 s N,2 s 1 s 2 s 3 s 4 s 1 s 2 s 3 s 4 Remapping w 1 : weight (normalization) s 1 s 2 Σ s 3 s 4 w 2 Combination: Arithmetic mean Geometric mean Harmonic mean WFST(combined) DECODER Words Decoding 16
17 Acous4c Model Combina4on GPU Accelerated Speech Recogni4on - Talked at GTC 2013 & Speech recognition contains many highly parallel tasks GPU processors + optimized for parallel = computing HYDRA an ASR engine designed specifically for GPUs 17
18 Acous4c Model Combina4on 18
19 Experimental evalua4ons Carnegie Mellon University Data: IARPA BABEL Program Vietnamese language collecbon: babel107b- v0.7 [1] Limited language pack (10 hrs training, 20 hrs test) Features: LMEL: Log Mel filter bank coefficients MFCC: Mel Frequency Cepstral Coefficients BNF: BoQlenect features FFV: Fundamental Frequency VariaBon feature Pitch: Pitch tracking feature Features Dim. Source feature Input frames BNF th lmel + FFV 11 BNF th lmel + FFV + Pitch 11 BNF th MFCC + FFV 11 [1] IARPA, Iarpa babel program, 19
20 Experimental evalua4ons Carnegie Mellon University Model Feature Tree WER (%) ATWV GMM 1 BNF GMM 2 BNF 2 Tree GMM 3 BNF DNN 1 BNF DNN 2 BNF 2 Tree DNN 3 BNF Baseline system performance Trained 6 acous4c models (3 GMMs, 3 DNNs) with 3 different feautres 20
21 Experimental evalua4ons Carnegie Mellon University Combination scheme WER ATWV Best single system (DNN 1 ) Arithmetic mean 63.6 (-3.7) (+30.3%) Geometric mean 65.4 (-2.9) (+23.9%) Harmonic mean 66.2 (-1.1) (+21.0%) WER and ATWV for different combination schemes Combined 6 acousbc models (GMM DNN ) ArithmeBc mean showed the most improved performance. 3.7% absolute WER improvement 30.3% relabve ATWV improvement 21
22 Experimental evalua4ons ATWV RTF ATWV Carnegie Mellon University Real-Time Factor Model 1-Model 1-Model 3-Models 6-Models CPU GPU-search GPU-based AM computation 0.0
23 Experimental evalua4ons Carnegie Mellon University State-level combination obtains best WER vs. Lattice Comb., Rover Note: same phone-states used across all models CombMNZ obtains better ATWV when combining more than 2 models However 5x - 10x slower At comparable RTF: Multi-stream=0.23 > CombMNZ=0.20 Models Multistream CombMNZ 1 model 67.3% (0.14) models 64.7% (0.17) models 63.6% (0.18) models 63.6% (0.18) models (large lattice) 64.7% (0.23)
24 Conclusion Proposed MulB- stream Model combinabon in GPU accelerated speech recognibon framework MulB- stream combinabon gives comparable performance with efficient runbme Future work More combina4on schemes: Weighted model combinabon (Model, HMM state level weights) DNN- based combinabon Faster decoding speed: Use of CUDA mulb- stream technique. 24
25 Q&A Thank you for your attention. 25
Pitch Prediction from Mel-frequency Cepstral Coefficients Using Sparse Spectrum Recovery
Pitch Prediction from Mel-frequency Cepstral Coefficients Using Sparse Spectrum Recovery Achuth Rao MV, Prasanta Kumar Ghosh SPIRE LAB Electrical Engineering, Indian Institute of Science (IISc), Bangalore,
More informationAutomatic Speech Recognition (ASR)
Automatic Speech Recognition (ASR) February 2018 Reza Yazdani Aminabadi Universitat Politecnica de Catalunya (UPC) State-of-the-art State-of-the-art ASR system: DNN+HMM Speech (words) Sound Signal Graph
More informationWhy DNN Works for Speech and How to Make it More Efficient?
Why DNN Works for Speech and How to Make it More Efficient? Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering, York University, CANADA Joint work with Y.
More informationBo#leneck Features from SNR- Adap9ve Denoising Deep Classifier for Speaker Iden9fica9on
Bo#leneck Features from SNR- Adap9ve Denoising Deep Classifier for Speaker Iden9fica9on TAN Zhili & MAK Man-Wai APSIPA 2015 Department of Electronic and Informa2on Engineering The Hong Kong Polytechnic
More informationA Scalable Speech Recognizer with Deep-Neural-Network Acoustic Models
A Scalable Speech Recognizer with Deep-Neural-Network Acoustic Models and Voice-Activated Power Gating Michael Price*, James Glass, Anantha Chandrakasan MIT, Cambridge, MA * now at Analog Devices, Cambridge,
More informationDiscriminative training and Feature combination
Discriminative training and Feature combination Steve Renals Automatic Speech Recognition ASR Lecture 13 16 March 2009 Steve Renals Discriminative training and Feature combination 1 Overview Hot topics
More informationAn In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization
An In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization Yun Wang and Florian Metze Language Technologies Institute, Carnegie Mellon University Pittsburgh, PA, U.S.A.
More informationSpeech Technology Using in Wechat
Speech Technology Using in Wechat FENG RAO Powered by WeChat Outline Introduce Algorithm of Speech Recognition Acoustic Model Language Model Decoder Speech Technology Open Platform Framework of Speech
More informationVariable-Component Deep Neural Network for Robust Speech Recognition
Variable-Component Deep Neural Network for Robust Speech Recognition Rui Zhao 1, Jinyu Li 2, and Yifan Gong 2 1 Microsoft Search Technology Center Asia, Beijing, China 2 Microsoft Corporation, One Microsoft
More informationA Comparison of Sequence-Trained Deep Neural Networks and Recurrent Neural Networks Optical Modeling For Handwriting Recognition
A Comparison of Sequence-Trained Deep Neural Networks and Recurrent Neural Networks Optical Modeling For Handwriting Recognition Théodore Bluche, Hermann Ney, Christopher Kermorvant SLSP 14, Grenoble October
More informationJoint Optimisation of Tandem Systems using Gaussian Mixture Density Neural Network Discriminative Sequence Training
Joint Optimisation of Tandem Systems using Gaussian Mixture Density Neural Network Discriminative Sequence Training Chao Zhang and Phil Woodland March 8, 07 Cambridge University Engineering Department
More informationAn Optimization of Deep Neural Networks in ASR using Singular Value Decomposition
An Optimization of Deep Neural Networks in ASR using Singular Value Decomposition Bachelor Thesis of Igor Tseyzer At the Department of Informatics Institute for Anthropomatics (IFA) Reviewer: Second reviewer:
More informationSVD-based Universal DNN Modeling for Multiple Scenarios
SVD-based Universal DNN Modeling for Multiple Scenarios Changliang Liu 1, Jinyu Li 2, Yifan Gong 2 1 Microsoft Search echnology Center Asia, Beijing, China 2 Microsoft Corporation, One Microsoft Way, Redmond,
More informationXing Fan, Carlos Busso and John H.L. Hansen
Xing Fan, Carlos Busso and John H.L. Hansen Center for Robust Speech Systems (CRSS) Erik Jonsson School of Engineering & Computer Science Department of Electrical Engineering University of Texas at Dallas
More informationarxiv: v1 [cs.cl] 30 Jan 2018
ACCELERATING RECURRENT NEURAL NETWORK LANGUAGE MODEL BASED ONLINE SPEECH RECOGNITION SYSTEM Kyungmin Lee, Chiyoun Park, Namhoon Kim, and Jaewon Lee DMC R&D Center, Samsung Electronics, Seoul, Korea {k.m.lee,
More informationLecture 7: Neural network acoustic models in speech recognition
CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 7: Neural network acoustic models in speech recognition Outline Hybrid acoustic modeling overview Basic
More informationCombination of FST and CN Search in Spoken Term Detection
Combination of FST and CN Search in Spoken Term Detection Justin Chiu 1, Yun Wang 1, Jan Trmal 2, Daniel Povey 2, Guoguo Chen 2, Alexander Rudnicky 1 1 Language Technologies Institute, Carnegie Mellon
More informationSPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION
Far East Journal of Electronics and Communications Volume 3, Number 2, 2009, Pages 125-140 Published Online: September 14, 2009 This paper is available online at http://www.pphmj.com 2009 Pushpa Publishing
More informationGMM-FREE DNN TRAINING. Andrew Senior, Georg Heigold, Michiel Bacchiani, Hank Liao
GMM-FREE DNN TRAINING Andrew Senior, Georg Heigold, Michiel Bacchiani, Hank Liao Google Inc., New York {andrewsenior,heigold,michiel,hankliao}@google.com ABSTRACT While deep neural networks (DNNs) have
More informationDynamic Time Warping
Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. Dynamic Time Warping Dr Philip Jackson Acoustic features Distance measures Pattern matching Distortion penalties DTW
More informationWHO WANTS TO BE A MILLIONAIRE?
IDIAP COMMUNICATION REPORT WHO WANTS TO BE A MILLIONAIRE? Huseyn Gasimov a Aleksei Triastcyn Hervé Bourlard Idiap-Com-03-2012 JULY 2012 a EPFL Centre du Parc, Rue Marconi 19, PO Box 592, CH - 1920 Martigny
More informationKernels vs. DNNs for Speech Recognition
Kernels vs. DNNs for Speech Recognition Joint work with: Columbia: Linxi (Jim) Fan, Michael Collins (my advisor) USC: Zhiyun Lu, Kuan Liu, Alireza Bagheri Garakani, Dong Guo, Aurélien Bellet, Fei Sha IBM:
More informationSemantic Word Embedding Neural Network Language Models for Automatic Speech Recognition
Semantic Word Embedding Neural Network Language Models for Automatic Speech Recognition Kartik Audhkhasi, Abhinav Sethy Bhuvana Ramabhadran Watson Multimodal Group IBM T. J. Watson Research Center Motivation
More informationImproving Bottleneck Features for Automatic Speech Recognition using Gammatone-based Cochleagram and Sparsity Regularization
Improving Bottleneck Features for Automatic Speech Recognition using Gammatone-based Cochleagram and Sparsity Regularization Chao Ma 1,2,3, Jun Qi 4, Dongmei Li 1,2,3, Runsheng Liu 1,2,3 1. Department
More informationLab 4: Hybrid Acoustic Models
v. 1.0 Lab 4: Hybrid Acoustic Models University of Edinburgh March 13, 2017 This is the final lab, in which we will have a look at training hybrid neural network acoustic models using a frame-level cross-entropy
More informationHello Edge: Keyword Spotting on Microcontrollers
Hello Edge: Keyword Spotting on Microcontrollers Yundong Zhang, Naveen Suda, Liangzhen Lai and Vikas Chandra ARM Research, Stanford University arxiv.org, 2017 Presented by Mohammad Mofrad University of
More informationModeling Phonetic Context with Non-random Forests for Speech Recognition
Modeling Phonetic Context with Non-random Forests for Speech Recognition Hainan Xu Center for Language and Speech Processing, Johns Hopkins University September 4, 2015 Hainan Xu September 4, 2015 1 /
More informationHidden Markov Models. Gabriela Tavares and Juri Minxha Mentor: Taehwan Kim CS159 04/25/2017
Hidden Markov Models Gabriela Tavares and Juri Minxha Mentor: Taehwan Kim CS159 04/25/2017 1 Outline 1. 2. 3. 4. Brief review of HMMs Hidden Markov Support Vector Machines Large Margin Hidden Markov Models
More informationLOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS
LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS Tara N. Sainath, Brian Kingsbury, Vikas Sindhwani, Ebru Arisoy, Bhuvana Ramabhadran IBM T. J. Watson
More informationEnd- To- End Speech Recogni0on with Recurrent Neural Networks
RTTH Summer School on Speech Technology: A Deep Learning Perspec0ve End- To- End Speech Recogni0on with Recurrent Neural Networks José A. R. Fonollosa Universitat Politècnica de Catalunya. Barcelona Barcelona,
More informationDeep Neural Networks in HMM- based and HMM- free Speech RecogniDon
Deep Neural Networks in HMM- based and HMM- free Speech RecogniDon Andrew Maas Collaborators: Awni Hannun, Peng Qi, Chris Lengerich, Ziang Xie, and Anshul Samar Advisors: Andrew Ng and Dan Jurafsky Outline
More informationLab 2: Training monophone models
v. 1.1 Lab 2: Training monophone models University of Edinburgh January 29, 2018 Last time we begun to get familiar with some of Kaldi s tools and set up a data directory for TIMIT. This time we will train
More informationCUED-RNNLM An Open-Source Toolkit for Efficient Training and Evaluation of Recurrent Neural Network Language Models
CUED-RNNLM An Open-Source Toolkit for Efficient Training and Evaluation of Recurrent Neural Network Language Models Xie Chen, Xunying Liu, Yanmin Qian, Mark Gales and Phil Woodland April 1, 2016 Overview
More informationMaximum Likelihood Beamforming for Robust Automatic Speech Recognition
Maximum Likelihood Beamforming for Robust Automatic Speech Recognition Barbara Rauch barbara@lsv.uni-saarland.de IGK Colloquium, Saarbrücken, 16 February 2006 Agenda Background: Standard ASR Robust ASR
More informationAndi Buzo, Horia Cucu, Mihai Safta and Corneliu Burileanu. Speech & Dialogue (SpeeD) Research Laboratory University Politehnica of Bucharest (UPB)
Andi Buzo, Horia Cucu, Mihai Safta and Corneliu Burileanu Speech & Dialogue (SpeeD) Research Laboratory University Politehnica of Bucharest (UPB) The MediaEval 2012 SWS task A multilingual, query by example,
More informationGating Neural Network for Large Vocabulary Audiovisual Speech Recognition
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. XX, NO. X, JUNE 2017 1 Gating Neural Network for Large Vocabulary Audiovisual Speech Recognition Fei Tao, Student Member, IEEE, and
More informationSYNTHESIZED STEREO MAPPING VIA DEEP NEURAL NETWORKS FOR NOISY SPEECH RECOGNITION
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) SYNTHESIZED STEREO MAPPING VIA DEEP NEURAL NETWORKS FOR NOISY SPEECH RECOGNITION Jun Du 1, Li-Rong Dai 1, Qiang Huo
More informationRLAT Rapid Language Adaptation Toolkit
RLAT Rapid Language Adaptation Toolkit Tim Schlippe May 15, 2012 RLAT Rapid Language Adaptation Toolkit - 2 RLAT Rapid Language Adaptation Toolkit RLAT Rapid Language Adaptation Toolkit - 3 Outline Introduction
More informationQuery-by-example spoken term detection based on phonetic posteriorgram Query-by-example spoken term detection based on phonetic posteriorgram
International Conference on Education, Management and Computing Technology (ICEMCT 2015) Query-by-example spoken term detection based on phonetic posteriorgram Query-by-example spoken term detection based
More informationSOUND EVENT DETECTION AND CONTEXT RECOGNITION 1 INTRODUCTION. Toni Heittola 1, Annamaria Mesaros 1, Tuomas Virtanen 1, Antti Eronen 2
Toni Heittola 1, Annamaria Mesaros 1, Tuomas Virtanen 1, Antti Eronen 2 1 Department of Signal Processing, Tampere University of Technology Korkeakoulunkatu 1, 33720, Tampere, Finland toni.heittola@tut.fi,
More informationLucida Sirius and DjiNN Tutorial
Lucida Sirius and DjiNN Tutorial Speakers: Johann Hauswald, Michael A. Laurenzano, Yunqi Zhang Organizers: Johann Hauswald, Michael A. Laurenzano, Yunqi Zhang, Lingjia Tang, Jason Mars Before We Begin
More informationLattice Rescoring for Speech Recognition Using Large Scale Distributed Language Models
Lattice Rescoring for Speech Recognition Using Large Scale Distributed Language Models ABSTRACT Euisok Chung Hyung-Bae Jeon Jeon-Gue Park and Yun-Keun Lee Speech Processing Research Team, ETRI, 138 Gajeongno,
More informationA long, deep and wide artificial neural net for robust speech recognition in unknown noise
A long, deep and wide artificial neural net for robust speech recognition in unknown noise Feipeng Li, Phani S. Nidadavolu, and Hynek Hermansky Center for Language and Speech Processing Johns Hopkins University,
More informationImplementing a Speech Recognition System on a GPU using CUDA. Presented by Omid Talakoub Astrid Yi
Implementing a Speech Recognition System on a GPU using CUDA Presented by Omid Talakoub Astrid Yi Outline Background Motivation Speech recognition algorithm Implementation steps GPU implementation strategies
More informationMaking Deep Belief Networks Effective for Large Vocabulary Continuous Speech Recognition
Making Deep Belief Networks Effective for Large Vocabulary Continuous Speech Recognition Tara N. Sainath 1, Brian Kingsbury 1, Bhuvana Ramabhadran 1, Petr Fousek 2, Petr Novak 2, Abdel-rahman Mohamed 3
More informationImplementing a Speech Recognition Algorithm with VSIPL++
Implementing a Speech Recognition Algorithm with VSIPL++ Don McCoy, Brooks Moses, Stefan Seefeld, Justin Voo Software Engineers Embedded Systems Division / HPC Group September 2011 Objective VSIPL++ Standard:
More informationAcoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing
Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing Samer Al Moubayed Center for Speech Technology, Department of Speech, Music, and Hearing, KTH, Sweden. sameram@kth.se
More informationSpeech User Interface for Information Retrieval
Speech User Interface for Information Retrieval Urmila Shrawankar Dept. of Information Technology Govt. Polytechnic Institute, Nagpur Sadar, Nagpur 440001 (INDIA) urmilas@rediffmail.com Cell : +919422803996
More informationSpeech Recogni,on using HTK CS4706. Fadi Biadsy April 21 st, 2008
peech Recogni,on using HTK C4706 Fadi Biadsy April 21 st, 2008 1 Outline peech Recogni,on Feature Extrac,on HMM 3 basic problems HTK teps to Build a speech recognizer 2 peech Recogni,on peech ignal to
More informationCHROMA AND MFCC BASED PATTERN RECOGNITION IN AUDIO FILES UTILIZING HIDDEN MARKOV MODELS AND DYNAMIC PROGRAMMING. Alexander Wankhammer Peter Sciri
1 CHROMA AND MFCC BASED PATTERN RECOGNITION IN AUDIO FILES UTILIZING HIDDEN MARKOV MODELS AND DYNAMIC PROGRAMMING Alexander Wankhammer Peter Sciri introduction./the idea > overview What is musical structure?
More informationNovel Methods for Query Selection and Query Combination in Query-By-Example Spoken Term Detection
Novel Methods for Query Selection and Query Combination in Query-By-Example Spoken Term Detection Javier Tejedor HCTLab, Universidad Autónoma de Madrid, Spain javier.tejedor@uam.es Igor Szöke Speech@FIT,
More informationMachine Learning for Speaker Recogni2on and Bioinforma2cs
Machine Learning for Speaker Recogni2on and Bioinforma2cs Man-Wai MAK Dept of Electronic and Informa8on Engineering, The Hong Kong Polytechnic University http://wwweiepolyueduhk/~mwmak/ UTS/PolyU Workshop
More informationTime series, HMMs, Kalman Filters
Classic HMM tutorial see class website: *L. R. Rabiner, "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition," Proc. of the IEEE, Vol.77, No.2, pp.257--286, 1989. Time series,
More informationThe Hitachi/JHU CHiME-5 system: Advances in speech recognition for everyday home environments using multiple microphone arrays
CHiME2018 workshop The Hitachi/JHU CHiME-5 system: Advances in speech recognition for everyday home environments using multiple microphone arrays Naoyuki Kanda 1, Rintaro Ikeshita 1, Shota Horiguchi 1,
More informationTWO-STEP SEMI-SUPERVISED APPROACH FOR MUSIC STRUCTURAL CLASSIFICATION. Prateek Verma, Yang-Kai Lin, Li-Fan Yu. Stanford University
TWO-STEP SEMI-SUPERVISED APPROACH FOR MUSIC STRUCTURAL CLASSIFICATION Prateek Verma, Yang-Kai Lin, Li-Fan Yu Stanford University ABSTRACT Structural segmentation involves finding hoogeneous sections appearing
More informationLeast Squares Signal Declipping for Robust Speech Recognition
Least Squares Signal Declipping for Robust Speech Recognition Mark J. Harvilla and Richard M. Stern Department of Electrical and Computer Engineering Carnegie Mellon University, Pittsburgh, PA 15213 USA
More informationGender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV
Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV Jan Vaněk and Josef V. Psutka Department of Cybernetics, West Bohemia University,
More informationReverberant Speech Recognition Based on Denoising Autoencoder
INTERSPEECH 2013 Reverberant Speech Recognition Based on Denoising Autoencoder Takaaki Ishii 1, Hiroki Komiyama 1, Takahiro Shinozaki 2, Yasuo Horiuchi 1, Shingo Kuroiwa 1 1 Division of Information Sciences,
More informationAutomatic Speech Recognition on Mobile Devices and over Communication Networks
Zheng-Hua Tan and Berge Lindberg Automatic Speech Recognition on Mobile Devices and over Communication Networks ^Spri inger g< Contents Preface Contributors v xix 1. Network, Distributed and Embedded Speech
More informationBaseball Game Highlight & Event Detection
Baseball Game Highlight & Event Detection Student: Harry Chao Course Adviser: Winston Hu 1 Outline 1. Goal 2. Previous methods 3. My flowchart 4. My methods 5. Experimental result 6. Conclusion & Future
More informationMemory-Efficient Heterogeneous Speech Recognition Hybrid in GPU-Equipped Mobile Devices
Memory-Efficient Heterogeneous Speech Recognition Hybrid in GPU-Equipped Mobile Devices Alexei V. Ivanov, CTO, Verbumware Inc. GPU Technology Conference, San Jose, March 17, 2015 Autonomous Speech Recognition
More informationA Gaussian Mixture Model Spectral Representation for Speech Recognition
A Gaussian Mixture Model Spectral Representation for Speech Recognition Matthew Nicholas Stuttle Hughes Hall and Cambridge University Engineering Department PSfrag replacements July 2003 Dissertation submitted
More informationTwo-Layered Audio-Visual Speech Recognition for Robots in Noisy Environments
The 2 IEEE/RSJ International Conference on Intelligent Robots and Systems October 8-22, 2, Taipei, Taiwan Two-Layered Audio-Visual Speech Recognition for Robots in Noisy Environments Takami Yoshida, Kazuhiro
More informationAn Ultra Low-Power Hardware Accelerator for Automatic Speech Recognition
An Ultra Low-Power Hardware Accelerator for Automatic Speech Recognition Reza Yazdani, Albert Segura, Jose-Maria Arnau, Antonio Gonzalez Computer Architecture Department, Universitat Politecnica de Catalunya
More informationSpeech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri
Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute eugenew@cs.nyu.edu Slide Credit: Mehryar Mohri Speech Recognition Components Acoustic and pronunciation model:
More informationSCORE NORMALIZATION AND SYSTEM COMBINATION FOR IMPROVED KEYWORD SPOTTING
SCORE NORMALIZATION AND SYSTEM COMBINATION FOR IMPROVED KEYWORD SPOTTING Damianos Karakos 1 Richard Schwartz 1 Stavros Tsakalidis 1 Le Zhang 1 Shivesh Ranjan 1 Tim Ng 1 Roger Hsiao 1 Guruprasad Saikumar
More informationHighlights Extraction from Unscripted Video
Highlights Extraction from Unscripted Video T 61.6030, Multimedia Retrieval Seminar presentation 04.04.2008 Harrison Mfula Helsinki University of Technology Department of Computer Science, Espoo, Finland
More informationDEEPFINNS. Group 3 Members FPGA IMPLEMENTATION OF A DEEP NEURAL NETWORK FOR SPEECH RECOGNITION. Fall 2016 LINDSAY DAVIS
Fall 2016 DEEPFINNS FPGA IMPLEMENTATION OF A DEEP NEURAL NETWORK FOR SPEECH RECOGNITION Group 3 Members LINDSAY DAVIS Electrical Engineering MICHAEL LOPEZ-BRAU Electrical Engineering ESTELLA GONG Computer
More informationEfficient acoustic detector of gunshots and glass breaking
Multimed Tools Appl (2016) 75:10441 10469 DOI 10.1007/s11042-015-2903-z Efficient acoustic detector of gunshots and glass breaking Martin Lojka 1 Matúš Pleva 1 Eva Kiktová 1 Jozef Juhár 1 Anton Čižmár
More informationNON-LINEAR DIMENSION REDUCTION OF GABOR FEATURES FOR NOISE-ROBUST ASR. Hitesh Anand Gupta, Anirudh Raju, Abeer Alwan
NON-LINEAR DIMENSION REDUCTION OF GABOR FEATURES FOR NOISE-ROBUST ASR Hitesh Anand Gupta, Anirudh Raju, Abeer Alwan Department of Electrical Engineering, University of California Los Angeles, USA {hiteshag,
More informationSimultaneous Design of Feature Extractor and Pattern Classifer Using the Minimum Classification Error Training Algorithm
Griffith Research Online https://research-repository.griffith.edu.au Simultaneous Design of Feature Extractor and Pattern Classifer Using the Minimum Classification Error Training Algorithm Author Paliwal,
More informationHIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION. Hung-An Chang and James R. Glass
HIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION Hung-An Chang and James R. Glass MIT Computer Science and Artificial Intelligence Laboratory Cambridge, Massachusetts, 02139,
More informationIn Silico Vox: Towards Speech Recognition in Silicon
In Silico Vox: Towards Speech Recognition in Silicon Edward C Lin, Kai Yu, Rob A Rutenbar, Tsuhan Chen Electrical & Computer Engineering {eclin, kaiy, rutenbar, tsuhan}@ececmuedu RA Rutenbar 006 Speech
More informationAn Arabic Optical Character Recognition System Using Restricted Boltzmann Machines
An Arabic Optical Character Recognition System Using Restricted Boltzmann Machines Abdullah M. Rashwan, Mohamed S. Kamel, and Fakhri Karray University of Waterloo Abstract. Most of the state-of-the-art
More informationAnalysis of Local Appearance-based Face Recognition on FRGC 2.0 Database
Analysis of Local Appearance-based Face Recognition on FRGC 2.0 Database HAZIM KEMAL EKENEL (ISL),, Germany e- mail: ekenel@ira.uka.de web: http://isl.ira.uka.de/~ekenel/ http://isl.ira.uka.de/face_recognition/
More informationAvailable online Journal of Scientific and Engineering Research, 2016, 3(4): Research Article
Available online www.jsaer.com, 2016, 3(4):417-422 Research Article ISSN: 2394-2630 CODEN(USA): JSERBR Automatic Indexing of Multimedia Documents by Neural Networks Dabbabi Turkia 1, Lamia Bouafif 2, Ellouze
More informationDeep Learning on Arm Cortex-M Microcontrollers. Rod Crawford Director Software Technologies, Arm
Deep Learning on Arm Cortex-M Microcontrollers Rod Crawford Director Software Technologies, Arm What is Machine Learning (ML)? Artificial Intelligence Machine Learning Deep Learning Neural Networks Additional
More informationA Deep Learning primer
A Deep Learning primer Riccardo Zanella r.zanella@cineca.it SuperComputing Applications and Innovation Department 1/21 Table of Contents Deep Learning: a review Representation Learning methods DL Applications
More informationSpeaker Verification with Adaptive Spectral Subband Centroids
Speaker Verification with Adaptive Spectral Subband Centroids Tomi Kinnunen 1, Bingjun Zhang 2, Jia Zhu 2, and Ye Wang 2 1 Speech and Dialogue Processing Lab Institution for Infocomm Research (I 2 R) 21
More informationThe Automatic Musicologist
The Automatic Musicologist Douglas Turnbull Department of Computer Science and Engineering University of California, San Diego UCSD AI Seminar April 12, 2004 Based on the paper: Fast Recognition of Musical
More informationGPU-Accelerated Large Vocabulary Continuous Speech Recognition for Scalable Distributed Speech Recognition. Jungsuk Kim Ian Lane
GPU-Accelerated Large Vocabulary Continuou Speech Recognition for Scalable Ditributed Speech Recognition Junguk Kim Ian Lane Electrical and Computer Engineering Carnegie Mellon Univerity March 20, 2015
More informationFraunhofer IAIS Audio Mining Solution for Broadcast Archiving. Dr. Joachim Köhler LT-Innovate Brussels
Fraunhofer IAIS Audio Mining Solution for Broadcast Archiving Dr. Joachim Köhler LT-Innovate Brussels 22.11.2016 1 Outline Speech Technology in the Broadcast World Deep Learning Speech Technologies Fraunhofer
More informationSirius: An Open End-to-End Voice and Vision Personal Assistant and Its Implications for Future Warehouse Scale Computers
Sirius: An Open End-to-End Voice and Vision Personal Assistant and Its Implications for Future Warehouse Scale Computers Johann Hauswald, Michael A. Laurenzano, Yunqi Zhang, Cheng Li, Austin Rovinski,
More informationSPEECH FEATURE DENOISING AND DEREVERBERATION VIA DEEP AUTOENCODERS FOR NOISY REVERBERANT SPEECH RECOGNITION. Xue Feng, Yaodong Zhang, James Glass
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) SPEECH FEATURE DENOISING AND DEREVERBERATION VIA DEEP AUTOENCODERS FOR NOISY REVERBERANT SPEECH RECOGNITION Xue Feng,
More informationAn Introduction to Pattern Recognition
An Introduction to Pattern Recognition Speaker : Wei lun Chao Advisor : Prof. Jian-jiun Ding DISP Lab Graduate Institute of Communication Engineering 1 Abstract Not a new research field Wide range included
More informationChapter 3. Speech segmentation. 3.1 Preprocessing
, as done in this dissertation, refers to the process of determining the boundaries between phonemes in the speech signal. No higher-level lexical information is used to accomplish this. This chapter presents
More informationA ROBUST SPEAKER CLUSTERING ALGORITHM
A ROBUST SPEAKER CLUSTERING ALGORITHM J. Ajmera IDIAP P.O. Box 592 CH-1920 Martigny, Switzerland jitendra@idiap.ch C. Wooters ICSI 1947 Center St., Suite 600 Berkeley, CA 94704, USA wooters@icsi.berkeley.edu
More informationGAUSSIAN MIXTURE MODEL EVALUATION FOR AUTOMATIC SPEECH RECOGNITION ON GPU
GAUSSIAN MIXTURE MODEL EVALUATION FOR AUTOMATIC SPEECH RECOGNITION ON GPU By Liana Nicklaus Advisor: Professor Deming Chen Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign,
More informationDeep Temporal Models (Benchmarks and Applica6ons Analysis)
Deep Temporal Models (Benchmarks and Applica6ons Analysis) Sek Chai SRI Interna6onal Presented at: NICE 2016, March 7, 2016 2016 SRI International Project Summary Goals Analyze Deep Temporal Models (DTMs).
More informationDeep Learning. Volker Tresp Summer 2014
Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there
More informationMultimedia Indexing. Lecture 12: EE E6820: Speech & Audio Processing & Recognition. Spoken document retrieval Audio databases.
EE E6820: Speech & Audio Processing & Recognition Lecture 12: Multimedia Indexing 1 Spoken document retrieval 2 Audio databases 3 Open issues Dan Ellis http://www.ee.columbia.edu/~dpwe/e6820/
More informationLecture 12: Multimedia Indexing. Spoken Document Retrieval (SDR)
EE E68: Speech & Audio Processing & Recognition Lecture : Multimedia Indexing 3 Spoken document retrieval Audio databases Open issues Dan Ellis http://www.ee.columbia.edu/~dpwe/e68/
More informationConfidence Measures: how much we can trust our speech recognizers
Confidence Measures: how much we can trust our speech recognizers Prof. Hui Jiang Department of Computer Science York University, Toronto, Ontario, Canada Email: hj@cs.yorku.ca Outline Speech recognition
More informationManifold Constrained Deep Neural Networks for ASR
1 Manifold Constrained Deep Neural Networks for ASR Department of Electrical and Computer Engineering, McGill University Richard Rose and Vikrant Tomar Motivation Speech features can be characterized as
More informationDetection of goal event in soccer videos
Detection of goal event in soccer videos Hyoung-Gook Kim, Steffen Roeber, Amjad Samour, Thomas Sikora Department of Communication Systems, Technical University of Berlin, Einsteinufer 17, D-10587 Berlin,
More informationEncrypted Deep Learning: A Guide to Privacy Preserving Speech Processing
Encrypted Deep Learning: A Guide to Privacy Preserving Speech Processing Nigel Cannings CTO nigel.cannings@intelligentvoice.com www.intelligentvoice.com @intelligentvox #GTC17 For $100 What is this encrypted
More information2014, IJARCSSE All Rights Reserved Page 461
Volume 4, Issue 1, January 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Real Time Speech
More informationIn Silico Vox: Towards Speech Recognition in Silicon
In Silico Vox: Towards Speech Recognition in Silicon Edward C Lin, Kai Yu, Rob A Rutenbar, Tsuhan Chen Electrical & Computer Engineering {eclin, kaiy, rutenbar, tsuhan}@ececmuedu RA Rutenbar 2006 Speech
More informationA Novel Template Matching Approach To Speaker-Independent Arabic Spoken Digit Recognition
Special Session: Intelligent Knowledge Management A Novel Template Matching Approach To Speaker-Independent Arabic Spoken Digit Recognition Jiping Sun 1, Jeremy Sun 1, Kacem Abida 2, and Fakhri Karray
More informationComparative Evaluation of Feature Normalization Techniques for Speaker Verification
Comparative Evaluation of Feature Normalization Techniques for Speaker Verification Md Jahangir Alam 1,2, Pierre Ouellet 1, Patrick Kenny 1, Douglas O Shaughnessy 2, 1 CRIM, Montreal, Canada {Janagir.Alam,
More information