A Collaborative Speech Transcription System for Live Streaming

Size: px
Start display at page:

Download "A Collaborative Speech Transcription System for Live Streaming"

Transcription

1 ustream Yourscribe PodCastle Yourscribe 2.0 A Collaborative Speech Transcription System for Live Streaming Shunsuke Ukita, 1 Jun Ogata, 2 Masataka Goto 2 and Tetsunori Kobayashi 1 In this paper, we propose a real-time transcription system, Yourscribe, that enables anonymous users to collaboratively transcribe speech in live video streaming like ustream. In previous approaches, transcription by human was laborsome and transcription by speech recognition was error-prone. PodCastle was developed to overcome such errors of speech recognition by having anonymous users correct errors, but was not appropriate for real-time transcription. To use Yourscribe, each user can just voluntarily type a short phrase that was heard while enjoying the live video without any interruption. Those phrases from many users were aggregated at all times and matched with results of realtime speech recognition to automatically form the transcription. This is a new instance of our research approach, Speech Recognition Research ustream 1 2 Web ustream twitter 3 Web 1)2)3)4)5)6) 100% Web PodCastle 7)8)9) ) PodCastle 1 Waseda University 2 National Institute of Advanced Industrial Science and Technology (AIST) c 2011 Information Processing Society of Japan

2 Yourscribe Yourscribe 2 Yourscribe Yourscribe Yourscribe Web Yourscribe Web Web ustream ustream twitter Yourscribe ustream 1!!!! Yourscribe : 1 Yourscribe Yourscribe Yourscribe 2 c 2011 Information Processing Society of Japan

3 Yourscribe 3. Yourscribe Yourscribe Web Web Web Yourscribe 1 ( ) 3.2 t e t e T delay t e T delay () DP DP (HMM: ) HMM 3 c 2011 Information Processing Society of Japan

4 Web 11) ( ) HMM t e T delay (HMM) HMM HMM HMM Viterbi ) HMM ustream Web Web 3 A B C Web % 75% PodCastle 12)13) CSJ tied-state cross-word triphone 39 PLP(12 PLP ) CMLLR 14). Web N-gram 11) Web CSJ HMM CSJ 32 monophone triphone monophone 4 c 2011 Information Processing Society of Japan

5 2 1 () + A 77.88% 85.66% B 47.42% 73.00% C 31.90% 60.25% ( ) () 25% 50% 75% A 80.47% 82.65% 85.66% B 58.57% 65.43% 73.00% C 42.18% 52.53% 60.25% T delay 10 1 C 3 50% monophone HMM 75% 50% 25% 75% 50% 25% )16)1)2) 15)16) 17) 18) 19) 3)20) 4)5)6) 21) 16) ( ) 7) PodCastle PodCastle Yourscribe PodCastle Web Web Yourscribe 5 c 2011 Information Processing Society of Japan

6 1 torne 2 (twitter ) 6. Yourscribe Web Yourscribe 2.0 7)8)9) PodCastle 1) Chen, S.S., Eide, E.M., Gales, M.J., Gopinath, R.A., Kanevsky, D. and Olsen, P. A.: Recent Improvements to IBM s Speech Recognition System for Automatic Transcription of Broadcast News, Proc. ICASSP 99, Vol.1, pp (1999). 2) Woodland, P.C., Gales, M.J., Pye, D. and Young, S.J.: Broadcast News Transcription Using HTK, Proc. ICASSP 97, Vol.2, pp (1997). 3) Glass, J., Hazen, T.J., Cyphers, S., Malioutov, I., Huynh, D. and Barzilay, R.: Recent Progress in the MIT Spoken Lecture Processing Project, Proc. of Interspeech 2007, pp (2007). 4) Janin, A., Baron, D., Edwards, J., Ellis, D., Gelbart, D., Morgan, N., Peskin, B., Pfau, T., Shriberg, E., Stolcke, A. and Wooters, C.: The ICSI Meeting Corpus, Proc. ICASSP 2003, Vol.1, pp (2003). 5) Metze, F., Waibel, A., Bett, M., Ries, K., Schaaf, T., Schultz, T., Soltau, H., Yu, H. and Zechner, K.: Advances in Automatic Meeting Record Creation and Access, Proc. ICASSP 2001, Vol.1, pp (2001). 6) Yu, H., Clark, C., Malkin, R. and Waibel, A.: Experiments in Automatic Meeting Transcription Using JRTk, Proc. ICASSP 98, Vol.2, pp (1998). 7) PodCastle: Vol.25, No.1, pp (2010). 8) Goto, M., Ogata, J. and Eto, K.: PodCastle: A Web 2.0 Approach to Speech Recognition Research, Proc. of Interspeech 2007, pp (2007). 9) Ogata, J. and Goto, M.: PodCastle: Collaborative Training of Acoustic Models on the Basis of Wisdom of Crowds for Podcast Transcription, Proc. of Interspeech 2009, pp (2009). 10) Goto, M. and Ogata, J.: PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions, Proc. of PACLIC 24, pp.3 11 (2010). 11) PodCastle: Web pp (2008). 12) Ogata, J., Goto, M. and Eto, K.: Automatic Transcription for a Web 2.0 Service to Search Podcasts, Proc. of Interspeech 2007, pp (2007). 13) PodCastle: 2010-SLP-84-2 (2010). 14) Gales, M. J.F.: Maximal likelihood linear transformations for HMM-Based speech recognition, Computer Speech & Language, Vol.12, pp (1998). 15) 2 pp (2008). 16) (D-II) Vol.J84-D-II, pp (2001). 17) Akita, Y., Mimura, M. and Kawahara, T.: Automatic Transcription System for Meetings of the Japanese National Congress, Proc. of Interspeech 2009 (2009). 18) Neubig, G SLP-84-3 (2010). 19) Kawahara, T., Nanjo, H., Shinozaki, T. and Furui, S.: Benchmark test for speech recognition using the corpus of spontaneous japanese, Proc. SSPR 2003 (2003). 20) 2 pp.7 14 (2008). 21) Ogata, J. and Goto, M.: Speech Repair: Quick Error Correction Just by Using Selection Operation for Speech Input Interfaces, Proc. of Interspeech 2005, pp (2005). 6 c 2011 Information Processing Society of Japan

PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions

PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions PACLIC 24 Proceedings 3 PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions Masataka Goto and Jun Ogata National Institute of Advanced Industrial Science and Technology

More information

Constrained Discriminative Training of N-gram Language Models

Constrained Discriminative Training of N-gram Language Models Constrained Discriminative Training of N-gram Language Models Ariya Rastrow #1, Abhinav Sethy 2, Bhuvana Ramabhadran 3 # Human Language Technology Center of Excellence, and Center for Language and Speech

More information

PodCastle and Songle: Crowdsourcing-Based Web Services for Retrieval and Browsing of Speech and Music Content

PodCastle and Songle: Crowdsourcing-Based Web Services for Retrieval and Browsing of Speech and Music Content PodCastle and Songle: Crowdsourcing-Based Web Services for Retrieval and Browsing of Speech and Music Content Masataka Goto Jun Ogata Kazuyoshi Yoshii Hiromasa Fujihara Matthias Mauch Tomoyasu Nakano National

More information

Recent Development of Open-Source Speech Recognition Engine Julius

Recent Development of Open-Source Speech Recognition Engine Julius Recent Development of Open-Source Speech Recognition Engine Julius Akinobu Lee and Tatsuya Kawahara Nagoya Institute of Technology, Nagoya, Aichi 466-8555, Japan E-mail: ri@nitech.ac.jp Kyoto University,

More information

Using Gradient Descent Optimization for Acoustics Training from Heterogeneous Data

Using Gradient Descent Optimization for Acoustics Training from Heterogeneous Data Using Gradient Descent Optimization for Acoustics Training from Heterogeneous Data Martin Karafiát Λ, Igor Szöke, and Jan Černocký Brno University of Technology, Faculty of Information Technology Department

More information

GMM-FREE DNN TRAINING. Andrew Senior, Georg Heigold, Michiel Bacchiani, Hank Liao

GMM-FREE DNN TRAINING. Andrew Senior, Georg Heigold, Michiel Bacchiani, Hank Liao GMM-FREE DNN TRAINING Andrew Senior, Georg Heigold, Michiel Bacchiani, Hank Liao Google Inc., New York {andrewsenior,heigold,michiel,hankliao}@google.com ABSTRACT While deep neural networks (DNNs) have

More information

Selection of Best Match Keyword using Spoken Term Detection for Spoken Document Indexing

Selection of Best Match Keyword using Spoken Term Detection for Spoken Document Indexing Selection of Best Match Keyword using Spoken Term Detection for Spoken Document Indexing Kentaro Domoto, Takehito Utsuro, Naoki Sawada and Hiromitsu Nishizaki Graduate School of Systems and Information

More information

Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV

Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV Jan Vaněk and Josef V. Psutka Department of Cybernetics, West Bohemia University,

More information

LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS

LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS Tara N. Sainath, Brian Kingsbury, Vikas Sindhwani, Ebru Arisoy, Bhuvana Ramabhadran IBM T. J. Watson

More information

N-gram Weighting: Reducing Training Data Mismatch in Cross-Domain Language Model Estimation

N-gram Weighting: Reducing Training Data Mismatch in Cross-Domain Language Model Estimation N-gram Weighting: Reducing Training Data Mismatch in Cross-Domain Language Model Estimation Bo-June (Paul) Hsu, James Glass MIT Computer Science and Artificial Intelligence Laboratory 32 Vassar Street,

More information

A Model for Meeting Content Storage and Retrieval

A Model for Meeting Content Storage and Retrieval A Model for Meeting Content Storage and Retrieval Saturnino Luz Department of Computer Science Trinity College, University of Dublin Dublin 2, Ireland luzs@cs.tcd.ie Masood Masoodian Department of Computer

More information

Writer Identification for Smart Meeting Room Systems

Writer Identification for Smart Meeting Room Systems Writer Identification for Smart Meeting Room Systems Marcus Liwicki 1, Andreas Schlapbach 1, Horst Bunke 1, Samy Bengio 2, Johnny Mariéthoz 2, and Jonas Richiardi 3 1 Department of Computer Science, University

More information

Scott Shaobing Chen & P.S. Gopalakrishnan. IBM T.J. Watson Research Center. as follows:

Scott Shaobing Chen & P.S. Gopalakrishnan. IBM T.J. Watson Research Center.   as follows: SPEAKER, ENVIRONMENT AND CHANNEL CHANGE DETECTION AND CLUSTERING VIA THE BAYESIAN INFORMATION CRITERION Scott Shaobing Chen & P.S. Gopalakrishnan IBM T.J. Watson Research Center email: schen@watson.ibm.com

More information

Lecture 7: Neural network acoustic models in speech recognition

Lecture 7: Neural network acoustic models in speech recognition CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 7: Neural network acoustic models in speech recognition Outline Hybrid acoustic modeling overview Basic

More information

Improving CTC-based Acoustic Model with Very Deep Residual Time-delay Neural Networks

Improving CTC-based Acoustic Model with Very Deep Residual Time-delay Neural Networks Interspeech 2018 2-6 September 2018, Hyderabad Improving CTC-based Acoustic Model with Very Deep Residual Time-delay Neural Networks Sheng Li 1, Xugang Lu 1, Ryoichi Takashima 1, Peng Shen 1, Tatsuya Kawahara

More information

THE most popular training method for hidden Markov

THE most popular training method for hidden Markov 204 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 12, NO. 3, MAY 2004 A Discriminative Training Algorithm for Hidden Markov Models Assaf Ben-Yishai and David Burshtein, Senior Member, IEEE Abstract

More information

A ROBUST SPEAKER CLUSTERING ALGORITHM

A ROBUST SPEAKER CLUSTERING ALGORITHM A ROBUST SPEAKER CLUSTERING ALGORITHM J. Ajmera IDIAP P.O. Box 592 CH-1920 Martigny, Switzerland jitendra@idiap.ch C. Wooters ICSI 1947 Center St., Suite 600 Berkeley, CA 94704, USA wooters@icsi.berkeley.edu

More information

NOMOS: A Semantic Web Software Framework for Annotation of Multimodal Corpora

NOMOS: A Semantic Web Software Framework for Annotation of Multimodal Corpora NOMOS: A Semantic Web Software Framework for Annotation of Multimodal Corpora John Niekrasz, Alexander Gruenstein Center for the Study of Language and Information Stanford University niekrasz@stanford.edu

More information

Development, Evaluation and Automatic Segmentation of Slovenian Broadcast News Speech Database

Development, Evaluation and Automatic Segmentation of Slovenian Broadcast News Speech Database Development, Evaluation and Automatic Segmentation of Slovenian Broadcast News Speech Database Janez Žibert, France Mihelič Faculty of Electrical Engineering University of Ljubljana Tržaška 25, 1000 Ljubljana,

More information

Using Document Summarization Techniques for Speech Data Subset Selection

Using Document Summarization Techniques for Speech Data Subset Selection Using Document Summarization Techniques for Speech Data Subset Selection Kai Wei, Yuzong Liu, Katrin Kirchhoff, Jeff Bilmes Department of Electrical Engineering University of Washington Seattle, WA 98195,

More information

SPIDER: A Continuous Speech Light Decoder

SPIDER: A Continuous Speech Light Decoder SPIDER: A Continuous Speech Light Decoder Abdelaziz AAbdelhamid, Waleed HAbdulla, and Bruce AMacDonald Department of Electrical and Computer Engineering, Auckland University, New Zealand E-mail: aabd127@aucklanduniacnz,

More information

Combining Audio and Video for Detection of Spontaneous Emotions

Combining Audio and Video for Detection of Spontaneous Emotions Combining Audio and Video for Detection of Spontaneous Emotions Rok Gajšek, Vitomir Štruc, Simon Dobrišek, Janez Žibert, France Mihelič, and Nikola Pavešić Faculty of Electrical Engineering, University

More information

Parallelizing Speaker-Attributed Speech Recognition for Meeting Browsing

Parallelizing Speaker-Attributed Speech Recognition for Meeting Browsing 2010 IEEE International Symposium on Multimedia Parallelizing Speaker-Attributed Speech Recognition for Meeting Browsing Gerald Friedland, Jike Chong, Adam Janin International Computer Science Institute

More information

PAPER Model Shrinkage for Discriminative Language Models

PAPER Model Shrinkage for Discriminative Language Models IEICE TRANS. INF. & SYST., VOL.E95 D, NO.5 MAY 2012 1465 PAPER Model Shrinkage for Discriminative Language Models Takanobu OBA, a), Takaaki HORI, Atsushi NAKAMURA, and Akinori ITO, Members SUMMARY This

More information

Implementing a Hidden Markov Model Speech Recognition System in Programmable Logic

Implementing a Hidden Markov Model Speech Recognition System in Programmable Logic Implementing a Hidden Markov Model Speech Recognition System in Programmable Logic S.J. Melnikoff, S.F. Quigley & M.J. Russell School of Electronic and Electrical Engineering, University of Birmingham,

More information

Discriminative Training of Decoding Graphs for Large Vocabulary Continuous Speech Recognition

Discriminative Training of Decoding Graphs for Large Vocabulary Continuous Speech Recognition Discriminative Training of Decoding Graphs for Large Vocabulary Continuous Speech Recognition by Hong-Kwang Jeff Kuo, Brian Kingsbury (IBM Research) and Geoffry Zweig (Microsoft Research) ICASSP 2007 Presented

More information

A Parallel Implementation of Viterbi Training for Acoustic Models using Graphics Processing Units

A Parallel Implementation of Viterbi Training for Acoustic Models using Graphics Processing Units A Parallel Implementation of Viterbi Training for Acoustic Models using Graphics Processing Units ABSTRACT Robust and accurate speech recognition systems can only be realized with adequately trained acoustic

More information

RLAT Rapid Language Adaptation Toolkit

RLAT Rapid Language Adaptation Toolkit RLAT Rapid Language Adaptation Toolkit Tim Schlippe May 15, 2012 RLAT Rapid Language Adaptation Toolkit - 2 RLAT Rapid Language Adaptation Toolkit RLAT Rapid Language Adaptation Toolkit - 3 Outline Introduction

More information

Combination of FST and CN Search in Spoken Term Detection

Combination of FST and CN Search in Spoken Term Detection Combination of FST and CN Search in Spoken Term Detection Justin Chiu 1, Yun Wang 1, Jan Trmal 2, Daniel Povey 2, Guoguo Chen 2, Alexander Rudnicky 1 1 Language Technologies Institute, Carnegie Mellon

More information

Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition

Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition ISCA Archive http://www.isca-speech.org/archive Auditory-Visual Speech Processing (AVSP) 2013 Annecy, France August 29 - September 1, 2013 Audio-visual interaction in sparse representation features for

More information

An In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization

An In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization An In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization Yun Wang and Florian Metze Language Technologies Institute, Carnegie Mellon University Pittsburgh, PA, U.S.A.

More information

Making Deep Belief Networks Effective for Large Vocabulary Continuous Speech Recognition

Making Deep Belief Networks Effective for Large Vocabulary Continuous Speech Recognition Making Deep Belief Networks Effective for Large Vocabulary Continuous Speech Recognition Tara N. Sainath 1, Brian Kingsbury 1, Bhuvana Ramabhadran 1, Petr Fousek 2, Petr Novak 2, Abdel-rahman Mohamed 3

More information

A long, deep and wide artificial neural net for robust speech recognition in unknown noise

A long, deep and wide artificial neural net for robust speech recognition in unknown noise A long, deep and wide artificial neural net for robust speech recognition in unknown noise Feipeng Li, Phani S. Nidadavolu, and Hynek Hermansky Center for Language and Speech Processing Johns Hopkins University,

More information

Discriminative training and Feature combination

Discriminative training and Feature combination Discriminative training and Feature combination Steve Renals Automatic Speech Recognition ASR Lecture 13 16 March 2009 Steve Renals Discriminative training and Feature combination 1 Overview Hot topics

More information

Informative Spectro-Temporal Bottleneck Features for Noise-Robust Speech Recognition

Informative Spectro-Temporal Bottleneck Features for Noise-Robust Speech Recognition Informatie Spectro-Temporal Bottleneck Features for Noise-Robust Speech Recognition Shuo-Yiin Chang 1,, Nelson Morgan 1, 1 EECS Department, niersity of California-Berkeley, Berkeley, CA, SA International

More information

THE RT04 EVALUATION STRUCTURAL METADATA SYSTEMS AT CUED. M. Tomalin and P.C. Woodland

THE RT04 EVALUATION STRUCTURAL METADATA SYSTEMS AT CUED. M. Tomalin and P.C. Woodland THE RT04 EVALUATION STRUCTURAL METADATA S AT CUED M. Tomalin and P.C. Woodland Cambridge University Engineering Department, Trumpington Street, Cambridge, CB2 1PZ, UK. Email: mt126,pcw @eng.cam.ac.uk ABSTRACT

More information

An Arabic Optical Character Recognition System Using Restricted Boltzmann Machines

An Arabic Optical Character Recognition System Using Restricted Boltzmann Machines An Arabic Optical Character Recognition System Using Restricted Boltzmann Machines Abdullah M. Rashwan, Mohamed S. Kamel, and Fakhri Karray University of Waterloo Abstract. Most of the state-of-the-art

More information

The NITE XML Toolkit meets the ICSI Meeting Corpus: import, annotation, and browsing. Jean Carletta and Jonathan Kilgour

The NITE XML Toolkit meets the ICSI Meeting Corpus: import, annotation, and browsing. Jean Carletta and Jonathan Kilgour The NITE XML Toolkit meets the ICSI Meeting Corpus: import, annotation, and browsing Jean Carletta and Jonathan Kilgour University of Edinburgh, HCRC Language Technology Group 2, Buccleuch Place, Edinburgh

More information

Processing and Linking Audio Events in Large Multimedia Archives: The EU inevent Project

Processing and Linking Audio Events in Large Multimedia Archives: The EU inevent Project Processing and Linking Audio Events in Large Multimedia Archives: The EU inevent Project H. Bourlard 1,2, M. Ferras 1, N. Pappas 1,2, A. Popescu-Belis 1, S. Renals 3, F. McInnes 3, P. Bell 3, S. Ingram

More information

Speech-based Information Retrieval System with Clarification Dialogue Strategy

Speech-based Information Retrieval System with Clarification Dialogue Strategy Speech-based Information Retrieval System with Clarification Dialogue Strategy Teruhisa Misu Tatsuya Kawahara School of informatics Kyoto University Sakyo-ku, Kyoto, Japan misu@ar.media.kyoto-u.ac.jp Abstract

More information

Tutorial of Building an LVCSR System

Tutorial of Building an LVCSR System Tutorial of Building an LVCSR System using HTK Shih Hsiang Lin( 林士翔 ) Department of Computer Science & Information Engineering National Taiwan Normal University Reference: Steve Young et al, The HTK Books

More information

The HTK Book. The HTK Book (for HTK Version 3.4)

The HTK Book. The HTK Book (for HTK Version 3.4) The HTK Book Steve Young Gunnar Evermann Mark Gales Thomas Hain Dan Kershaw Xunying (Andrew) Liu Gareth Moore Julian Odell Dave Ollason Dan Povey Valtcho Valtchev Phil Woodland The HTK Book (for HTK Version

More information

arxiv: v1 [cs.cl] 30 Jan 2018

arxiv: v1 [cs.cl] 30 Jan 2018 ACCELERATING RECURRENT NEURAL NETWORK LANGUAGE MODEL BASED ONLINE SPEECH RECOGNITION SYSTEM Kyungmin Lee, Chiyoun Park, Namhoon Kim, and Jaewon Lee DMC R&D Center, Samsung Electronics, Seoul, Korea {k.m.lee,

More information

HMM-Based On-Line Recognition of Handwritten Whiteboard Notes

HMM-Based On-Line Recognition of Handwritten Whiteboard Notes HMM-Based On-Line Recognition of Handwritten Whiteboard Notes Marcus Liwicki and Horst Bunke Institute of Computer Science and Applied Mathematics University of Bern, Neubrückstrasse 10, CH-3012 Bern,

More information

Query-by-example spoken term detection based on phonetic posteriorgram Query-by-example spoken term detection based on phonetic posteriorgram

Query-by-example spoken term detection based on phonetic posteriorgram Query-by-example spoken term detection based on phonetic posteriorgram International Conference on Education, Management and Computing Technology (ICEMCT 2015) Query-by-example spoken term detection based on phonetic posteriorgram Query-by-example spoken term detection based

More information

Semantic Word Embedding Neural Network Language Models for Automatic Speech Recognition

Semantic Word Embedding Neural Network Language Models for Automatic Speech Recognition Semantic Word Embedding Neural Network Language Models for Automatic Speech Recognition Kartik Audhkhasi, Abhinav Sethy Bhuvana Ramabhadran Watson Multimodal Group IBM T. J. Watson Research Center Motivation

More information

Introduction to HTK Toolkit

Introduction to HTK Toolkit Introduction to HTK Toolkit Berlin Chen 2003 Reference: - The HTK Book, Version 3.2 Outline An Overview of HTK HTK Processing Stages Data Preparation Tools Training Tools Testing Tools Analysis Tools Homework:

More information

Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri

Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute eugenew@cs.nyu.edu Slide Credit: Mehryar Mohri Speech Recognition Components Acoustic and pronunciation model:

More information

In Silico Vox: Towards Speech Recognition in Silicon

In Silico Vox: Towards Speech Recognition in Silicon In Silico Vox: Towards Speech Recognition in Silicon Edward C Lin, Kai Yu, Rob A Rutenbar, Tsuhan Chen Electrical & Computer Engineering {eclin, kaiy, rutenbar, tsuhan}@ececmuedu RA Rutenbar 2006 Speech

More information

Introduction to The HTK Toolkit

Introduction to The HTK Toolkit Introduction to The HTK Toolkit Hsin-min Wang Reference: - The HTK Book Outline An Overview of HTK HTK Processing Stages Data Preparation Tools Training Tools Testing Tools Analysis Tools A Tutorial Example

More information

Object Recognition by Integrated Information Using Web Images

Object Recognition by Integrated Information Using Web Images 2013 Second IAPR Asian Conference on Pattern Recognition Object Recognition by Integrated Information Using Web Images Hitoshi Nishimura, Yuko Ozasa, Yasuo Ariki, Mikio Nakano Graduate School of System

More information

Short-time Viterbi for online HMM decoding : evaluation on a real-time phone recognition task

Short-time Viterbi for online HMM decoding : evaluation on a real-time phone recognition task Short-time Viterbi for online HMM decoding : evaluation on a real-time phone recognition task Julien Bloit, Xavier Rodet To cite this version: Julien Bloit, Xavier Rodet. Short-time Viterbi for online

More information

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages Chirag Shah Dept. of CSE IIT Madras Chennai - 600036 Tamilnadu, India. chirag@speech.iitm.ernet.in A. Nayeemulla Khan Dept. of CSE

More information

COMBINING FEATURE SETS WITH SUPPORT VECTOR MACHINES: APPLICATION TO SPEAKER RECOGNITION

COMBINING FEATURE SETS WITH SUPPORT VECTOR MACHINES: APPLICATION TO SPEAKER RECOGNITION COMBINING FEATURE SETS WITH SUPPORT VECTOR MACHINES: APPLICATION TO SPEAKER RECOGNITION Andrew O. Hatch ;2, Andreas Stolcke ;3, and Barbara Peskin The International Computer Science Institute, Berkeley,

More information

JTrans, an open-source software for semi-automatic text-to-speech alignment

JTrans, an open-source software for semi-automatic text-to-speech alignment JTrans, an open-source software for semi-automatic text-to-speech alignment Christophe Cerisara, Odile Mella, Dominique Fohr To cite this version: Christophe Cerisara, Odile Mella, Dominique Fohr. JTrans,

More information

Baseball Game Highlight & Event Detection

Baseball Game Highlight & Event Detection Baseball Game Highlight & Event Detection Student: Harry Chao Course Adviser: Winston Hu 1 Outline 1. Goal 2. Previous methods 3. My flowchart 4. My methods 5. Experimental result 6. Conclusion & Future

More information

A study of large vocabulary speech recognition decoding using finite-state graphs 1

A study of large vocabulary speech recognition decoding using finite-state graphs 1 A study of large vocabulary speech recognition decoding using finite-state graphs 1 Zhijian OU, Ji XIAO Department of Electronic Engineering, Tsinghua University, Beijing Corresponding email: ozj@tsinghua.edu.cn

More information

WFSTDM Builder Network-based Spoken Dialogue System Builder for Easy Prototyping

WFSTDM Builder Network-based Spoken Dialogue System Builder for Easy Prototyping WFSTDM Builder Network-based Spoken Dialogue System Builder for Easy Prototyping Etsuo Mizukami and Chiori Hori Abstract This paper introduces a network-based spoken dialog system development tool kit:

More information

Learning N-gram Language Models from Uncertain Data

Learning N-gram Language Models from Uncertain Data Learning N-gram Language Models from Uncertain Data Vitaly Kuznetsov 1,2, Hank Liao 2, Mehryar Mohri 1,2, Michael Riley 2, Brian Roark 2 1 Courant Institute, New York University 2 Google, Inc. vitalyk,hankliao,mohri,riley,roark}@google.com

More information

Feature Engineering in Context-Dependent Deep Neural Networks for Conversational Speech Transcription

Feature Engineering in Context-Dependent Deep Neural Networks for Conversational Speech Transcription Feature Engineering in Context-Dependent Deep Neural Networks for Conversational Speech Transcription Frank Seide 1,GangLi 1, Xie Chen 1,2, and Dong Yu 3 1 Microsoft Research Asia, 5 Danling Street, Haidian

More information

SVD-based Universal DNN Modeling for Multiple Scenarios

SVD-based Universal DNN Modeling for Multiple Scenarios SVD-based Universal DNN Modeling for Multiple Scenarios Changliang Liu 1, Jinyu Li 2, Yifan Gong 2 1 Microsoft Search echnology Center Asia, Beijing, China 2 Microsoft Corporation, One Microsoft Way, Redmond,

More information

Confidence Measures: how much we can trust our speech recognizers

Confidence Measures: how much we can trust our speech recognizers Confidence Measures: how much we can trust our speech recognizers Prof. Hui Jiang Department of Computer Science York University, Toronto, Ontario, Canada Email: hj@cs.yorku.ca Outline Speech recognition

More information

From Multimedia Retrieval to Knowledge Management. Pedro J. Moreno JM Van Thong Beth Logan

From Multimedia Retrieval to Knowledge Management. Pedro J. Moreno JM Van Thong Beth Logan From Multimedia Retrieval to Knowledge Management Pedro J. Moreno JM Van Thong Beth Logan CRL 2002/02 March 2002 From Multimedia Retrieval to Knowledge Management Pedro J. Moreno JM Van Thong Beth Logan

More information

Discriminative Training and Adaptation of Large Vocabulary ASR Systems

Discriminative Training and Adaptation of Large Vocabulary ASR Systems Discriminative Training and Adaptation of Large Vocabulary ASR Systems Phil Woodland March 30th 2004 ICSI Seminar: March 30th 2004 Overview Why use discriminative training for LVCSR? MMIE/CMLE criterion

More information

SPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION

SPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION Far East Journal of Electronics and Communications Volume 3, Number 2, 2009, Pages 125-140 Published Online: September 14, 2009 This paper is available online at http://www.pphmj.com 2009 Pushpa Publishing

More information

A Survey on Spoken Document Indexing and Retrieval

A Survey on Spoken Document Indexing and Retrieval A Survey on Spoken Document Indexing and Retrieval Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University Introduction (1/5) Ever-increasing volumes of audio-visual

More information

Shrinking Exponential Language Models

Shrinking Exponential Language Models Shrinking Exponential Language Models Stanley F. Chen IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights, NY 10598 stanchen@watson.ibm.com Abstract In (Chen, 2009), we show that for a variety

More information

Graph-based Submodular Selection for Extractive Summarization

Graph-based Submodular Selection for Extractive Summarization Graph-based Submodular Selection for Extractive Summarization Hui Lin 1, Jeff Bilmes 1, Shasha Xie 2 1 Department of Electrical Engineering, University of Washington Seattle, WA 98195, United States {hlin,bilmes}@ee.washington.edu

More information

Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos

Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos K.Vijaya Kumar (09305081) under the guidance of Prof. Sridhar Iyer June 28, 2011 1 / 66 Outline Outline 1 Introduction 2 Motivation 3

More information

Novel Methods for Query Selection and Query Combination in Query-By-Example Spoken Term Detection

Novel Methods for Query Selection and Query Combination in Query-By-Example Spoken Term Detection Novel Methods for Query Selection and Query Combination in Query-By-Example Spoken Term Detection Javier Tejedor HCTLab, Universidad Autónoma de Madrid, Spain javier.tejedor@uam.es Igor Szöke Speech@FIT,

More information

HIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION. Hung-An Chang and James R. Glass

HIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION. Hung-An Chang and James R. Glass HIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION Hung-An Chang and James R. Glass MIT Computer Science and Artificial Intelligence Laboratory Cambridge, Massachusetts, 02139,

More information

LATENT SEMANTIC INDEXING BY SELF-ORGANIZING MAP. Mikko Kurimo and Chafic Mokbel

LATENT SEMANTIC INDEXING BY SELF-ORGANIZING MAP. Mikko Kurimo and Chafic Mokbel LATENT SEMANTIC INDEXING BY SELF-ORGANIZING MAP Mikko Kurimo and Chafic Mokbel IDIAP CP-592, Rue du Simplon 4, CH-1920 Martigny, Switzerland Email: Mikko.Kurimo@idiap.ch ABSTRACT An important problem for

More information

DANCEREPRODUCER: AN AUTOMATIC MASHUP MUSIC VIDEO GENERATION SYSTEM BY REUSING DANCE VIDEO CLIPS ON THE WEB

DANCEREPRODUCER: AN AUTOMATIC MASHUP MUSIC VIDEO GENERATION SYSTEM BY REUSING DANCE VIDEO CLIPS ON THE WEB DANCEREPRODUCER: AN AUTOMATIC MASHUP MUSIC VIDEO GENERATION SYSTEM BY REUSING DANCE VIDEO CLIPS ON THE WEB Tomoyasu Nakano 1 Sora Murofushi 3 Masataka Goto 2 Shigeo Morishima 3 1 t.nakano[at]aist.go.jp

More information

The HTK Book. Steve Young Gunnar Evermann Dan Kershaw Gareth Moore Julian Odell Dave Ollason Dan Povey Valtcho Valtchev Phil Woodland

The HTK Book. Steve Young Gunnar Evermann Dan Kershaw Gareth Moore Julian Odell Dave Ollason Dan Povey Valtcho Valtchev Phil Woodland The HTK Book Steve Young Gunnar Evermann Dan Kershaw Gareth Moore Julian Odell Dave Ollason Dan Povey Valtcho Valtchev Phil Woodland The HTK Book (for HTK Version 3.2) c COPYRIGHT 1995-1999 Microsoft Corporation.

More information

DECODING VISEMES: IMPROVING MACHINE LIP-READING. Helen L. Bear and Richard Harvey

DECODING VISEMES: IMPROVING MACHINE LIP-READING. Helen L. Bear and Richard Harvey DECODING VISEMES: IMPROVING MACHINE LIP-READING Helen L. Bear and Richard Harvey School of Computing Sciences, University of East Anglia, Norwich, NR4 7TJ, United Kingdom ABSTRACT To undertake machine

More information

In Silico Vox: Towards Speech Recognition in Silicon

In Silico Vox: Towards Speech Recognition in Silicon In Silico Vox: Towards Speech Recognition in Silicon Edward C Lin, Kai Yu, Rob A Rutenbar, Tsuhan Chen Electrical & Computer Engineering {eclin, kaiy, rutenbar, tsuhan}@ececmuedu RA Rutenbar 006 Speech

More information

Spoken Document Retrieval and Browsing. Ciprian Chelba

Spoken Document Retrieval and Browsing. Ciprian Chelba Spoken Document Retrieval and Browsing Ciprian Chelba OpenFst Library C++ template library for constructing, combining, optimizing, and searching weighted finite-states transducers (FSTs) Goals: Comprehensive,

More information

Speech recognition on an FPGA using discrete and continuous hidden Markov models Melnikoff, Stephen; Quigley, Steven; Russell, Martin

Speech recognition on an FPGA using discrete and continuous hidden Markov models Melnikoff, Stephen; Quigley, Steven; Russell, Martin Speech recognition on an FPGA using discrete and continuous hidden Markov models Melnikoff, Stephen; Quigley, Steven; Russell, Martin Document Version Peer reviewed version Citation for published version

More information

Fusion of LVCSR and Posteriorgram Based Keyword Search

Fusion of LVCSR and Posteriorgram Based Keyword Search INTERSPEECH 2015 Fusion of LVCSR and Posteriorgram Based Keyword Search Leda Sarı, Batuhan Gündoğdu, Murat Saraçlar Boğaziçi University, Bebek, Istanbul, 34342, Turkey {leda.sari, batuhan.gundogdu, murat.saraclar}@boun.edu.tr

More information

Reverberant Speech Recognition Based on Denoising Autoencoder

Reverberant Speech Recognition Based on Denoising Autoencoder INTERSPEECH 2013 Reverberant Speech Recognition Based on Denoising Autoencoder Takaaki Ishii 1, Hiroki Komiyama 1, Takahiro Shinozaki 2, Yasuo Horiuchi 1, Shingo Kuroiwa 1 1 Division of Information Sciences,

More information

SUBBAND ACOUSTIC WAVEFORM FRONT-END FOR ROBUST SPEECH RECOGNITION USING SUPPORT VECTOR MACHINES

SUBBAND ACOUSTIC WAVEFORM FRONT-END FOR ROBUST SPEECH RECOGNITION USING SUPPORT VECTOR MACHINES SUBBAND ACOUSTIC WAVEFORM FRONT-END FOR ROBUST SPEECH RECOGNITION USING SUPPORT VECTOR MACHINES Jibran Yousafzai 1, Zoran Cvetković 1, Peter Sollich 2 Department of Electronic Engineering 1 & Department

More information

A Novel Approach to On-Line Handwriting Recognition Based on Bidirectional Long Short-Term Memory Networks

A Novel Approach to On-Line Handwriting Recognition Based on Bidirectional Long Short-Term Memory Networks A Novel Approach to n-line Handwriting Recognition Based on Bidirectional Long Short-Term Memory Networks Marcus Liwicki 1 Alex Graves 2 Horst Bunke 1 Jürgen Schmidhuber 2,3 1 nst. of Computer Science

More information

REAL-TIME ASR FROM MEETINGS

REAL-TIME ASR FROM MEETINGS RESEARCH IDIAP REPORT REAL-TIME ASR FROM MEETINGS Philip N. Garner John Dines Thomas Hain Asmaa El Hannani Martin Karafiat Danil Korchagin Mike Lincoln Vincent Wan Le Zhang Idiap-RR-15-2009 JUNE 2009 Centre

More information

Real-Time Keyword Extraction from Conversations

Real-Time Keyword Extraction from Conversations Real-Time Keyword Extraction from Conversations Polykarpos Meladianos École Polytechnique & AUEB pmeladianos@aueb.gr Antoine J.-P. Tixier École Polytechnique antoine.tixier-1@colorado.edu Giannis Nikolentzos

More information

Car Information Systems for ITS

Car Information Systems for ITS Car Information Systems for ITS 102 Car Information Systems for ITS Kozo Nakamura Ichiro Hondo Nobuo Hataoka, Ph.D. Shiro Horii OVERVIEW: For ITS (intelligent transport systems) car information systems,

More information

Duration and Interval Hidden Markov Model for Sequential Data Analysis

Duration and Interval Hidden Markov Model for Sequential Data Analysis Duration and Interval Hidden Markov Model for Sequential Data Analysis arxiv:1508.04928v1 [cs.ai] 20 Aug 2015 Hiromi Narimatsu Graduate School of Information Systems The University of Electro-Communications

More information

Variable-Component Deep Neural Network for Robust Speech Recognition

Variable-Component Deep Neural Network for Robust Speech Recognition Variable-Component Deep Neural Network for Robust Speech Recognition Rui Zhao 1, Jinyu Li 2, and Yifan Gong 2 1 Microsoft Search Technology Center Asia, Beijing, China 2 Microsoft Corporation, One Microsoft

More information

Research Article Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images

Research Article Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images Hindawi Publishing Corporation EURASIP Journal on Audio, Speech, and Music Processing Volume 2007, Article ID 64506, 9 pages doi:10.1155/2007/64506 Research Article Audio-Visual Speech Recognition Using

More information

Conundrums in Unsupervised Keyphrase Extraction: Making Sense of the State-of-the-Art

Conundrums in Unsupervised Keyphrase Extraction: Making Sense of the State-of-the-Art Beijing, China, August 2, pp. 365--373. Conundrums in Unsupervised Keyphrase Extraction: Making Sense of the State-of-the-Art Kazi Saidul Hasan and Vincent Ng Human Language Technology Research Institute

More information

Robust DTW-based Recognition Algorithm for Hand-held Consumer Devices

Robust DTW-based Recognition Algorithm for Hand-held Consumer Devices C. Kim and K.-d. Seo: Robust DTW-based Recognition for Hand-held Consumer Devices Robust DTW-based Recognition for Hand-held Consumer Devices 699 Chanwoo Kim and Kwang-deok Seo, Member, IEEE Abstract This

More information

A Gaussian Mixture Model Spectral Representation for Speech Recognition

A Gaussian Mixture Model Spectral Representation for Speech Recognition A Gaussian Mixture Model Spectral Representation for Speech Recognition Matthew Nicholas Stuttle Hughes Hall and Cambridge University Engineering Department PSfrag replacements July 2003 Dissertation submitted

More information

Learning and Modeling Unit Embeddings for Improving HMM-based Unit Selection Speech Synthesis

Learning and Modeling Unit Embeddings for Improving HMM-based Unit Selection Speech Synthesis Interspeech 2018 2-6 September 2018, Hyderabad Learning and Modeling Unit Embeddings for Improving HMM-based Unit Selection Synthesis Xiao Zhou, Zhen-Hua Ling, Zhi-Ping Zhou, Li-Rong Dai National Engineering

More information

A Comparison of Visual Features for Audio-Visual Automatic Speech Recognition

A Comparison of Visual Features for Audio-Visual Automatic Speech Recognition A Comparison of Visual Features for Audio-Visual Automatic Speech Recognition N. Ahmad, S. Datta, D. Mulvaney and O. Farooq Loughborough Univ, LE11 3TU Leicestershire, UK n.ahmad@lboro.ac.uk 6445 Abstract

More information

THE THISL BROADCAST NEWS RETRIEVAL SYSTEM. Dave Abberley (1), David Kirby (2), Steve Renals (1) and Tony Robinson (3)

THE THISL BROADCAST NEWS RETRIEVAL SYSTEM. Dave Abberley (1), David Kirby (2), Steve Renals (1) and Tony Robinson (3) ISCA Archive THE THISL BROADCAST NEWS RETRIEVAL SYSTEM Dave Abberley (1), David Kirby (2), Steve Renals (1) and Tony Robinson (3) (1) University of Sheffield, Department of Computer Science, UK (2) BBC,

More information

Phoneme Analysis of Image Feature in Utterance Recognition Using AAM in Lip Area

Phoneme Analysis of Image Feature in Utterance Recognition Using AAM in Lip Area MIRU 7 AAM 657 81 1 1 657 81 1 1 E-mail: {komai,miyamoto}@me.cs.scitec.kobe-u.ac.jp, {takigu,ariki}@kobe-u.ac.jp Active Appearance Model Active Appearance Model combined DCT Active Appearance Model combined

More information

UNSUPERVISED SUBMODULAR SUBSET SELECTION FOR SPEECH DATA :EXTENDED VERSION. Kai Wei Yuzong Liu Katrin Kirchhoff Jeff Bilmes

UNSUPERVISED SUBMODULAR SUBSET SELECTION FOR SPEECH DATA :EXTENDED VERSION. Kai Wei Yuzong Liu Katrin Kirchhoff Jeff Bilmes UNSUPERVISED SUBMODULAR SUBSET SELECTION FOR SPEECH DATA :EXTENDED VERSION Kai Wei Yuzong Liu Katrin Kirchhoff Jeff Bilmes Department of Electrical Engineering, University of Washington Seattle, WA 98195,

More information

Vulnerability of Voice Verification System with STC anti-spoofing detector to different methods of spoofing attacks

Vulnerability of Voice Verification System with STC anti-spoofing detector to different methods of spoofing attacks Vulnerability of Voice Verification System with STC anti-spoofing detector to different methods of spoofing attacks Vadim Shchemelinin 1,2, Alexandr Kozlov 2, Galina Lavrentyeva 2, Sergey Novoselov 1,2

More information

D6.1.2: Second report on scientific evaluations

D6.1.2: Second report on scientific evaluations D6.1.2: Second report on scientific evaluations UPVLC, XEROX, JSI-K4A, RWTH, EML and DDS Distribution: Public translectures Transcription and Translation of Video Lectures ICT Project 287755 Deliverable

More information

SAUTRELA: A HIGHLY MODULAR OPEN SOURCE SPEECH RECOGNITION FRAMEWORK. Mikel Penagarikano and German Bordel

SAUTRELA: A HIGHLY MODULAR OPEN SOURCE SPEECH RECOGNITION FRAMEWORK. Mikel Penagarikano and German Bordel SAUTRELA: A HIGHLY MODULAR OPEN SOURCE SPEECH RECOGNITION FRAMEWORK Mikel Penagarikano and German Bordel Department of Electricity and Electronics University of the Basque Country, 48940 Leioa, Spain E-mail:

More information

Voice-driven Interaction: Harnessing the capacity of human voice for controlling computer interfaces

Voice-driven Interaction: Harnessing the capacity of human voice for controlling computer interfaces Voice-driven Interaction: Harnessing the capacity of human voice for controlling computer interfaces Susumu Harada Computer Science and Engineering Jacob O. Wobbrock The Information School James A. Landay

More information