A ROBUST SPEAKER CLUSTERING ALGORITHM

Similar documents
Speaker Diarization System Based on GMM and BIC

Client Dependent GMM-SVM Models for Speaker Verification

Scott Shaobing Chen & P.S. Gopalakrishnan. IBM T.J. Watson Research Center. as follows:

Optimizing feature representation for speaker diarization using PCA and LDA

application of learning vector quantization algorithms. In Proceedings of the International Joint Conference on

Hands On: Multimedia Methods for Large Scale Video Analysis (Lecture) Dr. Gerald Friedland,

Introduction to HTK Toolkit

THE most popular training method for hidden Markov

Some questions of consensus building using co-association

A SCANNING WINDOW SCHEME BASED ON SVM TRAINING ERROR RATE FOR UNSUPERVISED AUDIO SEGMENTATION

Chapter 3. Speech segmentation. 3.1 Preprocessing

Optimization of Observation Membership Function By Particle Swarm Method for Enhancing Performances of Speaker Identification

EM Algorithm with Split and Merge in Trajectory Clustering for Automatic Speech Recognition

Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV

Multimedia Event Detection for Large Scale Video. Benjamin Elizalde

The Expected Performance Curve: a New Assessment Measure for Person Authentication

The Expected Performance Curve: a New Assessment Measure for Person Authentication

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages

SRE08 system. Nir Krause Ran Gazit Gennady Karvitsky. Leave Impersonators, fraudsters and identity thieves speechless

Multifactor Fusion for Audio-Visual Speaker Recognition

An Introduction to Pattern Recognition

Person Authentication from Video of Faces: A Behavioral and Physiological Approach Using Pseudo Hierarchical Hidden Markov Models

TWO-STEP SEMI-SUPERVISED APPROACH FOR MUSIC STRUCTURAL CLASSIFICATION. Prateek Verma, Yang-Kai Lin, Li-Fan Yu. Stanford University

Noisy Text Clustering

Content Based Classification of Audio Using MPEG-7 Features

Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri

Xing Fan, Carlos Busso and John H.L. Hansen

Towards Audio-Visual On-line Diarization Of Participants In Group Meetings

SPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION

CHROMA AND MFCC BASED PATTERN RECOGNITION IN AUDIO FILES UTILIZING HIDDEN MARKOV MODELS AND DYNAMIC PROGRAMMING. Alexander Wankhammer Peter Sciri

Confidence Measures: how much we can trust our speech recognizers

2-2-2, Hikaridai, Seika-cho, Soraku-gun, Kyoto , Japan 2 Graduate School of Information Science, Nara Institute of Science and Technology

Introduction to Mobile Robotics

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Comparative Evaluation of Feature Normalization Techniques for Speaker Verification

Generation of Sports Highlights Using a Combination of Supervised & Unsupervised Learning in Audio Domain

Genetic algorithm with deterministic crossover for vector quantization

Modeling time series with hidden Markov models

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

Speaker Localisation Using Audio-Visual Synchrony: An Empirical Study

Multi-Modal Human Verification Using Face and Speech

HIERARCHICAL LARGE-MARGIN GAUSSIAN MIXTURE MODELS FOR PHONETIC CLASSIFICATION. Hung-An Chang and James R. Glass

LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

A Novel Template Matching Approach To Speaker-Independent Arabic Spoken Digit Recognition

Using Gradient Descent Optimization for Acoustics Training from Heterogeneous Data

SOUND EVENT DETECTION AND CONTEXT RECOGNITION 1 INTRODUCTION. Toni Heittola 1, Annamaria Mesaros 1, Tuomas Virtanen 1, Antti Eronen 2

Dynamic Thresholding for Image Analysis

Ergodic Hidden Markov Models for Workload Characterization Problems

Applied Clustering for Automatic Speaker-Based Segmentation of Audio Material

Variable-Component Deep Neural Network for Robust Speech Recognition

Applications of Keyword-Constraining in Speaker Recognition. Howard Lei. July 2, Introduction 3

1 1 λ ( i 1) Sync diagram is the lack of a synchronization stage, which isthe main advantage of this method. Each iteration of ITSAT performs ex

Jing Gao 1, Feng Liang 1, Wei Fan 2, Chi Wang 1, Yizhou Sun 1, Jiawei i Han 1 University of Illinois, IBM TJ Watson.

Parallelizing Speaker-Attributed Speech Recognition for Meeting Browsing

Video Summarization Using MPEG-7 Motion Activity and Audio Descriptors

Dynamic Time Warping

The Approach of Mean Shift based Cosine Dissimilarity for Multi-Recording Speaker Clustering

ORT EP R RCH A ESE R P A IDI! " #$$% &' (# $!"

Text-Independent Speaker Identification

UNSUPERVISED LOCATION-BASED SEGMENTATION OF MULTI-PARTY SPEECH. G. Lathoud, I.A. McCowan and J.M. Odobez

Evaluating String Comparator Performance for Record Linkage William E. Yancey Statistical Research Division U.S. Census Bureau

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

Discriminative training and Feature combination

Available online Journal of Scientific and Engineering Research, 2016, 3(4): Research Article

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Maximum Likelihood Beamforming for Robust Automatic Speech Recognition

Introduction to The HTK Toolkit

Highlights Extraction from Unscripted Video

OPTIMIZING CHEMICAL SENSOR ARRAY SIZES

Document Image Restoration Using Binary Morphological Filters. Jisheng Liang, Robert M. Haralick. Seattle, Washington Ihsin T.

An In-Depth Comparison of Keyword Specific Thresholding and Sum-to-One Score Normalization

Bus Detection and recognition for visually impaired people

6. Dicretization methods 6.1 The purpose of discretization

Short-time Viterbi for online HMM decoding : evaluation on a real-time phone recognition task

DUPLICATE DETECTION AND AUDIO THUMBNAILS WITH AUDIO FINGERPRINTING

Modeling Phonetic Context with Non-random Forests for Speech Recognition

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning

INF 4300 Classification III Anne Solberg The agenda today:

Random projection for non-gaussian mixture models

THE RT04 EVALUATION STRUCTURAL METADATA SYSTEMS AT CUED. M. Tomalin and P.C. Woodland

Speaker Verification with Adaptive Spectral Subband Centroids

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Probabilistic scoring using decision trees for fast and scalable speaker recognition

Pitch Prediction from Mel-frequency Cepstral Coefficients Using Sparse Spectrum Recovery

Face Verification Using Adapted Generative Models

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

Least Squares Signal Declipping for Robust Speech Recognition

Telecommunication and Informatics University of North Carolina, Technical University of Gdansk Charlotte, NC 28223, USA

Improving Speaker Verification Performance in Presence of Spoofing Attacks Using Out-of-Domain Spoofed Data

Conditional Random Field for tracking user behavior based on his eye s movements 1

Aditi Upadhyay Research Scholar, Department of Electronics & Communication Engineering Jaipur National University, Jaipur, Rajasthan, India

Hands On: Multimedia Methods for Large Scale Video Analysis (Lecture) Dr. Gerald Friedland,

Background Subtraction in Video using Bayesian Learning with Motion Information Suman K. Mitra DA-IICT, Gandhinagar

Intelligent Hands Free Speech based SMS System on Android

A Study on Clustering Method by Self-Organizing Map and Information Criteria

Voiced-Unvoiced-Silence Classification via Hierarchical Dual Geometry Analysis

the number of states must be set in advance, i.e. the structure of the model is not t to the data, but given a priori the algorithm converges to a loc

A Formal Approach to Score Normalization for Meta-search

Transcription:

A ROBUST SPEAKER CLUSTERING ALGORITHM J. Ajmera IDIAP P.O. Box 592 CH-1920 Martigny, Switzerland jitendra@idiap.ch C. Wooters ICSI 1947 Center St., Suite 600 Berkeley, CA 94704, USA wooters@icsi.berkeley.edu ABSTRACT In this paper, we present a novel speaker segmentation and clustering algorithm. The algorithm automatically performs both speaker segmentation and clustering without any prior knowledge of the identities or the number of speakers. Our algorithm uses standard speech processing components and techniques such as HMMs, agglomerative clustering, and the Bayesian Information Criterion. However, we have combined and modi ed these so as to produce an algorithm with the following advantages: No threshold adjustment requirements No need for training/development data Robustness to different data conditions This paper also reports the performance of this algorithm on different datasets released by the U.S. National Institute of Standards and Technology (NIST) with different initial conditions and parameter settings. The consistently low speaker-diarization error rate clearly indicates the robustness and utility of the algorithm. 1. INTRODUCTION The goal of a speaker segmentation system is to divide a speech signal into a sequence of speaker-homogeneous regions. Thus, the output of such a system provides the answer to the question, Who spoke when? Knowing when each speaker is speaking is useful as a pre-processing step in speech-to-text (STT) systems to improve the quality of the output. Such pre-processing may include vocal tract length normalization (VTLN) and/or speaker adaptation. Automatic speaker segmentation may also be useful in information retrieval and as part of the indexing information of audio archives. Dividing an audio recording into speaker-homogeneous regions presents many challenges. One challenge is to identify the locations of the boundaries between speakers the speaker segmentation problem. Another challenge is to identify which portions of the recording belong to which speakers the speaker clustering problem. Additonally, the speaker clustering problem requires that we correctly identify how many unique speakers occur in the recording. Speech researchers have proposed many techniques for solving the Who spoke when? problem. Most of these methods rst segment and then cluster the data. The segmentation is either assumed to be known a priori [1, 2, 3] or is performed automatically prior to clustering [4, 5]. However, approaches such as these, in which the segmentation and clustering are performed sequentially, have limitations. In the former case, the correct segmentation is rarely known a priori for practical applications. In the latter case, the errors made in the segmentation step can degrade the performance of the subsequent clustering step. Sequential systems such as those mentioned above can be improved by iterating the segmentation and clustering steps. However, they still require some means of guessing the appropriate number of clusters, or a threshold to decide when the appropriate number of clusters has been attained. The approach we are proposing also iterates the segmentation and clustering steps. However, we have developed a fully automatic stopping criterion that produces the optimal number of clusters and the optimal segmentation without the use of thresholds. The proposed algorithm for speaker segmentation deduces the number of clusters automatically while optimizing a likelihood-based objective function. The algorithm runs iteratively, where the likelihood of the data along the best segmentation path (Viterbi score) increases until it reaches an optimal point, then begins decreasing. (See Fig. 1.) We stop the process at the maximum value in the likelihood function. An important property of this algorithm is that it does not have a threshold term to adjust, for which a development test set is generally required. The algorithm is quite robust to different initial conditions and choice of acoustic feature vectors. These properties are demonstrated with the help of several experiments.

1 2 k-1 k states (clusters) Fig. 1. Segmentation as a function of the number of clusters. The likelihood function has a maximum at the correct number of clusters, which in this case is three. sub-states 1 2 S-1 S 2. SPEAKER CLUSTERING ALGORITHM As shown in Fig. 2, our algorithm is based on an ergodic hidden Markov model (HMM) formalism where the number of states in the HMM is equal to the initial number of clusters. Each state is composed of a set of S sub-states. These sub-states impose a minimum duration on the model. Thus, each state of the HMM is a cluster and is expected to represent a single speaker. The probability density function (PDF) of each state is assumed to be a Gaussian mixture model (GMM) with M Gaussian components, which are shared among all sub-states. In the absence of any a priori information about the number of speakers (and hence number of clusters), we start by over-clustering the data. Over-clustering refers to the process of segmenting the data into an initial number of clusters K, where K is greater than the expected number of speakers in the audio le. A result of over-clustering is that data from a single speaker is likely to be assigned to different clusters. Thus, our goal is to identify clusters that have been assigned data from the same source and merge them. It will be shown later that the algorithm is not sensitive to the exact value of K as long as that value is large relative to the actual number of speakers. The rst step in the algorithm is to initialize the parameters of the HMM. We perform this initialization using a uniform segmentation of the data in terms of K clusters and estimating the parameters of the cluster GMMs over these segments. We also tested the K-means algorithm for initialization. The results were not signi cantly different, probably because of the large minimum duration that we impose in our cluster models. The next step is to train the initial HMMs using the standard Expectation-Maximization (EM) algorithm. In the E- step, a segmentation of the data is obtained to maximize the Fig. 2. HMM Topology Used for Clustering GMM likelihood of the data, given the parameters of the GMMs. This is followed by an M-step where the parameters of the GMMs are re-estimated/updated based on this segmentation. The nal step in the algorithm is cluster merging. A consequence of over-clustering is that data from a single speaker may be assigned to different clusters. Thus, the algorithm requires a component that identi es clusters containing data from the same speaker and merges them. The details of cluster merging are explained in Section 2.1 below. Once we have merged a pair of clusters, we return to the segmentation and training step. We repeat the process of segmentation-training-merging until there are no more clusters to be merged. 2.1. Cluster Merging We begin with a high-level description of the overall problem: If X = {x 1,x 2,...,x N } 1 is the audio data to be segmented, we want to nd the optimal number of clusters (k ) and their acoustic models (Λ k ) that produce the best segmentation of the data (X) according to: 1 Generally this is a sequence of acoustic feature vectors extracted from the audio waveform at a regular time interval e.g. every 10ms.

Λ k,k = arg max p(x,q best Λ k,k) (1) Λ k,k where q best is the Viterbi segmentation path with the highest likelihood. Thus, we want to nd the set of clusters and their acoustic models that maximize the likelihood of the data and the associated segmentation based on this HMM topology. Since we do not want to consider all possible values for k, we begin by choosing a maximum value (k = K). Then, through the process of cluster merging, we reduce the value of k until we nd an optimal number of clusters (k ) and their acoustic models (Λ k ) according to (1). However, when two clusters are merged, the total number of parameters in the HMM decreases. Modeling the same amount of data using fewer parameters yields a lower likelihood score. Given that the merging process can only result in monotonically decreasing likelihoods, we will not observe a maximum in the likelihood function at any point other than the starting point. Therefore we need to choose a likelihood threshold to tell us when to stop merging. Ideally, we would like to nd a method of selecting clusters for merging such that a correct merge (i.e. a merge involving clusters of data from the same speaker) will produce an increase in the objective function (1) and an incorrect merge will result in a decrease. A common method of selecting between competing models is to use the Bayesian Information Criterion (BIC) [1]. BIC imposes a trade-off between model quality and model complexity. Using BIC as a merging criterion, two clusters would become a candidate for merging if the following is true: log p(d θ) 1 2 λklogn log p(d a θ a ) + log p(d b θ b ) where: (2) D a and D b represent the data in two clusters and θ a and θ b represent the parameters of the PDFs of these two clusters respectively. D is the data from D a D b and θ represents the parameters of PDF of D. λ is a tweakable parameter that is ideally set to 1.0 N is the number of data points in {D} K is the difference in the number of parameters between θ a and θ b BIC provides a simple way to decide when to stop merging. However, we found in our experiments that we had to tune λ to get the best results. Moreover, the optimal value of λ changes depending on the data conditions and nding the optimal value requires the use of a development dataset that closely resembles the test dataset. In general, the algorithm would be more robust if thresholds such as λ could be eliminated. 2.1.1. New Merging Criterion If we use BIC, but keep the number of parameters constant, we can eliminate the penalty term ( 1 2λKlogN) in (2). Thus, when we merge two clusters, we simply model the PDF of the new cluster using a model containing a number of parameters equal to the sum of the number of parameters of the two merged clusters. As with BIC, the objective function in (1) increases for correct merging (merging of two clusters having data from the same source) and decreases for incorrect merging. We de ne our merging criterion as follows: Let M a and M b represent the number of parameters (Gaussian components) in θ a and θ b respectively. Let us hypothesize a new cluster having data D = D a D b with a PDF modeled by a GMM parameterized by θ with M a +M b number of Gaussian components. Given these conditions, a pair of clusters (D a and D b ) becomes a candidate for merging if the following is true: log p(d θ) log p(d a θ a ) + log p(d b θ b ) (3) This is similar to BIC, except that the number of parameters in θ is equal to the sum of the number of parameters in θ a and θ b. By keeping the number of parameters constant from one iteration to the next, we have eliminated the need for the penalty term. We have veri ed empirically that selecting candidates for merging using this criterion does indeed result in an increase in the objective function associated with (1). After every new segmentation-training step, we look for the best pair satisfying (3). In the case of many such candidate pairs, we choose the pair that maximizes the difference of the terms of left hand side and right hand side of (3). The merging is stopped when there are no suitable candidates satisfying (3). This method provides a fully automatic stopping criterion that does not require the use of any tunable parameters. However, there are a few hyper-parameters in this algorithm, namely the initial number of clusters (K), the initial number of Gaussian components in each cluster (M), the type of initialization used to create the clusters, and the set of acoustic features used to represent the signal. In Section 3 we present the results of several experiments in which we explore the effects of varying the hyper-parameters.

3. EXPERIMENTS AND RESULTS 3.1. Evaluation Criterion Evaluation of the algorithm was done using NIST s RT-03S scoring script (SpkrSegEval-v21.pl). This calculates a time-based score (error) that is the percentage of speaker time not attributed correctly to a reference speaker. Thus, a score of 0.0 would represent a perfect segmentation. Also, it is possible to have an error > 100.0 because of the inclusion of false alarms. The error is: S {dur(s) (max(n ref (s),n sys (s)) N correct (s))} s=1 S {dur(s) N ref (s)} s=1 (4) where the speech data is divided into contiguous segments whose boundaries are de ned by all speaker change points (including both reference and hypothesized speakers) and where, for each segment s: dur(s) = the duration of s N ref (s) = the # of reference speakers speaking in s N sys (s) = the # of system speakers speaking in s N correct (seg) = the # of reference speakers speaking in s for whom their matching (mapped) system speakers are also speaking in s. 3.2. Data The algorithm was tested on three different datasets released by NIST, namely dryrun data (data used for preliminary experiments), devdata (data used as development data) and evaldata (data used in the nal RT-03S evaluation). 2 The dryrun data consists of six 10-minute excerpts from English broadcast news shows, while the devdata and evaldata each consisted of three half-hour English broadcast news audio segments. In the evaldata alone, there are 47 Male and 12 Female speakers. 3.3. Baseline System We used default values shown in Table 1 for all the hyperparameters in each of the experiments listed below, unless otherwise noted. The performance of the baseline system is shown in Table 2. The results on evaldata were submitted 3 as part of the 2 Not all the experiments were carried out on all three datasets. We experimented with different hyper-parameters at different stages on different datasets. Also, all results are shown for complete datasets only (rather than individual les) to avoid presenting too many numbers. 3 The results submitted to NIST also involved another postprocessing stage of speech/non-speech discrimination. Since this module is not a focus of this paper, the results shown here are without any speech/non-speech discrimination, i.e. only based on speaker clustering. Initialization Uniform Initial number of Gaussians (M) 5 Initial number of clusters (K) 15(dryrun), 40(devdata), 40 (evaldata) Minimum duration (S) 2 seconds Feature type LPC Cepstrum (LPCC) Feature vector frequency 100 Hz Table 1. Default values for the hyper-parameters Dataset Error dryrun 28.85% devdata 26.11% evaldata 21.40% Table 2. Baseline results for the three datasets. RT-03S 4 evaluation and the performance of the system was highly competitive compared to other submitted systems. However, as seen in Table 1, there are a number of hyperparameters, which can be seen as tunable parameters. We veri ed with the help of a series of experiments that the algorithm is not highly sensitive to any of these parameters. This together with experiments on different datasets, demonstrates the robustness of the algorithm. These experiments are summarized in next subsections. 3.4. Initialization As mentioned earlier, two different initializations were performed on the dryrun dataset. Table 3 presents the results for this experiment: Initialization Error Uniform 28.85% K-means 29.56% Table 3. Results for two different initialization schemes on dryrun dataset We observed that the type of initialization does not have a signi cant impact on the nal results. We speculate that this is due to the minimum duration constraint and the use of an iterative EM algorithm. Thus, for the subsequent experiments, we used only uniform initialization. 3.5. Experiments with different acoustic features Table 4 presents the results of using different acoustic features in the segmentation. In addition to the default 12-4 http://www.nist.gov/speech/tests/rt/rt2003/spring

LPCC features, we tried 19-Mel-frequency cepstral coef cients (MFCC). Dataset F eaturet ype Error dryrun LPCC 28.85% dryrun MFCC 29.22% devdata LPCC 26.11% devdata MFCC 25.13% evaldata LPCC 21.40% evaldata MFCC 20.79% Table 4. Results obtained with alternate features. As expected, the performance of LPCC and MFCC features in all the cases are comparable. However, while analyzing the performance on individual les of each dataset, we noticed that MFCCs worked better in noisy conditions, while LPCCs worked better during clean speech. 3.6. Experiments with minimum duration Table 5 presents the results obtained by varying the minimum duration of each cluster. These experiments were also carried out on devdata. Minimum Duration (secs) Error 2 26.11% 3 26.55% 4 26.18% Table 5. Results obtained with different minimum durations Results in Table 5 show that the algorithm is not sensitive to the minimum duration that we impose. However, the sensitivity also depends on the average duration of the speaker segments. If there are many short segments, a large minimum duration may result in a higher error rate. Thus, if we have any a priori information about average speaker durations, it can be used in the algorithm. 3.7. Experiments with the number of initial clusters Table 6 presents results for varying the number of initial clusters (K). These results show that the performance of the algorithm improves as we increase the value of K. Since the algorithm is based on agglomerative clustering, K should be larger than the expected number of speakers. We have found that a good rule of thumb (for English broadcast news) is to choose K equal to or greater than the number of minutes of audio data. However, it should be noted that as the value of K increases, so does the computational complexity of the algorithm. This is due to the number of pairwise cluster comparisons that must be made during merging. dataset K Error dryrun 15 28.85% dryrun 30 28.35% devdata 30 29.28% devdata 40 26.11% devdata 50 25.80% Table 6. Results obtained by varying the number of initial clusters. 3.8. Experiments with the number of initial Gaussians Table 7 presents the results of experiments on the devdata in which we varied the number of Gaussian components in each initial cluster (M). Note that we use a small number of Gaussians relative to those used in the speaker recognition framework. In the speaker recognition framework, the goal is to make a robust model for each speaker, and hence thousands of Gaussians per speaker are employed. However, the goal here is to make discriminative models for the speakers in a single audio stream. Thus, we need far fewer Gaussian components to estimate the PDFs of each cluster. M Error 5 26.11% 10 27.44% Table 7. Results obtained for different number of Gaussians (M) for devdata In the majority of our experiments (including those submitted to NIST), we used ve gaussians per initial cluster (M = 5). We determined that the performance of the algorithm does not vary signi cantly depending on the choice of M. In fact, the performance degrades slightly for larger values as can be seen in Table 7. 4. CONCLUSION In this paper, we have presented a novel speaker clustering algorithm and demonstrated its robustness to different data conditions. The algorithm uses an iterative, agglomerative clustering technique based on an HMM framework. Pairs of clusters are merged in successive iterations and merging stops automatically when the optimal number of clusters has been determined. We de ned a BIC-like merging criterion that we use to choose which clusters to merge and when to stop merging. This merging criterion produces an increase in a likelihood-based objective function for correct merges and a decrease for incorrect merges. An important property of our merging criterion is that it does not use an adjustable threshold. The absence of such tweakable parameters not only inproves the algorithm s robustness but also eliminates the need for a development dataset.

Acknowledgements This work was supported by project MULTI (Swiss NSF project no. 2000-068231.02/1), project M4 (EC-IST project no. 2001-34485) and the Swiss National Center of Competence in Research (NCCR) on Interactive Multimodal Information Management (IM)2. The NCCR is managed by the Swiss NSF on behalf of the federal authorities. This Material is also based upon work supported by the Defense Advanced Research Projects Agency Information Awareness Of ce EARS program. ARPA Order No. N614, Program Code No. 2E20, issued by DARPA/CMO under Contract No. MDA972-02-C-0038 as part of a subcontract to ICSI by SRI International. 5. REFERENCES [1] S. S. Chen and P. S. Gopalakrishnan, Speaker, environment and channel change detection and clustering via the Bayesian information criterion, Tech. Rep., IBM T.J. Watson Research Center, 1998. [2] M. Sugiyama, J. Murakami, and H. Watanabe, Speech segmentation and clustering based on speaker features, IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. 395 398, 1993. [3] A. Solomonoff, A. Mielke, M. Schmidt, and H. Gish, Clustering speakers by their voices, IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. 757 760, 1998. [4] M. A. Siegler, U. Jain, B. Raj, and R. M. Stern, Automatic segmentation, classi cation and clustering of broadcast news audio, DARPA Speech Recognition Workshop, Chantilly, pp. 97 99, Feb 1997. [5] T Hain, S. E. Johnson, A. Turek, P. C. Woodland, and S. J. Young, Segment generation and clustering in the HTK broadcast news transcription system, Proc. DARPA Broadcast News Transcription and Understanding Workshop, pp. 133 137, 1998.