Semi-blind Block Channel Estimation and Signal Detection Using Hidden Markov Models

Similar documents
Expectation and Maximization Algorithm for Estimating Parameters of a Simple Partial Erasure Model

Mixture Models and EM

Hidden Markov Models. Slides adapted from Joyce Ho, David Sontag, Geoffrey Hinton, Eric Xing, and Nicholas Ruozzi

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Package HMMCont. February 19, 2015

Constraints in Particle Swarm Optimization of Hidden Markov Models

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

Speech Recognition Lecture 8: Acoustic Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri

Introduction to Machine Learning CMU-10701

Ergodic Hidden Markov Models for Workload Characterization Problems

On the Parameter Estimation of the Generalized Exponential Distribution Under Progressive Type-I Interval Censoring Scheme

THE most popular training method for hidden Markov

DATA DETECTION WITH FUZZY C-MEANS BASED ON EM APPROACH BY MIMO SYSTEM FOR DIFFERENT CHANNEL S ESTIMATION IN WIRELESS CELLULAR SYSTEM

Simultaneous Design of Feature Extractor and Pattern Classifer Using the Minimum Classification Error Training Algorithm

Assignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018

A Review on Analysis on Codes using Different Algorithms

Treba: Efficient Numerically Stable EM for PFA

Efficient Markov Chain Monte Carlo Algorithms For MIMO and ISI channels

Biology 644: Bioinformatics

ModelStructureSelection&TrainingAlgorithmsfor an HMMGesture Recognition System

application of learning vector quantization algorithms. In Proceedings of the International Joint Conference on

Time series, HMMs, Kalman Filters

Expectation Maximization. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University

Hidden Markov Model for Sequential Data

A Search Algorithm for Global Optimisation

Channel Performance Improvement through FF and RBF Neural Network based Equalization

An embedded system of Face Recognition based on ARM and HMM

Pseudo-Maximum-Likelihood Data Estimation Algorithm and Its Application over Band-Limited Channels

Clustering Sequences with Hidden. Markov Models. Padhraic Smyth CA Abstract

Intelligent Hands Free Speech based SMS System on Android

Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models

Modeling time series with hidden Markov models

Hidden Markov Models in the context of genetic analysis

A Non-Iterative Approach to Frequency Estimation of a Complex Exponential in Noise by Interpolation of Fourier Coefficients

Advance Convergence Characteristic Based on Recycling Buffer Structure in Adaptive Transversal Filter

Parallel HMMs. Parallel Implementation of Hidden Markov Models for Wireless Applications

Regularization and Markov Random Fields (MRF) CS 664 Spring 2008

Software/Hardware Co-Design of HMM Based Isolated Digit Recognition System

HMM Training by a Hybrid of Chaos Optimization and Baum-Welch Algorithms for Discrete Speech Recognition

Expectation Maximization (EM) and Gaussian Mixture Models

The EM Algorithm Lecture What's the Point? Maximum likelihood parameter estimates: One denition of the \best" knob settings. Often impossible to nd di

CONTENT analysis of video is to find meaningful structures

Blind Deconvolution with Sparse Priors on the Deconvolution Filters

MR IMAGE SEGMENTATION

Effect of Initial HMM Choices in Multiple Sequence Training for Gesture Recognition

Telecommunications Engineering Course Descriptions

NOVEL HYBRID GENETIC ALGORITHM WITH HMM BASED IRIS RECOGNITION

HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation

STA 4273H: Statistical Machine Learning

THE general task in emission computed tomography (CT)

Variational Methods for Graphical Models

Hybrid Quasi-Monte Carlo Method for the Simulation of State Space Models

Optimization of HMM by the Tabu Search Algorithm

Skill. Robot/ Controller

NONLINEAR BACK PROJECTION FOR TOMOGRAPHIC IMAGE RECONSTRUCTION

Noise Canceller Using a New Modified Adaptive Step Size LMS Algorithm

CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs

Stephen Scott.

CSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates

CS 532c Probabilistic Graphical Models N-Best Hypotheses. December

Passive Differential Matched-field Depth Estimation of Moving Acoustic Sources

Introduction to Image Super-resolution. Presenter: Kevin Su

Evaluation of Model-Based Condition Monitoring Systems in Industrial Application Cases

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

ICA mixture models for image processing

A Performance Evaluation of HMM and DTW for Gesture Recognition

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

TRACKING PERFORMANCE OF THE MMAX CONJUGATE GRADIENT ALGORITHM. Bei Xie and Tamal Bose

A LOW COMPLEXITY ITERATIVE CHANNEL ESTIMATION AND EQUALISATION SCHEME FOR (DATA-DEPENDENT) SUPERIMPOSED TRAINING

DESIGN AND IMPLEMENTATION OF VARIABLE RADIUS SPHERE DECODING ALGORITHM

Jonathan H. Manton and Ba-Ngu Vo

BLIND EQUALIZATION BY SAMPLED PDF FITTING

Research on Emotion Recognition for Facial Expression Images Based on Hidden Markov Model

Optimization of Noisy Fitness Functions by means of Genetic Algorithms using History of Search with Test of Estimation

Biostatistics 615/815 Lecture 23: The Baum-Welch Algorithm Advanced Hidden Markov Models

LOW-DENSITY PARITY-CHECK (LDPC) codes [1] can

Procedia Computer Science

Ambiguity Detection by Fusion and Conformity: A Spectral Clustering Approach

Hierarchical Mixture Models for Nested Data Structures

Speech User Interface for Information Retrieval

Person Authentication from Video of Faces: A Behavioral and Physiological Approach Using Pseudo Hierarchical Hidden Markov Models

Automatic State Machine Induction for String Recognition

Missing Data Analysis for the Employee Dataset

DSW Feature Based Hidden Marcov Model: An Application on Object Identification

A new way to optimize LDPC code in gaussian multiple access channel

Fuzzy-Kernel Learning Vector Quantization

Graphical Models & HMMs

From Gaze to Focus of Attention

Hidden Markov Models. Gabriela Tavares and Juri Minxha Mentor: Taehwan Kim CS159 04/25/2017

= f (a, b) + (hf x + kf y ) (a,b) +

Adaptive Multiple-Frame Image Super- Resolution Based on U-Curve

Contact State Estimation using Multiple Model Estimation and Hidden Markov Models

One Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition

Conditional Random Fields : Theory and Application

Lecture 8: The EM algorithm

Missing Data Analysis for the Employee Dataset

The maximum reachable astrometric precision. The Cramér-Rao Limit

An Introduction to Pattern Recognition

Comparative Study of the Baum-Welch and Viterbi Training Algorithms Applied to Read and Spontaneous Speech Recognition

Transcription:

Semi-blind Block Channel Estimation and Signal Detection Using Hidden Markov Models Pei Chen and Hisashi Kobayashi Department of Electrical Engineering Princeton University Princeton, New Jersey 08544, USA Abstract - We propose two maximum likelihood based semiblind block channel estimation and signal detection algorithms for multipath channels with additive Gaussian noise. The algorithms are based on the Baum-Welch algorithm and the segmental k-means algorithm for Hidden Markov Models (HMMs). By making use of a training signal, the algorithms are applied block-wise to sequential disjoint subintervals of the whole observation interval. We study the effects of block length in terms of the bit error rate (BER), the mean square error (MSE) of the estimated channel impulse response, and its Cramer-Rao lower bound. Our simulation results show that the BER performance does not suffer even for a short block length when a good initial estimate is available. 1 INTRODUCTION In order to cope with an unknown and possibly timevarying radio channel, the problem of channel estimation is recognized as a key research problem, because a large class of interference cancellation algorithms and decoding algorithms require complete or partial knowledge of channel characteristics. The existing and proposed digital communication systems support the use of known training signals to estimate the channel. By incorporating the training signals into blind schemes, semi-blind channel estimation schemes are anticipated to improve the performance over pure blind or non-blind schemes. Hidden Markov Models (HMMs), on the other hand, have been extensively studied and successfully applied to problems of sequential pattern recognition such as speech recognition and handwriting recognition [7]. The parameter estimation problem for HMMs can be solved efficiently by the Baum-Welch algorithm [l]. The reestimation formulas of the Baum-Welch algorithm can be interpreted as an application of the Expectation- Maximization (EM) [3] algorithm to HMMs, and the computational efficiency of the Baum-Welch algorithm comes from the efficient procedure of the forwardbackward algorithm [7]. Recently the Baum-Welch algorithm for HMMs has been applied to maximum likelihood channel estimation [2,5]. In this paper, we propose a semi-blind block channel estimation and signal detection scheme which gives block-wise maximum likelihood channel estimate and MAP decoding by using the Baum-Welch algorithm in sequential disjoint subintervals of an observation. The advantage of block channel estimation lies in several aspects. Firstly, it can avoid a long delay to process the whole observation interval; secondly, it is suitable for a slow time varying channel; thirdly, it has lower computational complexity. We evaluate the performance in terms of the mean square error (MSE) and bit error rate (BER) for different block lengths. The segmental k-means algorithm for HMMs uses the state-optimized likelihood as its optimization criterion [4]. We apply the two iterative steps of the segmental k-means algorithm to channel estimation and signal detection and obtain a very simple structure which gives the maximum likelihood sequence estimate and decision directed channel estimation. This algorithm can also be applied to the received signal, block by block. We evaluate the effects of block length by simulation, and compare the results with those obtained by the Baum-Welch algorithm. The rest of the paper is organized as follows. Section 2 gives the channel model and its HMM description. Section 3 describes our semi-blind block channel estimation and signal detection scheme based on the Baum-Welch algorithm. Section 4 derives the two step iterative scheme based on the k-means algorithm. Section 5 evaluates the performance of the proposed schemes by BER, MSE and its Cramer-Rao lower bound for different block lengths. Section 6 presents our concluding remarks. 2 CHANNEL MODEL AND ITS HMM DESCRIPTION We consider the following multipath channel model: L Yn = hldn+l-l+ vn, (1) 1=1 0-7803-6451-1/00/$10.00 0 2000 IEEE 1051

where the channel has memory of length L; h = (hl, h2,..., h ~ is ) the channel impulse response; vn s are i.i.d. complex Gaussian random variables with zero mean and variance a ; d = (dl, d2,...,d~) is the transmitted baseband signal including a training part and a data part. For a given observation sequence y= (yi, YZ,..., YN), our goal is to find the maximumlikelihood estimate of the channel parameter vector @ = (hl, hz,..., ht, a ) that maximizes f(gl@). The above problem can be formulated as an HMM parameter estimation problem. An HMM models the case where the observation is a probabilistic function of the state of a Markov chain. An HMM is specified by three probability measures A, B and x, where A is the state transition probability matrix; B is a vector of conditional probability density functions (PDF) with each entry being the observation PDF in a state, and x is the initial state probability vector. We define state S, = (dn+l-~,...,&-i, d,). If each element takes on K possible values, the total number of states is KL. The parameters of the HMM can be found as : 0 State transition probability ai, = PISn+l = jlsn = i]: data part: aij = { 1/: training part: aij = { ii if i + j is permissible otherwise if i -+ j is permissible otherwise 0 Initial state probability xi = P[S1 = 21: - xi = { KL, if starting from data part 6(i - io), if starting from training part where io is the starting state. 0 Observation PDF in state i b;(y) = f(yis, = i,c9): 1 Iy - CL1 Wi,L+l-lI b.( ) - -exp(- - 2xu2 2a2 2 )I (2) over e. Here is the estimation of @ in the kth it- eration; s. = [SI, s,,..., SN] are the unobserved states corresponding to the observations y. From Bayed rule and the properties of the HMM, we obtain the following equations: N P(Sle) = n asn-isn I (5) N n=2 f(yls18) = n f(yrllsn18). (6) n=l Substituting Eqs.(4),(5) and (6) into Eq.(3), we have, where C is a constant independent of 0, and X,(i) is defined as X,(i) = f(g, s, = ip). (8) From Eqs.(2) and (7), the maximization problem is equivalent to minimizing over e. Therefore we can find which readily leads to by solving where &Lk+11 = (af+llb[k+ll a[k+ll [k+l1 T with ajk+ll 1 -*. L bl 1 and bjk+ll being the real and imaginary parts of hl in the k + lth estimation, respectively. G is a 2~ x 2~ matrix with (9) 3 BAUM-WELCH ALGORITHM 3.1 Baum- Welch Algorithm for Channel Estimation The Baum-Welch algorithm is a computationally efficient reestimation procedure to solve the parameter estimation problem for HMMs. It finds the parameter e that maximizes the likelihood function f(ylc9) by iteratively maximizing the Baum s auxiliary function as its (I,m)th subblock, where Di,lm is defined as and is a 2L x 1 vector with 1052

as its lth subvector. When h2[kf1] is available, tained as can be readily ob- and It is a 2L x 1 vector with the lth subvector ~2 1 Reid: + 1-1 Yn ~2 1 ~m{d:+ 1 - Yn > and ( ), L 1 Ni 62 = - I Y - ~ kdn+l-li2. 2Nt n=l k 1 (23) Comparing Eqs.(ll),(l2) and (22),(23), we can see that the channel estimation for the training part is a degenerated case of the Baum-Welch algorithm, since An(i) = f(y, S, = il@kl) - { j(yl@i), if i = it 0, otherwise, (24) where it is a part of the training sequence. with the boundary conditions for 1 5 i 5 KL. 3.2 Channel Estimation with Training Signal For the training part, the transmitted sequence known. The log-likelihood function is given by logf(yl!!) = -Nt log(2.1ra2) where Nt is the length of the training sequence. The maximum likelihood estimate of e can be simply found by v,logf(yle)le=g = 0. (21) Hence we obtain - il = G; 'It, is (22) where Gt is a 2L x 2L matrix with the (I, m)th subblock 3.3 Semi-blind Block Channel Estimation and Signal Detection The Baum-Welch algorithm discussed in Section 3.1 uses the entire observation to iteratively update the estimation of the channel parameters. It may cause a large time delay in decoding and is not suitable to time varying channels. To deal with these problems, we divide the observation interval into sequential disjoint subintervals and apply the Baum-Welch algorithm to each subinterval. Generally speaking, if the observation interval is too short, the Baum-Welch algorithm suffers from the problem that it may converge to a local minimum. However, we can avoid this problem by using the training signal. We can obtain a good estimate of the channel for the subinterval corresponding to the training signal by Eqs.(22)-(23), and then use the estimates as initial values in adjacent subintervals. If the channel is slowly varying, a good initial estimate is made available by using the estimates from the adjacent subintervals, and thus the problem of converging to a local minimum can be avoided. Notice that MAP decoding can be obtained as a byproduct of the Baum-Welch algorithm. For each subinterval, after the final iteration, we decode the transmitted signal by S, = argmuzix,(i). (25) Therefore, signal detection does not need any additional computation. 4 SEGMENTAL K-MEANS ALGORITHM Instead of the maximum likelihood criterion s - 1053

used in the Baum-Welch algorithm, the segmental k- means algorithm uses as the optimization criterion for estimation the parameters of the HMM. It focuses on the most likely state sequence as opposed to summing over all possible state sequences. The algorithm involves iteration of two fundamental steps: (1) segmentation step, which finds the sequence that maximizes the joint likelihood function f(y, sle) for the current parameter estimates; (2) optimeation step, which finds new estimates for e so as to maximize the above state-optimized likelihood. The convergence properties of the segmental k-means algorithm was proved in [4]. Applying the segmental k-means algorithm to our channel estimation and signal detection problem, we obtain a simple iterative structure as follows. At the ICth iteration, we have where Io; is the Fisher's information for 0; Using the identity a2 lei = --E{? aei log f(yl@)}. (31) f(yle) = P(S)f(Yle, S) = Ez{f(Yle, S)} (32) S - and Jensen's inequality we have a2 Io; 5 -EsE{y 1% f(yle?, S)}. dei (34) By substituting Eq.(2) into f(yl0,s) of the above expression, we find since all S's are equiprobable. 0 Optimization step: The segmentation step gives the maximum likelihood sequence estimation for each iteration of estimated channel parameters and can be implemented efficiently by the Viterbi algorithm. The optimization step estimates the channel in a decision directed mode, i.e. it updates the estimates of the channel by treating the current estimate of the transmitted signal as a known signal. As for the Baum-Welch algorithm, the segmental k- means algorithm can also be applied to the received signal block by block. We evaluate the performance of the algorithms for different block sizes in the next section. 5 PERFORMANCE EVALUATION 5.1 Cramer-Rao Lower Bound (CRLB) The CRLB gives a lower bound for the variance of an unbiased estimate. -- N' - (37) c74 where a; and bi are the real and imaginary parts of hi, respectively, and N' is the length of subintervals. We see that the CRLB is inversely proportional to the block size. 5.2 Simulation Results We evaluate the above semi-blind block channel estimation and signal detection schemes by simulation in terms of their BER performances and MSEs of the estimated channel impulse responses. We take an antipodal signal where d, E {+1,-1) and a multipath channel with impulse response h = [0.408 0.816 0.4081 (channel (b) in [6], chap.10). We choose two sets of parameters for initial estimates of h: hinitl = [0.577 0.577 0.5771 klnit2 = [0.3 0.9 0.31 hln;tl represents the case when we do not have any knowledge of the channel. In this case, we can only randomly choose an initial estimate, and the best we can do is to take an average value. klnit2 represents the case when a good initial estimate is available from an adjacent subinterval. We test the performance for different 1054

backward algorithm. Thus, the latter is much more computationally involved. In view of complexity, the segmental k-means algorithm is more attractive. 6 CONCLUDING REMARKS \ o I (b) : io io Bh( iw k >& Figure 1: MSE of lz, SNR= 13dB (a) Baum-Welch algorithm; (b) segmental k-means algorithm We have proposed two semi-blind block channel estimation and signal detection algorithms using the Baum- Welch algorithm and the segmental k-means algorithm for HMMs. The algorithms apply to the subintervals of the received signal instead of to the whole observation interval. This approach provides several advantages in terms of time delay, complexity and the ability to deal with a slowly fading channel. Our simulation results show that if a good initial estimate is made available by using a training signal, the algorithms with short block length suffer only a minor performance loss. Comparing with the Baum-Welch algorithm, the segmental k-means algorithm is more attractive when the complexity is a major issue. ACKNOWLEDGMENTS The present work has been supported by the National Science Foundation, the New Jersey Center for Wireless Telecommunications (NJCWT) and Mitsubishi Electric Research Laboratories (MERL). REFERENCES Figure 2: BER for different block lengths, SNR=13dB block lengths N = 13,26,52,104 with SNR varies from 7dB to 13dB. Fig. 1 and 2 illustrate the results for SNR= 13dB. The results for other SNR values exhibit similar patterns. From Fig. 2 we can see that for both the Baum-Welch algorithm and the segmental k-means algorithm if the channel is completely unknown, there is a substantial performance degradation to be incurred by using short blocks. However, if a good initial estimate is available, the block length can be shortened to a certain extent with only a minor performance loss. Also notice that with a good initial estimate, the performance degradation of the segmental k-means algorithm compared with the Baum-Welch algorithm is reasonably small. The segmental k-means algorithm uses the Viterbi algorithm, whereas the Baum-Welch algorithm involves the forward- L. E. Baum, An inequality and associated maximisation technique in statistical estimation for probabilistic functions of Markov processes, Inequalities, Vol. 3, pp. 1-8, 1972. H. A. Cirpan and M. K. Tsatsanis, Stochastic maximum likelihood methods for semi-blind channel estimation, IEEE Signal Processing Lett., Vol. 5, No. I, pp. 21-24, 1998. A. P. Dempster, N. M. Laird and D. B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, J. Roy. Stat. Soc., Vol. 39, No. 1, pp. 1-38, 1977. B. H. Juang and L. R. Rabiner, The segmental k-means algorithm for estimating parameters of Hidden Markov Models, IEEE Trans. Acoust., Speech, Signal Processing, Vol. 38, No. 9, pp. 1639-1641, 1990. G. K. Kaleh and R. Vallet, Joint parameter estimation and symbol detection for linear or nonlinear unknown channels, IEEE Trans. Commun., Vol. 42, No. 7, pp. 2406-2413, 1994. J. G. Proakis, Digital Communications, 3rd ed., McGraw-Hill, 1995. L. R. Rabiner, A tutorial on Hidden Markov Models and selected applications in speech recognition, Proceedings of the IEEE, Vol. 77, No. 2, pp. 257-285, 1989. 1055