Alignment of Viewing-Angle Dependent Ultrasound Images

Similar documents
Towards an Estimation of Acoustic Impedance from Multiple Ultrasound Images

Adaptive Multiscale Ultrasound Compounding Using Phase Information

Volumetric Ultrasound Panorama Basedon3DSIFT

Ultrasound Specific Similarity Measures for Three-Dimensional Mosaicing

Three-Dimensional Ultrasound Mosaicing

Sphere Extraction in MR Images with Application to Whole-Body MRI

Fast Ultrasound Image Simulation using the Westervelt Equation

A Non-Linear Image Registration Scheme for Real-Time Liver Ultrasound Tracking using Normalized Gradient Fields

3-D Compounding of B-Scan Ultrasound Images

Manifold Learning for Image-Based Breathing Gating with Application to 4D Ultrasound

Computerized Medical Imaging and Graphics

3D Ultrasound Volume Stitching Using Phase Symmetry and Harris Corner Detection for Orthopaedic Applications

Robust Spatio-Temporal Registration of 4D Cardiac Ultrasound Sequences

Computer-Aided Diagnosis in Abdominal and Cardiac Radiology Using Neural Networks

Object Identification in Ultrasound Scans

A Robust Meshing and Calibration Approach for Sensorless Freehand 3D Ultrasound

Chapter 3 Image Registration. Chapter 3 Image Registration

Depth-Layer-Based Patient Motion Compensation for the Overlay of 3D Volumes onto X-Ray Sequences

Patient-specific 3D Ultrasound Simulation Based on Convolutional Ray-tracing and Appearance Optimization

Methodological progress in image registration for ventilation estimation, segmentation propagation and multi-modal fusion

Iterative CT Reconstruction Using Curvelet-Based Regularization

GPU Ultrasound Simulation and Volume Reconstruction

ULTRASONIC TESTING AND FLAW CHARACTERIZATION. Alex KARPELSON Kinectrics Inc., Toronto, Canada

GE Healthcare. Agile Ultrasound. The Next Revolution in Ultrasound Imaging

연구용유방초음파질관리 원광대학병원김혜원

Image Thickness Correction for Navigation with 3D Intra-cardiac Ultrasound Catheter

UNIVERSITY OF SOUTHAMPTON

Comparison of Vessel Segmentations Using STAPLE

Precise Ultrasound Bone Registration with Learning-based Segmentation and Speed of Sound Calibration

Development and validation of a short-lag spatial coherence theory for photoacoustic imaging

Backward-Warping Ultrasound Reconstruction for Improving Diagnostic Value and Registration

Feature descriptors and matching

DUE to beam polychromacity in CT and the energy dependence

Whole Body MRI Intensity Standardization

Norbert Schuff VA Medical Center and UCSF

Image Guidance of Intracardiac Ultrasound with Fusion of Pre-operative Images

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Ensemble registration: Combining groupwise registration and segmentation

Simulation and Fully Automatic Multimodal Registration of Medical Ultrasound

Segmentation of 3-D medical image data sets with a combination of region based initial segmentation and active surfaces

Digital Volume Correlation for Materials Characterization

Nonrigid Registration using Free-Form Deformations

Hybrid Spline-based Multimodal Registration using a Local Measure for Mutual Information

Trajectory-based Deformation Correction in Ultrasound Images

Comparison of Vessel Segmentations using STAPLE

Image Segmentation and Registration

Fast 3D Mean Shift Filter for CT Images

Simplified statistical image reconstruction algorithm for polyenergetic X-ray CT. y i Poisson I i (E)e R } where, b i

Deformable Registration Using Scale Space Keypoints

MEDICAL IMAGE NOISE REDUCTION AND REGION CONTRAST ENHANCEMENT USING PARTIAL DIFFERENTIAL EQUATIONS

Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging

2 Michael E. Leventon and Sarah F. F. Gibson a b c d Fig. 1. (a, b) Two MR scans of a person's knee. Both images have high resolution in-plane, but ha

MEDICAL IMAGE ANALYSIS

Bioimage Informatics

Atlas-Based Segmentation of Abdominal Organs in 3D Ultrasound, and its Application in Automated Kidney Segmentation

Certificate in Clinician Performed Ultrasound (CCPU)

Room Acoustics. CMSC 828D / Spring 2006 Lecture 20

REMOVAL OF THE EFFECT OF COMPTON SCATTERING IN 3-D WHOLE BODY POSITRON EMISSION TOMOGRAPHY BY MONTE CARLO

Computer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.

Sparsity Based Spectral Embedding: Application to Multi-Atlas Echocardiography Segmentation!

Edge-Preserving Denoising for Segmentation in CT-Images

Efficient Image Compression of Medical Images Using the Wavelet Transform and Fuzzy c-means Clustering on Regions of Interest.

Local Image Registration: An Adaptive Filtering Framework

Nonrigid Registration Using a Rigidity Constraint

10/16/14. Ultrasound Physics & Instrumentation 5 th Edition. License Agreement. Chapter Outline. All Copyright Laws Apply.

3D Ultrasound System Using a Magneto-optic Hybrid Tracker for Augmented Reality Visualization in Laparoscopic Liver Surgery

Modeling Custom Surface Roughness with LucidShape 2D Scatter Curve BSDF Material

Three-Dimensional Face Recognition: A Fishersurface Approach

Three Dimensional Segmentation of Intravascular Ultrasound Data

Slide 1. Technical Aspects of Quality Control in Magnetic Resonance Imaging. Slide 2. Annual Compliance Testing. of MRI Systems.

Synthesis of Facial Images with Foundation Make-Up

Medical Image Analysis

Determination of Volume Characteristics of Cells from Dynamical Microscopic Image

Medical Image Segmentation Based on Mutual Information Maximization

arxiv: v1 [cs.cv] 1 May 2018

Temperature Distribution Measurement Based on ML-EM Method Using Enclosed Acoustic CT System

Motion Estimation for Video Coding Standards

A New Method for CT to Fluoroscope Registration Based on Unscented Kalman Filter

Schedule for Rest of Semester

Automatic Seed Placement for Breast Lesion Segmentation on US Images

CP Generalize Concepts in Abstract Multi-dimensional Image Model Component Semantics. David Clunie.

Experiments with Edge Detection using One-dimensional Surface Fitting

Improvement and Evaluation of a Time-of-Flight-based Patient Positioning System

Kidney Segmentation in Ultrasound Images Using Curvelet Transform and Shape Prior

Evaluation of Fourier Transform Coefficients for The Diagnosis of Rheumatoid Arthritis From Diffuse Optical Tomography Images

Assessing Accuracy Factors in Deformable 2D/3D Medical Image Registration Using a Statistical Pelvis Model

University of Cambridge Department of Engineering Trumpington Street Cambridge CB2 1PZ United Kingdom.

An Elliptical Level Set Method for Automatic TRUS Prostate Image Segmentation. Nezamoddin N. Kachouie, Paul Fieguth and Shahryar Rahnamayan

Image Registration with Local Rigidity Constraints

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging

Lesion Segmentation in Breast Sonography

Image Registration Lecture 4: First Examples

Segmentation and Classification of Breast Tumor Using Dynamic Contrast-Enhanced MR Images

Projection-Based Needle Segmentation in 3D Ultrasound Images

Globally Stabilized 3L Curve Fitting

Summary of Thesis: Statistical Segmentation and Registration of Medical Ultrasound Data

RESTORING ARTIFACT-FREE MICROSCOPY IMAGE SEQUENCES. Robotics Institute Carnegie Mellon University 5000 Forbes Ave, Pittsburgh, PA 15213, USA

Basics of treatment planning II

Bildverarbeitung für die Medizin 2007

Manifold Learning-based Data Sampling for Model Training

Transcription:

Alignment of Viewing-Angle Dependent Ultrasound Images Christian Wachinger and Nassir Navab Computer Aided Medical Procedures (CAMP), TUM, Munich, Germany {wachinge,navab}@cs.tum.edu Abstract. We address the problem of the viewing-angle dependency of ultrasound images for registration. The reflected signal from large scale tissue boundaries is dependent on the incident angle of the beam. This applies an implicit weighting on the ultrasound image, dependent on the viewing-angle, which negatively affects the registration process, especially when utilizing curved linear transducers. We show that a simple reweighting of the images, considering a common physical model for ultrasound imaging, is not feasible. We therefore introduce a new matching function, separating reflectivity and scattering regions, which are the results of two different types of physical interactions of the ultrasound beam with the tissue. We use the local phase for identifying regions of reflectivity, and consider it as one part of our matching function, combining feature- and intensity-based aspects. First experiments provide good results for this novel registration approach. 1 Introduction The fusion of several ultrasound images for the creation of a combined image with a larger field-of-view, referred to as ultrasound stitching, mosaicing or panorama imaging, is of clinical interest [1] and topic of on-going research. In the past, a variety of methods were proposed ranging from rigid to deformable [2], from intensity- to feature-based [3], and from pair- to groupwise [4]. While featurebased approaches may be an option for getting close to the correct alignment, the final alignment is best achieved with groupwise intensity-based methods [4]. The major drawback of intensity-based methods is the missing overlap invariance of the similarity measures, favoring a total overlap of the images [4]. The reasons for this phenomenon have not yet been investigated. They could either be rooted in the similarity measures themselves, like Cahill et al. [5] addressed, or in the ultrasound images. We show that the viewing-angle dependency of ultrasound images not only causes angle dependent artifacts like shadow, but also puts an implicit weighting on the images, favoring a total overlap. To address this issue we first evaluate methods for reweighting ultrasound images, making them viewing-angle independent. However, currently used ultrasonic imaging models are not accurate enough, as we will show, for allowing This work was partly funded by the European Project PASSPORT, ref. number 223904. The authors thank Andreas Keil and Reviewer 1 for valuable contributions. G.-Z. Yang et al. (Eds.): MICCAI 2009, Part I, LNCS 5761, pp. 779 786, 2009. c Springer-Verlag Berlin Heidelberg 2009

780 C. Wachinger and N. Navab such a reweighting. We instead propose a new ultrasound matching function, consisting of one term, measuring the alignment of reflectivity structures, and a second term, measuring the alignment of the remaining texture. This new matching function has advantages for US-US, as well as, US-CT registration. 2 Viewing Angle Dependency Structures, visible in ultrasound images, are the result of two different types of interaction of the ultrasound beam with the tissue: reflection and scattering. Reflection occurs at large scale tissue boundaries and is viewing-angle dependent. Scattering is caused by microscopic tissue inhomogeneities and provides the internal texture of the organs. Scattering is to a certain extent dependent on the direction of insonification because the ultrasound PSFs are not spherically symmetric. However, since we are not able to match single scattering responses from the resolution cells, it is more appropriate to consider entire scattering regions, and the echogeneity of these regions does not change with the viewing-angle. A common physical model for the reflection, as it is described in the textbooks of Hedrick et al. [6] or Zagzebski [7], and further used by Wein et al. [8],isto calculate the reflected signal R(x) as R(x) =ρ(x) I i (x) (cos φ(x)) m (1) with the incident angle φ(x) at position x, and the incident intensity I i.the reflection coefficient calculates as ( ) 2 z(x) z(x Δd) ρ(x) = (2) z(x)+z(x Δd) with the acoustic impedance z and Δd thedistancebetweenscanlinepoints. The exponent m models the heterogeneity of the tissue interface, ranging from rather specular reflections to diffuse reflectors. Typically, strong reflections can be observed at the diaphragm and the kidney surface. To illustrate the problems, that affect the registration process when working with viewing-angle dependent ultrasound images, we create two artificial images, acquired from a curved linear probe, see Figure 1. The images show a region of high reflectivity, having a cosine intensity profile, and a scattering region. Rayleigh distributed noise is added to the images (log-compressed) to simulate the speckle noise of the ultrasound images [9]. The correct alignment of the two images, which can be acquired with lateral displacement, is to overlap the scattering regions. However, as we show in the similarity plot of Figure 1, similarity measures like CD 2 [9] favor a total overlap of the images. The reason is the dominance of the reflectivity structure, leaving only a local minimum at the desired position. 2.1 Reweighting Ultrasound Images A first attempt, when thinking about the problem, is to identify structures of high reflectivity, calculate the incidence angle of the beam, and use the ultrasonic

Alignment of Viewing-Angle Dependent Ultrasound Images 781 1 0.8 0.6 0.4 0.2 0 CD2 50 0 50 Fig. 1. Artificial ultrasound images showing reflection, scatter, and speckle. Similarity plot using CD 2 [9] along lateral direction imaging model as shown in Equation (1) to reweight the image. The result would be the creation of normalized ultrasound images being less or not at all viewing angle dependent. In the case of our example, the reflectivity region would result in a constant similarity value for translations in lateral direction, so that the scattering region would dominate the registration process. In the following, we will investigate if the presented ultrasound model is accurate enough for reweighting the images. We acquired ultrasound B-mode and RF images with a curved linear transducer in a water bath with just one object in it, having a diffuse surface, usually used for single wall calibration, see Figure 2. The reflection coefficient ρ(x) is the same for every beam, as we have the common water-object boundary. The incident intensity I i is dependent on the attenuation of the beam along its way to the boundary. The attenuation coefficient, which includes absorption and scattering effects, is 2.2 10 3 db cm MHz for water. In comparison, the attenuation coefficients for Fat (0.6 db db cm MHz ), Liver (0.9 cm MHz ), and db Kidney (1.0 cm MHz ), are about three orders of magnitude larger. Further, the length of the beam in the middle and the one on the side vary about 2 cm (1.93 exactly). This leads to an attenuation of 22.8 10 3 db by using an acquisition frequency of 2.6 MHz and considering also the way back to the transducer. Since the ratio of change of intensity I1 I 0 iscloseto1forvaluescloseto0db,wecan neglect the effect of attenuation for our experiment. The received signal, therefore only depends on the incident angle of the beam. The subsequent cost function has to be minimized in order to find the optimal exponent m, characterizing best the type of the interface ( ) 2 R(x) C(m) = cos m (φ(x)) ρ(x) I i(x) dx (3) ( ) dc(m) R(x) dm = 2 cos m (φ(x)) ρ(x)i i(x) R(x) ln(cos(φ(x))) cos m (φ(x))dx with ρ(x) I i (x) being constant for all x on the surface and corresponding to the measured reflectivity of the middle beam, and R(x) the absolute value of the received signal. We plot the cost function in Figure 2(b), where a gradient descent optimization with the above stated gradient of the cost function leads to m =34.52. In Figure 2(c), we show the intensity profile of the RF data, together with

782 C. Wachinger and N. Navab Error 200 180 160 140 120 1 0.8 0.6 0.4 0.2 RF signal cos 34.52 (a) B-Mode US image 100 0 10 20 30 40 Exponent (b)costfunction 0 0 50 100 150 200 250 (c) Intensity Profile and Best Approximation (d) RF Profile (x-axis: lateral) (e) Reweighted RF Profile Fig. 2. Reweighting experiment on ultrasound RF data the optimal cosine variant. The advantage of working with RF data is that no further processing steps or filters are applied on the data, making a meaningful evaluation of the model possible. When working with B-mode images, especially the log-compression has to be considered, turning the multiplicative cosine term of Equation (1) into an additive one. In Figure 2(e), we show the reweighted RF signal with cos 34.52. As can be seen, the result is not satisfying because instead of having a constant intensity profile, we have a W shaped one. For other interfaces it may be possible to find an exponent m so that reweighting works, however, this would only characterize this specific type of interface and not be generally applicable. We conclude that the presented ultrasonic imaging model in Equation (1) is usually not accurate enough to allow for reweighting ultrasound images, in order to make them viewing-angle independent. This finding also affects a recently introduced algorithm for US-CT registration [8], which bases upon the simulation of ultrasound images, applying this ultrasonic imaging model. The the next section, we introduce a new ultrasound matching scheme, usable for US-US and US-CT registration, to address this issue. 3 A New Ultrasound Matching Scheme In this section we present a novel matching scheme for ultrasound images, dealing with the viewing angle dependency, discussed in the last section. As we

Alignment of Viewing-Angle Dependent Ultrasound Images 783 concluded, a correction of the angle dependency of the images is with the current model not feasible. Our approach is therefore to separate reflection and scatter, which are the results of two different physical interactions of the beam with the tissue, combined in one image. In the following, we first present the method that we use for the crucial identification of regions of reflectivity in the ultrasound images, and then detail the new matching function. 3.1 Reflectivity Regions We use local phase information for identifying regions of reflectivity because it provides us with structural information independent of the brightness and contrast [10]. This independence is very important for the extraction of reflectivity regions, since the absolute intensity of the reflection varies with the incident angle. The local phase was already used by Mulet-Prada and Noble [11] for boundary detection in echocardiography. Recently, Hacihaliloglu et al. [12] achieved good results in segmenting bone in ultrasound images employing the local phase. Mellor and Brady [13] apply mutual information on local phase images for multi-modal image registration. Grau et al. [10] use the local phase for aligning 3D echocardiographic sequences. For 1-D signals the phase is constructed from the original signal and its Hilbert transform. Different approaches exist to extend this concept to N-D. In the following, we use the monogenic signal introduced by Felsberg et al. [14]. Ituses a generalization of the Hilbert transform, the Riesz transform, to calculate phase information in N-D. The image is filtered by N filters, which are given in the Fourier domain by f i H i (f 1,...,f N )= N (4) j=1 f 2 j with f 1,...,f N the Fourier domain coordinates. We follow [10] in applying log- Gabor filters prior to the calculation of the monogenic signal of the image to extract frequency and spatial localization. The monogenic signal provides us with information about the local phase P (x) each pixel x. We threshold the phase image with a value τ to obtain a binary mask, indicating reflectivity regions. Throughout all the experiments we use τ =0.7, showing that this is not a crucial parameter. Alternatively, the local phase values could directly be used as weights, resulting in a fuzzy like mask. 3.2 Matching Function Since the local phase image provides us with structural information independent of brightness, building a matching function upon this like in [10] would already significantly reduce the effects of the viewing-angle dependency on the registration. However, this also leads to the following drawbacks: First, we would be completely dependent on the outcome of the algorithm calculating the local phase. And second, texture information in the form of scattering regions is not integrated. Focusing on boundaries seems totally fine for echcardiographic

784 C. Wachinger and N. Navab applications, where not much texturing exists, but for other applications, like abdominal ones, this is not optimal. Consequently, the matching function we propose considers both, feature- and intensity-based aspects, by using reflection and texture. Considering two images u and v, an image grid Ω, regions of reflectivity Γ u = {x Ω P u (x) >τ} and Γ v = {x Ω P v (x) >τ}, furtherγ = Γ u Γ v, the matching function is M(u, v) =M T Ω\Γ (u, v)+λ MR Ω(P u,p v ) (5) with M T Ω\Γ measuring the similarity of the texture regions, excluding reflectivity regions Ω \ Γ,andM R Ω (P u,p v ) measuring the similarity of the phase images on the whole grid Ω. The matching function can be easily extended to the multivariate case by using the multivariate similarity framework proposed in [4], summing up pairwise terms. Suitable similarity measures for M T are, next to standard measures like sum of squared differences (SSD), normalized cross-correlation (NCC), and mutual information (MI) assuming a Gaussian distributed noise, the similarity measures proposed by Cohen and Dinstein [9], assuming a Rayleigh distributed noise. Especially CD 2 was reported to have good performance [15]. Suitable similarity measures for M R are once again the standard measures, and the similarity measure proposed in [10], taking the cosine of the phase difference. The disadvantage, when choosing two different similarity measures, is that the correct selection of the weighting term λ is important, since it has to map the similarity score into a comparable range. Even when working with normalized images, the outcomes of the similarity measures are not directly comparable, so that we perform the experiments with the same similarity measures for both terms and simply set λ =1. This matching framework is also directly applicable to the US-CT registration [8], where texture and reflection parts are simulated separately and finally combined during the registration. In our matching approach, reflectivity and texture regions are extracted from the ultrasound image, and can directly be matched with the simulated ones. 4 Experiments In Figure 3, we show the results of the new matching function, using SSD as similarity metric for texture and reflection, for the data set illustrated in Figure 1. Comparing the similarity plots for CD 2 and M T, which both use the texture, we can see the improvement of not considering the reflectivity region Γ in the calculation. The plot of the term M R, measuring the similarity of the phase images is very smooth, however, does not indicate the correct alignment at 30. Combining both curves, M, leads to an accurate cost function with wider capture range than M T.

Alignment of Viewing-Angle Dependent Ultrasound Images 785 8 x 10 3 Texture SSD 8 x 10 3 Phase SSD 0.014 Mixing SSD 7.5 7 0.013 6 0.012 7 5 0.011 6.5 4 0.01 6 3 0.009 0.008 5.5 2 0.007 5 50 40 30 20 10 0 10 20 30 40 50 (a) M T 1 50 40 30 20 10 0 10 20 30 40 50 (b) M R 0.006 50 40 30 20 10 0 10 20 30 40 50 (c) M Fig. 3. Similarity plots for images from Figure 1 along lateral direction using SSD 0.04 SSD 0.038 0.036 0.034 0.032 0.03 0.028 0.026 0.024 0.022 50 100 150 200 250 300 0.1 Texture NCC 0 Phase NCC 0.2 Mixing NCC 0.15 0.05 0.3 0.2 0.1 0.4 0.25 0.15 0.5 0.3 0.2 0.6 0.35 0.25 0.4 0.3 0.7 0.45 0.35 0.8 0.5 50 100 150 200 250 300 (d) M T 0.4 50 100 150 200 250 300 (e) M R 0.9 50 100 150 200 250 300 (f) M Fig. 4. Ultrasound images acquired from laterally displace positions. Similarity plots for SSD and the different parts from our new matching scheme. We also conducted experiments for US-US registration on ultrasound images acquired with a curved linear transducer, see Figure 4. The images are acquired from an ultrasound phantom, with a true displacement of 88.0 mm. The bony structure on the left side of the first image is depicted in the middle of the second image. The similarity plots for SSD, NCC, and CD 2 are comparable, with SSD indicating the correct alignment at 96.8 mm. The combined cost function indicates the correct alignment at 88.8 mm, with NCC as similarity measure for both terms. We performed further experiments on a data set, similar to the one presented by Penney et al. [16], simulating the US-CT registration problem. An illustration of the images together with the good results of our matching approach can be found in the supplementary material 1. 1 http://www.webcitation.org/5hezwaajc

786 C. Wachinger and N. Navab 5 Conclusion The alignment of ultrasound images is complicated by the viewing-angle dependent nature of the images. After an analysis of the effects on the registration process, we presented a method for reweighting ultrasound images. However, the commonly applied imaging model for ultrasound was shown to be not accurate enough, for allowing a correction of the viewing-angle dependency. Instead, we introduced a new matching scheme for the alignment of ultrasound images, which can be used for US-US as well as US-CT registration. It incorporates the usage of feature- and intensity-based metrics into one global measure. We evaluated the new registration approach on various data sets, with good results. References 1. Henrich, W., Schmider, A., Kjos, S., Tutschek, B., Dudenhausen, J.W.: Advantages of and applications for extended field-of-view ultrasound in obstetrics. Archives of Gynecology and Obstetrics 268, 121 127 (2003) 2. Poon, T., Rohling, R.: Three-dimensional extended field-of-view ultrasound. Ultrasound in Medicine and Biology 32(3), 357 369 (2005) 3. Ni, D., Qu, Y., Yang, X., Chui, Y.-P., Wong, T.-T., Ho, S.S.M., Heng, P.-A.: Volumetric ultrasound panorama based on 3D SIFT. In: Metaxas, D., Axel, L., Fichtinger, G., Székely, G. (eds.) MICCAI 2008, Part II. LNCS, vol. 5242, pp. 52 60. Springer, Heidelberg (2008) 4. Wachinger, C., Wein, W., Navab, N.: Three-dimensional ultrasound mosaicing. In: Ayache, N., Ourselin, S., Maeder, A. (eds.) MICCAI 2007, Part II. LNCS, vol. 4792, pp. 327 335. Springer, Heidelberg (2007) 5. Cahill, N., Schnabel, J., Noble, J., Hawkes, D.: Revisiting overlap invariance in medical image alignment. In: MMBIA (2008) 6. Hedrick, W.R., Hykes, D.L., Starchman, D.E.: Ultrasound Physics and Instrumentation. Mosby, 4th edn. (2004) 7. Zagzebski, J.: Essentials Of Ultrasound Physics, 1st edn., Mosby (1996) 8. Wein, W., Brunke, S., Khamene, A., Callstrom, M.R., Navab, N.: Automatic ctultrasound registration for diagnostic imaging and image-guided intervention. Medical Image Analysis 12, 577 585 (2008) 9. Cohen, B., Dinstein, I.: New maximum likelihood motion estimation schemes for noisy ultrasound images. Pattern Recognition 35(2), 455 463 (2002) 10. Grau, V., Becher, H., Noble, J.: Registration of multiview real-time 3-d echocardiographic sequences. IEEE Transactions on Medical Imaging 26(9) (September 2007) 11. Mulet-Parada, M., Noble, J.: 2D+ T acoustic boundary detection in echocardiography. Medical Image Analysis 4(1), 21 30 (2000) 12. Hacihaliloglu, I., Abugharbieh, R., Hodgson, A.J., Rohling, R.: Bone segmentation and fracture detection in ultrasound using 3D local phase features. In: Metaxas, D., Axel, L., Fichtinger, G., Székely, G. (eds.) MICCAI 2008, Part I. LNCS, vol. 5241, pp. 287 295. Springer, Heidelberg (2008) 13. Mellor, M., Brady, M.: Phase mutual information as a similarity measure for registration. Medical Image Analysis 9(4), 330 343 (2005) 14. Felsberg, M., Sommer, G.: The monogenic signal. IEEE Transactions on Signal Processing 49(12), 3136 3144 (2001) 15. Boukerroui, D., Noble, J.A., Brady, M.: Velocity estimation in ultrasound images: a block matching approach.. Inf. Process Med. Imaging 18, 586 598 (2003) 16. Penney, G., Griffin, L., King, A., Hawkes, D.: A novel framework for multi-modal intensity-based similarity measures based on internal similarity. SPIE 6914 (2008)