Fusion of Multi-Modality Volumetric Medical Imagery

Similar documents
Fusion of Multi-Modality Volumetric Medical Imagery

Color and Shading. Color. Shapiro and Stockman, Chapter 6. Color and Machine Vision. Color and Perception

Nicolai Petkov Intelligent Systems group Institute for Mathematics and Computing Science

Nicolai Petkov Intelligent Systems group Institute for Mathematics and Computing Science

Computational Models of V1 cells. Gabor and CORF

A Study of Medical Image Analysis System

Babu Madhav Institute of Information Technology Years Integrated M.Sc.(IT)(Semester - 7)

2 Michael E. Leventon and Sarah F. F. Gibson a b c d Fig. 1. (a, b) Two MR scans of a person's knee. Both images have high resolution in-plane, but ha

Nicolai Petkov Intelligent Systems group Institute for Mathematics and Computing Science

EE795: Computer Vision and Intelligent Systems

Neurophysical Model by Barten and Its Development

Automatically Improving 3D Neuron Segmentations for Expansion Microscopy Connectomics. by Albert Gerovitch

Fourier analysis of low-resolution satellite images of cloud

Spectral Classification

CS4442/9542b Artificial Intelligence II prof. Olga Veksler

Texture Segmentation Using Gabor Filters

Extensions of One-Dimensional Gray-level Nonlinear Image Processing Filters to Three-Dimensional Color Space

Fuzzy Set Theory in Computer Vision: Example 3, Part II

Spatial Enhancement Definition

Artifacts and Textured Region Detection

Surface Projection Method for Visualizing Volumetric Data

Active Tracking of Surface Targets in Fused Video 1

(Refer Slide Time 00:17) Welcome to the course on Digital Image Processing. (Refer Slide Time 00:22)

One image is worth 1,000 words

Lecture 4: Spatial Domain Transformations

Digital Image Processing. Image Enhancement - Filtering

Image segmentation. Václav Hlaváč. Czech Technical University in Prague

Lucy Phantom MR Grid Evaluation

Vivekananda. Collegee of Engineering & Technology. Question and Answers on 10CS762 /10IS762 UNIT- 5 : IMAGE ENHANCEMENT.

Digital Image Processing ERRATA. Wilhelm Burger Mark J. Burge. An algorithmic introduction using Java. Second Edition. Springer

Visual object classification by sparse convolutional neural networks

Biomedical Image Processing

Medical Image Fusion using Rayleigh Contrast Limited Adaptive Histogram Equalization and Ant Colony Edge Method

Reading. 2. Color. Emission spectra. The radiant energy spectrum. Watt, Chapter 15.

PATENT COUNSEL 1176 HOWELL ST. CODE 00OC, BLDG. 11 NEWPORT, RI 02841

CS4442/9542b Artificial Intelligence II prof. Olga Veksler

Annales UMCS Informatica AI 1 (2003) UMCS. Registration of CT and MRI brain images. Karol Kuczyński, Paweł Mikołajczak

Quantifying Three-Dimensional Deformations of Migrating Fibroblasts

INTRODUCTION TO DEEP LEARNING

Prostate Detection Using Principal Component Analysis

Realtime 3D Computer Graphics Virtual Reality

Exercise: Simulating spatial processing in the visual pathway with convolution

Copyright 2005 Center for Imaging Science Rochester Institute of Technology Rochester, NY

Data: a collection of numbers or facts that require further processing before they are meaningful

Data Fusion. Merging data from multiple sources to optimize data or create value added data

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness

On the road to invariant object recognition: How cortical area V2 transforms absolute to relative disparity during 3D vision

NIH Public Access Author Manuscript Proc Soc Photo Opt Instrum Eng. Author manuscript; available in PMC 2014 October 07.

Object Identification in Ultrasound Scans

Perceptual Effects in Real-time Tone Mapping

Computational Foundations of Cognitive Science

Computational Algorithm for Estimating the Gradient Isocenter of an MRI Scanner

Locating ego-centers in depth for hippocampal place cells

IMAGE SEGMENTATION. Václav Hlaváč

Robust contour extraction and junction detection by a neural model utilizing recurrent long-range interactions

A Hierarchial Model for Visual Perception

Comparative Analysis in Medical Imaging

TUMOR DETECTION IN MRI IMAGES

Extraction of Color and Texture Features of an Image

Dynamic visual attention: competitive versus motion priority scheme

Visual Representation from Semiology of Graphics by J. Bertin

IMPORTANT INSTRUCTIONS

Chapter 4. Clustering Core Atoms by Location

IDL Tutorial. Working with Images. Copyright 2008 ITT Visual Information Solutions All Rights Reserved

A Model-Independent, Multi-Image Approach to MR Inhomogeneity Correction

Machine Learning 13. week

2006 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media,

Module 3. Illumination Systems. Version 2 EE IIT, Kharagpur 1

Representing the World

Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data

Illumination and Shading

Methodological progress in image registration for ventilation estimation, segmentation propagation and multi-modal fusion

Image Enhancement in Spatial Domain. By Dr. Rajeev Srivastava

Nonlinear Multiresolution Image Blending

Medical Image Processing using MATLAB

Single-Frame Image Processing Techniques for Low-SNR Infrared Imagery

Introduction to Medical Image Processing

Motivation. Intensity Levels

CT NOISE POWER SPECTRUM FOR FILTERED BACKPROJECTION AND ITERATIVE RECONSTRUCTION

SIVIC GUI Overview. SIVIC GUI Layout Overview

SYDE 575: Introduction to Image Processing

CIE L*a*b* color model

Spatially variant dimensionality reduction for the visualization of multi/hyperspectral images

An Algorithm for Blurred Thermal image edge enhancement for security by image processing technique

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Computational Medical Imaging Analysis Chapter 4: Image Visualization

NEAR-IR BROADBAND POLARIZER DESIGN BASED ON PHOTONIC CRYSTALS

Function approximation using RBF network. 10 basis functions and 25 data points.

CS635 Spring Department of Computer Science Purdue University

Color, Edge and Texture

By Colin Childs, ESRI Education Services. Catalog

EEM 463 Introduction to Image Processing. Week 3: Intensity Transformations

Visualizing Flow Fields by Perceptual Motion

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

Volume visualization. Volume visualization. Volume visualization methods. Sources of volume visualization. Sources of volume visualization

CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS

Compression of 3-Dimensional Medical Image Data Using Part 2 of JPEG 2000

Histograms of Oriented Gradients

Image Compression. -The idea is to remove redundant data from the image (i.e., data which do not affect image quality significantly)

Lecture Image Enhancement and Spatial Filtering

Transcription:

Fusion of Multi-Modality Volumetric Medical Imagery Mario Aguilar Knowledge ystems Laboratory Mathematical, omputing and Information ciences Department Jacksonville tate University marioa@ksl.jsu.edu Abstract - Ongoing efforts at our laboratory have targeted the development of techniques for fusing medical imagery of various modalities (i.e. MRI, T, PET, PET, etc.) into single image products. Past results have demonstrated the potential for user performance improvements and workload reduction. While these are positive results, a need exists to address the threedimensional nature of most medical image data sets. In particular, image fusion of three-dimensional imagery (e.g. MRI slices) must account for information content not only within a given slice but also across adjacent slices. In this paper, we describe extensions made to our 2D image fusion system that utilize 3D convolution kernels to determine locally relevant fusion parameters. Representative examples are presented for fusion of MRI and PET imagery. We also present these examples in the context of a GUI platform under development aimed at improving user-computer interaction for exploration and mining of medical data. Keywords: Image fusion, sensor fusion, medical imaging, multi-dimensional visualization, decorrelation, 3D kernels. 1 Introduction The widespread availability of medical imaging devices has led to an explosion in the amount of information that medical specialists must confront in diagnosing and planning treatment. To avoid situations where the imaging information can overwhelm medical personnel or can degrade patient care, methods for combining multi-source imaging information must be developed. At Fusion 2001, we presented our work [1] on utilizing a neurophysiologically-motivated fusion architecture for combining multi-modality medical imagery. The approach is based on techniques originally developed for fusion of low-light visible and thermal infrared for a night-vision system [2]. While our original report presented positive results regarding fusion of 2-, 3-, and 4-modality medical imagery, the approach was not capable of accounting for volumetric information present in medical imaging modalities such as MRI and PET. In particular, these modalities are Joshua R. New Knowledge ystems Laboratory Mathematical, omputing and Information ciences Department Jacksonville tate University newj@ksl.jsu.edu recorded as multi-plane (slice) images across the organ of interest. While it is natural to associate the individual images as being the main tool used by radiologists in diagnosis, it is also true that medical specialists rely heavily on the information available across adjacent slices. Hence, it is routine practice for radiologists to utilize multislice displays in order to gain volumetric contextual information and assess the extent of features in all three dimensions. Furthermore, we identified the need to perform image enhancement across slices so as to obtain better definition of tissue in the context of their extent in all three dimensions. The work reported here consists of our initial results in extending the neurophysiologcallymotivated fusion architecture to three dimensions. 1.1 Neurophysiologically-motivated fusion The visual system of primates and humans contains three types of light sensors known as cones, which have overlapping sensitivities (short, medium, and long wavelengths). It is through the combination of these three sensor sets that we obtain our perception of color. ircuitry in the retina is functionally divided into two stages. The first one utilizes non-linear neural activations and lateral inhibition within bands to enhance and normalize the inputs. The second stage utilizes similar neural components in an arrangement of connections that lead to between-band competition that in turn produces a number of combinations of the three original bands [5]. This last stage of processing enhances the complementary information that exists in each of the bands (e.g. a spectral decorrelation operation). The fusion architecture is motivated by this basic connectivity. Each of the processing stages is implemented via a non-linear neural network known as the shunt (described in the next section). The resulting nonlinear combinations lead to information decorrelation not unlike what is usually targeted by principal component analysis techniques. However, in neuro-physiological systems and in our architecture, the non-linear operator has a very narrow spatial window providing a better-tuned decorrelation. In addition, the operator is modulated by more globally defined statistical characteristics of the input 1206

that produce normalization, smoothing, and between-band calibration. 2 Methods 2.1 2D hunt The basic building block of the fusion architecture is a non-linear neural network known as a shunting neural network [4]. This neural network, which acts like a filter, models the dynamics of neural activation in response to three different signals: an excitatory one-to-one input corresponding to the current pixel, an inhibitory input from surrounding neurons, and passive activation decay (i.e. a leaky integrator). The expression that captures these interactions in a dynamical system is defined in terms of the following differential equation: ( B x )[ I ] ( D + x)[ G I ] x& = Ax + (1) Here, x is the activation of each cell receiving input from each of the pixels in the input image. A is a decay rate, B and D are the maximum and minimum activation levels respectively and are set to 1 in the simulations. and G s, as represented by the following equation, G 2 2 ( x + y ) 1 2 2σ ( x, y) e 2 = (2) 2πσ serve to weigh the excitatory input (I c ) vs. lateral inhibitory (I s ) image inputs. The neural network consists of a two dimensional array of these shunting neurons with dimensions corresponding to the width and height of the input image. When the input is applied, the network rapidly reaches equilibrium, which produces the resulting output image. This equilibrium state can be understood in terms of the possible values of x after the neuron has reached a steady state as shown in the following equation: x = A + [ I G I ] + [ I + G I ] Here, it is straightforward to understand the numerator as a contrast enhancement operation since it represents a difference of Gaussians. The denominator serves to normalize the activation of x with respect to the activity of its neighborhood. In effect, the combination of these two operations leads to the dynamic range compression of the input image in conjunction with contrast enhancement. Parameter A serves to control the characteristics of the (3) operator, from ratio processing (when A is small with respect to the local statistics) to linear filtering (when A is comparatively large). In the case in which the operator is used to combine two bands, one of the bands is mapped to the center and the other to the surround. In the case where band 1 is mapped to the center, each of the pixels from band 1 is used to drive the excitatory input of its corresponding shunting operator. Then, a corresponding broader area of the image from band 2 is used as the surround input of the same shunt operator. The result is the contrast enhancement of information in band 1 as matched against band 2. The relationship between this operation and decorrelation has been previously documented [5]. 2.2 3D hunt As discussed previously, modern medical imaging techniques such as MRI and PET produce multi-slice image volumes. For this reason, we have extended the shunting operator to three dimensions in order to account for information content between slices. Furthermore, it is desirable to process such image sets and perform image fusion while exploiting the information present in the third dimension. To accomplish this, it is necessary to modulate contrast enhancement and information decorrelation as a function of the local context in all three dimensions. The shunting equation that performs this function is as follows: ( B x ) G [ I ] ( D+ x )[ G I ] x& = Ax + (4) c Here, the activity of cell x (i.e. the resulting pixel value) is defined in terms of the contrast enhancement and dynamicrange normalization defined by a three-dimensional neighborhood around the cell. In this equation, we introduce two Gaussian terms: the first one modulating the excitatory inputs to cell x, and the second one modulating its inhibitory inputs. This extension required the use of a three-dimensional Gaussian kernel as follows: G 2 2 2 ( x + y + z ) 1 2 2σ ( x, y, z) e 2 3 = (5) 4π σ The combination of both kernels (G c and G s ) in equation 4 leads to a three-dimensional operator that we represent by the following icon: 1207

2.3 olor-opponency for image fusion The basic fusion architecture consists of two distinct processing stages. In the first one, as in the retina, we utilize a shunting operation to obtain within-band image enhancement and normalization. This produces contrast enhancement, dynamic range calibration, and normalization of input images. In this version however, the input to the 3D shunting operator is a stack of image slices. Here, one slice is processed at a time with the shunt operating on the current slice as well as those slices immediately above and below. The result is a slice with enhanced contrast characteristics due to between-slice context information. The second stage adopts the use of the same non-linear neural network operator to produce between-band decorrelation, information enhancement, and fusion. These stages for the two-band fusion of MRI are illustrated in Figure 1. Here, the icon for the 3D shunting operator is colored coded so as to indicate the source of its inputs. The shunt combinations of the second stage, as shown in Figure 1, provide two unique sets of information-rich images. These combinations perform paired decorrelations between the input image modalities. For instance, in the case of the top combination (red-center and greensurround in Fig. 1) the result is the decorrelation of the first modality (T1) from the second modality (T2). In other words, it enhances information that is present in band 1 but not in band 2. The resulting images from both decorrelations are then mapped to the red and blue channels of the color display. The enhanced T1 image was chosen as the input to the green channel in order to highlight information from that modality. Another processing stage may be introduced prior to producing the final color image that remaps the color assignments from those derived by the fusion process. Here, a mapping from RGB to HV space allows the user to manipulate the appearance of the image (e.g. hue remap) to obtain a more natural coloring scheme. The modified HV values are mapped back to RGB to be used in generating the final color fused image. The architecture for three-band MRI fusion is illustrated in Figure 2. Here, the first stage of processing is as before, where each of the input bands is separately contrast enhanced and normalized. Then, two between-band shunting operations produce distinct fusion products. The first one decorrelates the information between bands 1 (T1) and 2 (PD). The second does it for bands 3 (T2) and 2. The resulting fused images are then mapped to the I and Q (also known as red-green and blue-yellow) components of the YIQ color space of the image. In this case, the resulting images highlight information that is unique to bands 1 and 3 in red and blue hues, while presenting unique information to band one in green and yellow hues. The Y or achromatic component is derived from the enhanced band 2 image that provides the most faithful structural details. The YIQ components are then mapped to RGB space. Figure 2. Three-band MRI fusion architecture. ee text for details. Figure 1. Two-band fusion architecture used for processing functional MRI imagery. oncentric spherical volumes represent a shunting neural network operator. olor coding indicates polarity of the operations. ee Text for details. Figure 3. Four-band MRI/PET fusion architecture. ee text for details. 1208

The architecture for four-band fusion between MRI and PET is shown in Figure 3. Here, the second stage of processing produces the decorrelation between T1- weighted and PET, as well as between T2-weighted and PET. Notice that the decorrelation is done in both directions for each of the pairs. The most noticeable difference is the addition of a third processing stage. This additional between-band competition leads to further color contrast enhancement and decorrelation as suggested by connectivity in the primary visual cortex of primates. The two resulting decorrelated images are mapped to the chromatic I and Q channels. Once again, to preserve the high resolution of the MRI imagery, the structural modality MRI-PD image is mapped to the Y, or achromatic channel. 3 Results The fusion architecture was applied to medical imagery obtained from the Brain Atlas web site (see [1] for reference). The imagery included 3-modality MRI and PET slices. We present the results of processing the imagery using the new shunting operator in three of the stages of the fusion process: within-band enhancement, between-band decorrelation, and color fusion. 3.1 Within-band shunting. The 3D shunting operator (eq. 3) was applied to each of the 3 original MRI image modalities (PD, T1-weighted, and T2-weighted) (see [1] for a description of the modalities). The resulting contrast-enhancement and normalization were compared to the results obtained using the original 2D shunting operator. Figure 4 presents this comparison. We include the original unprocessed imagery to provide a sense as to the success of both methods in bringing out significant amount of information. Here, the parameters of both operators were the same for the two-dimensional extent of the neighborhoods. In addition, the threedimensional extent of the neighborhood for the 3D shunt was set to two slices above and two below. We have experimented with different shape Gaussians for the third dimension and have found that a symmetrical neighborhood works well for this data set. We anticipate that as resolution along the third dimension degrades (i.e. distance between slices) the Gaussian kernel would need to be made narrower accordingly. We hope to test this idea as soon as we obtain a relevant image set. 3.2 Between-band shunting The 3D shunting operator was again applied across slice stacks, this time utilizing one modality as the input to the center and the other to the surround of the operator. The result is a decorrelation of information between the bands as explained earlier. Furthermore, this decorrelation includes a modulation factor derived from the context of adjacent slices. The results for 2-band fusion are shown in Figure 5. lice 10 T1-weighted T2-weighted lice 11 lice 12 Figure 4. omparison of 2D vs. 3D within-band shunting of three adjacent image slices. Left column: original unprocessed imagery. Middle column: 2D shunted imagery. Left column: 3D shunted imagery. T1 vs. T2 decorrelation T2 vs. T1 decorrelation Figure 5. Results of applying a 3D shunt to combine and decorrelate T1-weighted and T2-weighted images. 1209

3.3 olor-fusion of 2-, 3-, and 4-modality imagery The imagery was further processed using the 3D fusion architecture to produce the color-fused results that combine information from all input modalities. Here, both color and brightness characteristics of the resulting image are used to convey information about the characteristics and source of the information in the original imagery. Figure 6 presents side-by-side, the result of fusing 2- and 3-band MRI imagery. Three slices are shown to provide across-slice context and highlight the efficacy of the method in enhancing information across all three dimensions. In figure 7, we present a comp arison of the original (2D) and updated (3D) 4-band fusion results as applied to slice 11. Here, the result on the left is as obtained using lice 10 lice 11 lice 12 2-band fusion 3-band fusion Figure 6. omparison of 2- and 3-band fusion results obtained on the MRI imagery. Left column : 2-band fusion of T1- and T2-weighted slices. Right column : 3-band fusion of PD, T1- and T2-weighted slices. 2D olor Fusion 3D olor Fusion Figure 7. omparison of 2D vs. 3D 4-band fusion of MRI and PET. Notice the enhanced delineation of PET information as indicated by its red hues. the original fusion architecture. On the right, the result of the 3D fusion illustrates enhanced information content derived from across-slice processing. In particular, the resulting 3D fused imagery presents better delineations of shape as a result of accounting for volumetric information across slices. Finally, to emphasize the benefits of the 3D operator in enhancing information across slices, Figure 8 presents a comparison of 3D imagery as obtained with the new fusion architecture (on the left) and the original architecture (on the righ). Here, we have utilized our 3D visualization tools to recreate three orientation planes and their corresponding fused imagery overlaid on them. These orientations (coronal, saggital, and axial) represent orthogonal views through the cranial volume. The results of the 3D fusion architecture demonstrate better definition of the highlighted volume (tumor) in all three dimensions. 4 onclusions We have extended our fusion architecture to address the three-dimensional nature of medical image sets. In particular, we have introduced a 3D shunting operator that contrast enhances and normalizes information present in the immediate neighborhood of each pixel including adjacent image slices. Our results compare favorably against the results of the original 2D fusion architecture. Better definition of image details and volumetric information make this approach a significant advance in medical image fusion. Ongoing efforts will help us assess the significance of the information gains and fine tune our fusion architecture. Acknowledgements This work was supported by a Faculty Research Grant awarded to the first author by the faculty research committee and Jacksonville tate University. Opinions, interpretations and conclusions are those of the authors and not necessarily endorsed by the committee or Jacksonville tate University. 1210

A D B E F Figure 8. Three-dimensional visualization of 3-band fusion results. Left column presents results obtained using the new 3D shunting operator. Images on the right column are the results obtained using the original shunt operator. A. Far view of the three orthogonal image planes obtained with the 3D fusion architecture. B. loser view of the intersection of the planes highlighting the volume of interest (tumor).. losest view of the area of interest where the contrast-enhancement across slices can be observed (slices change in the up-down direction). D. Far view of the three orthogonal image planes obtained with the 2D fusion architecture. E. loser view of the intersection of the planes highlighting the volume of interest (tumor). F. losest view of the area of interest where the contrastenhancement across slices is noticeably absent. 1211

References [1] M. Aguilar and A.L. Garrett, Neurophysiologicallymotivated sensor fusion for visualization and characterization of medical imagery, Proc. of the Fourth International onference on Information Fusion, 2001. [2] M. Aguilar, D.A. Fay, W.D. Ross, A.M. Waxman, D.B. Ireland, and J.P. Racamato, Real-time fusion of low-light D and uncooled IR imagery for color night vision, Proc. Of PIE onf. On Enhanced and ynthetic Vision, 3364, 1998. [3] P. chiller and N.K. Logothetis, The color-opponent and broad-band channels of the primate visual system, Trends in Neuroscience, 13, pp.392-398, 1990. [4]. Grossberg, Neural Networks and Natural Intelligence, ambridge, MA: MIT Press, 1988. [5] M. Aguilar and A.M. Waxman, omparison of opponentcolor neural processing and principal components analysis in the fusion of visible and thermal IR imagery. Proc. of the Vision, Recognition, and Action: Neural Models of Mind and Machine onference, Boston, MA, 1997.q 1212