Escolar Documentos
Profissional Documentos
Cultura Documentos
By
Jeremy Michael Nett
B.S.E.E., University of Louisville, 2000
A Thesis
Submitted to the Faculty of the
University of Louisville
Speed Scientific School
As Partial Fulfillment of the Requirements
For the Professional Degree
MASTER OF ENGINEERING
Department of Electrical and Computer Engineering
December, 2001
ii
THE STUDY OF MS USING MRI, IMAGE PROCESSING, AND VISUALIZATION
Submitted by:
Jeremy Michael Nett
A Thesis Approved on
By the Following Reading and Examination Committee:
Aly A. Farag, Thesis Director
Tom Cleaver
Kyung Kang
Robert Falk
Christina Kaufman
Stephen Hushek
iii
ACKNOWLEDGMENTS
The author would like to thank all those that made his education at the University
of Louisville possible, including the many contributors to the scholarship funds of which
he was profoundly honored to be a recipient.
The author would also like to thank Dr. Aly Farag and the CVIP Lab for an
interesting thesis topic, and support during the completion of this work. Additionally, the
author would like to thank Dr. Robert Falk of Jewish Hospital for spending much of his
valuable time guiding this work, commenting on its quality, and for many suggestions on
how to make the tools developed more applicable in the setting of medical research and
clinical use. The author would also like to thank the additional members of his thesis
committee for their review and critique of his work, including Dr. Stephen Hushek of
Norton Healthcare, Dr. Christina Kaufman of the Institute for Cellular Therapeutics of
the School of Medicine at the University of Louisville, Dr. Thomas Cleaver of the
Electrical and Computer Engineering Department, and Dr. Kyung Kang of the Chemical
Engineering Department.
Last but not least, the author would like to sincerely thank his parents, Michael
and Kathy Nett, for putting up with him, and his grandmother, Agnita Nett, for a place to
stay. Without their support, this work would have not been possible.
iv
ABSTRACT
Multiple sclerosis (MS), a well known disease of the nervous system in which the
myelin sheaths of axons are damaged, may be imaged using magnetic resonance imaging
(MRI) modalities. In this research, image analysis, volume registration, and scientific
visualization techniques are applied for quantitative and qualitative analysis of the
response of the disease to a proposed treatment regimen. Image analysis techniques are
applied, with the objective of automated and reliable quantitative evaluation of MS
lesions in the brain, through segmentation and classification of the brain and MS lesions.
To facilitate a time-series analysis of lesions, a volume registration technique is applied
to geometrically align MRI scans taken at different times over the course of treatment.
Additionally, scientific visualization techniques are utilized to facilitate three-
dimensional analysis of the disease pathology, and evaluation of changes in the structure
of lesions over a period of treatment. The result of these efforts is a preliminary system
for the study of MS using MRI.
v
TABLE OF CONTENTS
Page
APPROVAL PAGE..
ii
ACKNOWLEDGMENTS
iii
ABSTRACT. iv
NOMENCLATURE. viii
LIST OF TABLES xi
LIST OF FIGURES.. xii
I. INTRODUCTION... 1
A. General Introduction. 1
B. Introduction to Multiple Sclerosis 2
C. Introduction to Magnetic Resonance Imaging of MS.. 2
D. Introduction to Computer-Assisted Evaluation of MS 5
E. Previous Work in MS Studies Using MRI and Image
Processing Approaches..
7
F. Introduction to Lab Facilities and Data Acquisition. 9
G. Outline of the Components of the Researched Solution.. 11
H. Summary..
12
II. VOLUME REGISTRATION. 13
A. Introduction to Volume Registration... 13
B. Imaging Model. 19
C. Transformations... 26
vi
D. Volume Interpolation ... 30
E. Overview of Registration Approaches. 31
F. Registration Metric/Criteria. 32
G. Computation of the Mutual Information Metric.. 37
H. Search Algorithm. 43
I. Implementation.. 44
J. Results: MS Studies.. 44
K. Summary.. 50
III. BRAIN SEGMENTATION FROM THE HEAD. 51
A. Introduction and Necessity.. 51
B. Brain Segmentation Utilizing Registration.. 52
C. Results: Manual a priori Segmentation... 55
D. Results: Semi-Automatic a priori Segmentation. 59
E. Summary.. 62
IV. TISSUE SEGMENTATION. 63
A. Introduction and Necessity... 63
B. Feature Selection.. 64
C. Thresholding. 67
D. Segmentation by Image Enhancement. 69
E. Segmentation by Unsupervised Clustering.. 73
F. Bayesian Classification 77
G. Bayesian Classification: Gaussian Conditionals, Equal a priori
Probabilities...
80
vii
H. Bayesian Classification: Nonparametric Conditionals, a priori
Probabilities Modeled by Markov Random Fields
82
I. Quantification 89
J. Accuracy 90
K. Summary.. 91
V. VISUALIZATION. 92
A. Introduction.. 92
B. Visualization for the Study of Volume Registration 92
C. Visualization for the Study of Volume Segmentation. 96
D. Web-Based Presentation of Results..
98
E. Summary.. 99
VI. CONCLUSIONS AND RECOMMENDATIONS 101
REFERENCES.. 104
VITA. 109
viii
NOMENCLATURE
l = a volume coordinate vector in the image coordinate system
w = a volume coordinate vector in the world coordinate system
c =
coordinate vector locating the center of a volume
or
the number of states of nature
c(i) = a diffusion function
C =
transformation matrix for alignment of the center of a volume to
the origin in the world coordinate system
V =
transformation matrix for scaling a volume to account for
spatial sampling rates
= transformation matrix for accounting of a gantry angle
A
i,w
=
transformation matrix for conversion of image to world
coordinates
R = reference volume for registration
F =
floating volume for registration
or
a random field
T = transformation matrix for translation
R
x
, R
y
, R
z
= axis-angle rotation matrices
R = transformation matrix for rotation
T
FR
=
transformation matrix from image coordinates in the floating
volume to the reference volume
lower_left_N1,
lower_left_N2,
lower_left_N3
=
coordinates for alignment of a bounding cube for trilinear
volume interpolation
ix
V(sample
i
) =
interpolated sample from a volume for a sample coordinate of
sample
i
delta
x
, delta
y
, delta
z
=
translational differences between a coordinate to interpolate at,
and a point that forms an origin for trilinear interpolation
weight
i
= a weight and index i for trilinear interpolation
sample
i
= a sample at index i in a bounding cube for trilinear interpolation
p(x), q(x) = a probability distribution function
D(p||q) = relative entropy
I(X, Y) = mutual information of the random variables X and Y
H(X) = entropy of the random variable X
T
\
|
=
= 1
2
1
2
1
2
1
2
1
(2)
Incorporation of the volume center as an origin for locating the volume in 3D
space is given by multiplication of the image coordinate vector l by the matrix C, given in
eq. 3.
(
(
(
(
=
1 0 0 0
1 0 0
0 1 0
0 0 1
z
y
x
c
c
c
C
(3)
Voxel sizes are the spatial sampling periods of the imaged volume. There are
three such sizes, v
x
, v
y
, and v
z
, representing the size of a voxel in each of the coordinate
dimensions. Incorporation of the voxel sizes in proper scaling of the image coordinates
to world coordinates is given by multiplication of the image coordinate vector l by the
matrix V, given below in eq. 4.
24
(
(
(
(
=
1 0 0 0
0 0 0
0 0 0
0 0 0
z
y
x
v
v
v
V
(4)
Finally, an additional term, a gantry angle , is taken into consideration. This
term measures the angle of the gantry as it is positioned in the scanner. This factor is
easily incorporated in conversion of an image coordinate vector l to world coordinates by
the matrix , given below in eq. 5. Note that this matrix is simply a rotation matrix, a
class of transformations to be discussed further below.
(
(
(
(
=
1 0 0 0
0 1 sin 0
0 0 cos 0
0 0 0 1
(5)
Therefore, with each of these matrices considered, the conversion from image to
world coordinates can be represented as the matrix product of the , V, and C matrices.
The resulting coordinate transformation is referenced as A
i,w
, and is given below in eq. 6.
25
(
(
(
(
= =
1 0 0 0
sin sin 0
cos 0 cos 0
0 0
,
y y z z y
y y y
x x x
w i
c v c v vz v
c v v
c v v
C V A
(6)
It will be of interest to be able to convert from world coordinates into image
coordinates. This can be simply done by inverting the A
i,w
matrix, to obtain A
i,w
-1
, as
given in eq. 7 below, in closed form.
(
(
(
(
(
(
(
1 0 0 0
1
cos
sin
0
0
cos
1
0
0 0
1
1
,
z
z z
y
y
x
x
w i
c
v v
c
v
c
v
A
(7)
Thus, the relationships given in eq. 8 allow for conversion between image and
world coordinates.
w A l
l A w
w i
w i
=
=
1
,
,
(8)
26
Each of the above parameters (d
x
, d
y
, d
z
, v
x
, v
y
, v
z
, ) are known a priori
parameters obtained from the DICOM header of a study.
Two volumes will be used in the registration, a reference volume R, and a floating
volume F. The purpose of registration is to find a transformation that aligns these two
volumes.
C. Transformations
For alignment, rotation and translation will be considered. Alignment
transformations that consist of only translation and rotation are known as rigid-body
transformations.
Only the translation and rotation transformations are considered here, as these
effects are observed in the data to be by far the predominant effects necessary for
registration of two or more studies of the same patient. Other transformations are judged
to be either negligible in effect, or not applicable. Included in these transformations are
scaling (or magnification) between two studies, as the voxel sizes (the metric dimensions
of a voxel) of each study are known from the scanner, and are known to be accurate
enough for use, given that the scanner is in good repair.
Translation quantifies a three-dimensional offset along coordinate axes between
two volumes in world coordinates. In homogeneous coordinates, with t
x
, t
y
, and t
z
representing translation along the x, y, and z coordinate axes, respectively, the translation
27
can be formulated by a matrix T, that multiplies a coordinate w to obtain a translated
coordinate vector. T is given below in eq. 9.
(
(
(
(
=
1 0 0 0
1 0 0
0 1 0
0 0 1
z
y
x
t
t
t
T
(9)
A rotation is quantified by three angles,
x
,
y
,
z
, that are axis-angle
parameterizations of rotation. These parameters specify a rotation around the unit
direction vectors of the world coordinate system. Corresponding to each parameter is a
rotation matrix, denoted as R
x
, R
y
, and R
z
, respectively. A complete rotation matrix,
taking into account each rotation angle, is given by the matrix product of R
x
, R
y
, and R
z
,
and forms the matrix R, given in eq. 13 below. R
x
, R
y
, and R
z
are given in eqs. 10
through 12, respectively.
(
(
(
(
=
1 0 0 0
0 cos sin 0
0 sin cos 0
0 0 0 1
x x
x x
x R
(10)
28
(
(
(
(
=
1 0 0 0
0 cos 0 sin
0 0 1 0
0 sin 0 cos
y y
y y
y R
(11)
(
(
(
(
=
1 0 0 0
0 1 0 0
0 0 cos sin
0 0 sin cos
z z
z z
z R
(12)
Rz Ry Rx R = (13)
Utilization of the matrices T and R constitutes a rigid-body transformation. The
complete transformation can be written as the product of these two matrices, and shall be
denoted as A. To calculate the rigid body transformation of a coordinate w, the
coordinate vector w is multiplied by the transformation matrix A, as shown in eq. 14,
where w
1
is a coordinate, and w
2
is the rigid body transformation of that coordinate.
1 2
w A w = (14)
29
Next, a formulation of the above rigid body transformation shall be given. Below,
w
F
, l
F
, w
R
, and l
R
will represent world and image coordinates in the floating volume, and
world and image coordinates in the reference volume, respectively. A
i,w,F
and A
i,w,R
will
represent the transformations from image to world coordinates for the floating and
reference volumes, respectively. The matrix A represents the transformation between
two volumes, as a function of the registration parameters. A
FR
will represent the
transformation between the floating and reference volume, in the image coordinate
system of each volume.
The below formation is a simple derivation based on eqs. 6, 7, 8, and 14.
Essentially, by taking the transformation between world coordinates, given by eq. 14, and
the relationships between image and world coordinates given by eqs. 6 through 8, eq. 15
below can be written.
( )
F FR R
F F w i R w i R
F F w i R R w i
F R
w T l
w T A T l
w T A l T
w A w
=
=
=
=
, ,
1
, ,
, , , ,
(15)
Therefore, using the results given in eq. 15, computation of the matrix T
FR
allows
only coordinates in the image to be dealt with explicitly.
30
D. Volume Interpolation
With the application of T
FR
to convert between image coordinates in the floating
volume to the reference volume, it is highly unlikely that the coordinate l
R
will fall on a
sampled point, and therefore, volume interpolation will be necessary to approximate the
value of the volume at the coordinate l
R
.
Here, trilinear interpolation is used, which is an extension of linear interpolation
in one dimension, and bilinear interpolation in two dimensions, to the three-dimensional
structure of a volume. An approximation of the value of the volume at the coordinate l
R
is given by a linear weighting of the surrounding eight voxels, that form the smallest
bounding cube around l
R
in the lattice.
Trilinear interpolation uses the quantities given in Table I below. As a reference,
the sample closest to the origin in the image coordinate system in the surrounding cube of
eight voxels is selected as a reference. The image coordinates of this voxel are denoted
by (lower_left_N1, lower_left_N2, lower_left_N3). The value of the volume at a
coordinate sample
i
is denoted by V(sample
i
). The interpolated value in the volume V,
denoted by interpolated_value, is given by the sum of the product of weight
i
and
V(sample
i
), for i from 0 to 7, indexing the eight surrounding voxels.
31
TABLE I
SAMPLES AND SAMPLE WEIGHTINGS FOR TRILINEAR INTERPOLATION.
3 _ _
2 _ _
1 _ _
N left lower z delta
N left lower y delta
N left lower x delta
Z
Y
X
=
=
=
Z Y X
Z Y X
Z Y X
Z Y X
Z Y X
Z Y X
Z Y X
Z Y X
delta delta delta weight
delta delta delta weight
delta delta delta weight
delta delta delta weight
delta delta delta weight
delta delta delta weigth
delta delta delta weight
delta delta delta weight
=
=
=
=
=
=
=
=
7
6
5
4
3
2
1
0
) 1 (
) 1 (
) 1 ( ) 1 (
) 1 (
) 1 ( ) 1 (
) 1 ( ) 1 (
) 1 ( ) 1 ( ) 1 (
) 1 3 _ _ , 1 2 _ _ , 1 1 _ _ (
) 1 3 _ _ , 1 2 _ _ , 1 _ _ (
) 1 3 _ _ , 2 _ _ , 1 1 _ _ (
) 1 3 _ _ , 2 _ _ , 1 _ _ (
) 3 _ _ , 1 2 _ _ , 1 1 _ _ (
) 3 _ _ , 1 2 _ _ , 1 _ _ (
) 3 _ _ , 2 _ _ , 1 1 _ _ (
) 3 _ _ , 2 _ _ , 1 _ _ (
7
6
5
4
3
2
1
0
+ + + =
+ + =
+ + =
+ =
+ + =
+ =
+ =
=
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
N left lower N left lower N left lower sample
interpolated_value =
=
7
0
) (
i
i i
sample V weight
E. Overview of Registration Approaches
Computer-assisted registration has received a large amount of attention in
medical imaging, owing to its importance, and the strong desire for an automated, non-
invasive approach. Consequently, there are a number of different paradigms that have
32
been introduced for 3D volume registration, including landmark-based methods,
segmentation-based methods, surface-based methods, and volumetric methods [18, 19].
In this work, registration is pursued using a technique based on volumetric
registration. The choice of this approach allows for the features used for registration to
be the samples comprising the volumes to be registered directly, with no pre-processing
or feature extraction applied. This technique, known as volume registration by
maximization of mutual information, relies on the evaluation of a metric function to
quantify the quality of alignment, of the floating and reference volumes, given a
registration parameter vector. The metric function used is the mutual information
function of the floating and reference volumes.
F. Registration Metric/Criteria
Relative entropy, also known as the Kullbak Leibler distance, between two
probability mass functions p(x) and q(x), is defined as the quantity D(p||q), given in eq.
16 below [21].
=
X
x q
x p
x p q p D
) (
) (
log ) ( ) || (
2
(16)
33
For two discrete random variables X and Y, with marginal probability mass
functions p
X
(x) and p
Y
(y), and a joint distribution p(x, y), the mutual information
function I(x,y) is the relative entropy between the joint distribution p(x, y), and the
distribution p
X
(x)p
Y
(y), the joint distribution when X and Y are independent random
variables. Thus, the mutual information of X and Y is given in eq. 17 below [17].
= =
Y X Y X
Y X
y p x p
y x p
y x p y p x p y x p D Y X I
,
2
) ( ) (
) , (
log ) , ( )) ( ) ( || ) , ( ( ) , (
(17)
If X and Y are independent random variables, then p(x, y) is given by the product
of the marginals p
X
(x) and p
Y
(y), and therefore, the quantity that is the argument of the
logarithm function in eq. 17 is one. As the logarithm of one is zero, the mutual
information I(X, Y), when X and Y are statistically independent, is zero.
There is a close relationship between entropy, a measure of information content,
and the mutual information quantity. The entropy H(X) of a discrete random variable X
is defined in eq. 18. The joint entropy H(X,Y) of the discrete random variables X and Y
is defined below in eq. 19. The conditional entropy H(Y|X) of the discrete random
variables X and Y is defined below in eq. 20 [21].
34
=
X
x p x p X H ) ( log ) ( ) (
2
(18)
=
Y X
y x p y x p Y X H
,
2
) , ( log ) , ( ) , (
(19)
=
Y X
x y p y x p X Y H
,
2
) | ( log ) , ( ) | (
(20)
Entropy is a common measurement of information content. Information content
is increased as entropy is increased, and decreased as entropy is decreased. The less
concentrated a probability density or mass function is, the more information content that
is encoded in the random variable. For example, consider a continuous random variable
X. If X is distributed as a uniform random variable, the information content of X is
greatest, because over the range of X, the probability of X taking on a value x
1
is equal in
all cases to X taking on a value of x
2
. For X distributed as a Gaussian random variable,
with a mean and variance
2
, X encodes less information, as the values of X around
are more likely than values far from , with the spread or concentration of probability
around quantified by the variance
2
.
By simple algebraic manipulation, and using the definition of conditional
probabilities, the mutual information function I(X,Y) can be related in a number of ways
to the entropy quantities given in eqs. 18 through 20. These relationships are given
below in eqs. 21 and 22.
35
) | ( ) ( ) | ( ) ( ) , ( X Y H Y H Y X H X H Y X I = =
(21)
) , ( ) ( ) ( ) , ( Y X H Y H X H Y X I + = (22)
For volumetric registration, there will be two volumes, a floating volume, and a
reference volume. The floating volume will be transformed by a transformation T
, a
transformation with registration parameters . Using the intensities of the volumes as the
features to use for registration, the random variable R shall denote the intensities
observed in the reference volume, the random variable F shall denote the intensities
observed in the floating volume, and the random variable T
F), from the definition of mutual information in eq. 17, and the
relationships of mutual information to entropy in eqs. 21 and 22, can be written as below,
in eqs. 23 through 25, respectively.
) ( ) (
) , (
log ) , ( ) , (
2
,
f T p r p
f T r p
f T r p F T R I
f T R F T R
=
(23)
) | ( ) ( ) , ( F T R H R H F T R I
=
(24)
36
) , ( ) ( ) ( ) , ( F T R H F T H R H F T R I
+ = (25)
In the process of registration, only the overlapping portions of the reference and
transformed floating volumes will be considered in computation of the mutual
information metric, as will be discussed below. Therefore, the quantities H(R) and
H(T
F) is equivalent to
minimization of H(R|T
F), given T
F, the
information content of R is to be minimized. Qualitatively, if T
F) has established a
relationship between the observations of R, and the observations of T
F, regardless of the
mathematical nature of the relationship. This property allows the mutual information
metric to be useful for multimodal registration, where other similarity metrics perform
poorly [22].
Considering the relationship between the mutual information metric and entropy
given in eq. 25, it can be observed that maximization of I(R, T
F) is equivalent to
minimization of H(R, T
f), this
corresponds to building concentrations of probability, as observed in a comparison
information content of a random variable distributed as a Gaussian to a random variable
distributed uniformly. This process allows for registration by favoring registration
37
parameters that establish a relationship between the reference and floating volumes, with
this relationship being the spatial alignment of anatomy, implicitly.
G. Computation of the Mutual Information Metric
Computation of the mutual information metric is based upon direct evaluation of
eq. 23. Furthermore, this computation is improved in terms of execution speed on the
Onyx2 supercomputer by dividing the task of computation of I(R, T
F) across several
processors. Computation of the metric, and parallelization of this computation is
discussed below.
Beginning with computation of the metric, from eq. 23, three quantities are
necessary: p(r, T
f), p
R
(r), and p
TF
(T
f) may be
obtained directly from the joint probability function p(r, T
f). Here,
normalization refers to scaling of the histogram, such that the sum of approximated
probabilities equals 1.0. The marginals are then approximated from h(r, T
f) by
summation over the rows of h(r, T
is applied, to arrive at a
coordinate set in the image coordinate system of the reference volume. If the
transformed coordinate is outside the measured reference volume, then the remaining
operations are not executed, and the process starts again with the next sample in the
floating volume. Otherwise, a sample in the reference volume at the transformed
38
coordinates is approximated using trilinear interpolation, and discretized. The two
samples, one from the floating volume, and one from the reference volume, are then
binned in the joint histogram.
In the studies considered here, each sample obtained from the MRI is an 8-bit
sample, allowing for 256 discrete levels. In the computation of the joint histogram
h(r,T
f), each discrete level is utilized in the joint and marginal histograms. Therefore,
there are 256 x 256 = 65,536 bins in the joint histogram, and 256 each in the marginal
histograms.
Computation of the joint histogram involves the processing of each sample in the
floating volume, application of a transformation to the coordinate of the sample in the
floating volume to obtain a coordinate in the reference volume, interpolation in the
reference volume, and binning in the joint histogram. For a typical 256 x 256 x 20 MRI
volume, there are thus 256 x 256 x 20 = 1,310,720 samples to process.
Following computation of the joint histogram, normalization and computation of
the marginal histograms must be performed. This involves one pass over the joint
histogram, therefore processing 256 x 256 = 65,536 elements from the joint histogram.
This processing consists of normalization, and summation to compute the marginal
histograms.
Following this operation, the mutual information metric itself may be computed.
This processing involves computation of the sum given eq. 23, and involves one pass
over the joint histogram, therefore processing 256 x 256 = 65,536 elements that compose
the sum given in eq. 23.
39
Therefore, computation of the joint histogram is by far the most computationally
costly component in computation of the mutual information metric. Therefore,
performance may be best increased by decreasing the execution time of the computation
of the joint histogram. Computation of the joint histogram is an amenable problem for
parallel execution, as computation of a part of the joint histogram does not depend on the
computational results of any other part of the joint histogram, allowing individual bins, or
entire regions of the joint histogram to be computed independently, and then merged to
form the total joint histogram.
The architecture of the SGI Onyx2 supercomputer allows for tasks running on
different processors to access memory anywhere within the 20 Gb of shared memory in
the machine. Reads from an arbitrary location in memory, then, are unfettered. Writes to
memory, however, should be carefully planned, so as to avoid potential problems with
performance decreases, due to cache refreshes. Therefore, a desired parallel solution
should permit reading of common input data from only one location (to avoid
unnecessary duplication and other overhead), but also keep outputs of the computation
local to a process.
For implementation of parallel execution, POSIX pthreads are used. The
implementation on the SGI Onyx2 allows for full utilization of the shared memory
capabilities of the machine. Furthermore, the use of POSIX pthreads allows for the
developed application to be easily ported to other architectures and operating systems,
such as a personal computer (PC) running a variant of the Linux operating system.
40
A solution to this problem, fitting within the above constraints, is a division of the
computation of the joint histogram into two tasks. The first task computes a number of
joint histograms over sub-volumes. The second task merges these sub-joint histograms to
form the total joint histogram. Below, figure 10 illustrates the architecture of the
solution.
41
F
I
G
U
R
E
1
0
:
I
l
l
u
s
t
r
a
t
i
o
n
o
f
t
h
e
p
a
r
a
l
l
e
l
c
o
m
p
u
t
a
t
i
o
n
o
f
t
h
e
j
o
i
n
t
h
i
s
t
o
g
r
a
m
n
e
c
e
s
s
a
r
y
f
o
r
c
o
m
p
u
t
a
t
i
o
n
o
f
t
h
e
m
u
t
u
a
l
i
n
f
o
r
m
a
t
i
o
n
m
e
t
r
i
c
.
42
Given a division of the floating volume into a set of sub-volumes, each process or
thread in task 1 computes a joint histogram over a sub-volume. Subsequent tasks cannot
proceed until all of the task 1 threads are finished in their execution, therefore, load
balancing between threads is accomplished by parameterizing the number of slices from
the floating volume to use per thread, and then by dynamic computation of the number of
threads that are necessary to execute. Each sub-joint histogram is computed as the joint
histogram is computed, as discussed above. The outputs of task 1 threads are a sub-joint
histogram corresponding to each sub-volume.
Given a set of sub-joint histograms, threads in task 2 compute the total joint
histogram over a region of the joint histogram. Each thread in task 2 has an assigned
region over which to compute the joint histogram. Given an element in the total joint
histogram to compute, the quantity is computed by summation over all sub-joint
histograms of the corresponding element. Load balancing between threads is
accomplished by dividing the total joint histogram into regions of equal numbers of
elements, and assigning a thread to compute a single region. The output of task 2 threads
is the total joint histogram.
Therefore, given a set of registration parameters, the mutual information metric
I(R,T
\
|
|
.
|
\
|
=
2
| |
exp ) (
i
i c
(27)
The diffusion constant controls the effect of the filter. The value of 8.0 was
used in the studies here for image enhancement. This value was chosen experimentally,
to give the best results visually for a study. It was observed that a constant value of this
parameter produced consistent results across different studies.
71
The filter operates iteratively, and operation is by selection of a number of
iterations to perform. Increasing the number of iterations results in improved image
enhancement, but results in longer computation times. It was observed experimentally
that 2,000 iterations provides an acceptable tradeoff between these conflicting factors,
and this parameter was used in the experiments presented here.
The filter is applied in an iterative fashion, across each pixel of the input image.
Computation at a time t, for a given pixel, depends only on the present value of the
image. Therefore, the filtering process may be easily parallelized for improved execution
time. Owing to the simple structure of the filter, this is accomplished on the SGI Onyx2
supercomputer using the MIPS compiler to automatically parallelize the execution of the
filter.
Figure 20 below gives two input FLAIR slices from two different MS patients.
The above filter was applied, prior to manual segmentation of the brain from the head.
The results of application of this filter to these input images are also given in figure 20.
72
(a)
(b)
(c)
(d)
FIGURE 20: Non-linear anisotropic diffusion filtering of MRI slices for segmentation.
Input slices (a and c) are FLAIR weighted imagery. The corresponding output slices (b
and d, respectively) show much enhanced images, but with the loss of some MS lesions.
Unfortunately, while promising, the results introduce several difficulties. First of
all, lesions that do not have a strong edge on all sides are smoothed into brain material,
which leads to large classification errors when subsequent labeling is applied. Secondly,
large numbers of iterations of the filter are required before regions of similar tissue have
nearly the same intensity. Even at this point, some type of thresholding is necessary, and
incurs similar difficulties as discussed above. Lastly, the resulting images do not entirely
contain the idealized three similar intensities. This is best illustrated by the input slice in
figure 20 (a), and the output slice in figure 20 (b), where intermediate intensity levels are
73
obtained, corresponding to blurred regions and scanning artifacts in the input slice.
Figure 21 below illustrates some of the types of errors encountered.
FIGURE 21: Illustration of types of errors encountered with segmentation via image
enhancement.
E. Segmentation by Unsupervised Clustering
Clustering methods are algorithms that operate on an input data set, grouping data
into clusters based on similarity of the data in like clusters. As such, clustering
algorithms are unsupervised classifiers, assigning states of nature without regard to any a
priori labeled training samples. Clustering algorithms are also useful for data
exploration, allowing a user to discover patterns or similarities in a data set [33].
74
Explored in this research are clustering techniques for unsupervised classification.
A well-known clustering algorithm, k-means clustering, was implemented, and tested on
sample slices of FLAIR weighted MRI slices from MS patients.
The k-means algorithm accepts as input the number of clusters to organize data
within, initial location of cluster centers, and a data set to cluster.
The number of clusters that the algorithm fits the data to, denoted by c, is
specified to the algorithm, and represents a parameter the user desires to experiment with,
or as a parameter expressing the expected or desired number of classes to discern from
the data. With the k-means algorithm proper, there are no conditions upon which c is
decreased or increased. Additionally, there are no conditions upon which data is
excluded or included in consideration of fit to classes; all data provided as input to the
algorithm is classified. A given piece of data, or feature measurement, is assigned
exclusively to one class (as compared to fuzzy k-means clustering, where a degree of
membership is assigned to each data item, for each class).
Initial cluster centers are provided here as random locations in feature space.
There is a cluster center
i
for each cluster, such that i = 1c. Each cluster center is a d-
component column vector, where d is the number of input features in the data set. For
the case of a single feature taken to be the intensity of the FLAIR study, d = 1.
The data set provided to the k-means algorithm consists of N items, with each
item a d-component column vector, denoted as x
i
, i = 1N.
75
The k-means algorithm is an iterative algorithm, assigning a class at each iteration
to each data element x
i
. Iterations of the algorithm cease when there is no change in the
classification solution during an iteration.
Each iteration consists of classifying the data set by comparison of the data set to
the current cluster centers. A data item is assigned to the same class as a cluster center
i
if the Euclidean distance between the data item and
i
is the least distance between the
data item and all cluster centers
i
, i = 1...c. Following class assignment, cluster centers
are updated, by computation of the centroid of the data set classified as the same class.
The k-means algorithm was implemented, and tested with a sample MRI slice of a
patient with MS. It was generally found that the k-means clustering approach found
structure in the data, however, it was not structure that is of interest in the classification
problem of normal brain tissue, diseased brain tissue, and CSF. It was initially expected
that for a slice containing MS lesions, that three clusters may be appropriate, in
consideration of the desired classification into three tissue classes. It was observed,
however, that best results were obtained when a larger number of clusters were permitted,
followed by manual re-labeling of clusters of like tissue classes.
Figure 22 below illustrates results of application of the k-means clustering
algorithm, using five classes. The input image in (a) is a FLAIR image. The k-means
algorithm is executed, using five clusters, to obtain the output, classified slice in (b). (c)
shows the results after manual re-labeling of class, such that tissues of the same type,
split into two or more clusters, are re-labeled to a single class.
76
(a)
(b)
(c)
FIGURE 22: Results of k-means clustering, using 5 clusters. The input slice is shown in
(a). The resulting clustered output image is shown in (b). By manual re-labeling of
clusters such that clusters of similar tissue belong to the same cluster, the output slice in
(c) is obtained.
Due to the observation that the number of clusters used is not necessarily related
to the number of classes desired in the study, and thus, the need for manual intervention,
classification via the use of a clustering algorithm such as k-means is not found to be an
acceptable means of segmentation of lesions from brain tissue. It should also be noted
that it is very likely that even though each patient will be known to have MS, individual
slices from the brain, taken individually, may not show evidence of MS. Therefore, the
number of clusters used would likely need to be adjusted based on if the slice contains
MS lesions, which requires further involvement by an expert.
It is observed, however, that a clustering algorithm, such as k-means, in
combination with the nonlinear anisotropic diffusion filter, may prove to be useful for
generation of initial training data for more complex classification paradigms.
77
F. Bayesian Classification
The above approaches, while generating tantalizing results, all fail with regard to
accuracy, ease of application, and logical relation to the classification problem. A
statistical classification approach addressees these concerns, providing an optimal
solution to the classification problem when tissue models are known or well
approximated. Additionally, statistical classification is well founded in probability
theory, and has enjoyed success in a number of applications, including classification of
remote sensing data, scene classification, and classification of medical imagery.
Statistical classification is fundamentally based upon the application of Bayes rule
for quantifying a decision procedure. Bayes rule, given in eq. 28 below, specifies how to
compute a posterior probability of a class, given feature measurements [33].
) (
) ( ) | (
) | (
x p
P x p
x P
j j
j
=
(28)
In eq. 28,
j
denotes a class, or state of nature. The vector x denotes a feature
vector, obtained from measurements of quantities of interest. For image analysis, the
feature vector x is composed for each pixel, and consists of the features of the pixel. For
the approach here, the feature x therefore shall consist of one component, the intensity of
the pixel from the FLAIR image. The quantity p(x|
j
) is known as the class conditional
78
probability density function for class
j
. This distribution quantifies the probability of
measurement of a feature, given a state of nature. The quantity P(
j
) is known as the a
priori probability, and quantifies the probability of observing a state of nature, regardless
of any feature measurement. The quantity p(x) is known as the evidence, and serves only
as a scale factor, such that the quantity in eq. 28 is indeed a true probability, with values
between zero and one. The quantity P(
j
|x) is known as the posterior probability, and
quantifies the probability of observing class
j
, given that the feature x was measured.
The quantity p(x) in eq. 28 can normally be ignored for implementing decision
theory, as it is only a constant for establishing a probability between zero and one, and
constant for all classes. Therefore, the quantity given below in eq. 29, known as the
maximum a posteriori (MAP) estimate of eq. 28, is used.
) ( ) | ( ) | (
j j j
P x p x P = (29)
Bayes decision rule, based upon the application of eq. 28 or 29, gives a
framework for how to make a sound decision. For c possible classes, Bayes decision rule
mandates the selection of class
j
for a feature x, if the a posteriori quantities in eq. 28 or
29 are maximum for class
j
, compared to
i
, i = 1c, i j. This rule is stated below in
eq. 30 [33].
79
Decide
j
if P(
j
| x) > P(
i
| x) for i = 1c, i j (30)
Bayes decision rule is optimal in the sense of minimization of the probability of
error [33].
Despite the optimality of the Bayesian classifier, the decision rule is optimal only
if the probability models for the class conditional distributions, and the a priori
probabilities are known. In the context of classification of brain tissue, the probability
models are not known, and therefore, must be approximated. The performance of the
Bayesian classifier is directly related to how well these distributions can be modeled.
In the development below, two Bayesian-based approaches are considered. The
first assumes Gaussian models for the class conditional probability distributions, and
equal a priori probabilities. The results obtained from this approach demonstrate the
need for additional complexity to be introduced in the modeling of the class conditional
and a posteriori probability models. Subsequently, the class conditional probability
models are improved using Parzen windowing, a nonparametric density estimation
technique. Additionally, a priori probability models are improved by approximation of
the imaged volume using a Markov random field model.
80
G. Bayesian Classification: Gaussian Conditionals, Equal a priori Probabilities
As an initial approach, class conditional probabilities are modeled using a
Gaussian distribution, and in the absence of any other information, equal a posteriori
probabilities are assumed [34].
The general form of a Gaussian probability distribution is given below in eq. 31.
In the context of the present problem, a single feature is used, being the intensity from the
FLAIR imagery. The form of a univariate distribution is given below in eq. 32.
In eq. 31, d is the dimension of the distribution, represents the d x d covariance
matrix, x is a d-component column vector, and is a d-component mean vector. In eq.
32, is the standard deviation, x is a scalar, and is the mean value of the distribution.
|
.
|
\
|
=
) ( ) (
2
1
exp
) 2 (
1
) | (
1
2
1
2
x x x p
t
d
j
(31)
|
|
.
|
\
|
|
.
|
\
|
=
2
2
1
exp
2
1
) | (
x
x p
j
(32)
As no information is known otherwise, the a posteriori probabilities for each class
are assumed to be equal. Therefore, for three classes, the probability P(
i
) is equal to 1/3
for each class. These probabilities are equal for each class, and therefore, do not affect
the result of the computation of eq. 28 or 29, and therefore, for the case considered here,
81
the posterior quantity upon which a decision is made is given completely by the result of
the computation of eq. 32.
From manually selected training samples, the mean and the standard deviation
are estimated using the maximum likelihood estimators (MLEs) for a univariate Gaussian
probability distribution function, given below in eqs. 33 and 34, respectively [35]. The
MLE estimates of the mean and standard deviation are and , respectively. The
quantity n represents the number of training samples available, and x
k
is the k
th
training
sample.
=
=
n
k
k
x
n
1
'
1
(33)
( )
=
=
n
k
k
x
n
1
2 '
'
2
) (
1
(34)
Below in figure 23, a sample result of the application of this scheme is presented.
In terms of performance, this classifier performs poorly. Upon close evaluation,
however, it is evident that the classifier works decently when the tissue is completely
unambiguous, and not in close proximity to boundaries with other tissues. For example,
the classification of normal brain material works well when not in close proximity to MS
lesions, and vice-versa. Based on the sample slice presented below in figure 23, as well
82
as other results obtained, it was observed that the classifier finds the lesions. Too much
of the surrounding tissue, however, is erroneously classified as diseased tissue.
(a)
(b)
FIGURE 23: Sample result of tissue classification in the brain using a simple Bayesian
classifier, with equal a priori probabilities, and Gaussian class conditional probabilities.
H. Bayesian Classification: Nonparametric Conditionals, a priori Probabilities Modeled
by Markov Random Fields
The above classification results leave much to be desired. Upon inspection of the
MAP estimator that is the basis of the Bayesian classifier, it is apparent that broadly, the
probability models for the class conditional and a priori distributions must be improved.
It is expected that improvement in these models will result in subsequent improvement in
the performance of the classifier.
In the above classifier, the class conditional probability model was assumed to
have a known parametric form. In this approach, the form of the distribution is known,
leaving only the parameters of the particular distribution to estimate. In the above
classifier, the parametric form assumed was that of a Gaussian. The parameters of the
83
Gaussian distribution, the mean and the standard deviation, were estimated using
maximum likelihood estimators.
A more general approach involves the use of nonparametric techniques for
density estimation [33]. Such approaches allow for direct estimation of the probability
distribution, eliminating the necessity of choosing a parametric form of the class
conditional distribution. Assumption of a parametric form is acceptable, if it is known
that the true distribution is Gaussian, or if a Gaussian is a good approximation to the true
class conditional distribution.
Use of a nonparametric technique, however, removes the need for the assumption
and estimation of the parameters of the assumed parametric model. Instead, the
probability distribution is estimated directly from the training data.
The nonparametric technique known as Parzen windowing is used here to
improve the model of the class conditional probabilities [33]. For each tissue class,
training samples are used to form an estimated probability distribution. The estimated
probability distribution is given by evaluation of eq. 35 below, where n is the number of
training samples, and (x) is a Parzen window function, taken to be a Gaussian kernel,
with a mean of zero, and a variance of one, here.
( )
=
=
n
i
i j
x x
n
x p
1
1
) | (
(35)
84
Estimation using the Parzen window estimate of eq. 35 can be made
computationally efficient by the generation of look-up tables for p(x|
j
). The use of
look-up tables avoids the computation of a sum, of a large number of training samples (n
of them), at each evaluation for some x and
j
.
The a priori probabilities P(
j
) were assumed in the simple Bayesian classifier,
above, to be equal for each tissue class. This assumption, however, is not well justified.
For example, in the input slice in figure 21 (a), it is very apparent that the probability of
normal brain material exceeds that of abnormal brain material. In the improvement of the
simple Bayesian classifier, then, it is also necessary to consider a more appropriate model
for the a priori probabilities.
It is reasonable to consider an a priori probability on a local scale, and consider
the spatial nature of the feature measurements obtained from MRI studies of the brain. A
feature measurement has a spatial location within the brain. If this feature measurement
is known to have a state of nature of
i
, then it is reasonable to assume that the feature
measurements surrounding this known voxel have an increased probability of also having
a state of nature of
i
. This assumption is reasonable, as there is order to anatomy
imaged, with abnormal brain material forming in regions, for example. The opposite of
this observation would be to take a slice, such as shown in figure 21 (a), and randomly
swap voxels until there is no region homogeneity in the image.
Markov random field (MRF) models [36, 37] provide such a modeling of
neighborhood influences that can be utilized to form an improved probability model for
the a priori probability P(
j
).
85
A Markov random field models an image or volume as a random field, a
structured collection of random variables, F = {F
1
, , F
m
}, and defined on the set S.
Here, each random variable is a lattice point in the sampled volume. A Markov random
field is a special case of a random field. There are two conditions imposed to allow a
random field to be labeled as a Markov random field [36]:
1. F f f p > , 0 ) ( , known as the positivity condition
2. ) | ( ) | (
} {
i
N i i S i
f f p f f p =
\
|
= ) (
1
exp
1
) ( f U
T Z
f p
(36)
|
.
|
\
|
=
F
f U
T
Z ) (
1
exp
(37)
=
C
c
f V f U ) ( ) (
(38)
A result known as the Hammersley-Clifford theorem states that if and only if a
random field F on S is a Markov random field with respect to a neighborhood system N,
then F is a Gibbs random field on S with respect to a neighborhood system N [36]. This
result allows the conditional probability given as the Markovianity condition of an MRF
to be converted to the non-conditional probability of a Gibbs distribution, given in eq. 36.
This allows for computation of the a priori class probabilities modeled as an MRF, and
therefore, incorporation of this model into a practical classifier.
In the classifier constructed here, a simple model for the energy function U(f) is
utilized [38]. This model consists of a linear combination of products of elements in the
cliques. For experimentation, a first order neighborhood system for image segmentation
is considered. The resulting model for U(f) is given below in eq. 39.
88
( ) ( ) ( )
1 , 1 , 2 , 1 , 1 1 , ,
) ( ) (
+ +
+ + + + = =
l k l k l k l k l k l k
f f f f f f U f U (39)
The quantity f
i,j
in eq. 39 refers to the sample at indices (i,j) in the labeled image.
The parameters ,
1
, and
2
are parameters that allow for adjusting relative weights or
contributions of neighborhood interactions. In the experiments here, is taken to be 1.0,
and
1
and
2
are both taken to be 0.75. These weightings allow the current classification
at index (i, j) to take an importance somewhat greater than the neighborhood
classifications, and gives equal weighting to the neighbors of the feature measured at (i,
j). Also, it should be noted that the temperature parameter T is taken to be 1.0.
The implemented algorithm accepts a slice of data, and training points. An initial
classification solution is obtained by assumption of equal a priori class probabilities.
Subsequent iterations implement modeling of a priori class probabilities using the MRF
model discussed above. Each iteration classifies the entire image, using the previous
classification as a basis to model neighborhood interactions. The iterative process
terminates when no changes are made in the classification. Class conditional
probabilities are modeled throughout from the training points initially supplied.
A sample of the results obtained is given below in figure 25. The classifier
successfully detects the lesions, and in general, is observed to perform well. Additional
comments on the behavior of the classifier are given below. In terms of computational
89
performance, each segmentation executes within seconds on the SGI Onyx2
supercomputer, as a non-threaded application.
(a)
(b)
(c)
(d)
FIGURE 25: Sample classification results obtained using a Bayesian classifier, with
class conditional probabilities modeled nonparametrically, and class a priori
probabilities obtained from a Markov random field model of the image. Input slices are
shown in (a) and (c), and the classified images are shown in (b) and (d). In (b) and (d),
normal brain material is indicated by the color yellow, CSF by gray, and lesions as red.
I. Quantification
With the classification results obtained, quantification of disease burden as judged
from MRI may be obtained by computing the total volume of lesions, and the total
volume of normal brain material. This can be simply accomplished by having the
90
computer count the number of labels assigned for each class, and then multiplying this
number by the volume of each voxel, which is known a priori from the scanner. Results
of this process are given in Table IV, below, for the output slices given in figure 25.
TABLE IV
QUANTIFICATION OF MS DISEASE BURDEN FROM THE SAMPLE RESULTS
GIVEN IN FIGURE 25.
Input
Slice
Number
of
normal
brain
tissue
voxels
Number
of
abnormal
brain
tissue
voxels
Volume
of normal
brain
tissue
(mm
3
)
Volume
of
abnormal
brain
tissue
(mm
3
)
Percentage
of normal
brain tissue
volume to
total brain
volume
Percentage
of
abnormal
brain tissue
volume to
total brain
volume
Figure.
23 (a)
12222 1,463 53709.960 6429.199 89.3 % 10.7 %
Figure
23 (c)
13920 1,053 61171.875 4627.441 93.0 % 7.0 %
J. Accuracy
The results obtained subjectively show good performance. Lesions are detected,
and highlighted. The classifier erroneously detects lesions around the outer perimeter of
the brain, mistaking scanning artifacts with MS lesions.
Results were shown to Dr. Robert Falk, a radiologist at Jewish Hospital. Dr.
Falks comments on the results obtained were positive.
91
K. Summary
Pattern recognition approaches were explored for classification of the tissues
present in the brain. Several approaches were attempted, including image processing
approaches using thresholding and image enhancement, unsupervised classification via
the use of the k-means algorithm, and statistical classification utilizing Bayes rule, with
Gaussian probability models for class conditional probability distributions.
Improvement of the statistical classifier by nonparametric modeling of the class
conditional probability distributions using Parzen windowing, and modeling of the a
priori class probabilities via the use of a Markov random field image model, provided the
best results obtained. Quantification of MS may then be implemented by computing the
total volume of normal and diseased brain material, from the resulting states of nature
assigned by the classifier and the voxel sizes used in the acquisition of the study.
92
V. VISUALIZATION
A. Introduction
Several scientific visualization techniques are applied here for the study of the
results of the registration and segmentation processes implemented. For the evaluation of
the volume registration results, a simple data fusion technique is applied, which overlaps
regions of the floating and re-sampled reference volumes. As well, a visualization
application is presented which allows for volume rendering, and re-slicing volumes along
arbitrary planes, allowing for more thorough evaluation of registration accuracy. For the
evaluation of segmentation results, a simple morphing or frame interpolation technique
for transitioning from the raw input image to the classified, colored output image is
explored. Finally, a web-based approach for presentation of results is considered.
Scientific visualization is of interest in this problem, due to the complexity of the
data sets involved, and the desire to facilitate discovery with regard to the pathology of
MS. Utilization of visualization procedures allows the raw data and processed results to
be manipulated on a higher level than otherwise available.
B. Visualization for the Study of Volume Registration
Following the process of volume registration using the software tool developed
for this research, one of the outputs available is a re-sampled version of the reference
volume, sampled at the locations corresponding to the sampling locations in the floating
93
volume. With the re-sampled reference volume, the reference and floating volumes may
then be compared on a slice-by-slice basis, as corresponding anatomy is registered in 3D
space.
As an aid to judging the quality of registration, a simple data fusion technique was
implemented for visualization. This technique takes regions from the re-sampled
reference volume and the floating volume to form a patchwork, composite image that
contains components of the imaged anatomy in both the reference and floating volumes.
This composite image, referred to as a checkerboard image, facilitates judgment of the
accuracy of the volume registration by allowing contours (such as the skull, or features
within the brain) to be followed between volumes, directly within one image. The
regions used here are 32 x 32 square pixel regions. Figure 26 below illustrates the
generation of the checkerboard images from the floating and re-sampled reference
volumes. Figure 27 below shows sample results obtained, from several registration
studies.
FIGURE 26: Generation of the checkerboard image, by formation of a composite image
of square pixel regions from the re-sampled reference and floating volumes.
94
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
FIGURE 27: Sample slices from different registration studies, showing the a slice from
the floating volume ((a), (d), and (g)), corresponding slices from the re-sampled,
registered reference volume ((b), (e),and (h), respectively), and the checkerboard slices
((c), (f), and (i), respectively).
An additional approach to the comparison of registration results was made with
the utilization of volume rendering, and arbitrary volume re-slicing. Figure 28, below,
shows an example of this application. The graphics components of this application were
95
developed using the Visualization Toolkit [43, 44], and implemented on the SGI Onyx2
supercomputer.
FIGURE 28: Volume rendering and arbitrary volume re-slicing for comparison of two
volumes that are registered via maximization of mutual information.
96
C. Visualization for the Study of Volume Segmentation
A very simple technique is considered for comparison of one volume to another,
on a slice-by-slice basis. This technique takes two inputs, an initial image, and a final
image. Between these two images, intermediate images are generated that represent an
interpolation between the initial and final images, as illustrated below in figure 29. This
process allows for a simple fade or morphing between two volumes, and can be used for
direct comparison of two raw scan volumes, or comparison of a raw scan volume to a
classified and colored volume. The initial, final, and intermediate volumes may be
combined in sequential order, and at a specified frame rate to form a movie that
progresses from the initial image/volume to the final image/volume.
FIGURE 29: Frame interpolation.
97
The interpolation scheme implemented in this research consists of interpolation
on a sample to sample basis. Thus, a sample from the initial image is interpolated until
the sample in the corresponding location in the final image is reached, independent of the
other samples in the image/volume.
Simple linear interpolation from a sample x
0
to x
1
is given in eq. 40, below. The
interpolated value, x(t), is a function of time t, a parameter for ordering the interpolated
images/volumes. The time parameter will range from 0, which generates the initial value
x
0
, to 1, which generates the final value x
1
.
0 0 1
) ( ) ( x t x x t x + = (40)
The above interpolation scheme may be generalized, by a time-dependent kernel
function k(t), other than t. Eq. 41 below generalizes the interpolation given in eq. 40.
0 0 1
) ( ) ( ) ( x t k x x t x + = (41)
Different kernel functions, including exponential, and quadratic functions were
experimented with. Kernels, such as an exponential function, with less abrupt transitions
98
from zero to one were found to produce the most visually appealing results, in
comparison to a linear interpolation scheme such as eq. 40.
For input and/or output images that are color, interpolation for each sample is
performed on the color components of the samples, using a red, green, blue (RGB) color
parameterization.
D. Web-Based Presentation of Results
Finally, web-based presentation of results was implemented, for presentation of
the results on the Internet. Therefore, the results may be viewed and evaluated from
virtually anyplace in the world (anyplace with Internet access available, at least). Figure
30 below shows an example of one of the web pages that was created for this presentation
format.
99
FIGURE 30: A sample of a web page developed for presentation of the results of this
study on the Internet.
E. Summary
Several simple visualization techniques were implemented and utilized for the
judgment of volume registration accuracy, and for facilitation of comparison of one
100
volume to another. These techniques are simple, but useful, and form groundwork for
more involved visualization to be applied for the study of MS.
101
VI. CONCLUSIONS AND RECOMMENDATIONS
The components of the preliminary system for the study of MS using MRI include
registration for volumetric alignment of patient studies over the course of time, brain
segmentation from the head facilitated by the use of registration, segmentation and
classification of brain tissue to the three tissue classes: CSF, normal brain material, and
diseased brain material, and the use of scientific visualization for the comparison of
patient scans and validation of the registration and segmentation processing.
While additional follow-up work is necessary, the developed approaches provide
a solid foundation on which to build future development in the study of MS using MRI,
as well as to future studies in medical imaging.
Based on the research and development implemented in this thesis, the following
recommendations for future work are made:
For better volumetric registration, enhancement of the mutual
information approach to incorporate additional features into the registration
for better spatial alignment [39], improvement in the metric itself [40], and
perhaps incorporation of non-rigid body registration[41]. The results
obtained for registration here are useful, however, there is room for
improvement.
For improved brain segmentation from the head, exploration of
techniques to directly segment the brain, or further development of the
registration approach explored here, to incorporate a high resolution, expert-
102
segmented model of the head, to provide the a priori segmentation. The
results obtained here show that the registration approach can be useful,
however, manual segmentation of the brain for an a priori segmented model
is far from ideal for use in a large scale study.
For improvement in the segmentation/classification of the extracted
brain material into CSF, normal brain tissue, and diseased brain tissue, a
means of correcting for inter-intensity variation in the MRI. The wide
variation in intensity levels encountered from the data available to this
research negates any possibility of application of a statistical classification
technique based upon training data from a slice.
For automation in the segmentation/classification of the extracted
brain material, generation of training data automatically. It is hypothesized
that by use of a filter (such as the non-linear anisotropic diffusion filter
applied in other areas here), an unsupervised classification algorithm (such as
the k-means clustering algorithm studied here), and a rule set, that generation
of training data may be possible. Inter- and intra-slice intensity correction
would assist in this effort, as well.
From an intensity-corrected study, and automated generation of
training data, 3D segmentation of lesions from the brain is possible, allowing
for more complex, 3D visualization tools to be applied for qualitative
discovery.
For multiple scans of a patient, taken at different time points, it may
103
be possible to use previous scans, prior to a given time, to improve in the
future segmentation and classification of brain material into CSF, normal
brain tissue, and diseased brain tissue. This would allow for, if for nothing
else, providing an initial solution for an iterative approach, and possibly
assisting in generating training data.
Improved quantification of error in the registration and segmentation
processing. At the present, the judgment of quality of the registration and
segmentation procedures largely depends upon qualitative evaluation by an
expert; it would be useful to quantify these errors numerically.
104
REFERENCES
[1] J. H. Noseworthy, C. Lucchinetti, M. Rodriguez, and B. G. Weinshenker,
Multiple sclerosis, The New England Journal of Medicine, vol. 343, no. 13, pp.
938-952, Sept. 28, 2000.
[2] National Institutes of Neurological Disorders and Stroke, Multiple Sclerosis:
Hope Through Research, NIH Publication No. 96-75, July 1, 2001.
[3] Z. P. Liang and P. C. Lauterbur, Principles of Magnetic Resonance Imaging: A
Signal Processing Perspective. New York: IEEE Press, 2000.
[4] D. H. Miller, P. S. Albert, F. Barkhof, G. Francis, J. A. Frank, S. Hodgkinson, F.
D. Liblin, D. W. Paty, S. C. Reingold, and J. Simon, Guidelines for the Use of
Magnetic Resonance Techniques in Monitoring the Treatment of Multiple
Sclerosis, Annals of Neurology, vol. 39, no. 1, Jan. 1996.
[5] C. R. G. Guttmann, R. Kikinis, M. C. Anderson, M. Jakab, S. K. Warfield, R. J.
Killiany, H. L. Weiner, F. A. Jolesz, Quantitative Follow-up of Patients with
Multiple Sclerosis: Reproducibility, JMRI, vol. 9, no. 4., pp. 509-518, June
1999.
[6] R. Kikinis, C. R. G. Guttmann, D. Metcalf, W. M. Wells, G. J. Ettinger, H. L.
Weiner, and F. A. Jolesz, Quantitative Follow-up of Patients with Multiple
Sclerosis: Technical Aspects, JMRI, vol. 9, no. 4, pp. 519-530, June 1999.
[7] W. M. Wells, W. E. L. Grimson, R. Kikinis, and F. A. Jolesz, Adaptive
Segmentation of MRI Data, IEEE Transactions on Medical Imaging, vol. 15, no.
4, pp. 429-442, August 1996.
[8] M. Kamber, R. Shinghal, D. L. Collins, G. S. Francis, and A. C. Evans, Model-
Based 3-D Segmentation of Multiple Sclerosis Lesions in Magnetic Resonance
Brain Images, IEEE Transactions on Medical Imaging, vol. 14., no. 3, pp. 442-
453, Sept. 1995.
105
[9] B. Johnston, M. S. Atkins, B. Mackiewich, and M. Anderson, Segmentation of
Multiple Sclerosis Lesions in Intensity Corrected Multispectral MRI, IEEE
Transactions on Medical Imaging, vol. 15, no. 2, pp. 154-169, April 1996.
[10] J. Besag, On the Statistical Analysis of Dirty Pictures, Journal of the Royal
Statistical Society B, vol. 48, no. 3, pp. 259-30, 1986.
[11] J. K. Udupa, L. Wei, S. Samarasekera, Y. Miki, M. A. van Buchem, and R. I.
Grossman, Multiple Sclerosis Lesion Quantifiction Using Fuzzy-Connectedness
Principles, IEEE Transactions on Medical Imaging, vol. 16, no. 5, pp. 598-609,
Oct. 1997.
[12] J. K. Udupa and S. Samarasekera, Fuzzy Connectedness and Object Definition:
Theory, Algorithms, and Applications in Image Segmentation, Graphical Models
and Image Processing, vol. 58, no. 3, pp. 246-261, May 19996.
[13] K. V. Leemput, F. Maes, D. Vandermeulen, A. Colchester, and P. Suetens,
Automated Segmentation of Multiple Sclerosis Lesions by Model Outlier
Detection, IEEE Transactions on Medical Imaging, vol. 20, no. 8, pp. 677-688,
August 2001.
[14] F. Maes, A. Collignon, D. Vandermeulen, G. Marchal, and P. Suetens,
Multimodality Image Registration by Maximization of Mutual Information,
IEEE Transactions on Medical Imaging, vol. 16, no. 2, pp. 187-198, April 1997.
[15] W. M. Wells, P. Viola, H. Atsumi, S. Nakajima, and R. Kikinis, Multi-modal
Volume Registration by Maximization of Mutual Information, Medical Image
Analysis, vol. 1, no. 1, pp. 35-51, 1996.
[16] C. R. Meyer, J. L. Boes, B. Kim, P. H. Bland, K. R. Zasadny, P. V. Kison, K.
Koral, K. A. Frey, and R. L. Wahl, Demonstration of Accuracy and Clinical
Versatility of Mutual Information for Automatic Multimodality Image Fusion
Using Affine and Thin-Plate Spline Warped Geometric Deformations, Medical
Image Analysis, vol. 1., no. 3, pp. 195-206, 1996/7.
106
[17] J. Nolte, The Human Brain: An Introduction to its Functional Anatomy. St.
Louis: Mosby Year Book, 3
rd
edition, 1993.
[18] J. B. A. Maintz and M. A. Viergever, A Survey of Medical Image Registration,
Medical Image Analysis, vol. 2., no. 1, pp. 1-36, 1998.
[19] M. A. Audette, F. P. Ferrie, and T. M. Peters, An Algorithmic Overview of
Surface Registration Techniques for Medical Imaging, Medical Image Analysis,
vol. 4, pp. 201-217, 2000.
[20] G. Wolberg, Digital Image Warping. Los Alamitos: IEEE Computer Society
Press, 1990.
[21] T. M. Cover and J. A. Thomas, Elements of Information Theory. New York:
John Wiley & Sons, 1991.
[22] G. P. Penney, J. Weese, J. A. Little, P. Desmedt, D. L. G. Hill, and D. J. Hawkes,
A Comparison of Similarity Measures for Use in 2D-3D Medical Image
Registration, IEEE Transactions on Medical Imaging, vol. 17, no. 4, pp. 586-
595, August 1998.
[23] W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery, Numerical
Recipes in C: The Art of Scientific Computing. Cambridge University Press,
Second Edition, 1992.
[24] F. Maes, D. Vandermeulen, and P. Suetens, Comparative Evaluation of
Multiresolution Optimization Strategies for Multimodality Image Registration by
Maximization of Mutual Information, Medical Image Analysis, vol. 3, no. 4, pp.
373-386, 1999.
[25] S. L. S. Jacoby, J. S. Kowalik, and J. T. Pizzo, Iterative Methods for Nonlinear
Optimization Problems. Englewood Cliffs: Prentice-Hall, 1972.
107
[26] M. Joliet and B. M. Mazoyer, Three-Dimensional Segmentation and
Interpolation of Magnetic Resonance Brain Images, IEEE Transactions on
Medical Imaging, vol. 12, no. 2, pp. 269-277, June 1993.
[27] M. E. Brummer, R. M. Mersereau, R. L. Eisner, and R. R. J. Lewine, Automatic
Detection of Brain Contours in MRI Data Sets, IEEE Transactions on Medical
Imaging, vol. 12, no. 2, pp. 153-166, June 1993.
[28] T. Kapur, W. E. L. Grimson, W. M. Wells, and R. Kikinis, Segmentation of
Brain Tissue from Magnetic Resonance Images, Medical Image Analysis, vol. 1,
no. 2, pp. 109-127, 1996.
[29] M. S. Atkins and B. T. Mackiewich, Fully Automatic Segmentation of the Brain
in MRI, IEEE Transactions on Medical Imaging, vol. 17, no. 1, pp. 98-107, Feb.
1998.
[30] C. Sites, A. Farag, and T. Moriarty, Visualization Algorithms for Medical
Imaging with Telemedicine and High Performance Computing Applications,
Technical Report, CVIP Laboratory, December 2000.
[31] N. Shareef, D. L. Wang, and R. Yagel, Segmentation of Medical Images Using
LEGION, IEEE Transactions on Medical Imaging, vol. 18, no. 1, pp. 74-91, Jan.
1999.
[32] G. Gerig, O. Kbler, R. Kikinis, and F. A. Jolesz, Nonlinear Anisotropic
Filtering of MRI Data, IEEE Transactions on Medical Imaging, vol. 11, no. 2,
pp. 221-232, June 1992.
[33] R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification. New York: John
Wiley & Sons, 2
nd
edition, 2001.
[34] H. E. Cline, W. E. Lorensen, R. Kikinis, and F. Jolesz, Three-Dimensional
Segmentation of MR Images of the Head Using Probability and Connectivity,
Journal of Computer Assisted Tomography, vol. 14, no. 6, pp. 1037-1045,
November/December 1990.
108
[35] R. V. Hogg and E. A. Tanis, Probability and Statistical Inference. Upper Saddle
River: Prentice-Hall, 6
th
edition, 2001.
[36] S. Z. Li, Markov Random Field Modeling in Image Analysis. Tokyo: Springer-
Verlag, 2001
[37] K. Held, E. R. Kops, B. J. Krause, W. M. Wells, R. Kikinis, and H. W. Mller-
Grtner, Markov Random Field Segmentation of Brain MR Images, IEEE
Transactions on Medical Imaging, vol. 16, no. 6, pp. 878-886, Dec. 1997.
[38] C. W. Therrien, Decision, Estimation, and Classification. New York: John Wiley
& Sons, 1989.
[39] J. P. W. Pluim, J. B. A. Maintz, and M. A. Viergever, Image Registration by
Maximization of Combined Mutual Information and Gradient Information, IEEE
Transactions on Medical Imaging, vol. 19, no. 8, pp. 809-814, August 2000.
[40] C. Studholme, D. L. G. Hill, and D. J. Hawkes, An Overlap Invariant Entropy
Measure of 3D Medical Image Alignment, Pattern Recognition, vol. 32, pp. 71-
86, 1999.
[41] B. Likar and F. Pernus, A Hierarchical Approach to Elastic Registration Based
on Mutual Information, Image and Vision Computing, vol. 19, pp. 33-44, 2001.
[42] American College of Radiology, National Electrical Manufacturers Association,
"Digital Imaging and Communications in Medicine (DICOM): Version 3.0",
Draft Standard, ACR-NEMA Committee, Working Group VI, Washington, DC,
1993.
[43] W. J. Schroeder, L. S. Avila, K. M. Martin, W. A. Hoffman, and C. C. Law, The
VTK Users Guide. Kitware, 2001.
[44] W. Schroeder, K. Martin, and B. Lorensen, The Visualization Toolkit: An Object-
Oriented Approach to 3D Graphics. Upper Saddle River: Prentice-Hall, 2
nd
Edition, 1998.
109
VITA
Jeremy Michael Nett, son of Michael and Kathy Nett, was born on July 9, 1978 in
Louisville, Kentucky. He graduated as class valedictorian from Tates Creek High School
in Lexington, Kentucky, in 1996. In the fall of 1996, he entered the Speed Scientific
School at the University of Louisville. In spring of 2000, he graduated summa cum laude
from the University of Louisville with a Bachelor of Science degree in Electrical
Engineering. During his undergraduate studies, he completed three co-operative
internships with Thomson Consumer Electronics, in Indianapolis, IN. In the summer of
2000, he began his studies towards a Masters of Engineering degree at the University of
Louisville's Computer Vision and Image Processing Lab, in the Electrical and Computer
Engineering Department.