Biomedical Imaging - Reiner Salzer - E-Book

Biomedical Imaging E-Book

Reiner Salzer

0,0
120,99 €

-100%
Sammeln Sie Punkte in unserem Gutscheinprogramm und kaufen Sie E-Books und Hörbücher mit bis zu 100% Rabatt.

Mehr erfahren.
Beschreibung

This book presents and describes imaging technologies that can be used to study chemical processes and structural interactions in dynamic systems, principally in biomedical systems. The imaging technologies, largely biomedical imaging technologies such as MRT, Fluorescence mapping, raman mapping, nanoESCA, and CARS microscopy, have been selected according to their application range and to the chemical information content of their data. These technologies allow for the analysis and evaluation of delicate biological samples, which must not be disturbed during the profess. Ultimately, this may mean fewer animal lab tests and clinical trials.

Sie lesen das E-Book in den Legimi-Apps auf:

Android
iOS
von Legimi
zertifizierten E-Readern

Seitenzahl: 754

Veröffentlichungsjahr: 2012

Bewertungen
0,0
0
0
0
0
0
Mehr Informationen
Mehr Informationen
Legimi prüft nicht, ob Rezensionen von Nutzern stammen, die den betreffenden Titel tatsächlich gekauft oder gelesen/gehört haben. Wir entfernen aber gefälschte Rezensionen.



Table of Contents

Title Page

Copyright

Preface

Contributors

Chapter 1: Evaluation of Spectroscopic Images

1.1 Introduction

1.2 Data Analysis

1.3 Applications

References

Chapter 2: Evaluation of Tomographic Data

2.1 Introduction

2.2 Image Reconstruction

2.3 Image Data Representation: Pixel Size and Image Resolution

2.4 Consequences of Limited Spatial Resolution

2.5 Tomographic Data Evaluation: Tasks

2.6 Summary

References

Chapter 3: X-Ray Imaging

3.1 Basics

3.2 Instrumentation

3.3 Clinical Applications

3.4 Radiation Exposure to Patients and Employees

References

Chapter 4: Computed Tomography

4.1 Basics

4.2 Instrumentation

4.3 Measurement Techniques

4.4 Applications

4.5 Outlook

References

Chapter 5: Magnetic Resonance Technology

5.1 Introduction

5.2 Magnetic Nuclei Spin in a Magnetic Field

5.3 Image Creation

5.4 Image Reconstruction

5.5 Image Resolution

5.6 Noise in the Image—SNR

5.7 Image Weighting and Pulse Sequence Parameters TE and TR

5.8 A Menagerie of Pulse Sequences

5.9 Enhanced Diagnostic Capabilities of MRI—Contrast Agents

5.10 Molecular MRI

5.11 Reading the Mind—Functional MRI

5.12 Magnetic Resonance Spectroscopy

5.13 MR Hardware

5.14 MRI Safety

5.15 Imaging Artefacts in MRI

5.16 Advanced MR Technology and Its Possible Future

5.17 Acknowledgments

References

Chapter 6: Toward A 3D View of Cellular Architecture: Correlative Light Microscopy and Electron Tomography

6.1 Introduction

6.2 Historical Perspective

6.3 Stains for CLEM

6.4 Probes for CLEM

6.5 Clem Applications

6.6 Future Perspective

References

Chapter 7: Tracer Imaging

7.1 Introduction

7.2 Instrumentation

7.3 Measurement Techniques

7.4 Applications

7.5 Acknowledgements

References

Chapter 8: Fluorescence Imaging

8.1 Introduction

8.2 Contrast Mechanisms

8.3 Direct Methods: Fluorescent Probes

8.4 Indirect Methods: Fluorescent Proteins

8.5 Microscopy

8.6 Macroscopic Imaging/Tomography

8.7 Planar Imaging

8.8 Tomography

8.9 Conclusion

8.10 Acknowledgments

References

Chapter 9: Infrared and Raman Spectroscopic Imaging

9.1 Introduction

9.2 Instrumentation

9.3 Raman Imaging

9.4 Sampling Techniques

9.5 Data Analysis and Image Evaluation

9.6 Applications

References

Chapter 10: Coherent Anti-Stokes Raman Scattering Microscopy

10.1 Basics

10.2 Theory

10.3 CARS Microscopy in Practice

10.4 Instrumentation

10.5 Laser Sources

10.6 Data Acquisition

10.7 Measurement Techniques

10.8 Applications

10.9 Conclusions

10.10 Acknowledgements

References

Chapter 11: Biomedical Sonography

11.1 Basic Principles

11.2 Instrumentation of Real-Time Ultrasound Imaging

11.3 Measurement Techniques of Real-Time Ultrasound Imaging

11.4 Application Examples of Biomedical Sonography

References

Chapter 12: Acoustic Microscopy for Biomedical Applications

12.1 Sound Waves and Basics of Acoustic Microscopy

12.2 Types of Acoustic Microscopy

12.3 Biomedical Applications of Acoustic Microscopy

12.4 Examples of Tissue Investigations using SAM

12.5 Acknowledgements

References

Index

For further information visit: the book web page http://www.openmodelica.org, the Modelica Association web page http://www.modelica.org, the authors research page http://www.ida.liu.se/labs/pelab/modelica, or home page http://www.ida.liu.se/~petfr/, or email the author at [email protected]. Certain material from the Modelica Tutorial and the Modelica Language Specification available at http://www.modelica.org has been reproduced in this book with permission from the Modelica Association under the Modelica License 2 Copyright © 1998–2011, Modelica Association, see the license conditions (including the disclaimer of warranty) at http://www.modelica.org/modelica-legal-documents/ModelicaLicense2.html. Licensed by Modelica Association under the Modelica License 2.

Modelica© is a registered trademark of the Modelica Association. MathModelica© is a registered trademark of MathCore Engineering AB. Dymola© is a registered trademark of Dassault Syst`emes. MATLAB© and Simulink© are registered trademarks of MathWorks Inc. Java is a trademark of Sun MicroSystems AB. Mathematica© is a registered trademark of Wolfram Research Inc.

Copyright © 2011 by the Institute of Electrical and Electronics Engineers, Inc.

Published by John Wiley & Sons, Inc., Hoboken, New Jersey. All rights reserved.

Published simultaneously in Canada.

No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning, or otherwise, except as permitted under Section 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 750-4744. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permission.

Limit of Liability/Disclaimer of Warranty: While the publisher and author have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives or written sales materials. The advice and strategies contained herein may not be suitable for your situation. You should consult with a professional where appropriate. Neither the publisher nor author shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages.

For general information on our other products and services or for technical support, please contact our Customer Care Department within the United States at (800) 762-2974, outside the United States at (317) 572-3993 or fax (317) 572-4002.

Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic formats. For more information about Wiley products, visit our web site at www.wiley.com.

Library of Congress Cataloging-in-Publication Data:

Biomedical imaging : principles and applications / [edited by] Reiner Salzer.

p. ; cm.

ISBN 978-0-470-64847-6 (hardback)

I. Salzer, Reiner, 1942-

[DNLM: 1. Diagnostic Imaging. 2. Spectrum Analysis. WN 180]

616.07'54–dc23

2011042654

Preface

Biomedical imaging is improving healthcare and helps selecting the most efficient therapy. Imaging technologies provide snapshots of biomarkers and diseases such as cancer. Imaging can take this information even a step further, showing the activity of these markers in vivo and how their location changes over time. Advances in experimental and clinical imaging are likely to enable doctors not only to locate and delineate the disease but also to assess the activity of the biological processes and to provide localized treatment. New imaging technologies are increasingly being used to understand the biological complexity, diversity, and the in vivo behaviour. Imaging is considered an important bridge between basic research and bed-side application.

A wide range of technologies is already available for in vivo, ex vivo, and in vitro imaging. The introduction of new imaging instrumentation requires the combination of know-how in medicine and biology, in data processing, in engineering, and in science. Biologists and MDs are interested in technical basics and methods of measurement. Engineers need detailed descriptions of the biomedical basis of the measured data. Scientists want more background information on instrumentation and measurement techniques. Different imaging modalities always have specific strengths and weaknesses. For each modality, the basics of how it works, important information parameters, and the state-of-the-art instrumentation are described in this book. Examples of imaging applications are presented.

X-rays, gamma rays, radiofrequency signals, and ultrasound waves are standard probes, but others such as visible and infrared light, microwaves, terahertz rays, and intrinsic and applied electric and magnetic fields are being explored. Some of the younger technologies, such as molecular imaging, may enhance existing imaging modalities; however, they also, in combination with nanotechnology, biotechnology, bioinformatics, and new forms of computational hardware and software, may well lead to novel approaches to clinical imaging. This review provides a brief overview of the current state of image-based diagnostic medicine and offers comments on the directions in which some of its subfields may be heading.

Visualization can augment our ability to reason about complex data, thereby increasing the efficiency of manual analyses. In some cases, the appropriate image makes the solution obvious. The first two chapters give an overview of existing methods and tools for visualization and highlight some of their limitations and challenges. The next chapters describe technology and applications of established imaging modalities such as X-ray imaging, CT (Computed Tomography), MRI (Magnetic Resonance Imaging), and tracer imaging. The final part deals with imaging technologies using light (fluorescence imaging, infrared and Raman imaging, CARS microscopy) or sound (biomedical sonography and acoustic microscopy).

Thanks go to all authors for their efforts and commitments to the publication of this volume. The support by the publisher WILEY in the final composition and edition of the book should be acknowledged as well. The greatest debt of gratitude goes to our families for their patience and encouragement.

Contributors

Jürgen Bereiter-Hahn, Institut für Zellbiologie und Neurowissenschaft, Johann-Wolfgang-Goethe-Universität, Frankfurt/M, Germany
Stefan Bernet, Department für Physiologie und Medizinische Physik, Medizinische Universität Innsbruck, Innsbruck, Austria
Christian Brackmann, Department of Chemical and Biological Engineering, Chalmers University of Technology, Göteborg, Sweden
Lutgarde M.C. Buydens, Radboud University Nijmegen, Institute for Molecules and Materials, Department of Analytical Chemistry/Chemometrics, Nijmegen, The Netherlands
Nikolaos C. Deliolanis, Institute for Biological and Medical Imaging (IBMI), Helmholtz Zentrum München and Technische Universität München, Munich, Germany
Annika Enejder, Department of Chemical and Biological Engineering, Chalmers University of Technology, Göteborg, Sweden
Thomas Flohr, Computed Tomography Department, Siemens Healthcare, Forchheim, Germany
Christoph Heinrich, Department für Physiologie und Medizinische Physik, Medizinische Universität Innsbruck, Innsbruck, Austria
Volker Hietschold, Department of Radiology, University Hospital, “Carl Gustav Carus,” Dresden, Germany
Rainer Hinz, Wolfson Molecular Imaging Centre, University of Manchester, Manchester, UK
Karen A. Jansen, Department of Molecular Cell Biology, Leiden University Medical Center, Leiden, The Netherlands
Abraham J. Koster, Department of Molecular Cell Biology, Leiden University Medical Center, Leiden, The Netherlands
Patrick W.T. Krooshof, Radboud University Nijmegen, Institute for Molecules and Materials, Department of Analytical Chemistry/Chemometrics, Nijmegen, The Netherlands
Willem J. Melssen, Radboud University Nijmegen, Institute for Molecules and Materials, Department of Analytical Chemistry/Chemometrics, Nijmegen, The Netherlands
Vasilis Ntziachristos, Institute for Biological and Medical Imaging (IBMI), Helmholtz Zentrum München and Technische Universität München, Munich, Germany
Geert J. Postma, Radboud University Nijmegen, Institute for Molecules and Materials, Department of Analytical Chemistry/Chemometrics, Nijmegen, The Netherlands
Monika Ritsch-Marte, Department für Physiologie und Medizinische Physik, Medizinische Universität Innsbruck, Innsbruck, Austria
Reiner Salzer, Department of Chemistry and Food Chemistry, Technische Universität Dresden, Dresden, Germany
Georg Schmitz, Department for Electrical Engineering and Information Technology, Medical Engineering, Ruhr-University, Bochum, Germany
Christian P. Schultz, Siemens Medical and Center for Molecular Imaging Research, Massachusetts General Hospital, Charlestown, MA, USA
Jonathan C. Sharp, Institute for Biodiagnostics (West), National Research Council of Canada, Calgary, AB, Canada
Gerald Steiner, Clinical Sensoring and Monitoring, Medical Faculty Carl Gustav Carus, Dresden University of Technology, Dresden, Germany
Boguslaw Tomanek, Institute for Biodiagnostics (West), National Research Council of Canada, Calgary, AB, Canada
Stefan Ulzheimer, Computed Tomography Department, Siemens Healthcare, Forchheim, Germany
Jack A. Valentijn, Department of Molecular Cell Biology, Leiden University Medical Center, Leiden, The Netherlands
Karine M. Valentijn, Department of Molecular Cell Biology, Leiden University Medical Center, Leiden, The Netherlands
Jörg van den Hoff, Department of Positron Emission Tomography, Institute of Radiopharmacy, Helmholtz-Zentrum Dresden-Rossendorf, Dresden, Germany
Linda F. van Driel, Department of Molecular Cell Biology, Leiden University Medical Center, Leiden, The Netherlands

Chapter 1

Evaluation of Spectroscopic Images

Patrick W.T. Krooshof, Geert J. Postma, Willem J. Melssen, and Lutgarde M.C. Buydens

Radboud University Nijmegen, Institute for Molecules and Materials, Department of Analytical Chemistry/Chemometrics, Nijmegen, The Netherlands

1.1 Introduction

Many sophisticated techniques are currently used for an accurate recognition and diagnosis of different diseases. Advanced imaging techniques are useful in studying medical conditions in a noninvasive manner. Common imaging methodologies to visualize and study anatomical structures include Computed Tomography (CT, Chapter 4), Magnetic Resonance Imaging (MRI, Chapter 5), and Positron Emission Tomography (PET, Chapter 7). Recent developments are focused on understanding the molecular mechanisms of diseases and the response to therapy. Magnetic Resonance Spectroscopy (MRS) (Section 5.12), for example, provides chemical information about particular regions within an organism or sample. This technique has been used on patients with a wide range of neurological and psychiatric disorders, such as stroke, epilepsy, multiple sclerosis, dementia, and schizophrenia.

Examination of the images, obtained by any of the imaging techniques to visualize and study anatomical structures, is a straightforward task. In many situations, abnormalities are clearly visible in the acquired images and often the particular disease can also be identified by the clinician. However, in some cases, it is more difficult to make the diagnosis. The spectral data obtained from MRS can then assist, to a large extent, in the noninvasive diagnosis of diseases. However, the appearance of the spectral data is different compared to the image data (Fig. 1.5). Although spectra obtained from diseased tissues are different from spectra obtained from normal tissue, the complexity of the data limits the interpretability. Furthermore, the amount of spectral data can be overwhelming, which makes the data analysis even more difficult and time consuming. In order to use the information in MR spectra effectively, a (statistical) model is required, which reduces the complexity and provides an output that can easily be interpreted by clinicians. Preferably, the output of the model should be some kind of an image that can be compared to MRI images to obtain a better diagnosis.

In this chapter, the application of a chemometric approach to facilitate the analysis of image data is explained. This approach is based on a similarity measure between data obtained from a patient and reference data by searching for patterns in the data. In particular, when the amount of data is large, the use of such a mathematical approach has proved to be useful. The basics of pattern recognition methods are discussed in Section 1.2. A distinction is made between commonly used methods and several advantages and disadvantages are discussed. The application of a useful pattern recognition technique is presented in Section 1.3. The required data processing and quantitation steps are mentioned, and subsequently, data of different patients is classified. Finally, results are shown to illustrate the applicability of pattern recognition techniques.

1.2 Data Analysis

Chemometrics is a field in chemistry that helps improve the understanding of chemical data (1, 2, 3). With the use of mathematical and statistical methods, chemical data is studied to obtain maximum information and knowledge about the data. Chemometrics is typically used to explore patterns in data sets, that is, to discover relations between samples. In particular, when the data is complex and the amount of data is large, chemometrics can assist in data analysis. A technique that is frequently applied to compress the information into a more comprehensible form is Principal Component Analysis (PCA) (2, 4). Another application of chemometrics is to predict properties of a sample on the basis of the information in a set of known measurements. Such techniques are found very useful in process monitoring and process control to predict and make decisions about product quality (3, 5). Finally, chemometrics can be used to make classification models that divide samples into several distinct groups (6, 7).

This last mentioned application of chemometrics could be very helpful, for example, in the discrimination of the different and complex spectra acquired by MRS examinations. Patient diagnosis and treatment can be improved if chemometric techniques can automatically distinguish diseased tissue from normal tissue.

The aim of such pattern recognition techniques is to search for patterns in the data. Individual measurements are grouped into several categories on the basis of a similarity measure (7, 8, 9). If class membership is used in the grouping of objects, the classification is called supervised pattern recognition (6, 8). Pattern recognition can also be unsupervised, where no predefined classes are available. The grouping of objects is then obtained by the data itself. Unsupervised pattern recognition is also called clustering (7, 8, 9).

The resulting clusters obtained by pattern recognition contain objects, for example, MR spectra, which are more similar to each other compared to objects in the other clusters. If the spectra of a patient with brain tumor are considered, the data could be divided into two groups: one group contains normal spectra and the other group contains spectra acquired from the tumorous tissue. If the group that contains normal spectra and the group that contains tumorous spectra can be identified, this grouping can be used for classification. The tissue from a region of the brain can be classified as normal or tumorous by matching its spectra to the class that contains the most similar spectra.

1.2.1 Similarity Measures

Most essential in pattern recognition is the definition of the similarity measure. Usually, the (dis)similarity between a set of objects is calculated using a distance measure, of which the Euclidean distance is most popular (6, 7, 8, 10). The dissimilarity between two objects xi and xj is calculated as in Equation 1.1.

1.1

1.2

Another widely used distance measure is the Mahalanobis distance, which incorporates the correlations between variables in the calculations (6, 7, 11). To calculate the distance between an object xi and the centroid (mean) of a group of objects, μk, it takes the covariance matrix Ck of the cluster into account, that is, the size and shape of the cluster. The squared Mahalanobis distance is given by

1.3

Several variants of these distance measures exist, such as the Manhattan or Bhattacharyya distance (3, 6, 7, 8), but are not commonly used in practice.

Instead of clustering the data using distances as a similarity measure, the data can also be modeled by several distributions such as the normal distribution. In that case, the likelihood, which is discussed in Section 1.2.2.1, is used as a criterion function (12, 13).

1.2.2 Unsupervised Pattern Recognition

A large variety of clustering methods have been developed for different kinds of problems (6, 7, 8, 14). A distinction between these different approaches can be made on the basis of their definition of a cluster. The techniques can be categorized into three main types: partitional (15, 16), hierarchical (17, 18), and density based (19).

1.2.2.1 Partitional Clustering

The type of clustering techniques that are most widely applied obtain a single partition of the data. These partitional clustering methods try to divide the data into a predefined number of clusters. Usually, the techniques divide the data into several clusters by optimizing a criterion or cost function. In the popular K-means algorithm (15, 16, 20), for example, the sum of squares of within-cluster distances is minimized (Equation 1.4). This is obtained by iteratively transferring objects between clusters, until the data is partitioned into well-separated and compact clusters. Because compact clusters contain objects with a relatively small distance to the mean of the cluster, these clusters result in a small value for the criterion function.

1.4

The K-means algorithm starts with a random selection of K cluster centers. In the next step, each object of the data set is assigned to the closest cluster. To determine the closest cluster, the distances of a particular object to the cluster centers, d(xi, μk), are calculated using one of the similarity measures. Subsequently, the cluster centers are updated, and this process is repeated until a stop criterion is met, such as a threshold for the criterion function. In the end, each object is assigned to one cluster.

This clustering algorithm requires short computation time and is therefore suitable to handle large data sets. A major disadvantage, however, is the sensitivity to the cluster centers chosen initially, which makes the clustering results hard to reproduce. Another drawback is that the number of clusters has to be defined in advance (6, 7).

It is also possible to associate each object to every cluster using a membership function. Membership reflects the probability that the object belongs to the particular cluster. The K-means variant, which results in a fuzzy clustering by including cluster memberships, is fuzzy c-means (16, 21). The membership function uik, which is used in fuzzy c-means, is given in Equation 1.5 and represents the probability of object xi belonging to cluster k. This membership is dependent on the distance of the object to the cluster center. If the distance to a cluster center is small, the membership for this cluster becomes relatively large (22).

1.5

In addition to the distances, the membership also depends on the fuzziness index γ, which is 2 in most situations. By taking smaller values for the index, the membership for clusters close to object xi is increased. If two clusters overlap in variable space, the membership of an object will be low for both clusters because the uncertainty of belonging to a particular cluster is high. This is an attractive property of fuzzy methods. Because the problem with overlapping clusters is common in cluster analysis, fuzzy clustering algorithms are frequently applied.

The partitional methods that have been described earlier are based on a distance measure to calculate cluster similarities. Another variant to determine clusters in a data set is based on a statistical approach and is called model-based clustering or mixture modeling (7, 13, 23). It describes the data by mixtures of multivariate distributions. The density of objects in a particular cluster can, for example, be described by a P-dimensional Gaussian distribution. The formula of the distribution is given in Equation 1.6, where μk and Ck are the mean and covariance matrix of the data in cluster k, respectively.

1.6

If the sum of the Gaussians, which describe the density of objects in the individual clusters, exactly fits the original data, the entire data set is perfectly described. Each Gaussian can then be considered as one cluster. For the one-dimensional data set presented in Figure 1.1a three Gaussian distributions (dashed lines) are required to obtain a proper fit of the density of objects in the original data (solid line) (24).

Figure 1.1 Example of modeling the density of objects by three Gaussian distributions for (a) one-dimensional and (b) two-dimensional data sets (25).

With the use of multidimensional Gaussian distributions, more complicated data sets can be modeled. The density of objects in a data set consisting of two variables can be modeled by two-dimensional distributions, as illustrated in Figure 1.1b. In this situation also, three clusters are present, and therefore, three distributions are required to obtain a good fit of the object density in the entire data.

The goodness of fit is evaluated by the log-likelihood criterion function, which is given in Equation 1.7 (13). The distributions are weighted by the mixture proportion τk, which corresponds to the fraction of objects in the particular cluster. The log-likelihood depends also on the cluster memberships of each of the N objects. In Equation 1.7, this is expressed as uik, which represents the probability that object xi belongs to cluster k, similar to the membership function given in Equation 1.5. The data set is optimally described by the distributions when the criterion function is maximized (13).

1.7

The optimal partitioning of the data is usually obtained by the Expectation-Maximization (EM) algorithm (4, 7, 26). In the first step of EM, the probabilities uik are estimated by an initial guess of some statistical parameters of each cluster. These parameters are the means, covariances, and mixture proportions of the cluster. Subsequently, the statistical parameters are recalculated using these estimated probabilities (4). This process is iterated until convergence of the log-likelihood criterion. The data is eventually clustered according to the calculated probabilities of each object to belong to the particular clusters. It is an advantage that model-based clustering yields cluster memberships instead of assigning each object to one particular cluster. However, because of the random initialization of the parameters, the results of mixture modeling are not robust. Furthermore, the number of distributions to describe the data has to be defined ahead of the clustering procedure (7, 24).

1.2.2.2 Hierarchical Clustering

Another approach to clustering is to obtain a clustering structure instead of a single partitioning of the data. Such a hierarchical clustering can be agglomerative or divisive. The agglomerative strategy starts with assigning each object to an individual cluster (16, 17, 27). Subsequently, the two most similar clusters are iteratively merged, until the data is grouped in one single cluster. Once an object is merged to a cluster, it cannot join another cluster. Divisive hierarchical clustering is similar to the agglomerative strategy, but starts with one cluster that is divided into two clusters that have least similarity. This process is repeated until all clusters contain only one object. Repeated application of hierarchical clustering will result in identical merging or splitting sequences, and thus the results are reproducible (12).

Agglomerative methods are more commonly used. On the basis of the definition of the similarity measure, several variants exist: single, complete, average, and centroid linkage (28, 29). In single linkage, the (updated) distance between the objects of a particular cluster (e.g. c1 and c2) and an object xi is the minimum distance (dmin) between xi and the objects of the cluster. The maximum (dmax) and average (davg) of the distances between the particular object xi and the objects of the cluster is used in complete and average linkage, respectively. In centroid linkage, the distance between an object and the centroid of the cluster (dcen) is used. This is schematically represented in Figure 1.2.

Figure 1.2 Distances between clusters used in hierarchical clustering. (a) Single linkage. (b) Complete linkage. (c) Average linkage. (d) Centroid linkage.

Analogous to the methods based on distance measures, hierarchical clustering can also be performed by model-based clustering. The hierarchical approach is an adaptation from the partitional approach of model-based clustering (12, 23). Model-based agglomerative clustering also starts with individual objects, but merges the pair of objects that lead to the largest increase in the log-likelihood criterion (see Eq. 1.7). This process then continues until all objects are grouped into one cluster (23).

The sequence of merging or splitting can be visualized in a dendrogram, representing, in a treelike manner, the similarity levels at which clusters are merged. The dendrogram can be cut at a particular level to obtain a clustering with a desired number of clusters. The results with different number of clusters can then be easily compared. The dendrogram obtained by average linkage, applied to MR spectra of a patient, is given in Figure 1.3. If, for example, the data should be clustered into four groups, the dendrogram should be cut at a distance of 11,700. This threshold is indicated by the red line in Figure 1.3.

Figure 1.3 Dendrogram showing the sequence of merging MR spectral data by hierarchical clustering. The red line indicates the threshold to obtain four clusters.

Hierarchical clustering methods are not sensitive to outliers because outliers will be assigned to distinct clusters (6). A possible drawback is the computation time. Hierarchical clustering of large data sets will require the merging of many objects: at each merging step, the similarities between pairs of objects need to be recalculated (6, 12, 23, 30).

1.2.2.3 Density-Based Clustering

The third type of clustering methods is based on the density of objects in variable space (7, 19, 31). Clusters are formed by high density areas, and the boundaries of the clusters are given by less dense regions. These densities are determined by a threshold. Another parameter that has to be defined is the size of the volume for which the density is estimated. Objects are then assigned to a cluster when the density within this volume exceeds the predefined threshold. The number of areas with high density indicates the number of clusters in the clustering result. Objects that are not assigned to a cluster are considered as noise or outliers.

DBSCAN is a well-known method to cluster data into regions using high density constraints (19, 22). The algorithm scans an area within a certain radius from a particular object and determines the number of other objects within this neighborhood. The size of the area and the minimum number of objects in the neighborhood have to be defined in advance. If the neighborhood contains more objects than the threshold, then every object in the neighborhood is assigned to one cluster. Subsequently, the neighborhood of another object in the particular cluster is scanned to expand the cluster. When the cluster does not grow anymore, the neighborhood of another object, not belonging to this cluster, is considered. If this object is also located in a dense region, a second cluster is found, and the whole procedure is repeated. With fewer objects in the neighborhood than the threshold, an object is assigned to a group of noisy objects.

Originally, density-based clustering was developed to detect clusters in a data set with exceptional shapes and to exclude noise and outliers. However, the method fails to simultaneously detect clusters with different densities (22). Clusters with a relatively low density will then be considered as noise. Another limitation is the computation time for calculating the density estimation for each object. Moreover, it can be difficult to determine proper settings for the size of the neighborhood and the threshold for the number of objects.

1.2.3 Supervised Pattern Recognition

Pattern recognition can also be supervised, by including class information in the grouping of objects (6, 8). Predefined classes are used by this type of pattern recognition for the classification of unknown objects. For example, the distance of an unidentified object to all the objects in the reference data set can be calculated by a particular distance measure, to determine the most similar (closest) object. The unknown object can then be assigned to the class to which this nearest neighbor belongs. This is the basic principle of the k-nearest neighbor (kNN) method (3, 6, 7). A little more sophisticated approach is to extend the number of neighbors. In that case, there is a problem if the nearest neighbors are from different classes. Usually, a majority rule is applied to assign the object to the class to which the majority of the nearest neighbors belong. If the majority rule cannot be applied because there is a tie, that is, the number of nearest neighbors of several classes is equal, another approach is required. The unknown object can, for example, randomly be assigned to a predefined class. Another method is to assign the object, in this situation, to the class to which its nearest neighbor belongs (7).

Another type of supervised pattern recognition is discriminant analysis (3, 7, 32). These methods are designed to find boundaries between classes. One of the best-known methods is Linear Discriminant Analysis (LDA). With the assumption that the classes have a common covariance matrix, it describes the boundaries by straight lines. More generally, an unknown object is assigned to the class for which the Mahalanobis distance (Eq. 1.3) is minimal. Because the covariance matrices of the classes are assumed to be equal, the pooled covariance matrix is used to calculate the Mahalanobis distances:

1.8

where Ck and nk are the covariance matrix and the number of objects in cluster k, K is the number of predefined classes, and n is the total number of objects in the data set.

In Quadratic Discriminant Analysis (QDA), the covariance matrices of the classes are not assumed to be equal. Each class is described by its own covariance matrix (3, 7, 32). Similar to LDA, QDA calculates the Mahalanobis distances of unknown objects to the predefined classes and assigns the objects to the closest class. Other more sophisticated techniques also exist, such as support vector machines (33) or neural networks (34), but these approaches are beyond the scope of this chapter.

1.2.3.1 Probability of Class Membership

To reflect the reliability of the classification, the probabilities of class membership could be calculated. This is especially useful to detect overlapping classes. An object will then have a relatively high probability to belong to two or more classes. Furthermore, the probabilities can be used to find new classes, which are not present in the reference data set. If the class membership is low for the predefined classes, the unknown object probably belongs to a totally different class.

The probabilities of class membership can be estimated on the basis of the distribution of the objects in the classes (6, 7). The density of objects at a particular distance from a class centroid is a direct estimator of the probability that an object at this distance belongs to the class. If it is assumed that the data follows a normal distribution, the density of objects can be expressed as in Equation 1.6. If the distance of a new object to a class centroid is known, the density and thus the probability of class membership can be calculated on the basis of Equation 1.6 (3, 7, 32).

A more straightforward approach is based on the actual distribution of the objects, without the assumption that the data can be described by a theoretical distribution (35). In this method, the Mahalanobis distances of the objects in a particular class, say class A, with respect to the centroid of this class are calculated. Also, the Mahalanobis distances of the other objects (not belonging to class A) with respect to the centroid of class A are calculated. This procedure is repeated for every class present in the data set. Eventually, the distances are used to determine the number of objects within certain distances from the centroid of each class. These distributions of objects can be visualized in a plot as presented in Figure 1.4 (36). The solid line represents the percentage of objects from class A within certain Mahalanobis distances (d) from the centroid of class A. Every object of class A is within a distance of 6d from the particular centroid. The dotted line represents the percentage of other objects (not from class A) within certain distances from the centroid of class A. In this example, there is little overlap between objects from class A and the other objects, indicating that class A is well separated from the other classes.

Figure 1.4 Distribution of objects belonging to class A (solid line) and objects belonging to other classes (dotted line) with respect to the centroid of class A (36).

Figure 1.5 Example of the data obtained by MRI and MRS. (a) An MRI image of the brain that clearly shows the presence of abnormal tissue. The grid on the MRI image indicates the resolution of the spectroscopic data. (b) Part of the spectroscopic data, showing the spectra obtained by MRS from several regions of the brain.

The percentage of objects within a particular distance from a class centroid reflects the object density of the class at this distance. Therefore, these percentages can be used to estimate the probabilities of class membership for the classes that are present in the data set. At a distance 3d from the centroid of class A, for example, about 30% of the objects belonging to class A are within this distance. This is illustrated in Figure 1.4. If the Mahalanobis distance of an unknown object with respect to the centroid of class A is 3d, the estimated probability is then 70%. By comparing the probabilities of class membership for each class, the unknown objects can be classified and conclusions can be drawn about the reliability of classification (35).

1.3 Applications

Pattern recognition techniques can be applied to magnetic resonance data to improve the noninvasive diagnosis of brain tumors (37, 38, 39, 40, 41). Because the spectra obtained by MRS are complex, statistical models can facilitate data analysis. The application of pattern recognition techniques to MR spectra and MRI image data is illustrated using research performed on a widely studied data set (24, 35). This data set was constructed during a project called INTERPRET, which was funded by the European Commission to develop new methodologies for automatic tumor type recognition in the human brain (42).

1.3.1 Brain Tumor Diagnosis

Uncontrolled growth of cells is a major issue in medicine, as it results in a malignant or benign tumor. If the tumor spreads to vital organs, such as the brain, tumor growth can even be life threatening (43). Brain tumors are the leading cause of cancer death in children and third leading cause of cancer death in young adults. Only one-third of people diagnosed with a brain tumor survive more than 5 years from the moment of diagnosis (44).

Two commonly used techniques to diagnose brain tumors are magnetic resonance imaging (MRI, Chapter 5) and magnetic resonance spectroscopy (MRS, Section 5.12). MRI provides detailed pictures of organs and soft tissues within the human body (45, 46). This technique merely shows the differences in the water content and composition of various tissues. Because tumorous tissues have a composition (and water content) different from that of normal tissues, MRI can be used to detect tumors, as shown in Figure 1.5a. Even different types of tissue within the same organ, such as white and gray matter in the brain, can easily be distinguished (46).

Magnetic resonance spectroscopy (MRS) is another technique that can be used for diagnosing brain tumors (47, 48, 49). It allows the qualitative and quantitative assessment of the biochemical composition in specific brain regions (50). A disadvantage of the technique is that interpretation of the resulting spectra representing the compounds present in the human tissue is difficult and time consuming. Several spectra acquired from a tumorous region in the brain are presented in Figure 1.5b to illustrate the complexity of the data. To compare the differences in resolution, the MR spectra are visualized on top of the corresponding region of the MRI image. Another limitation of MRS is that the size of the investigated region, for example, of the brain, might be larger than the suspected lesion. The heterogeneity of the tissue under examination will then disturb the spectra, making characterization of the region more difficult (51).

1.3.2 MRS Data Processing

Before chemometrics can be applied to the complex spectra obtained by MRS, these spectra require some processing. Owing to time constraints, the quality of the acquired MR spectra is often very poor. The spectra frequently contain relatively small signals and a large amount of noise: the so-called signal-to-noise ratio is low. Furthermore, several artifacts are introduced by the properties of the MR system. For example, magnetic field inhomogeneities result in distortion of the spectra. Also, patient movement during the MR examinations will introduce artifacts. Another characteristic of the spectra is the appearance of broad background signals from macromolecules and the presence of a large water peak.

1.3.2.1 Removing MRS Artifacts

In order to remove the previously mentioned artifacts, several processing steps need to be performed (26). Different software packages are commercially available to process and analyze MR spectra (43, 52, 53) and, in general, they apply some commonly used correction methods. These methods include eddy current correction (54), residual water filtering (55), phase and frequency shift correction (56), and a baseline correction method (26).

Eddy current correction is performed to correct for magnetic field inhomogeneities, induced in the magnetic system during data acquisition (54, 57). One method to correct for these distortions is to measure the field variation as a function of time. This can be achieved by measuring the phase of the much stronger signal of water. The actual signal can then be divided by this phase factor in order to remove the effect of field variation (57).

Residual water filtering is required to remove the intense water peak that is still present in the spectra after correction for eddy current distortions. A useful filtering method is based on Hankel-Lanczos Singular Value Decomposition (HLSVD) (58). Resonances between 4.1 and 5.1 ppm, as determined by the HLSVD algorithm, are subtracted from the spectra. Water resonates at approximately 4.7 ppm, and therefore, the water peak and its large tails are removed from the spectra without affecting the peak areas of other compounds (55, 58).

Several small phase differences between the peaks in a spectrum may still be present after eddy current correction. In addition, frequency shifts between spectra of different regions, for example, of the brain, may also be present. These peak shifts may be induced by patient movement. A correction method based on PCA can be applied to eliminate the phase and frequency shift variations of a single resonance peak across a series of spectra. PCA methodology is used to model the effects of phase and frequency shifts, and this information can then be used to remove the variations (56, 59).

Broad resonances of large molecules or influences from the large water peak may contribute to baseline distortions, which make the quantification of the resonances of small compounds more difficult. The removal of these broad resonances improves the accuracy of quantification and appearance of the spectrum. Usually, the correction is performed by estimating the baseline using polynomial functions, followed by subtraction from the original signal (26).

1.3.2.2 MRS Data Quantitation

After processing the spectra obtained by MRS, the data can be interpreted. As the spectra contain information from important brain metabolites, deviation in the spectra, and thus in metabolite concentrations, might be indicative of the presence of abnormal tissue. Two different MR spectra are shown in Figure 1.6. The spectrum in Figure 1.6a is acquired from a normal region of the brain and the spectrum in Figure 1.6b originates from a malignant region. The differences between these spectra are obvious and MRS could therefore be used to detect abnormal tissues. Several metabolites are particularly useful for tumor diagnosis, and some of these are creatine (resonates at 3.95 and 3.02 ppm), glutamate (3.75 and 2.20 ppm), glutamine (3.75 and 2.20 ppm), myoinositol (3.56 ppm), choline (3.20 ppm), N-acetyl aspartate (NAA, 2.02 ppm), lactate (1.33 ppm), and fatty acids (1.3 and 0.90 ppm) (25, 60, 61).

Figure 1.6 Two MR spectra, illustrating the difference between spectra obtained from a normal region (a) and a malignant region (b) of the brain. The signals of some metabolites are indicated in the figure: myoinositol (mI) at 3.56 ppm, choline (Cho) at 3.20 ppm, creatine (Cr) at 3.02 ppm, NAA at 2.02 ppm, and lactate (Lac) at 1.33 ppm.

To use the metabolic information in the MR spectra for diagnosis of brain tumors, the intensity of the signals in the spectra requires quantitation (62). A simple approach is to integrate several spectral regions, assuming that each region contains information from one single metabolite. Because some metabolites show overlap in the spectra, for example, glutamate and glutamine, more sophisticated methods could be applied. More accurate methods fit the spectrum by a specific lineshape function, using a reference set of model spectra. A method that has been introduced for the analysis of MR spectra is the LCModel (52, 62, 63). This method analyzes a spectrum as linear combinations of a set of model spectra from individual metabolites in solution. Another powerful tool for processing and quantitation of MR spectra is MRUI (64, 65). This program has a graphical user interface and is able to analyze the MR spectra and present the results in an accessible manner.

Differences in the quantitated metabolite levels are often used to diagnose the malignancy of a tumor. But even when only the peak intensities of the previously mentioned metabolites are quantitated and interpreted, the amount of data is still large. Especially if MRS is applied to a large region of the brain, to obtain multiple spectra, many metabolite concentrations have to be compared. To facilitate the data analysis, the relative metabolite concentrations within different regions (or actually volumes) could be presented in an image. These regions are referred to as voxels. The resulting metabolic maps visualize the spatial distribution of the concentration of several metabolic compounds, and this can be used to localize or diagnose brain tumors. This is shown in Figure 1.7, in which the relative metabolite concentrations of choline, creatine, and NAA are presented. As shown, an increased concentration of choline is detected in the tumorous region, and reduced concentrations of creatine and NAA are found. Another application of such metabolic maps is to study tumor heterogeneity since this is important for an accurate diagnosis (66, 67, 68).

Figure 1.7 Metabolic maps constructed from MRS data. (a) The MRI image shows the presence of a tumor in the lower right corner of the image. The differences in metabolic concentration are illustrated for (b) choline, (c) creatine, and (d) NAA. Bright pixels represent a high concentration of the particular metabolite.

1.3.3 MRI Data Processing

As mentioned in Section 5.7, the echo time (TE) and the repetition time (TR) are acquisition parameters that determine the T1- and T2-sensitivity of the acquired images. By using different settings for these parameters, different MRI images are obtained (45, 46). Examples of these images are given in Figure 1.8. In the T1- and T2-weighted images (Fig. 1.8a and b), the differences in contrast reveal the ventricles while this is less visible in the PD-weighted image (Fig. 1.8c).

Figure 1.8 MRI images obtained by different acquisition parameters. (a) T1-weighted image. (b) T2-weighted image. (c) Proton density image. (d) T1-weighted image after administration of a gadolinium tracer.

T1-, T2-, and PD-weighted images are commonly acquired with different combinations of TR and TE to be able to discriminate between different tissues. For tumor diagnosis, a contrast medium can be used to improve the tissue differentiation. Usually, gadolinium (Gd) is used as a contrast agent to enhance lesions where the blood–brain barrier is defective. An example of a T1-weighted image enhanced by gadolinium is presented in Figure 1.8d (69).

When the different images need to be compared, they should be aligned with respect to each other. This is necessary because when patients move during the acquisition of the different MRI images, artifacts may be introduced, which complicates the data analysis.

1.3.3.1 Image Registration

Image registration is the alignment of the different images obtained by MRI examinations. This alignment compensates for differences in the position or orientation of the brain in the images due to patient movement. If the images are taken from a series of MRI examinations to study tumor growth or shrinkage after radiation treatment, differences in image size or resolution may be obtained. Image registration should then be applied to match the areas of interest in the images (36).

Although manual alignment of images is possible, it is time consuming and not always reproducible. Automated procedures are therefore preferred. Numerous approaches are available for medical images (69, 70), and in general, they are based on a similarity measure between an image and a corresponding reference image. Because patient movement results in small shifts between different images, a simple cross-correlation method can be used to correct for this artifact (71). However, sensitivity to large intensity differences in different contrast images limits the use of cross-correlation methods. To perform the alignment by matching specific features in the images, which are insensitive to changes in tissue or acquisition, the registration can be improved. These features are, for example, edges and corners in normal brain tissue regions. The chemometric technique Procrustes analysis can then be applied to match the specific features by means of translation, rotation, and uniform scaling transformations. The best match is found when the least-squares solution is obtained by minimizing the distance between all pairs of points in the two images (3, 72).

1.3.4 Combining MRI and MRS Data

If the magnetic resonance data is properly processed, tumorous tissue may be distinguished from normal or other abnormal nonmalignant tissue types, such as necrosis, edema, or scar tissue, by examining the data. However, from the standard acquired MRI images, it is often difficult to properly discriminate tissues within the tumorous region. This can be improved by the contrast-enhanced image, which reveals the lesion where the blood–brain barrier is defective. But because this barrier may variably be affected, the size of the tumor may be under- or overestimated. This has been observed in some brain tumor studies, where the contrast-enhanced lesion was found to be much smaller than the region of abnormal metabolism (73). In addition, MRS examinations have shown that metabolite levels are highly variable for particular tissue types. Also, overlap of metabolite levels between different tumor grades has been observed (73). These findings indicate that MRI and MRS are two complementary techniques for brain tumor diagnosis. MRS provides metabolic information on a low spatial resolution, and MRI provides morphological information on a high spatial resolution. Therefore, the analysis of MRI images and metabolic maps should be performed simultaneously. One of the possibilities is to display a specific metabolic map over an MRI image. Examination of these overlay images facilitates the differentiation between tissue types. For example, the delineation of a viable tumor can be detected more accurately. This is important for studying the effect of chemotherapy or radiation treatment. The combination of MRI and MRS data will improve patient diagnosis and treatment or will facilitate the sampling of biopsies to regions of tumorous tissue.

1.3.4.1 Reference Data Set

For pattern recognition, the data from several patients with a brain tumor and several volunteers was selected from the INTERPRET database (42). After reaching consensus about the histopathology, three tumor types were identified according to the World Health Organization classification system. These three classes contained glial tumors with different grades, indicating the level of malignancy. If a tumor could be assigned to more than one grade, the highest malignancy level determined the grade for the entire tumor, even if most of the tumor is of a lower grade (74). The brain tumors of six patients were diagnosed as grade II, of four patients as grade III, and of five patients as grade IV glial tumors. In addition, classes of normal tissue and Cerebrospinal Fluid (CSF) were created from the patient and volunteer data.

For each predefined class, a selection of voxels from different patients was made. Only voxels from regions that clearly consisted of tissue belonging to the particular class were selected. The data for the normal class was selected from the volunteers or from the contralateral brain region of the patients. The CSF class contained only data from CSF regions that were not in close contact with the tumor region (35).

From the MR spectra of these selected voxels, only the regions where important metabolites appear were selected, as these regions contain most information for the discrimination between brain tumors. The remaining regions contain mainly spectral noise. Besides excluding noise, this also results in a significant reduction in the amount of data.

The estimated metabolite levels should be subsequently combined with the MRI data to improve data analysis (75, 76, 77). However, the spatial resolution of MRI images is much higher than the resolution of MRS data. Therefore, to combine MRI with MRS data, the resolution of the MRI images was lowered to the resolution of the MRS data by averaging the pixel intensities within each spectroscopic voxel. To equally weigh the influence from both data modalities in the classification process, the averaged intensities were scaled to the range of the spectral data.

Different (statistical) methods are available to classify the processed data (see Section 1.2). Application of these methods to the data will result in a classification for each individual voxel. It is also possible to calculate an accuracy measure to indicate the reliability of the classification outcome. An example of a classification method with a statistical basis and the possibility of determining the reliability are discussed in the following sections.

1.3.5 Probability of Class Memberships

After processing the data of the reference set, the distribution of objects in each class can be investigated. By determining the percentage of objects within certain distances from the class centroids, as explained in Section 1.2.3.1, separated and overlapping classes can be found. To investigate which classes are overlapping, the Mahalanobis distance of the objects in each surrounding tissue class with respect to a class centroid was calculated. The surrounding of an isolated class will contain no objects from other classes, whereas the neighborhood of overlapping classes will contain objects belonging to multiple classes. The distributions of the classes in the reference data set are presented in Figure 1.9 (35). Within a distance of about 6d from the centroid of the normal class, every object belonging to this class and some objects from the cerebrospinal fluid class are found. The CSF class shows some overlap with the normal class, but is well separated from the other classes, as depicted in Figure 1.9b. This indicates that the normal and CSF classes are isolated from the tumor classes, and thus, a classification method should discriminate well between normal tissue and malignant tumors.

Figure 1.9 Distribution of the objects in different classes with respect to the centroid of the (a) normal, (b) CSF, (c) grade II, (d) grade III, and (e) grade IV class. The vertical dotted lines indicate thresholds that are used in the classification.

More overlap exists between grade II and III classes. Although at a distance of 5.5d from the centroid of grade II class, almost every object from this class is found; also, about 25% of the objects from grade III class appear within this distance. Less overlap is found with grade IV class. Therefore, discrimination between grade II and grade IV is possible, but grade II class is difficult to be separated from grade III class. The same can be concluded from the distribution of grade III class (Fig. 1.9d). Because grade IV class is isolated from the other classes, this type of tumor can probably be classified with a high reliability (Fig. 1.9e).

1.3.6 Class Membership of Individual Voxels

The distribution of objects can be used to estimate probabilities of class membership for a new patient, as discussed in Section 1.2.3.1. Four patients (A–D) were analyzed; histopathological examination of the brain tissue resulted in the diagnosis of grade III (patient A and D), grade II (patient B), and grade IV (patient C) tumors. The probability maps calculated for the patients are presented in Figure 1.10. Dark voxels in the map represent a high probability that the underlying tissue belongs to the particular class. The probability maps need to be compared regionwise because the probabilities are relative, that is, the intensities are scaled to its maximum value. If a region in the map, for example, of grade II, is dark, while it is bright in the other maps, the estimated probability for grade II class is high (close to 100%). If the region is dark in several maps, the estimated probability of the class membership is lower, and a black pixel does not correspond with a probability of 100%. The probabilities may, therefore, also assist in the assessment of tumor heterogeneity. If the probability is high exclusively for one class, then the tissue will be homogeneous. If the probability is high for two tissue classes, the region might be a mixture of these tissues and might therefore be heterogeneous. However, if a mixture of these two tissues is not possible from a clinical point of view, the high probabilities for both classes indicate a low classification reliability.

Figure 1.10 Probability maps of four patients (A–D). For comparison, an MRI image of each patient is shown on the left. The probability of each voxel to belong to the normal, CSF, grade II, grade III, and grade IV class are shown. Patients A and D are diagnosed with a grade III tumor, patient B with a grade II tumor, and patient C with a grade IV tumor.

The probability maps for the four patients show that large regions have a high probability to belong to the normal class while these regions have a low probability to belong to another class. The CSF and grade IV regions of the patients show a similar behavior. The voxels with a high probability for the CSF or grade IV class have low probabilities to belong to the other classes. This indicates that these classes are well separated from other classes and that the reliability of classification is high.

Different results are obtained for grade II and grade III classes. The regions where the estimated probability is high for belonging to grade II class also have a high probability to belong to grade III class. In particular, patients A and B show clearly the overlapping problem of these two tissue classes. This could indicate the presence of heterogeneous tumor tissue in these regions and that the reliability of the classification is low.

1.3.7 Classification of Individual Voxels

The estimated probabilities of class membership can be used for classification by assigning each voxel to the class for which it has the highest probability. To define specific classification accuracies, thresholds are set for each class. For a correct classification of 99% of the normal objects, and thus an α-error of 1%, the threshold was set at the position shown in Figure 1.9a. However, with the threshold set at 6d, about 3% of the objects will be classified incorrectly (β-error) as normal. Similar to the normal class, the threshold for the CSF objects was also set at 6d to obtain a low α-error. To minimize the chance that normal or CSF objects are classified as grade IV, the specific threshold for grade IV class was set at 5.5d even though the α-error is then relatively large. Because of the overlap between grade II and III classes, additional thresholds were defined. To obtain a low β-error, an object is directly assigned to grade II when the distance to the particular centroid is very small. With larger distances to grade II class (between 2.5d and 5.5d), the object is assigned to this class only if the distance to grade III class is 1.3 times larger than to grade II class. Otherwise, the object is located in the area of overlap between both classes and the object is assigned to an “undecided” class. The value of 1.3 is obtained after optimization. Identical thresholds are used when the object is closest to grade III class (Fig. 1.9d). If the distance of an object exceeds the specific thresholds for each individual class, the object probably does not belong to any of the classes in the reference data set, and the object is therefore assigned to an “unknown” class.

By applying these thresholds to the classification of the four patients A–D, good results are obtained, as shown in Figure 1.11. The results of pattern recognition are projected over different MRI images to compare the results with morphological information. Patient A was diagnosed with a grade III tumor, and the classification result shows the presence of regions of grade II and grade III tumors. But because a tumor is diagnosed by the most malignant tissue present, the result of the statistical method is in agreement with histopathology. From the maps in Figure 1.10, it was clear that some voxels in the tumorous region have equal probability of belonging to grade II and grade III classes. This indicates the presence of a heterogeneous tumor. Therefore, the voxels in the overlapping region of grade II and III tumors are assigned to the undecided class. The voxels in the center of the tumor have the highest probability of belonging to grade III class and are therefore assigned to this class. The tumors of patients B and C are classified as grade II and grade IV tumor in correspondence with the histopathology. Although the tumorous voxels of patient B also have a relatively high probability to belong to grade III class, the estimated probability for grade II class is much higher, and the voxels are classified as grade II. One region of patient C is classified as unknown. This is probably because the particular class is not represented in the reference data set. One voxel of patient D has been misclassified as a grade IV tumor while the patient was diagnosed with a grade III tumor. The reason for this could be that the voxel is a corner voxel, where the MRS signal is low because of the characteristics of the MRS data acquisition.

Figure 1.11 Classification results of patients A–D. Classification was performed by applying specific thresholds to define classification accuracies.

1.3.8 Clustering into Segments

More robust classification results can be obtained by grouping similar voxels in advance of the classification. The voxels that are grouped in one cluster will then be assigned to the same class. The voxels within a region covering a brain tumor, for example, will be grouped in one cluster, and this cluster will subsequently be classified as tumorous tissue. Because a tumor is often larger than the size of a voxel, it could be advantageous to combine the voxels within the tumorous region for classification purposes. However, when the information of the voxels within the tumor is averaged, it is not possible to discover any spatial heterogeneity of the tumor. Therefore, a clustering method that groups only similar voxels instead of grouping neighboring voxels should be applied. One of the methods that could be used to cluster similar voxels is mixture modeling (see Section 1.2.2.1), which describes the spread of the data by a set of distributions. Similar objects will be described by the same distribution and the resulting distributions can be considered as individual clusters. This method is based on a statistical approach, which makes it possible to calculate the probability of an object to belong to a particular cluster. The fit of the data can be evaluated by the likelihood criterion given in Equation 1.7. However, the number of distributions, and thus clusters, has to be defined in advance.

The voxels of a patient are clustered to obtain a partitioning with three, four, five, and six segments. The results for patient A are given in Figure 1.12