Proceedings Volume 6914

Medical Imaging 2008: Image Processing

cover
Proceedings Volume 6914

Medical Imaging 2008: Image Processing

View the digital version of this volume at SPIE Digital Libarary.

Volume Details

Date Published: 8 April 2008
Contents: 22 Sessions, 171 Papers, 0 Presentations
Conference: Medical Imaging 2008
Volume Number: 6914

Table of Contents

icon_mobile_dropdown

Table of Contents

All links to SPIE Proceedings will open in the SPIE Digital Library. external link icon
View Session icon_mobile_dropdown
  • Front Matter: Volume 6914
  • Segmentation I: Methodology
  • Atlases and Population Studies
  • Registration I: Applications
  • Neurological Applications
  • Classification and Pattern Recognition
  • Registration II: Methodology
  • Cardiovascular Applications
  • Image Restoration and Enhancement
  • Liver Applications
  • Pulmonary Applications
  • Segmentation II: Applications
  • Posters: Classification and Pattern Recognition
  • Posters: Image Restoration and Enhancement
  • Posters: Motion Analysis
  • Posters: MRI
  • Posters: Multiresolution and Wavelets
  • Posters: Registration
  • Posters: Segmentation
  • Posters: Shape
  • Posters: Texture
  • Posters: Validation
Front Matter: Volume 6914
icon_mobile_dropdown
Front Matter: Volume 6914
This PDF file contains the front matter associated with SPIE Proceedings Volume 6914, including the Title Page, Copyright information, Table of Contents, Introduction (if any), and the Conference Committee listing.
Segmentation I: Methodology
icon_mobile_dropdown
Area prior constrained level set evolution for medical image segmentation
Ismail Ben Ayed, Shuo Li, Ali Islam, et al.
The level set framework has proven well suited to medical image segmentation1-6 thanks to its ability of balancing the contribution of image data and prior knowledge in a principled, flexible and transparent way. It consists of evolving a curve toward the target object boundaries. The curve evolution equation is sought following the optimization of a cost functional containing two types of terms: data terms, which measure the fidelity of segmentation to image intensities, and prior terms, which traduce learned prior knowledge. Without priors many algorithms are likely to fail due to high noise, low contrast and data incompleteness. Different priors have been investigated such as shape1 and appearance priors.7 In this study, we propose a simple type of priors: the area prior. This prior embeds knowledge of an approximate object area and has two positive effects. First, It speeds up significantly the evolution when the curve is far from the target object boundaries. Second, it slows down the evolution when the curve is close to the target. Consequently, it reinforces curve stability at the desired boundaries when dealing with low contrast intensity edges. The algorithm is validated with several experiments using Magnetic Resonance (MR) images and Computed Tomography (CT) images. A comparison with another level set method illustrates the positive effects of the area prior.
A resistive-network model for image segmentation
Peter J. Yim
A novel physical analogy and associated algorithm is proposed for image segmentation. The physical analogy is that of an electrical network. The network is composed of discrete resistive elements that are interconnected, according to adjacency relations in the image. The resistances vary linearly with the image gradient and exponentially with the voltage across the elements. Segmentation is obtained by applying a voltage between a voltage source and ground that are located at points on the inside and outside of an object, respectively. Those points are identified in an interactive manner. Provided that a sufficiently high voltage is applied, a dichotomization of the voltages in the network occurs. In practice, the solution can be obtained in an iterative manner by gradually increasing the applied voltage. At each step as the source voltage is increased, the resistances are held constant and the Kirchoff's-law system of equations is solved algebraically. It is not clear yet whether image segmentation obtained by this algorithm is globally optimal with respect to a cost-function, as is the case for minimum s-t cut image segmentation. However, like minimum s-t cut algorithm, the resistive-network algorithm does tend to favor segmentation boundaries with shorter, and thus smoother, boundaries. A potential advantage of the resistive-network algorithm is that it is amenable to parallelization. The algorithm was found to provide reasonable results for segmentation of a synthetic image of step boundary degraded by blurring and Gaussian noise, low-contrast computed tomography (CT) of a ball, and CT of a liver lesion.
A new distribution metric for image segmentation
Romeil Sandhu, Tryphon Georgiou, Allen Tannenbaum
In this paper, we present a new distribution metric for image segmentation that arises as a result in prediction theory. Forming a natural geodesic, our metric quantifies "distance" for two density functionals as the standard deviation of the difference between logarithms of those distributions. Using level set methods, we incorporate an energy model based on the metric into the Geometric Active Contour framework. Moreover, we briefly provide a theoretical comparison between the popular Fisher Information metric, from which the Bhattacharyya distance originates, with the newly proposed similarity metric. In doing so, we demonstrate that segmentation results are directly impacted by the type of metric used. Specifically, we qualitatively compare the Bhattacharyya distance and our algorithm on the Kaposi Sarcoma, a pathology that infects the skin. We also demonstrate the algorithm on several challenging medical images, which further ensure the viability of the metric in the context of image segmentation.
Segmenting images analytically in shape space
Yogesh Rathi, Samuel Dambreville, Marc Niethammer, et al.
This paper presents a novel analytic technique to perform shape-driven segmentation. In our approach, shapes are represented using binary maps, and linear PCA is utilized to provide shape priors for segmentation. Intensity based probability distributions are then employed to convert a given test volume into a binary map representation, and a novel energy functional is proposed whose minimum can be analytically computed to obtain the desired segmentation in the shape space. We compare the proposed method with the log-likelihood based energy to elucidate some key differences. Our algorithm is applied to the segmentation of brain caudate nucleus and hippocampus from MRI data, which is of interest in the study of schizophrenia and Alzheimer's disease. Our validation (we compute the Hausdorff distance and the DICE coefficient between the automatic segmentation and ground-truth) shows that the proposed algorithm is very fast, requires no initialization and outperforms the log-likelihood based energy.
A unified framework for joint registration and segmentation
Accurate image registration is a necessary prerequisite for many diagnostic and therapy planning procedures where complementary information from different images has to be combined. The design of robust and reliable non-parametric registration schemes is currently a very active research area. Modern approaches combine the pure registration scheme with other image processing routines such that both ingredients may benefit from each other. One of the new approaches is the combination of segmentation and registration ("segistration"). Here, the segmentation part guides the registration to its desired configuration, whereas on the other hand the registration leads to an automatic segmentation. By joining these image processing methods it is possible to overcome some of the pitfalls of the individual methods. Here, we focus on the benefits for the registration task. In the current work, we present a novel unified framework for non-parametric registration combined with energy-based segmentation through active contours. In the literature, one may find various ways to combine these image processing routines. Here, we present the most promising approaches within the general framework. It is based on a single variational formulation of both the registration and the segmentation part. The performance tests are carried out for magnetic resonance (MR) images of the brain, and they demonstrate the potential of the proposed methods.
Atlases and Population Studies
icon_mobile_dropdown
Adaptive local multi-atlas segmentation: application to heart segmentation in chest CT scans
Atlas-based segmentation is a popular generic technique for automated delineation of structures in volumetric data sets. Several studies have shown that multi-atlas based segmentation methods outperform schemes that use only a single atlas, but running multiple registrations on large volumetric data is too time-consuming for routine clinical use. We propose a generally applicable adaptive local multi-atlas segmentation method (ALMAS) that locally decides how many and which atlases are needed to segment a target image. Only the selected parts of atlases are registered. The method is iterative and automatically stops when no further improvement is expected. ALMAS was applied to segmentation of the heart on chest CT scans and compared to three existing atlas-based methods. It performed significantly better than single-atlas methods and as good as multi-atlas methods at a much lower computational cost.
Robust registration between cardiac MRI images and atlas for segmentation propagation
X. Zhuang, D. J. Hawkes, W. R. Crum, et al.
We propose a new framework to propagate the labels in a heart atlas to the cardiac MRI images for ventricle segmentations based on image registrations. The method employs the anatomical information from the atlas as priors to constrain the initialisation between the atlas and the MRI images using region based registrations. After the initialisation which minimises the possibility of local misalignments, a fluid registration is applied to fine-tune the labelling in the atlas to the detail in the MRI images. The heart shape from the atlas does not have to be representative of that of the segmented MRI images in terms of morphological variations of the heart in this framework. In the experiments, a cadaver heart atlas and a normal heart atlas were used to register to in-vivo data for ventricle segmentation propagations. The results have shown that the segmentations based on the proposed method are visually acceptable, accurate (surface distance against manual segmentations is 1.0 ± 1.0 mm in healthy volunteer data, and 1.6 ± 1.8 mm in patient data), and reproducible (0.7 ± 1.0 mm) for in-vivo cardiac MRI images. The experiments also show that the new initialisation method can correct the local misalignments and help to avoid producing unrealistic deformations in the nonrigid registrations with 21% quantitative improvement of the segmentation accuracy.
The SRI24 multichannel brain atlas: construction and applications
Torsten Rohlfing, Natalie M. Zahr, Edith V. Sullivan, et al.
We present a new standard atlas of the human brain based on magnetic resonance images. The atlas was generated using unbiased population registration from high-resolution images obtained by multichannel-coil acquisition at 3T in a group of 24 normal subjects. The final atlas comprises three anatomical channels (T1-weighted, early and late spin echo), three diffusion-related channels (fractional anisotropy, mean diffusivity, diffusion-weighted image), and three tissue probability maps (CSF, gray matter, white matter). The atlas is dynamic in that it is implicitly represented by nonrigid transformations between the 24 subject images, as well as distortion-correction alignments between the image channels in each subject. The atlas can, therefore, be generated at essentially arbitrary image resolutions and orientations (e.g., AC/PC aligned), without compounding interpolation artifacts. We demonstrate in this paper two different applications of the atlas: (a) region definition by label propagation in a fiber tracking study is enabled by the increased sharpness of our atlas compared with other available atlases, and (b) spatial normalization is enabled by its average shape property. In summary, our atlas has unique features and will be made available to the scientific community as a resource and reference system for future imaging-based studies of the human brain.
A generalization of voxel-wise procedures for high-dimensional statistical inference using ridge regression
Karl Sjöstrand, Valerie A. Cardenas, Rasmus Larsen, et al.
Whole-brain morphometry denotes a group of methods with the aim of relating clinical and cognitive measurements to regions of the brain. Typically, such methods require the statistical analysis of a data set with many variables (voxels and exogenous variables) paired with few observations (subjects). A common approach to this ill-posed problem is to analyze each spatial variable separately, dividing the analysis into manageable subproblems. A disadvantage of this method is that the correlation structure of the spatial variables is not taken into account. This paper investigates the use of ridge regression to address this issue, allowing for a gradual introduction of correlation information into the model. We make the connections between ridge regression and voxel-wise procedures explicit and discuss relations to other statistical methods. Results are given on an in-vivo data set of deformation based morphometry from a study of cognitive decline in an elderly population.
The evaluation of a population based diffusion tensor image atlas using a ground truth method
Wim Van Hecke, Alexander Leemans, Emiliano D'Agostino, et al.
Purpose: Voxel based morphometry (VBM) is increasingly being used to detect diffusion tensor (DT) image abnormalities in patients for different pathologies. An important requisite for these VBM studies is the use of a high-dimensional, non-rigid coregistration technique, which is able to align both the spatial and the orientational information. Recent studies furthermore indicate that high-dimensional DT information should be included during coregistration for an optimal alignment. In this context, a population based DTI atlas is created that preserves the orientational DT information robustly and contains a minimal bias towards any specific individual data set. Methods: A ground truth evaluation method is developed using a single subject DT image that is deformed with 20 deformation fields. Thereafter, an atlas is constructed based on these 20 resulting images. Thereby, the non-rigid coregistration algorithm is based on a viscous fluid model and on mutual information. The fractional anisotropy (FA) maps as well as the DT elements are used as DT image information during the coregistration algorithm, in order to minimize the orientational alignment inaccuracies. Results: The population based DT atlas is compared with the ground truth image using accuracy and precision measures of spatial and orientational dependent metrics. Results indicate that the population based atlas preserves the orientational information in a robust way. Conclusion: A subject independent population based DT atlas is constructed and evaluated with a ground truth method. This atlas contains all available orientational information and can be used in future VBM studies as a reference system.
Multivariate longitudinal statistics for neonatal-pediatric brain tissue development
Shun Xu, Martin Styner, John Gilmore, et al.
The topic of studying the growth of human brain development has become of increasing interest in the neuroimaging community. Cross-sectional studies may allow comparisons between means of different age groups, but they do not provide a growth model that integrates the continuum of time, nor do they present any information about how individuals/population change over time. Longitudinal data analysis method arises as a strong tool to address these questions. In this paper, we use longitudinal analysis methods to study tissue development in early brain growth. A novel approach of multivariate longitudinal analysis is applied to study the associations between the growth of different brain tissues. In this paper, we present the methodologies to statistically study scalar (univariate) and vector (multivariate) longitudinal data, and demonstrate exploratory results in a neuroimaging study of early brain tissue development. We obtained growth curves as a quadratic function of time for all three tissues. The quadratic terms were tested to be statistically significant, showing that there was indeed a quadratic growth of tissues in early brain development. Moreover, our result shows that there is a positive correlation between repeated measurements of any single tissue, and among those of different tissues. Our approach is generic in natural and thus can be applied to any longitudinal data with multiple outcomes, even brain structures. Also, our joint mixed model is flexible enough to allow incomplete and unbalanced data, i.e. subjects do not need to have the same number of measurements, or be measured at the exact time points.
Registration I: Applications
icon_mobile_dropdown
Mosaicing of single plane illumination microscopy images using groupwise registration and fast content-based image fusion
Stephan Preibisch, Torsten Rohlfing, Michael P. Hasak, et al.
Single Plane Illumination Microscopy (SPIM; Huisken et al., Nature 305(5686):1007-1009, 2004) is an emerging microscopic technique that enables live imaging of large biological specimens in their entirety. By imaging the living biological sample from multiple angles SPIM has the potential to achieve isotropic resolution throughout even relatively large biological specimens. For every angle, however, only a relatively shallow section of the specimen is imaged with high resolution, whereas deeper regions appear increasingly blurred. In order to produce a single, uniformly high resolution image, we propose here an image mosaicing algorithm that combines state of the art groupwise image registration for alignment with content-based image fusion to prevent degrading of the fused image due to regional blurring of the input images. For the registration stage, we introduce an application-specific groupwise transformation model that incorporates per-image as well as groupwise transformation parameters. We also propose a new fusion algorithm based on Gaussian filters, which is substantially faster than fusion based on local image entropy. We demonstrate the performance of our mosaicing method on data acquired from living embryos of the fruit fly, Drosophila, using four and eight angle acquisitions.
Ultrasound specific similarity measures for three-dimensional mosaicing
The introduction of 2D array ultrasound transducers enables the instantaneous acquisition of ultrasound volumes in the clinical practice. The next step coming along is the combination of several scans to create compounded volumes that provide an extended field-of-view, so called mosaics. The correct alignment of multiple images, which is a complex task, forms the basis of mosaicing. Especially the simultaneous intensity-based registration has many properties making it a good choice for ultrasound mosaicing in comparison to the pairwise one. Fundamental for each registration approach is a suitable similarity measure. So far, only standard measures like SSD, NNC, CR, and MI were used for mosaicing, which implicitly assume an additive Gaussian distributed noise. For ultrasound images, which are degraded by speckle patterns, alternative noise models based on multiplicative Rayleigh distributed noise were proposed in the field of motion estimation. Setting these models into the maximum likelihood estimation framework, which enables the mathematical modeling of the registration process, led us to ultrasound specific bivariate similarity measures. Subsequently, we used an extension of the maximum likelihood estimation framework, which we developed in a previous work, to also derive multivariate measures. They allow us to perform ultrasound specific simultaneous registration for mosaicing. These measures have a higher potential than afore mentioned standard measures since they are specifically designed to cope with problems arising from the inherent contamination of ultrasound images by speckle patterns. The results of the experiments that we conducted on a typical mosaicing scenario with only partly overlapping images confirm this assumption.
Three-dimensional image registration of MR proximal femur images for the analysis of trabecular bone parameters
This study investigated the feasibility of automatic image registration of MR high-spatial resolution proximal femur trabecular bone images as well as the effects of gray-level interpolation and volume of interest (VOI) misalignment on MR-derived trabecular bone structure parameters. For six subjects, a baseline scan and a follow-up scan of the proximal femur were acquired on the same day. An automatic image registration technique, based on mutual information, utilized a baseline and a follow-up scan to compute transform parameters that aligned the two images. These parameters were subsequently used to transform the follow-up image with three different gray-level interpolators. Nearest neighbor interpolation and b-spline approximation did not significantly alter bone parameters, while linear interpolation significantly modified bone parameters (p<0.01). Improvement in image alignment due to the automatic registration was determined by visually inspecting difference images and 3D renderings. This work demonstrates the first application of automatic registration, without prior segmentation, of high-spatial resolution trabecular bone MR images of the proximal femur. Additionally, effects due to imprecise analysis volume alignment are investigated. Inherent heterogeneity in trabecular bone structure and imprecise positioning of the VOI along the slice (A/P) direction resulted in significant changes in bone parameters (p<0.01). Results suggest that automatic mutual information registration using nearest-neighbor gray-level interpolation to transform the final image ensures VOI alignment between baseline and follow-up images and does not compromise the integrity of MR-derived trabecular bone parameters.
Vertebral surface registration using ridgelines/crestlines
Sovira Tan, Jianhua Yao, Lawrence Yao, et al.
The Iterative Closest Point (ICP) algorithm is an efficient and popular technique for surface registration. It however suffers from the well-known problem of local minima that make the algorithm stop before it reaches the desired global solution. ICP can be improved by the use of landmarks or features. We recently developed a level set capable of evolving on the surface of an object represented by a triangular mesh. This level set permits the segmentation of portions of a surface based on curvature features. The boundary of a segmented portion forms a ridgeline/crestline. We show that the ridgelines/crestlines and corresponding enclosed surfaces extracted by the algorithm can substantially improve ICP registration. We compared the performance of an ICP algorithm in three setups: 1) ICP without landmarks. 2) ICP using ridgelines. 3) ICP using ridgelines and corresponding enclosed surfaces. Our material consists of vertebral body surfaces extracted for a study about the progression of Ankylosing Spondylitis. Same vertebrae scanned at intervals of one or two years were rigidly registered. Vertebral body rims and the end plate surfaces they enclose were used as landmarks. The performance measure was the mean error distance between the registered surfaces. From the one hundred registrations that we performed the average mean error was respectively 0.503mm, 0.335mm and 0.254mm for the three setups. Setup 3 almost halved the average error of setup 1. Moreover the error range is dramatically reduced from [0.0985, 2.19]mm to just [0.0865, 0.532]mm, making the algorithm very robust.
Bi-planar 2D-to-3D registration in Fourier domain for stereoscopic x-ray motion tracking
Dominique Zosso, Benoît Le Callennec, Meritxell Bach Cuadra, et al.
In this paper we present a new method to track bone movements in stereoscopic X-ray image series of the knee joint. The method is based on two different X-ray image sets: a rotational series of acquisitions of the still subject knee that allows the tomographic reconstruction of the three-dimensional volume (model), and a stereoscopic image series of orthogonal projections as the subject performs movements. Tracking the movements of bones throughout the stereoscopic image series means to determine, for each frame, the best pose of every moving element (bone) previously identified in the 3D reconstructed model. The quality of a pose is reflected in the similarity between its theoretical projections and the actual radiographs. We use direct Fourier reconstruction to approximate the three-dimensional volume of the knee joint. Then, to avoid the expensive computation of digitally rendered radiographs (DRR) for pose recovery, we develop a corollary to the 3-dimensional central-slice theorem and reformulate the tracking problem in the Fourier domain. Under the hypothesis of parallel X-ray beams, the heavy 2D-to-3D registration of projections in the signal domain is replaced by efficient slice-to-volume registration in the Fourier domain. Focusing on rotational movements, the translation-relevant phase information can be discarded and we only consider scalar Fourier amplitudes. The core of our motion tracking algorithm can be implemented as a classical frame-wise slice-to-volume registration task. Results on both synthetic and real images confirm the validity of our approach.
Neurological Applications
icon_mobile_dropdown
Cortical thickness measurement from magnetic resonance images using partial volume estimation
Measurement of the cortical thickness from 3D Magnetic Resonance Imaging (MRI) can aid diagnosis and longitudinal studies of a wide range of neurodegenerative diseases. We estimate the cortical thickness using a Laplacian approach whereby equipotentials analogous to layers of tissue are computed. The thickness is then obtained using an Eulerian approach where partial differential equations (PDE) are solved, avoiding the explicit tracing of trajectories along the streamlines gradient. This method has the advantage of being relatively fast and insure unique correspondence points between the inner and outer boundaries of the cortex. The original method is challenged when the thickness of the cortex is of the same order of magnitude as the image resolution since partial volume (PV) effect is not taken into account at the gray matter (GM) boundaries. We propose a novel way to take into account PV which improves substantially accuracy and robustness. We model PV by computing a mixture of pure Gaussian probability distributions and use this estimate to initialize the cortical thickness estimation. On synthetic phantoms experiments, the errors were divided by three while reproducibility was improved when the same patients was scanned three consecutive times.
Parallel optimization of tumor model parameters for fast registration of brain tumor images
Evangelia I. Zacharaki, Cosmina S. Hogea, Dinggang Shen, et al.
The motivation of this work is to register MR brain tumor images with a brain atlas. Such a registration method can make possible the pooling of data from different brain tumor patients into a common stereotaxic space, thereby enabling the construction of statistical brain tumor atlases. Moreover, it allows the mapping of neuroanatomical brain atlases into the patient's space, for segmenting brains and thus facilitating surgical or radiotherapy treatment planning. However, the methods developed for registration of normal brain images are not directly applicable to the registration of a normal atlas with a tumor-bearing image, due to substantial dissimilarity and lack of equivalent image content between the two images, as well as severe deformation or shift of anatomical structures around the tumor. Accordingly, a model that can simulate brain tissue death and deformation induced by the tumor is considered to facilitate the registration. Such tumor growth simulation models are usually initialized by placing a small seed in the normal atlas. The shape, size and location of the initial seed are critical for achieving topological equivalence between the atlas and patient's images. In this study, we focus on the automatic estimation of these parameters, pertaining to tumor simulation. In particular, we propose an objective function reflecting feature-based similarity and elastic stretching energy and optimize it with APPSPACK (Asynchronous Parallel Pattern Search), for achieving significant reduction of the computational cost. The results indicate that the registration accuracy is high in areas around the tumor, as well as in the healthy portion of the brain.
Spatial normalization of diffusion tensor images based on anisotropic segmentation
Jinzhong Yang, Dinggang Shen, Chandan Misra, et al.
A comprehensive framework is proposed for the spatial normalization of diffusion tensor (DT) brain images using tensor-derived tissue attributes. In this framework, the brain tissues are first classified into three categories: the white matter (WM), the gray matter (GM), and the cerebral-spinal fluid (CSF) using the anisotropy and diffusivity information derived from the full tensor. The tissue attributes obtained from this anisotropic segmentation are then incorporated into a very-high-dimensional elastic registration method to produce a spatial deformation field. Finally, the rotational component in the deformation field, together with the estimated underlying fiber direction, is used to determine an appropriate tensor reorientation. This framework has been assessed quantitatively and qualitatively based on a sequence of experiments. A simulated experiment has been performed to evaluate the accuracy of the spatial warping by examining the variation between deformation fields. To verify the tensor reorientation, especially, in the anisotropic microstructures of WM fiber tissues, an experiment has been designed to compare the fiber tracts generated from the DT template and the normalized DT subjects in some regions of interest (ROIs). Finally, this method has been applied to spatially normalize 31 subjects to a common space, the case in which there exist large deformations between subjects and the existing approaches are normally difficult to achieve satisfactory results. The average across the individual normalized DT images shows a significant improvement in signal-to-noise ratio (SNR).
Fusion of rat brain histology and MRI using weighted multi-image mutual information
Christoph Palm, Graeme P. Penney, William R. Crum, et al.
Introduction - Fusion of histology and MRI is frequently demanded in biomedical research to study in vitro tissue properties in an in vivo reference space. Distortions and artifacts caused by cutting and staining of histological slices as well as differences in spatial resolution make even the rigid fusion a difficult task. State-of- the-art methods start with a mono-modal restacking yielding a histological pseudo-3D volume. The 3D information of the MRI reference is considered subsequently. However, consistency of the histology volume and consistency due to the corresponding MRI seem to be diametral goals. Therefore, we propose a novel fusion framework optimizing histology/histology and histology/MRI consistency at the same time finding a balance between both goals. Method - Direct slice-to-slice correspondence even in irregularly-spaced cutting sequences is achieved by registration-based interpolation of the MRI. Introducing a weighted multi-image mutual information metric (WI), adjacent histology and corresponding MRI are taken into account at the same time. Therefore, the reconstruction of the histological volume as well as the fusion with the MRI is done in a single step. Results - Based on two data sets with more than 110 single registrations in all, the results are evaluated quantitatively based on Tanimoto overlap measures and qualitatively showing the fused volumes. In comparison to other multi-image metrics, the reconstruction based on WI is significantly improved. We evaluated different parameter settings with emphasis on the weighting term steering the balance between intra- and inter-modality consistency.
Comparison of EM-based and level set partial volume segmentations of MR brain images
Hemant D. Tagare, Yunmei Chen, Robert K. Fulbright
EM and level set algorithms are competing methods for segmenting MRI brain images. This paper presents a fair comparison of the two techniques using the Montreal Neurological Institute's software phantom. There are many flavors of level set algorithms for segmentation into multiple regions (multi-phase algorithms, multi-layer algorithms). The specific algorithm evaluated by us is a variant of the multi-layer level set algorithm. It uses a single level set function for segmenting the image into multiple classes and can be run to completion without restarting. The EM-based algorithm is standard. Both algorithms have the capacity to model a variable number of partial volume classes as well as image inhomogeneity (bias field). Our evaluation consists of systematically changing the number of partial volume classes, additive image noise, and regularization parameters. The results suggest that the performances of both algorithms are comparable across noise, number of partial volume classes, and regularization. The segmentation errors of both algorithms are around 5 - 10% for cerebrospinal fluid, gray and white matter. The level set algorithm appears to have a slight advantage for gray matter segmentation. This may be beneficial in studying certain brain diseases (Multiple Sclerosis or Alzheimer's disease) where small changes in gray matter volume are significant.
3D MRI brain image segmentation based on region restricted EM algorithm
This paper presents a novel algorithm of 3D human brain tissue segmentation and classification in magnetic resonance image (MRI) based on region restricted EM algorithm (RREM). The RREM is a level set segmentation method while the evolution of the contours was driven by the force field composed by the probability density functions of the Gaussian models. Each tissue is modeled by one or more Gaussian models restricted by free shaped contour so that the Gaussian models are adaptive to the local intensities. The RREM is guaranteed to be convergency and achieving the local minimum. The segmentation avoids to be trapped in the local minimum by the split and merge operation. A fuzzy rule based classifier finally groups the regions belonging to the same tissue and forms the segmented 3D image of white matter (WM) and gray matter (GM) which are of major interest in numerous applications. The presented method can be extended to segment brain images with tumor or the images having part of the brain removed with the adjusted classifier.
Automatic segmentation of the facial nerve and chorda tympani using image registration and statistical priors
Jack H. Noble, Frank M. Warren M.D., Robert F. Labadie M.D., et al.
In cochlear implant surgery, an electrode array is permanently implanted in the cochlea to stimulate the auditory nerve and allow deaf people to hear. A minimally invasive surgical technique has recently been proposed--percutaneous cochlear access--in which a single hole is drilled from the skull surface to the cochlea. For the method to be feasible, a safe and effective drilling trajectory must be determined using a pre-operative CT. Segmentation of the structures of the ear would improve trajectory planning safety and efficiency and enable the possibility of automated planning. Two important structures of the ear, the facial nerve and chorda tympani, present difficulties in intensity based segmentation due to their diameter (as small as 1.0 and 0.4 mm) and adjacent inter-patient variable structures of similar intensity in CT imagery. A multipart, model-based segmentation algorithm is presented in this paper that accomplishes automatic segmentation of the facial nerve and chorda tympani. Segmentation results are presented for 14 test ears and are compared to manually segmented surfaces. The results show that mean error in structure wall localization is 0.2 and 0.3 mm for the facial nerve and chorda, proving the method we propose is robust and accurate.
Classification and Pattern Recognition
icon_mobile_dropdown
Feature selection and classification of multiparametric medical images using bagging and SVM
Yong Fan, Susan M. Resnick, Christos Davatzikos
This paper presents a framework for brain classification based on multi-parametric medical images. This method takes advantage of multi-parametric imaging to provide a set of discriminative features for classifier construction by using a regional feature extraction method which takes into account joint correlations among different image parameters; in the experiments herein, MRI and PET images of the brain are used. Support vector machine classifiers are then trained based on the most discriminative features selected from the feature set. To facilitate robust classification and optimal selection of parameters involved in classification, in view of the well-known "curse of dimensionality", base classifiers are constructed in a bagging (bootstrap aggregating) framework for building an ensemble classifier and the classification parameters of these base classifiers are optimized by means of maximizing the area under the ROC (receiver operating characteristic) curve estimated from their prediction performance on left-out samples of bootstrap sampling. This classification system is tested on a sex classification problem, where it yields over 90% classification rates for unseen subjects. The proposed classification method is also compared with other commonly used classification algorithms, with favorable results. These results illustrate that the methods built upon information jointly extracted from multi-parametric images have the potential to perform individual classification with high sensitivity and specificity.
Bleeding detection in wireless capsule endoscopy using adaptive colour histogram model and support vector classification
Michal W. Mackiewicz, Mark Fisher, Crawford Jamieson
Wireless Capsule Endoscopy (WCE) is a colour imaging technology that enables detailed examination of the interior of the gastrointestinal tract. A typical WCE examination takes ~ 8 hours and captures ~ 40,000 useful images. After the examination, the images are viewed as a video sequence, which generally takes a clinician over an hour to analyse. The manufacturers of the WCE provide certain automatic image analysis functions e.g. Given Imaging offers in their Rapid Reader software: The Suspected Blood Indicator (SBI), which is designed to report the location in the video of areas of active bleeding. However, this tool has been reported to have insufficient specificity and sensitivity. Therefore it does not free the specialist from reviewing the entire footage and was suggested only to be used as a fast screening tool. In this paper we propose a method of bleeding detection that uses in its first stage Hue-Saturation-Intensity colour histograms to track a moving background and bleeding colour distributions over time. Such an approach addresses the problem caused by drastic changes in blood colour distribution that occur when it is altered by gastrointestinal fluids and allow detection of other red lesions, which although are usually "less red" than fresh bleeding, they can still be detected when the difference between their colour distributions and the background is large enough. In the second stage of our method, we analyse all candidate blood frames, by extracting colour (HSI) and texture (LBP) features from the suspicious image regions (obtained in the first stage) and their neighbourhoods and classifying them using Support Vector Classifier into Bleeding, Lesion and Normal classes. We show that our algorithm compares favourably with the SBI on the test set of 84 full length videos.
Statistical modeling and MAP estimation for body fat quantification with MRI ratio imaging
We are developing small animal imaging techniques to characterize the kinetics of lipid accumulation/reduction of fat depots in response to genetic/dietary factors associated with obesity and metabolic syndromes. Recently, we developed an MR ratio imaging technique that approximately yields lipid/{lipid + water}. In this work, we develop a statistical model for the ratio distribution that explicitly includes a partial volume (PV) fraction of fat and a mixture of a Rician and multiple Gaussians. Monte Carlo hypothesis testing showed that our model was valid over a wide range of coefficient of variation of the denominator distribution (c.v.: 0-0:20) and correlation coefficient among the numerator and denominator (&rgr; 0-0.95), which cover the typical values that we found in MRI data sets (c.v.: 0:027-0:063, &rgr;: 0:50-0:75). Then a maximum a posteriori (MAP) estimate for the fat percentage per voxel is proposed. Using a digital phantom with many PV voxels, we found that ratio values were not linearly related to PV fat content and that our method accurately described the histogram. In addition, the new method estimated the ground truth within +1.6% vs. +43% for an approach using an uncorrected ratio image, when we simply threshold the ratio image. On the six genetically obese rat data sets, the MAP estimate gave total fat volumes of 279 ± 45mL, values ≈ 21% smaller than those from the uncorrected ratio images, principally due to the non-linear PV effect. We conclude that our algorithm can increase the accuracy of fat volume quantification even in regions having many PV voxels, e.g. ectopic fat depots.
A variational method for automatic localization of the most pathological ROI in the knee cartilage
Osteoarthritis (OA) is a degenerative joint disease characterized by degradation of the articular cartilage, and is a major cause of disability. At present, there is no cure for OA and currently available treatments are directed towards relief of symptoms. Recently it was shown that cartilage homogeneity visualized by MRI and representing the biochemical changes undergoing in the cartilage is a potential marker for early detection of knee OA. In this paper based on homogeneity we present an automatic technique, embedded in a variational framework, for localization of a region of interest in the knee cartilage that best indicates where the pathology of the disease is dominant. The technique is evaluated on 283 knee MR scans. We show that OA affects certain areas of the cartilage more distinctly, and these are more towards the peripheral region of the cartilage. We propose that this region in the cartilage corresponds anatomically to the area covered by the meniscus in healthy subjects. This finding may provide valuable clues in the pathology and the etiology of OA and thereby may improve treatment efficacy. Moreover our method is generic and may be applied to other organs as well.
Motion blur detection in radiographs
Hui Luo, William J. Sehnert, Jacquelyn S. Ellinwood, et al.
Image blur introduced by patient motion is one of the most frequently cited reasons for image rejection in radiographic diagnostic imaging. The goal of the present work is to provide an automated method for the detection of anatomical motion blur in digital radiographic images to help improve image quality and facilitate workflow in the radiology department. To achieve this goal, the method first reorients the image to a predetermined hanging protocol. Then it locates the primary anatomy in the radiograph and extracts the most indicative region for motion blur, i.e., the region of interest (ROI). The third step computes a set of motion-sensitive features from the extracted ROI. Finally, the extracted features are evaluated by using a classifier that has been trained to detect motion blur. Preliminary experiments show promising results with 86% detection sensitivity, 72% specificity, and an overall accuracy of 76%.
Registration II: Methodology
icon_mobile_dropdown
Hybrid physics-based elastic image registration using approximating splines
Stefan Wörz, Karl Rohr
We introduce a new hybrid physics-based approach for elastic image registration using approximating splines. As underlying deformation model we employ Gaussian elastic body splines (GEBS), which are an analytic solution of the Navier equation under Gaussian forces and are represented by matrix-valued basis functions. Our approach is formulated as an energy-minimizing functional that incorporates both landmark and intensity information as well as a regularization based on GEBS. We also include landmark localization uncertainties represented by weight matrices. Since the approach is based on a physical deformation model, cross-effects in elastic deformations can be handled. We demonstrate the applicability of our scheme based on MR images of the human brain. It turns out that the new scheme is superior to a pure landmark-based as well as a pure intensity-based scheme.
On the development of a new non-rigid image registration using deformation based grid generation
In this paper, we present the latest results of the development of a novel non-rigid image registration method (NiRuDeGG) using a well-established mathematical framework known as the deformation based grid generation. The deformation based grid generation method is able to generate a grid with desired grid density distribution which is free from grid folding. This is achieved by devising a positive monitor function describing the anticipated grid density in the computational domain. Based on it, we have successfully developed a new non-rigid image registration method, which has many advantages. Firstly, the functional to be optimized consists of only one term, a similarity measure. Thus, no regularization functional is required in this method. In particular, there is no weight to balance the regularization functional and the similarity functional as commonly required in many non-rigid image registration methods. Nevertheless, the regularity (no mesh folding) of the resultant deformation is theoretically guaranteed by controlling the Jacobian determinant of the transformation. Secondly, since no regularization term is introduced in the functional to be optimized, the resultant deformation field is highly flexible that large deformation frequently experienced in inter-patient or image-atlas registration tasks can be accurately estimated. Detailed description of the deformation based grid generation, a least square finite element (LSFEM) solver for the underlying div-curl system, and a fast div-curl solver approximating the LSFEM solution using inverse filtering, along with several 2D and 3D experimental results are presented.
A novel framework for multi-modal intensity-based similarity measures based on internal similarity
We present a novel framework for describing intensity-based multi-modal similarity measures. Our framework is based around a concept of internal, or self, similarity. Firstly the locations of multiple regions or patches which are "similar" to each other are identified within a single image. The term "similar" is used here to represent a generic intra-modal similarity measure. Then if we examine a second image in the same locations, and this image is registered to the first image, we should find that the patches in these locations are also "similar", though the actual features in the patches when compared between the images could be very different. We propose that a measure based on this principle could be used as an inter-modal similarity measure because, as the two images become increasingly misregistered then the patches within the second image should become increasingly dissimilar. Therefore, our framework results in an inter-modal similarity measure by using two intra-modal similarity measures applied separately within each image. In this paper we describe how popular multi-modal similarity measures such as mutual information can be described within this framework. In addition the framework has the potential to allow the formation of novel similarity measures which can register using regional information, rather than individual pixel/voxel intensities. An example similarity measure is produced and its ability to guide a registration algorithm is investigated. Registration experiments are carried out using three datasets. The pairs of images to be registered were specifically chosen as they were expected to challenge (i.e. result in misregistrations) standard intensity-based measures, such as mutual information. The images include synthetic data, cadaver data and clinical data and cover a range of modalities. Our experiments show that our proposed measure is able to achieve accurate registrations where standard intensity-based measures, such as mutual information, fail.
Volume preserving image registration via a post-processing stage
Reinhard Hameeteman, Jifke F. Veenland, Wiro J. Niessen
In this paper a method to remove the divergence from a vector field is presented. When applied to a displacement field, this will remove all local compression and expansion. The method can be used as a post-processing step for (unconstrained) registered images, when volume changes in the deformation field are undesired. The method involves solving Poisson's equation for a large system. Algorithms to solve such systems include Fourier analysis and Cyclic Reduction. These solvers are vastly applied in the field of fluid dynamics, to compensate for numerical errors in calculated velocity fields. The application to medical image registration as described in this paper, has to our knowledge not been done before. To show the effect of the method, it is applied to the registration of both synthetic data and dynamic MR series of the liver. The results show that the divergence in the displacement field can be reduced by a factor of 10-1000 and that the accuracy of the registration increases.
Improved CT and MR image registration with the introduction of a dual-modality contrast agent: performance assessment using quantitative and information theoretic methods
Jeremy D. P. Hoisak, Jinzi Zheng, Christine Allen, et al.
The ability of computed tomography (CT) and magnetic resonance (MR) imaging to visualize and discriminate between normal and diseased tissues is improved with contrast agents, which are designed to differentially accumulate in tissues and modify their inherent imaging signal. Conventional contrast agents are limited to a single modality and require fast acquisitions due to rapid clearance following injection. Encapsulation of iohexol and gadoteridol within a nano-engineered liposome has been achieved and can increase their in vivo half-life to several days. We hypothesize that the persistence of this contrast agent in vivo, and the simultaneous co-localized contrast enhancement across modalities will improve longitudinal image registration. This work investigates the in vivo registration performance of the dual-modality contrast agent under realistic conditions. Previous characterizations of single-modality contrast agents were limited to qualitative inspections of signal intensity enhancement. We present quantitative, information theoretic methods for assessing image registration performance. The effect of increased localized contrast upon the mutual information of the MR and CT image sets was shown to increase post-injection. Images registered post- injection had a decreased registration error compared with pre-contrast images. Performance was maintained over extended time frames, contrast agent concentrations, and with decreased field-of-view. This characterization allows optimization of the contrast agent against desired performance for a given imaging task. The ability to perform robust longitudinal image registration is essential for pre-clinical investigations of tumor development, monitoring of therapy response, and therapy guidance over multiple fractions where registration of online cone-beam CT to planning CT and MR is necessary.
Conditional statistical model building
We present a new statistical deformation model suited for parameterized grids with different resolutions. Our method models the covariances between multiple grid levels explicitly, and allows for very efficient fitting of the model to data on multiple scales. The model is validated on a data set consisting of 62 annotated MR images of Corpus Callosum. One fifth of the data set was used as a training set, which was non-rigidly registered to each other without a shape prior. From the non-rigidly registered training set a shape prior was constructed by performing principal component analysis on each grid level and using the results to construct a conditional shape model, conditioning the finer parameters with the coarser grid levels. The remaining shapes were registered with the constructed shape prior. The dice measures for the registration without prior and the registration with a prior were 0.875 ± 0.042 and 0.8615 ± 0.051, respectively.
Cardiovascular Applications
icon_mobile_dropdown
Nonrigid registration of carotid ultrasound and MR images using a "twisting and bending" model
Nuwan D. Nanayakkara, Bernard Chiu, Abbas Samani, et al.
Atherosclerosis at the carotid bifurcation resulting in cerebral emboli is a major cause of ischemic stroke. Most strokes associated with carotid atherosclerosis can be prevented by lifestyle/dietary changes and pharmacological treatments if identified early by monitoring carotid plaque changes. Plaque composition information from magnetic resonance (MR) carotid images and dynamic characteristics information from 3D ultrasound (US) are necessary for developing and validating US imaging tools to identify vulnerable carotid plaques. Combining these images requires nonrigid registration to correct the non-linear miss-alignments caused by relative twisting and bending in the neck due to different head positions during the two image acquisitions sessions. The high degree of freedom and large number of parameters associated with existing nonrigid image registration methods causes several problems including unnatural plaque morphology alteration, computational complexity, and low reliability. Our approach was to model the normal movement of the neck using a "twisting and bending model" with only six parameters for nonrigid registration. We evaluated our registration technique using intra-subject in-vivo 3D US and 3D MR carotid images acquired on the same day. We calculated the Mean Registration Error (MRE) between the segmented vessel surfaces in the target image and the registered image using a distance-based error metric after applying our "twisting bending model" based nonrigid registration algorithm. We achieved an average registration error of 1.33±0.41mm using our nonrigid registration technique. Visual inspection of segmented vessel surfaces also showed a substantial improvement of alignment with our non-rigid registration technique.
Deformable registration of 3D vessel structures to a single projection image
Darko Zikic, Martin Groher, Ali Khamene, et al.
Alignment of angiographic preoperative 3D scans to intraoperative 2D projections is an important issue for 3D depth perception and navigation during interventions. Currently, in a setting where only one 2D projection is available, methods employing a rigid transformation model present the state of the art for this problem. In this work, we introduce a method capable of deformably registering 3D vessel structures to a respective single projection of the scene. Our approach addresses the inherent ill-posedness of the problem by incorporating a priori knowledge about the vessel structures into the formulation. We minimize the distance between the 2D points and corresponding projected 3D points together with regularization terms encoding the properties of length preservation of vessel structures and smoothness of deformation. We demonstrate the performance and accuracy of the proposed method by quantitative tests on synthetic examples as well as real angiographic scenes.
3D inters-subject cardiac registration using 4D information
Alfredo Lopez, Karl D. Fritscher, Thomas Trieb, et al.
In this paper we present a new approach for the registration of cardiac 4D image sequences of different subjects, where we assume that a temporal association between the sequences is given. Moreover, we allow for one (or two) selected pair(s) of associated points in time of both sequences, which we call the bridging points in time, the use of additional information such as the semi-automatic segmentation of the investigated structure. We establish the 3D inter-subject registration for all other pairs of points in time exploiting (1) the inter-subject registration for the bridging pair of points in time, (2) the intra-subject motion calculation in both sequences with respect to the bridging pair, and (3) the concatenation of the obtained transformations. We formulate a cost functional integrating the similarity measures comparing the images of the bridging pair(s) of points in time and of the current pair of points in time, respectively. We evaluated our algorithm on 8 healthy volunteers leading to 28 inter-subject combinations and we analyze the behaviour for different parameter settings weighting differently the involved pairs of points in time. The approach based on the bridging pairs outperforms a direct 3D registration of corresponding points in time, in particular in the right ventricle we gain up to 33% in registration accuracy. Starting with a cost functional taking into account the similarity at the first bridging point in time, the results improve stepwise by integrating, firstly, information from the current pair of points in time and secondly, from a second bridging point in time. Our results also show a steep rise of the importance of regularization on the registration accuracy when registering the current point in time with our procedure (17% gain in accuracy) with respect to a direct registration in the bridging point (less than 1%). However, regularization during intra-sequence registration had only minor effects on the accuracy of our registration procedure.
Level set segmentation of the heart from 4D phase contrast MRI
Blood flow properties in the heart can be examined non invasively by means of Phase Contrast MRI (PC MRI), an imaging technique that provides not only morphology images but also velocity information. We present a novel feature combination for level set segmentation of the heart's cavities in multidirectional 4D PC MRI data. The challenge in performing the segmentation task successfully in this context is first of all the bad image quality, as compared to classical MRI. As generally in heart segmentation, the intra and inter subject variability of the heart has to be coped with as well. The central idea of our approach is to integrate a set of essentially differing sources of information into the segmentation process to make it capable of handling qualitatively bad and highly varying data. To the best of our knowledge our system is the first to concurrently incorporate a flow measure as well as a priori shape knowledge into a level set framework in addition to the commonly used edge and curvature information. The flow measure is derived from PC MRI velocity data. As shape knowledge we use a 3D shape of the respective cavity. We validated our system design by a series of qualitative performance tests. The combined use of shape knowledge and a flow measure increases segmentation quality compared to results obtained by using only one of those features. A first clinical study was performed on two 4D datasets, from which we segmented the left ventricle and atrium. The segmentation results were examined by an expert and judged suitable for use in clinical practice.
Segmentation of myocardial perfusion MR sequences with multi-band active appearance models driven by spatial and temporal features
N. Baka, J. Milles, E. A. Hendriks, et al.
This work investigates knowledge driven segmentation of cardiac MR perfusion sequences. We build upon previous work on multi-band AAMs to integrate into the segmentation both spatial priors about myocardial shape as well as temporal priors about characteristic perfusion patterns. Different temporal and spatial features are developed without a strict need for temporal correspondence across the image sequences. We also investigate which combination of spatial and temporal features yields the best segmentation performance. Our evaluation criteria were boundary errors wrt manual segmentations, area overlap, and convergence envelope. From a quantitative evaluation on 19 perfusion studies, we conclude that a combination of the maximum intensity projection feature and gradient orientation map yields the best segmentation performance, with an average point-to-curve error of 0.9-1 pixel wrt manual contours. We also conclude that addition of different temporal features does not necessarily increase performance.
Four-chamber heart modeling and automatic segmentation for 3D cardiac CT volumes
Multi-chamber heart segmentation is a prerequisite for quantification of the cardiac function. In this paper, we propose an automatic heart chamber segmentation system. There are two closely related tasks to develop such a system: heart modeling and automatic model fitting to an unseen volume. The heart is a complicated non-rigid organ with four chambers and several major vessel trunks attached. A flexible and accurate model is necessary to capture the heart chamber shape at an appropriate level of details. In our four-chamber surface mesh model, the following two factors are considered and traded-off: 1) accuracy in anatomy and 2) easiness for both annotation and automatic detection. Important landmarks such as valves and cusp points on the interventricular septum are explicitly represented in our model. These landmarks can be detected reliably to guide the automatic model fitting process. We also propose two mechanisms, the rotation-axis based and parallel-slice based resampling methods, to establish mesh point correspondence, which is necessary to build a statistical shape model to enforce priori shape constraints in the model fitting procedure. Using this model, we develop an efficient and robust approach for automatic heart chamber segmentation in 3D computed tomography (CT) volumes. Our approach is based on recent advances in learning discriminative object models and we exploit a large database of annotated CT volumes. We formulate the segmentation as a two step learning problem: anatomical structure localization and boundary delineation. A novel algorithm, Marginal Space Learning (MSL), is introduced to solve the 9-dimensional similarity transformation search problem for localizing the heart chambers. After determining the pose of the heart chambers, we estimate the 3D shape through learning-based boundary delineation. Extensive experiments demonstrate the efficiency and robustness of the proposed approach, comparing favorably to the state-of-the-art. This is the first study reporting stable results on a large cardiac CT dataset with 323 volumes. In addition, we achieve a speed of less than eight seconds for automatic segmentation of all four chambers.
Segmentation of the heart and major vascular structures in cardiovascular CT images
Segmentation of organs in medical images can be successfully performed with shape-constrained deformable models. A surface mesh is attracted to detected image boundaries by an external energy, while an internal energy keeps the mesh similar to expected shapes. Complex organs like the heart with its four chambers can be automatically segmented using a suitable shape variablility model based on piecewise affine degrees of freedom. In this paper, we extend the approach to also segment highly variable vascular structures. We introduce a dedicated framework to adapt an extended mesh model to freely bending vessels. This is achieved by subdividing each vessel into (short) tube-shaped segments ("tubelets"). These are assigned to individual similarity transformations for local orientation and scaling. Proper adaptation is achieved by progressively adapting distal vessel parts to the image only after proximal neighbor tubelets have already converged. In addition, each newly activated tubelet inherits the local orientation and scale of the preceeding one. To arrive at a joint segmentation of chambers and vasculature, we extended a previous model comprising endocardial surfaces of the four chambers, the left ventricular epicardium, and a pulmonary artery trunk. Newly added are the aorta (ascending and descending plus arch), superior and inferior vena cava, coronary sinus, and four pulmonary veins. These vessels are organized as stacks of triangulated rings. This mesh configuration is most suitable to define tubelet segments. On 36 CT data sets reconstructed at several cardiac phases from 17 patients, segmentation accuracies of 0.61-0.80mm are obtained for the cardiac chambers. For the visible parts of the newly added great vessels, surface accuracies of 0.47-1.17mm are obtained (larger errors are asscociated with faintly contrasted venous structures).
AdaBoost classification for model-based segmentation of the outer wall of the common carotid artery in CTA
A novel 2D slice based automatic method for model based segmentation of the outer vessel wall of the common carotid artery in CTA data set is introduced. The method utilizes a lumen segmentation and AdaBoost, a fast and robust machine learning algorithm, to initially classify (mark) regions outside and inside the vessel wall using the distance from the lumen and intensity profiles sampled radially from the gravity center of the lumen. A similar method using the distance from the lumen and the image intensity as features is used to classify calcium regions. Subsequently, an ellipse shaped deformable model is fitted to the classification result. The method has achieved smaller detection error than the inter observer variability, and the method is robust against variation of the training data sets.
Image Restoration and Enhancement
icon_mobile_dropdown
Image intensity standardization in 3D rotational angiography and its application to vascular segmentation
Knowledge-based vascular segmentation methods typically rely on a pre-built training set of segmented images, which is used to estimate the probability of each voxel to belong to a particular tissue. In 3D Rotational Angiography (3DRA) the same tissue can correspond to different intensity ranges depending on the imaging device, settings and contrast injection protocol. As a result, pre-built training sets do not apply to all images and the best segmentation results are often obtained when the training set is built specifically for each individual image. We present an Image Intensity Standardization (IIS) method designed to ensure a correspondence between specific tissues and intensity ranges common to every image that undergoes the standardization process. The method applies a piecewise linear transformation to the image that aligns the intensity histogram to the histogram taken as reference. The reference histogram has been selected from a high quality image not containing artificial objects such as coils or stents. This is a pre-processing step that allows employing a training set built on a limited number of standardized images for the segmentation of standardized images which were not part of the training set. The effectiveness of the presented IIS technique in combination with a well-validated knowledge-based vasculature segmentation method is quantified on a variety of 3DRA images depicting cerebral arteries and intracranial aneurysms. The proposed IIS method offers a solution to the standardization of tissue classes in routine medical images and effectively improves automation and usability of knowledge-based vascular segmentation algorithms.
Fast bias field reduction by localized Lloyd-Max quantization
R. Hanel, K. J. Batenburg, S. De. De Backer, et al.
Bias field reduction is a common problem in medical imaging. A bias field usually manifests itself as a smooth intensity variation across the image. The resulting image inhomogeneity is a severe problem for posterior image processing and analysis techniques such as registration or segmentation. In this paper, we present a fast debiasing technique based on localized Lloyd-Max quantization. Thereby, the local bias is modelled as a multiplicative field and is assumed to be slowly varying. The method is based on the assumption that the local, undegraded histogram is characterized by a limited number of gray values. The goal is then to find the discrete intensity values such that spreading those values according to the local bias field reproduces the global histogram as good as possible. We show that our method is capable of efficiently reducing (even strong) bias fields in 3D volumes in only a few seconds.
Feature-preserving artifact removal from dermoscopy images
Howard Zhou, Mei Chen, Richard Gass, et al.
Dermoscopy, also called surface microscopy, is a non-invasive imaging procedure developed for early screening of skin cancer. With recent advance in skin imaging technologies and image processing techniques, there has been increasing interest in computer-aided diagnosis of skin cancer from dermoscopy images. Such diagnosis requires the identification of over one hundred cutaneous morphological features. However, computer procedures designed for extracting and classifying these intricate features can be distracted by the presence of artifacts like hair, ruler markings, and air bubbles. Therefore, reliable artifact removal is an important pre-processing step for improving the performance of computer-aided diagnosis of skin cancer. In this paper, we present a new scheme that automatically detects and removes hairs and ruler markings from dermoscopy images. Moreover, our method also addresses the issue of preserving morphological features during artifact removal. The key components of this method include explicit curvilinear structure detection and modeling, as well as feature guided exemplar-based inpainting. We experiment on a number of dermoscopy images and demonstrate that our method produces superior results compared to existing techniques.
A quantitative performance measure for a clinical evaluation of comb structure removal algorithms in flexible endoscopy
Modern techniques for technical inspection as well as medical diagnostics and therapy in keyhole-surgery scenarios make use of flexible endoscopes. Common to both application fields are very small natural or manmade entry points to the observed scene, as well as the complexity of the hollow itself. These make the use of rigid lens-based endoscopes or tip chip videoscopes impossible. Due to the fact that the fiber-optic image guide of a flexible endoscope introduces a comb structure to the acquired images, many research has been devoted to algorithms for an effective removal of such artifacts. Oftentimes, this research has been motivated by the fact, that the comb structure prevents an application of some well-established methods offered by the computer vision and image processing community. Unfortunately, the performance of the presented approaches are commonly visually evaluated or with respect to proprietary, non-standardized metrics. Thus, the performances of individual algorithms are hard to compare with each other. For this reasons, we propose a performance measure for fiber-optic imaging devices that has been motivated by the physics of optics. In this field, an optical system is frequently described by linear systems theory and the system's quality can be expressed by its transfer function. The determination of this transfer function has been standardized by the ISO for lens based imaging systems and represents a widely accepted measure for the quality of such systems. In this contribution, we present methods that account for fiber-optic imaging systems and thus enable a standardized performance evaluation. Finally, we demonstrate its use by comparing two recent state of the art comb structure removal algorithms, each of them being a representative of a spatial and a frequency domain method, respectively.
Digital mammogram enhancement based on ROI enhancement and background suppression
This paper presents a framework for mammogram enhancement that is based on a selective enhancement technique. Several enhancement algorithms under this framework are developed, which include weighted mean gray value- and fuzzy cross-over point-based thresholding methods, algorithm fusion, iterative enhancement method, and statistical decision theory-based techniques. Using various abnormal mammograms, the presented algorithms prove to be more robust and yield superior performance when compared with six representative enhancement approaches available in the literature.
Liver Applications
icon_mobile_dropdown
Adaptive directional region growing segmentation of the hepatic vasculature
Accurate analysis of the hepatic vasculature is of great importance for many medical applications, such as liver surgical planning and diagnosis of tumors and/or vascular diseases. Vessel segmentation is a pivotal step for the morphological and topological analysis of the vascular systems. Physical imaging limitations together with the inherent geometrical complexity of the vessels make the problem challenging. In this paper, we propose a series of methods and techniques that separate and segment the portal vein and the hepatic vein from CT images, and extract the centerlines of both vessel trees. We compare the results obtained with our iterative segmentation-and-reconnection approach with those obtained with a traditional region growing method, and we show that our results are substantially better.
Quantitative growth measurement of lesions in hepatic interval CT exams
Saradwata Sarkar, Ramkrishnan Narayanan, Hyunjin Park, et al.
Standard clinical radiological techniques for determining lesion volume changes in interval exams are, as far as we know, quantitatively non-descriptive or approximate at best. We investigate two new registration based methods that help sketch an improved quantitative picture of lesion volume changes in hepatic interval CT exams. The first method, Jacobian Integration, employs a constrained Thin Plate Spline warp to compute the deformation of the lesion of interest over the intervals. The resulting jacobian map of the deformation is integrated to yield the net lesion volume change. The technique is fast, accurate and requires no segmentation, but is sensitive to misregistration. The second scheme uses a Weighted Gray Value Difference image of two registered interval exams to estimate the change in lesion volume. A linear weighting and trimming curve is used to accurately account for the contribution of partial voxels. This technique is insensitive to slight misregistration and useful in analyzing simple lesions with uniform contrast or lesions with insufficient mutual information to allow the computation of an accurate warp. The methods are tested on both synthetic and in vivo liver lesions and results are evaluated against estimates obtained through careful manual segmentation of the lesions. Our findings so far have given us reason to believe that the estimators are reliable. Further experiments on numerous in vivo lesions will probably establish the improved efficacy of these methods in supporting earlier detection of new disease or conversion from stable to progressive disease in comparison to existing clinical estimation techniques.
Liver segmentation combining Gabor filtering and traditional vector field snake
This paper presents a study of a more accurately propagating deformable contour for outlining the liver in a Computed Tomography image of the abdomen, relying on the idea that a deformable parametric snake will propagate more accurately to the correct edges of an image when applied to textural information of the image as opposed to simple gray level information. The texture information is quantified using a set of Gabor filters and various methods of curve deformation are investigated, including a traditional vector field, gradient vector flow, and an expanding level-set method. Given the relative similarity in gray values of adjacent soft tissues, we found that a deformation algorithm that provides too large a capture range would be easily distracted by nearby values and therefore unsuitable for the particular task of segmenting the liver. Our results demonstrate both a general increase in performance of snake segmentation across the dataset as well as a significant regional improvement in accuracy, particularly in images corresponding with the top of the liver.
Pulmonary Applications
icon_mobile_dropdown
Lung lobe modeling and segmentation with individualized surface meshes
Thomas Blaffert, Hans Barschdorf, Jens von Berg, et al.
An automated segmentation of lung lobes in thoracic CT images is of interest for various diagnostic purposes like the quantification of emphysema or the localization of tumors within the lung. Although the separating lung fissures are visible in modern multi-slice CT-scanners, their contrast in the CT-image often does not separate the lobes completely. This makes it impossible to build a reliable segmentation algorithm without additional information. Our approach uses general anatomical knowledge represented in a geometrical mesh model to construct a robust lobe segmentation, which even gives reasonable estimates of lobe volumes if fissures are not visible at all. The paper describes the generation of the lung model mesh including lobes by an average volume model, its adaptation to individual patient data using a special fissure feature image, and a performance evaluation over a test data set showing an average segmentation accuracy of 1 to 3 mm.
Robust system for human airway-tree segmentation
Robust and accurate segmentation of the human airway tree from multi-detector computed-tomography (MDCT) chest scans is vital for many pulmonary-imaging applications. As modern MDCT scanners can detect hundreds of airway tree branches, manual segmentation and semi-automatic segmentation requiring significant user intervention are impractical for producing a full global segmentation. Fully-automated methods, however, may fail to extract small peripheral airways. We propose an automatic algorithm that searches the entire lung volume for airway branches and poses segmentation as a global graph-theoretic optimization problem. The algorithm has shown strong performance on 23 human MDCT chest scans acquired by a variety of scanners and reconstruction kernels. Visual comparisons with adaptive region-growing results and quantitative comparisons with manually-defined trees indicate a high sensitivity to peripheral airways and a low false-positive rate. In addition, we propose a suite of interactive segmentation tools for cleaning and extending critical areas of the automatically segmented result. These interactive tools have potential application for image-based guidance of bronchoscopy to the periphery, where small, terminal branches can be important visual landmarks. Together, the automatic segmentation algorithm and interactive tool suite comprise a robust system for human airway-tree segmentation.
Voxel classification based airway tree segmentation
This paper presents a voxel classification based method for segmenting the human airway tree in volumetric computed tomography (CT) images. In contrast to standard methods that use only voxel intensities, our method uses a more complex appearance model based on a set of local image appearance features and Kth nearest neighbor (KNN) classification. The optimal set of features for classification is selected automatically from a large set of features describing the local image structure at several scales. The use of multiple features enables the appearance model to differentiate between airway tree voxels and other voxels of similar intensities in the lung, thus making the segmentation robust to pathologies such as emphysema. The classifier is trained on imperfect segmentations that can easily be obtained using region growing with a manual threshold selection. Experiments show that the proposed method results in a more robust segmentation that can grow into the smaller airway branches without leaking into emphysematous areas, and is able to segment many branches that are not present in the training set.
4DCT image-based lung motion field extraction and analysis
Respiratory motion is a complicating factor in radiation therapy, tumor ablation, and other treatments of the thorax and upper abdomen. In most cases, the treatment requires a demanding knowledge of the location of the organ under investigation. One approach to reduce the uncertainty of organ motion caused by breathing is to use prior knowledge of the breathing motion. In this work, we extract lung motion fields of seven patients in 4DCT inhale-exhale images using an iterative shape-constrained deformable model approach. Since data was acquired for radiotherapy planning, images of the same patient over different weeks of treatment were available. Although, respiratory motion shows a repetitive character, it is well-known that patient's variability in breathing pattern impedes motion estimation. A detailed motion field analysis is performed in order to investigate the reproducibility of breathing motion over the weeks of treatment. For that purpose, parameters being significant for breathing motion are derived. The analysis of the extracted motion fields provides a basis for a further breathing motion prediction. Patient-specific motion models are derived by averaging the extracted motion fields of each individual patient. The obtained motion models are adapted to each patient in a leave-one-out test in order to simulate motion estimation to unseen data. By using patient-specific mean motion models 60% of the breathing motion can be captured on average.
The evaluation of a highly automated mixture model based technique for PET tumor volume segmentation
Michalis Aristophanous, Charles A. Pelizzari
PET-based tumor volume segmentation techniques are under investigation in recent years due to the increased utilization of FDG-PET imaging in radiation therapy. We have taken the approach of using a Gaussian mixture model (GMM) to model the image intensity distribution of a selected 3D region that completely covers the tumor, called the "analysis region". The modeling is performed with a predetermined number of Gaussian classes and results in a classification of every voxel into one of these classes. The classes are then grouped together to obtain the tumor volume. The only user interaction required is the selection of the "analysis region" and then the algorithm proceeds automatically to initialize the parameters of the different classes and finds the maximum likelihood estimate with expectation maximization. We used 13 clinical and 19 phantom cases to evaluate the precision and accuracy of the segmentation. Reproducibility was within 10% of the average tumor volume estimate and accuracy was ±35% of the true tumor volume and better when compared to two other proposed techniques. The GMM segmentation is extremely user friendly with good precision and accuracy. It has shown great potential to be used in the clinical environment.
Segmentation II: Applications
icon_mobile_dropdown
Shape priors for segmentation of the cervix region within uterine cervix images
Shelly Lotenberg, Shiri Gordon, Hayit Greenspan
The work focuses on a unique medical repository of digital Uterine Cervix images ("Cervigrams") collected by the National Cancer Institute (NCI), National Institute of Health, in longitudinal multi-year studies. NCI together with the National Library of Medicine is developing a unique web-based database of the digitized cervix images to study the evolution of lesions related to cervical cancer. Tools are needed for the automated analysis of the cervigram content to support the cancer research. In recent works, a multi-stage automated system for segmenting and labeling regions of medical and anatomical interest within the cervigrams was developed. The current paper concentrates on incorporating prior-shape information in the cervix region segmentation task. In accordance with the fact that human experts mark the cervix region as circular or elliptical, two shape models (and corresponding methods) are suggested. The shape models are embedded within an active contour framework that relies on image features. Experiments indicate that incorporation of the prior shape information augments previous results.
Use of a CT statistical deformation model for multi-modal pelvic bone segmentation
We present a segmentation algorithm using a statistical deformation model constructed from CT data of adult male pelves coupled to MRI appearance data. The algorithm allows the semi-automatic segmentation of bone for a limited population of MRI data sets. Our application is pelvic bone delineation from pre-operative MRI for image guided pelvic surgery. Specifically, we are developing image guidance for prostatectomies using the daVinci telemanipulator. Hence the use of male pelves only. The algorithm takes advantage of the high contrast of bone in CT data, allowing a robust shape model to be constructed relatively easily. This shape model can then be applied to a population of MRI data sets using a single data set that contains both CT and MRI data. The model is constructed automatically using fluid based non-rigid registration between a set of CT training images, followed by principal component analysis. MRI appearance data is imported using CT and MRI data from the same patient. Registration optimisation is performed using differential evolution. Based on our limited validation to date, the algorithm may outperform segmentation using non-rigid registration between MRI images without the use of shape data. The mean surface registration error achieved was 1.74 mm. The algorithm shows promise for use in segmentation of pelvic bone from MRI, though further refinement and validation is required. We envisage that the algorithm presented could be extended to allow the rapid creation of application specific models in various imaging modalities using a shape model based on CT data.
Prostate segmentation from 3D transrectal ultrasound using statistical shape models and various appearance models
T. Heimann, M. Baumhauer, T. Simpfendörfer, et al.
Due to the high noise and artifacts typically encountered in ultrasound images, segmenting objects from this modality is one of the most challenging tasks in medical image analysis. Model-based approaches like statistical shape models (SSMs) incorporate prior knowledge that supports object detection in case of incomplete evidence from the image data. How well the model adapts to an unseen image is primarily determined by the suitability of the used appearance model, which evaluates the goodness of fit during model evolution. In this paper, we compare two gradient profile models with a region-based approach featuring local histograms to detect the prostate in 3D transrectal ultrasound (TRUS) images. All models are used within an SSM segmentation framework with optimal surface detection for outlier removal. Evaluation was performed using cross-validation on 35 datasets. While the histogram model failed in 10 cases, both gradient models had only 2 failures and reached an average surface distance of 1.16 ± 0.38 mm in comparison with interactively generated reference contours.
Fuzzy pulmonary vessel segmentation in contrast enhanced CT data
Jens N. Kaftan, Atilla P. Kiraly, Annemarie Bakai, et al.
Pulmonary vascular tree segmentation has numerous applications in medical imaging and computer-aided diagnosis (CAD), including detection and visualization of pulmonary emboli (PE), improved lung nodule detection, and quantitative vessel analysis. We present a novel approach to pulmonary vessel segmentation based on a fuzzy segmentation concept, combining the strengths of both threshold and seed point based methods. The lungs of the original image are first segmented and a threshold-based approach identifies core vessel components with a high specificity. These components are then used to automatically identify reliable seed points for a fuzzy seed point based segmentation method, namely fuzzy connectedness. The output of the method consists of the probability of each voxel belonging to the vascular tree. Hence, our method provides the possibility to adjust the sensitivity/specificity of the segmentation result a posteriori according to application-specific requirements, through definition of a minimum vessel-probability required to classify a voxel as belonging to the vascular tree. The method has been evaluated on contrast-enhanced thoracic CT scans from clinical PE cases and demonstrates overall promising results. For quantitative validation we compare the segmentation results to randomly selected, semi-automatically segmented sub-volumes and present the resulting receiver operating characteristic (ROC) curves. Although we focus on contrast enhanced chest CT data, the method can be generalized to other regions of the body as well as to different imaging modalities.
Vessel segmentation in 3D spectral OCT scans of the retina
The latest generation of spectral optical coherence tomography (OCT) scanners is able to image 3D cross-sectional volumes of the retina at a high resolution and high speed. These scans offer a detailed view of the structure of the retina. Automated segmentation of the vessels in these volumes may lead to more objective diagnosis of retinal vascular disease including hypertensive retinopathy, retinopathy of prematurity. Additionally, vessel segmentation can allow color fundus images to be registered to these 3D volumes, possibly leading to a better understanding of the structure and localization of retinal structures and lesions. In this paper we present a method for automatically segmenting the vessels in a 3D OCT volume. First, the retina is automatically segmented into multiple layers, using simultaneous segmentation of their boundary surfaces in 3D. Next, a 2D projection of the vessels is produced by only using information from certain segmented layers. Finally, a supervised, pixel classification based vessel segmentation approach is applied to the projection image. We compared the influence of two methods for the projection on the performance of the vessel segmentation on 10 optic nerve head centered 3D OCT scans. The method was trained on 5 independent scans. Using ROC analysis, our proposed vessel segmentation system obtains an area under the curve of 0.970 when compared with the segmentation of a human observer.
Lymph node segmentation on CT images by a shape model guided deformable surface method
Daniel Maleike, Michael Fabel, Ralf Tetzlaff, et al.
With many tumor entities, quantitative assessment of lymph node growth over time is important to make therapy choices or to evaluate new therapies. The clinical standard is to document diameters on transversal slices, which is not the best measure for a volume. We present a new algorithm to segment (metastatic) lymph nodes and evaluate the algorithm with 29 lymph nodes in clinical CT images. The algorithm is based on a deformable surface search, which uses statistical shape models to restrict free deformation. To model lymph nodes, we construct an ellipsoid shape model, which strives for a surface with strong gradients and user-defined gray values. The algorithm is integrated into an application, which also allows interactive correction of the segmentation results. The evaluation shows that the algorithm gives good results in the majority of cases and is comparable to time-consuming manual segmentation. The median volume error was 10.1% of the reference volume before and 6.1% after manual correction. Integrated into an application, it is possible to perform lymph node volumetry for a whole patient within the 10 to 15 minutes time limit imposed by clinical routine.
A novel shape prior based segmentation of touching or overlapping ellipse-like nuclei
Cell nuclei segmentation is a key issue in automatic cell image analysis for nuclear malignancy. However, due to the complexity of microscopic images, it is usually not easy to obtain satisfied segmentation results, especially on the separation of touching or overlapping nuclei. We propose a method to separate overlapping nuclei whose shapes are similar to ellipses, even if they are tightly clustered and no edge is present where they touch. As a class-specific approach, it introduces a statistical shape model as an extra constraint within the energy functional that measures the homogeneity of regional intensity. The desired contours of each nucleus can be obtained by minimizing this energy functional. The proposed algorithm has been tested on human cervical nuclei images. Experiment results show that our method can separate touching or overlapping ellipse-like nuclei from each other accurately, and the tests on noisy and textured nuclei images also demonstrate its robustness. The resulting segmentation contours are ellipses in different sizes and directions, therefore the shapes of the nuclei have been preserved to a certain degree. The algorithm can be naturally extended to color images, and also has the potential to deal with the separation for overlapping nuclei of other shapes.
Posters: Classification and Pattern Recognition
icon_mobile_dropdown
A machine learning approach for body part recognition based on CT images
Keigo Nakamura, Yuanzhong Li, Wataru Ito, et al.
Body part recognition based on CT slice images is very important for many applications in PACS and CAD systems. In this paper, we propose a novel approach that can recognize which body part a slice image belongs to robustly. We focus on how to effectively express and use the unique statistical information of the correlation between the CT value and the position information of each body part. We apply the machine learning method AdaBoost to express and use this statistical information. Our approach consists of a training process and a recognition process. In the training process, we first define the whole body using a set of specific classes to ensure that training images in the same class have a high similarity, and prepare a training image set (positive samples and negative samples) for each class. Second, the training images are normalized to a fixed size and rotation in each class. Third, features are calculated for each normalized training image. Finally, AdaBoosted histogram classifiers are trained. After the training process, each class has its own classifiers. In the recognition process, given a series of CT images, the scores of all classes for each slice image are calculated based on the classifiers obtained in the training process. Then, based on the scores of each slice and a simple model of body part sequence continuity, we use dynamic programming (DP) to eliminate false recognition results. Experimental results on 440 unknown series including lesions show that our approach has high a recognition rate.
Personal identification based on blood vessels of retinal fundus images
Biometric technique has been implemented instead of conventional identification methods such as password in computer, automatic teller machine (ATM), and entrance and exit management system. We propose a personal identification (PI) system using color retinal fundus images which are unique to each individual. The proposed procedure for identification is based on comparison of an input fundus image with reference fundus images in the database. In the first step, registration between the input image and the reference image is performed. The step includes translational and rotational movement. The PI is based on the measure of similarity between blood vessel images generated from the input and reference images. The similarity measure is defined as the cross-correlation coefficient calculated from the pixel values. When the similarity is greater than a predetermined threshold, the input image is identified. This means both the input and the reference images are associated to the same person. Four hundred sixty-two fundus images including forty-one same-person's image pairs were used for the estimation of the proposed technique. The false rejection rate and the false acceptance rate were 9.9×10-5% and 4.3×10-5%, respectively. The results indicate that the proposed method has a higher performance than other biometrics except for DNA. To be used for practical application in the public, the device which can take retinal fundus images easily is needed. The proposed method is applied to not only the PI but also the system which warns about misfiling of fundus images in medical facilities.
Efficient classifier generation and weighted voting for atlas-based segmentation: two small steps faster and closer to the combination oracle
Atlas-based segmentation has proven effective in multiple applications. Usually, several reference images are combined to create a representative average atlas image. Alternatively, a number of independent atlas images can be used, from which multiple segmentations of the image of interest are derived and later combined. One of the major drawbacks of this approach is its large computational burden caused by the high number of required registrations. To address this problem, we introduce One Registration, Multiple Segmentations (ORMS), a procedure to obtain multiple segmentations with a single online registration. This can be achieved by pre-computing intermediate transformations from the initial atlas images to an average image. We show that, compared to the usual approach, our method reduces time considerably with little or no loss in accuracy. On the other hand, optimum combination of these segmentations remains an unresolved problem. Different approaches have been adopted, but they are all far from the upper bound of any combination strategy. This is given by the Combination Oracle, which classifies a voxel correctly if any individual segmentation coincides with the ground truth. We present here a novel combination approach, based on weighting the different segmentations according to the mutual information between the test image and the atlas image after registration. We compare this method with other existing combination strategies using microscopic MR images of mouse brains, achieving statistically significant improvement in segmentation accuracy.
Scene analysis with structural prototypes for content-based image retrieval in medicine
Benedikt Fischer, Michael Sauren, Mark O. Güld, et al.
The content of medical images can often be described as a composition of relevant objects with distinct relationships. Each single object can then be represented as a graph node, and local features of the objects are associated as node attributes, e.g. the centroid coordinates. The relations between these objects are represented as graph edges with annotated relational features, e.g. their relative size. Nodes and edges build an attributed relational graph (ARG). For a given setting, e.g. a hand radiograph, a generalization of the relevant objects, e.g. individual bone segments, can be obtained by the statistical distributions of all attributes computed from training images. These yield a structural prototype graph consisting of one attributed node per relevant object and of their relations represented as attributed edges. In contrast to the ARG, the mean and standard deviation of each local or relational feature are used to annotate the prototype nodes or edges, respectively. The prototype graph can then be used to identify the generalized objects in new images. As new image content is represented by hierarchical attributed region adjacency graphs (HARAGs) which are obtained by region-growing, the task of object or scene identification corresponds to the problem of inexact sub-graph matching between a small prototype and the current HARAG. For this purpose, five approaches are evaluated in an example application of bone-identification in 96 radiographs: Nested Earth Mover's Distance, Graph Edit Distance, a Hopfield Neural Network, Pott's Mean Field Annealing and Similarity Flooding. The discriminative power of 34 local and 12 relational features is judged for each object by sequential forward selection. The structural prototypes improve recall by up to 17% in comparison to the approach without relational information.
A co-occurrence texture semi-invariance to direction, distance, and patient size
Texture-based models are intensively used in medical image processing to quantify the homogeneity and consistency of soft tissues across different patients. Several research studies have shown that the co-occurrence texture model and its Haralick descriptors can be successfully applied to capture the statistical properties of the soft tissues' patterns. Given that the calculation of the co-occurrence texture model is a computationally-intensive task, in this paper we investigate the usefulness of using all possible angles and all displacements for capturing the texture properties of an organ of interest, specifically, the liver. Based on the Analysis of Variance (ANOVA) technique and multiple pair-wise comparisons, we found that using only the "near" and "far" displacements is enough to capture the spatial properties of the texture for the liver.
Tissue classification using cluster features for lesion detection in digital cervigrams
Xiaolei Huang, Wei Wang, Zhiyun Xue, et al.
In this paper, we propose a new method for automated detection and segmentation of different tissue types in digitized uterine cervix images using mean-shift clustering and support vector machines (SVM) classification on cluster features. We specifically target the segmentation of precancerous lesions in a NCI/NLM archive of 60,000 cervigrams. Due to large variations in image appearance in the archive, color and texture features of a tissue type in one image often overlap with that of a different tissue type in another image. This makes reliable tissue segmentation in a large number of images a very challenging problem. In this paper, we propose the use of powerful machine learning techniques such as Support Vector Machines (SVM) to learn, from a database with ground truth annotations, critical visual signs that correlate with important tissue types and to use the learned classifier for tissue segmentation in unseen images. In our experiments, SVM performs better than un-supervised methods such as Gaussian Mixture clustering, but it does not scale very well to large training sets and does not always guarantee improved performance given more training data. To address this problem, we combine SVM and clustering so that the features we extracted for classification are features of clusters returned by the mean-shift clustering algorithm. Compared to classification using individual pixel features, classification by cluster features greatly reduces the dimensionality of the problem, thus it is more efficient while producing results with comparable accuracy.
Posters: Image Restoration and Enhancement
icon_mobile_dropdown
An image reconstruction method based on machine learning for dual-energy subtraction radiography
Yoshiro Kitamura, Masahiko Yamada, Wataru Ito
We propose a novel image reconstruction method for dual-energy subtraction radiography. When one of the dual-energy images is obtained at a low dose, a bone image generated with a dual-energy subtraction technique is degraded due to noise, especially high frequency noise. Our method restores the degraded bone image using a regression filter trained by support vector regression. The regression filter is trained based on the input of degraded bone images against an output of corresponding noiseless bone images. Due to strong correlation between the high frequency and low frequency signals of bone, the high frequency signal can be accurately generated based on the observed low frequency signal. However, learning such correlation directly is generally difficult. Therefore our technique first generates a "2-class bone model" that explicitly expresses a bone structure that should be restored. Then while utilizing this model, regression filtering is applied. The accuracy of regression learning is largely improved with this approach. Verification tests show that our method works well: a soft-tissue image obtained by subtracting a restored bone image from a standard radiograph reveals that the rib structure has been thoroughly removed and that the sharpness of the soft-tissue signal is improved in general and among the fine vessels. In conclusion, the proposed method can provide superior dose reduction as well as a better reflection of the anatomical structures in an image. With these advantages, the proposed method can offer high clinical value for the detection of lung lesions.
Development of adaptive noise reduction filter algorithm for pediatric body images in a multi-detector CT
Eiji Nishimaru, Katsuhiro Ichikawa, Izumi Okita, et al.
Recently, several kinds of post-processing image filters which reduce the noise of computed tomography (CT) images have been proposed. However, these image filters are mostly for adults. Because these are not very effective in small (< 20 cm) display fields of view (FOV), we cannot use them for pediatric body images (e.g., premature babies and infant children). We have developed a new noise reduction filter algorithm for pediatric body CT images. This algorithm is based on a 3D post-processing in which the output pixel values are calculated by nonlinear interpolation in z-directions on original volumetric-data-sets. This algorithm does not need the in-plane (axial plane) processing, so the spatial resolution does not change. From the phantom studies, our algorithm could reduce SD up to 40% without affecting the spatial resolution of x-y plane and z-axis, and improved the CNR up to 30%. This newly developed filter algorithm will be useful for the diagnosis and radiation dose reduction of the pediatric body CT images.
Retinal vessel enhancement based on directional field
Jian Chen, Jie Tian
Motivated by the goal of improving detection of micro-vessels with low contrast, a new technique based on directional field is presented for enhancing vessels in retinal images. This technique consists of two steps: the estimation of directional field and the enhancement. We will make enhancement along the vascular direction and normalize the brightness of image in a single step. After estimating the directional field, the mean and variance values in a local neighborhood were calculated pixel-by-pixel, accordingly normalize and enhance the local neighborhood. In order to eliminate the artificial boundary between two adjacent areas, an anisotropic Gaussian kernel was introduced to weight the enhancement. The proposed method can obviously increase the contrast of retinal vessels. 20 retinal images were tested in our experiments, and the results demonstrate an effective vessel enhancement algorithm.
Fast multiscale vessel enhancement filtering
Dong Hye Ye, Dongjin Kwon, Il Dong Yun, et al.
This paper describes a fast multi-scale vessel enhancement filter in 3D medical images. For efficient review of the vascular information, clinicians need rendering the 3D vascular information as a 2D image. Generally, the maximum intensity projection (MIP) is a useful and widely used technique for producing a 2D image from the 3D vascular data. However, the MIP algorithm reduces the conspicuousness for small and faint vessels owing to the overlap of non-vascular structures. To overcome this invisibility, researchers have examined the multi-scale vessel enhancement filter based on a combination of the eigenvalues of the 3D Hessian matrix. This multi-scale vessel enhancement filter produces higher contrast. However, it is time-consuming and requires high cost computation due to large volume of data and complex 3D convolution. For fast vessel enhancement, we propose a novel multi-scale vessel enhancement filter using 3D integral images and 3D approximated Gaussian kernel. This approximated kernel looks like cube but it is not exact cube. Each layer of kernel is approximated 2D Gaussian second order derivative by dividing it into three rectangular regions whose sum is integer. 3D approximated kernel is a pile of these 2D box kernels which are normalized by Frobenius norm. Its size fits to vessel width in order to achieve better visualization of the small vessel. Proposed method is approximately five times faster and produces comparable results with previous multi-scale vessel enhancement filter.
Adaptive kernel algorithm for FPGA-based speckle reduction
Gerhard Tech, Robert Schwann, Goetz Kappen, et al.
Image quality from ultrasound and optical coherence tomography (OCT) is degraded by speckle patterns, which limit the detection of small features and cause a loss of image contrast. To reduce speckle patterns, a novel adaptive kernel algorithm suitable for linearly scaled and log-compressed OCT and ultrasound images is presented. This algorithm combines region growing with a stick based approach. For each direction from the center of a square window to its border, a stick length is selected depending on a homogeneity criterion. Such a set of sticks forms an individual filter kernel for each image pixel. The current kernel size is observed to detect outliers in speckle. If the kernel size drops below a threshold, an outlier is assumed and the filter output is corrected by using a median filter in a second filtering stage. A new homogeneity model is presented that incorporates two existing models and can be fit to actual image statistics. In addition, methods to compute filter parameters for different speckle correlation lengths and imaging systems are presented. An FPGA real-time implementation is proposed and discussed. Measured and simulated speckle images are processed and results compared to existing FPGA based speckle reduction methods. The proposed filter provides good results for various kinds of medical speckle images. The flexible nature of the proposed kernel guarantees suitability for highly correlated as well as for uncorrelated speckle patterns.
Enhancing regional lymph nodes from endoscopic ultrasound images
Esophageal ultrasound (EUS) is particularly useful for isolating lymph nodes in the N-staging of esophageal cancer, a disease with very poor overall prognosis. Although EUS is relatively low-cost and real time, and it provides valuable information to the clinician, its usefulness to less trained "users" including opportunities for computer-aided diagnosis is still limited due to the strong presence of spatially correlated interference noise called speckles. To this end, in this paper, we present a technique for enhancing lymph nodes in EUS images by first reducing the spatial correlation of the specular noise and then using a modified structured tensor-based anisotropic filter to complete the speckle reduction process. We report on a measure of the enhancement and also on the extent of automatic processing possible, after the speckle reduction process has taken place. Also, we show the limitations of the enhancement process by extracting relevant lymph node features from the despeckled images. When tested on five representative classes of esophageal lymph nodes, we found the despeckling process to greatly reduce the specularity of the original EUS images, therefore proving very useful for visualization purposes. But it still requires additional work for the complete automation of the lymph node characterizing process.
Clinical validation and performance evaluation of enhancement methods acquired from interventional C-ARM x-ray
Liyang Wei, Dinesh Kumar, Animesh Khemka, et al.
Digital Subtraction Angiography (DSA) is a well-established powerful modality for the visualization of stenosis and blood vessels in general. This paper presents two novel approaches which address image quality. In the first approach we combine anisotropic diffusion with nonlinear normalization. The second approach consists of an introduction of a regularization strategy followed by a classification procedure to improve the enhancement. The performances of two strategies are evaluated based on a database of 73 subjects using SNR, CNR and Tenengrad's metric. Compared with conventional DSA, Eigen's diffusion embedded nonlinear enhancement strategies can improve image quality 95.25% in terms of SNR. The regularization embedded linear enhancement strategy can also improve SNR 51.46% compared with conventional DSA. Similar results are obtained by CNR and Tenengrad's metric measurements. Our system runs on a PC-based workstation using C++ in Windows environment.
Dermascopic hair disocclusion using inpainting
Inpainting, a technique originally used to restore film and photographs, is used to disocclude hair from dermascopic images of skin lesions. The technique is compared to the conventional software DullRazor, which uses linear interpolation to perform disocclusion. Comparison was performed by simulating occluding hair on a dermascopic image, applying DullRazor and inpainting and calculating the error induced. Inpainting is found to perform approximately 33% better than DullRazor's linear interpolation, and is more stable under heavy occlusion. The results are also compared to published results from two other alternatives: auto-regressive (AR) model signal extrapolation and band-limited (BL) signal interpolation.
Denoising of brain MRI images using modified PDE model based on pixel similarity
Renchao Jin, Enmin Song, Lijuan Zhang, et al.
Although various image denoising methods such as PDE-based algorithms have made remarkable progress in the past years, the trade-off between noise reduction and edge preservation is still an interesting and difficult problem in the field of image processing and analysis. A new image denoising algorithm, using a modified PDE model based on pixel similarity, is proposed to deal with the problem. The pixel similarity measures the similarity between two pixels. Then the neighboring consistency of the center pixel can be calculated. Informally, if a pixel is not consistent enough with its surrounding pixels, it can be considered as a noise, but an extremely strong inconsistency suggests an edge. The pixel similarity is a probability measure, its value is between 0 and 1. According to the neighboring consistency of the pixel, a diffusion control factor can be determined by a simple thresholding rule. The factor is combined into the primary partial differential equation as an adjusting factor for controlling the speed of diffusion for different type of pixels. An evaluation of the proposed algorithm on the simulated brain MRI images was carried out. The initial experimental results showed that the new algorithm can smooth the MRI images better while keeping the edges better and achieve higher peak signal to noise ratio (PSNR) comparing with several existing denoising algorithms.
Pyramidal flux in an anisotropic diffusion scheme for enhancing structures in 3D images
Pyramid based methods in image processing provide a helpful framework for accelerating the propagation of information over large spatial domains, increasing the efficiency for large scale applications. Combined with an anisotropic diffusion scheme tailored to preserve the boundaries at a given level, an efficient way for enhancing large structures in 3D images is presented. In our approach, the partial differential equation defining the evolution of the intensity in the image is solved in an explicit scheme at multiple resolutions in an ascending-descending cycle. Intensity 'flux' between distant voxels is allowed, while preserving borders relative to the scale. Experiments have been performed both with phantoms and with real data from 3D Transrectal Ultrasound Imaging. The effectiveness of the method to remove speckle noise and to enhance large structures such as the prostate has been demonstrated. For instance, using two scales reduces the computation time by 87% as compared to a single scale. Furthermore, we show that the boundaries of the prostate are mainly preserved, by comparing with manually outlined edges.
Informative frame detection from wireless capsule video endoscopic images
Md. Khayrul Bashar, Kensaku Mori, Yasuhito Suenaga, et al.
Wireless capsule endoscopy (WCE) is a new clinical technology permitting the visualization of the small bowel, the most difficult segment of the digestive tract. The major drawback of this technology is the high amount of time for video diagnosis. In this study, we propose a method for informative frame detection by isolating useless frames that are substantially covered by turbid fluids or their contamination with other materials, e.g., faecal, semi-processed or unabsorbed foods etc. Such materials and fluids present a wide range of colors, from brown to yellow, and/or bubble-like texture patterns. The detection scheme, therefore, consists of two stages: highly contaminated non-bubbled (HCN) frame detection and significantly bubbled (SB) frame detection. Local color moments in the Ohta color space are used to characterize HCN frames, which are isolated by the Support Vector Machine (SVM) classifier in Stage-1. The rest of the frames go to the Stage-2, where Laguerre gauss Circular Harmonic Functions (LG-CHFs) extract the characteristics of the bubble-structures in a multi-resolution framework. An automatic segmentation method is designed to extract the bubbled regions based on local absolute energies of the CHF responses, derived from the grayscale version of the original color image. Final detection of the informative frames is obtained by using threshold operation on the extracted regions. An experiment with 20,558 frames from the three videos shows the excellent average detection accuracy (96.75%) by the proposed method, when compared with the Gabor based- (74.29%) and discrete wavelet based features (62.21%).
Posters: Motion Analysis
icon_mobile_dropdown
Automated motion correction based on target tracking for dynamic nuclear medicine studies
Xinhua Cao, Tracy Tetrault, Fred Fahey, et al.
Nuclear medicine dynamic studies of kidneys, bladder and stomach are important diagnostic tools. Accurate generation of time-activity curves from regions of interest (ROIs) requires that the patient remains motionless for the duration of the study. This is not always possible since some dynamic studies may last from several minutes to one hour. Several motion correction solutions have been explored. Motion correction using external point sources is inconvenient and not accurate especially when motion results from breathing, organ motion or feeding rather than from body motion alone. Centroid-based motion correction assumes that activity distribution is only inside the single organ (without background) and uniform, but this approach is impractical in most clinical studies. In this paper, we present a novel technique of motion correction that first tracks the organ of interest in a dynamic series then aligns the organ. The implementation algorithm for target tracking-based motion correction consists of image preprocessing, target detection, target positioning, motion estimation and prediction, tracking (new search region generation) and target alignment. The targeted organ is tracked from the first frame to the last one in the dynamic series to generate a moving trajectory of the organ. Motion correction is implemented by aligning the organ ROIs in the image series to the location of the organ in the first image. The proposed method of motion correction has been applied to several dynamic nuclear medicine studies including radionuclide cystography, dynamic renal scintigraphy, diuretic renography and gastric emptying scintigraphy.
Multi-object tracking of human spermatozoa
Lauge Sørensen, Jakob Østergaard, Peter Johansen, et al.
We propose a system for tracking of human spermatozoa in phase-contrast microscopy image sequences. One of the main aims of a computer-aided sperm analysis (CASA) system is to automatically assess sperm quality based on spermatozoa motility variables. In our case, the problem of assessing sperm quality is cast as a multi-object tracking problem, where the objects being tracked are the spermatozoa. The system combines a particle filter and Kalman filters for robust motion estimation of the spermatozoa tracks. Further, the combinatorial aspect of assigning observations to labels in the particle filter is formulated as a linear assignment problem solved using the Hungarian algorithm on a rectangular cost matrix, making the algorithm capable of handling missing or spurious observations. The costs are calculated using hidden Markov models that express the plausibility of an observation being the next position in the track history of the particle labels. Observations are extracted using a scale-space blob detector utilizing the fact that the spermatozoa appear as bright blobs in a phase-contrast microscope. The output of the system is the complete motion track of each of the spermatozoa. Based on these tracks, different CASA motility variables can be computed, for example curvilinear velocity or straight-line velocity. The performance of the system is tested on three different phase-contrast image sequences of varying complexity, both by visual inspection of the estimated spermatozoa tracks and by measuring the mean squared error (MSE) between the estimated spermatozoa tracks and manually annotated tracks, showing good agreement.
Tracking the hyoid bone in videofluoroscopic swallowing studies
Patrick M. Kellen, Darci Becker, Joseph M. Reinhardt, et al.
Difficulty swallowing, or dysphagia, has become a growing problem. Swallowing complications can lead to malnutrition, dehydration, respiratory infection, and even death. The current gold standard for analyzing and diagnosing dysphagia is the videofluoroscopic barium swallow study. In these studies, a fluoroscope is used to image the patient ingesting barium solutions of different volumes and viscosities. The hyoid bone anchors many key muscles involved in swallowing and plays a key role in the process. Abnormal hyoid bone motion during a swallow can indicate swallowing dysfunction. Currently in clinical settings, hyoid bone motion is assessed qualitatively, which can be subject to intra-rater and inter-rater bias. This paper presents a semi-automatic method for tracking the hyoid bone that makes quantitative analysis feasible. The user defines a template of the hyoid on one frame, and this template is tracked across subsequent frames. The matching phase is optimized by predicting the position of the template based on kinematics. An expert speech pathologist marked the position of the hyoid on each frame of ten studies to serve as the gold standard. Results from performing Bland-Altman analysis at a 95% confidence interval showed a bias of 0.0±0.08 pixels in x and -0.08±0.09 pixels in y between the manually-defined gold standard and the proposed method. The average Pearson's correlation between the gold standard and the proposed method was 0.987 in x and 0.980 in y. This paper also presents a method for automatically establishing a patient-centric coordinate system for the interpretation of hyoid motion. This coordinate system corrects for upper body patient motion during the study and identifies superior-inferior and anterior-posterior motion components. These tools make the use of quantitative hyoid motion analysis feasible in clinical and research settings.
Posters: MRI
icon_mobile_dropdown
Towards user-independent DTI quantification
Jan Klein, Hannes Stuke, Jan Rexilius, et al.
Quantification of diffusion tensor imaging (DTI) parameters has become an important role in the neuroimaging, neurosurgical, and neurological community as a method to identify major white matter tracts afflicted by pathology or tracts at risk for a given surgical approach. We introduce a novel framework for a reliable and robust quantification of DTI parameters, which overcomes problems of existing techniques introduced by necessary user inputs. In a first step, a hybrid clustering method is proposed that allows for extracting specific fiber bundles in a robust way. Compared to previous methods, our approach considers only local proximities of fibers and is insensitive to their global geometry. This is very useful in cases where a fiber tracking of the whole brain is not available. Our technique determines the overall number of clusters iteratively using a eigenvalue thresholding technique to detect disjoint clusters of independent fiber bundles. Afterwards, possible finer substructures based on an eigenvalue regression are determined within each bundle. In a second step, a quantification of DTI parameters of the extracted bundle is performed. We propose a method that automatically determines a 3D image where the voxel values encode the minimum distance to a reconstructed fiber. This image allows for calculating a 3D mask where each voxel within the mask corresponds to a voxel that lies in an isosurface around the fibers. The mask is used for an automatic classification between tissue classes (fiber, background, and partial volume) so that the quantification can be performed on one or more of such classes. This can be done per slice or a single DTI parameter can be determined for the whole volume which is covered by the isosurface. Our experimental tests confirm that major white matter fiber tracts may be robustly determined and can be quantified automatically. A great advantage of our framework is its easy integration into existing quantification applications so that uncertainties can be reduced, and higher intrarater- as well as interrater reliabilities can be achieved.
An exploration of spatial similarities in temporal noise spectra in fMRI measurements
D. H. J. Poot, J. Sijbers, A. J. den Dekker
In this paper, we describe a method to evaluate similarities in estimated temporal noise spectra of functional Magnetic Resonance Imaging (fMRI) time series. Accurate noise spectra are needed for reliable activation detection in fMRI. Since these spectra are a-priori unknown, they have to be estimated from the fMRI data. A noise model can be estimated for each voxel separately, but when noise spectra of neighboring voxels are (almost) equal, the power of the activation detection test can be improved by estimating the noise model from a set of neighboring voxels. In this paper, a method is described to evaluate the similarity of noise spectra of neighboring voxels. Noise spectrum similarities are studied in simulation as well as experimental fMRI datasets. The parameters of the model describing the voxel time series are estimated by a Maximum Likelihood (ML) estimator. The similarity of the ML estimated noise processes is assessed by the Model Error (ME), which is based on the Kullback Leibler divergence. Spatial correlations in the fMRI data reduce the ME between the noise spectra of (neighboring) voxels. This undesired effect is quantified by simulation experiments where spatial correlation is introduced. By plotting the ME as a function of the distance between voxels, it is observed that the ME increases as a function of this distance. Additionally, by using the theoretical distribution of the ME, it is observed that neighboring voxels indeed have similar noise spectra and these neighbors can be used to improve the noise model estimate.
White matter tractographies registration using Gaussian mixture modeling
Orly Zvitia, Arnaldo Mayer, Hayit Greenspan
This paper proposes a novel and robust approach to the registration (matching) of intra-subject white matter (WM) fiber sets extracted from DT-MRI scans by Tractography. For each fiber, a feature space representation is obtained by appending the sequence of its 3D coordinates. Clustering by non-parametric adaptive mean shift provides a representative fiber for each cluster hereafter termed the fiber-mode (FM). For each FM, the parameters of a multivariate Gaussian are computed from its fiber population, leading to a mixture of Gaussians (MoG) for the whole fiber set. The number of Gaussians used for a fiber set equals the number of FM representing the set. The alignment of two fiber sets is then treated as the alignment between two MoGs, and is solved by maximizing the correlation ratio between them. Initial results are presented for real intrasubject fiber sets and synthetic transformations.
Tensor distribution function
Alex D. Leow, Siwei Zhu
Diffusion weighted MR imaging is a powerful tool that can be employed to study white matter microstructure by examining the 3D displacement profile of water molecules in brain tissue. By applying diffusion-sensitizing gradients along a minimum of 6 directions, second-order tensors (represetnted by 3-by-3 positive definiite matrices) can be computed to model dominant diffusion processes. However, it has been shown that conventional DTI is not sufficient to resolve more complicated white matter configurations, e.g. crossing fiber tracts. More recently, High Angular Resolution Diffusion Imaging (HARDI) seeks to address this issue by employing more than 6 gradient directions. To account for fiber crossing when analyzing HARDI data, several methodologies have been introduced. For example, q-ball imaging was proposed to approximate Orientation Diffusion Function (ODF). Similarly, the PAS method seeks to reslove the angular structure of displacement probability functions using the maximum entropy principle. Alternatively, deconvolution methods extract multiple fiber tracts by computing fiber orientations using a pre-specified single fiber response function. In this study, we introduce Tensor Distribution Function (TDF), a probability function defined on the space of symmetric and positive definite matrices. Using calculus of variations, we solve for the TDF that optimally describes the observed data. Here, fiber crossing is modeled as an ensemble of Gaussian diffusion processes with weights specified by the TDF. Once this optimal TDF is determined, ODF can easily be computed by analytical integration of the resulting displacement probability function. Moreover, principle fiber directions can also be directly derived from the TDF.
Susceptibility correction for improved tractography using high field DT-EPI
W. Pintjens, D. H. J. Poot, M. Verhoye, et al.
Diffusion Tensor Magnetic Resonance Imaging (DTI) is a well known technique that can provide information about the neuronal fiber structure of the brain. However, since DTI requires a large amount of data, a high speed MRI acquisition technique is needed to acquire these data within a reasonable time. Echo Planar Imaging (EPI) is a technique that provides the desired speed. Unfortunately, the advantage of speed is overshadowed by image artifacts, especially at high fields. EPI artifacts originate from susceptibility differences in adjacent tissues and correction techniques are required to obtain reliable images. In this work, the fieldmap method, which tries to measure distortion effects, is optimized by using a non linear least squares estimator for calculating pixel shifts. This method is tested on simulated data and proves to be more robust against noise compared to previously suggested methods. Another advantage of this new method is that other parameters like relaxation and the odd/even phase difference are estimated. This new way of estimating the field map is demonstrated on a hardware phantom, which consists of parallel bundles made of woven strands of Micro Dyneema fibers. Using a modified EPI-sequence, reference data was measured for the calculation of fieldmaps. This allows one to reposition the pixels in order to obtain images with less distortions. The correction is applied to non-diffusion weighted images as well as diffusion weighted images and fiber tracking is performed on this corrected data.
A Bayesian method with reparameterization for diffusion tensor imaging
Diwei Zhou, Ian L. Dryden, Alexey Koloydenko, et al.
A multi-tensor model with identifiable parameters is developed for diffusion weighted MR images. A new parameterization method guarantees the symmetric positive-definiteness of the diffusion tensor. We set up a Bayesian method for parameter estimation. To investigate properties of the method, Monte Carlo simulated data from three distinct DTI direction schemes have been analyzed. The multi-tensor model with automatic model selection has also been applied to a healthy human brain dataset. Standard tensor-derived maps are obtained when the single-tensor model is fitted to a region of interest with a single dominant fiber direction. High anisotropy diffusion flows and main diffusion directions can be shown clearly in the FA map and diffusion ellipsoid map. For another region containing crossing fiber bundles, we estimate and display the ellipsoid map under the single tensor and double-tensor regimes of the multi-tensor model, suitably thresholding the Bayes factor for model selection.
Automatic regional analysis of DTI properties in the developmental macaque brain
Martin Styner, Rebecca Knickmeyer, Christopher Coe, et al.
Many neuroimaging studies are applied to monkeys as pathologies and environmental exposures can be studied in well-controlled settings and environment. In this work, we present a framework for the use of an atlas based, fully automatic segmentation of brain tissues, lobar parcellations, subcortical structures and the regional extraction of Diffusion Tensor Imaging (DTI) properties. We first built a structural atlas from training images by iterative, joint deformable registration into an unbiased average image. On this atlas, probabilistic tissue maps, a lobar parcellation and subcortical structures were determined. This information is applied to each subjects structural image via affine, followed by deformable registration. The affinely transformed atlas is employed for a joint T1 and T2 based tissue classification. The deformed parcellation regions mask the tissue segmentations to define the parcellation for white and gray matter separately. Each subjects structural image is then non-rigidly matched with its DTI image by normalized mutual information, b-spline based registration. The DTI property histograms were then computed using the probabilistic white matter information for each lobar parcellation. We successfully built an average atlas using a developmental training datasets of 18 cases aged 16-34 months. Our framework was successfully applied to over 50 additional subjects in the age range of 9 70 months. The probabilistically weighted FA average in the corpus callosum region showed the largest increase over time in the observed age range. Most cortical regions show modest FA increase, whereas the cerebellums FA values remained stable. The individual methods used in this segmentation framework have been applied before, but their combination is novel, as is their application to macaque MRI data. Furthermore, this is the first study to date looking at the DTI properties of the developing macaque brain.
Posters: Multiresolution and Wavelets
icon_mobile_dropdown
Short basis functions for constant-variance interpolation
An interpolation model is a necessary ingredient of intensity-based registration methods. The properties of such a model depend entirely on its basis function, which has been traditionally characterized by features such as its order of approximation and its support. However, as has been recently shown, these features are blind to the amount of registration bias created by the interpolation process alone; an additional requirement that has been named constant-variance interpolation is needed to remove this bias. In this paper, we present a theoretical investigation of the role of the interpolation basis in a registration context. Contrarily to published analyses, ours is deterministic; it nevertheless leads to the same conclusion, which is that constant-variance interpolation is beneficial to image registration. In addition, we propose a novel family of interpolation bases that can have any desired order of approximation while maintaining the constant-variance property. Our family includes every constant-variance basis we know of. It is described by an explicit formula that contains two free functional terms: an arbitrary 1-periodic binary function that takes values from {-1, 1}, and another arbitrary function that must satisfy the partition of unity. These degrees of freedom can be harnessed to build many family members for a given order of approximation and a fixed support. We provide the example of a symmetric basis with two orders of approximation that is supported over [-3/2, 3/2] this support is one unit shorter than a basis of identical order that had been previously
Efficient random access high resolution region-of-interest (ROI) image retrieval using backward coding of wavelet trees (BCWT)
Efficient retrieval of high quality Regions-Of-Interest (ROI) from high resolution medical images is essential for reliable interpretation and accurate diagnosis. Random access to high quality ROI from codestreams is becoming an essential feature in many still image compression applications, particularly in viewing diseased areas from large medical images. This feature is easier to implement in block based codecs because of the inherent spatial independency of the code blocks. This independency implies that the decoding order of the blocks is unimportant as long as the position for each is properly identified. In contrast, wavelet-tree based codecs naturally use some interdependency that exploits the decaying spectrum model of the wavelet coefficients. Thus one must keep track of the decoding order from level to level with such codecs. We have developed an innovative multi-rate image subband coding scheme using "Backward Coding of Wavelet Trees (BCWT)" which is fast, memory efficient, and resolution scalable. It offers far less complexity than many other existing codecs including both, wavelet-tree, and block based algorithms. The ROI feature in BCWT is implemented through a transcoder stage that generates a new BCWT codestream containing only the information associated with the user-defined ROI. This paper presents an efficient technique that locates a particular ROI within the BCWT coded domain, and decodes it back to the spatial domain. This technique allows better access and proper identification of pathologies in high resolution images since only a small fraction of the codestream is required to be transmitted and analyzed.
Posters: Registration
icon_mobile_dropdown
Semi-automatic matching of OCT and IVUS images for image fusion
Olivier Pauly, Gozde Unal, Greg Slabaugh, et al.
Medical imaging is essential in the diagnosis of atherosclerosis. In this paper, we propose the semi-automatic matching of two promising and complementary intravascular imaging techniques, Intravascular Ultrasound (IVUS) and Optical Coherence Tomography (OCT), with the ultimate goal of producing hybrid images with increased diagnostic value for assessing arterial health. If no ECG gating has been performed on the IVUS and OCT pullbacks, there is typically an anatomical shuffle (displacement in time and space) in the image sequences due to the catheter motion in the artery during the cardiac cycle, and thus, this is not possible to perform a 3D registration. Therefore, the goal of our work is to detect semi-automatically the corresponding images in both modalities as a preprocessing step for the fusion. Our method is based on the characterization of the lumen shape by a set of Gabor Jets features. We also introduce different correction terms based on the approximate position of the slice in the artery. Then we train different support vector machines based on these features to recognize these correspondences. Experimental results demonstrate the usefulness of our approach, which achieves up to 95% matching accuracy for our data.
Nonlinear elastic model for image registration and soft tissue simulation based on piecewise St. Venant-Kirchhoff material approximation
Evgeny Gladilin, Roland Eils
Linear elastic model widely applied for simulation of soft tissue deformations in biomedical imaging applications is basically limited to the range of small deformations and rotations. Thus, computation of large deformations and rotations using linear elastic approximation and its derivatives is associated with substantial error. More realistic modeling of mechanical behavior of soft tissue requires handling of different types of nonlinearities. This paper presents a framework for more accurate modeling of deformable structures based on the St. Venant-Kirchhoff law with the nonlinear Green-Lagrange strain tensor and variable material constants, which considers both material and geometric nonlinearities. We derive the governing partial differential equation of nonlinear elasticity, which represents consistent extension of the Lame-Navier PDE of linear elasticity, and describe two alternative numerical schemes for solving this nonlinear PDE via the Newton's and fixed point method, respectively. The results of our comparative studies demonstrate the advantages of nonlinear elastic model for accurate computing of large deformations and rotations in comparison to the linear elastic approximation.
Validation and comparison of registration methods for free-breathing 4D lung CT
Torbjørn Vik, Sven Kabus, Jens von Berg, et al.
We have compared and validated image registration methods with respect to the clinically relevant use-case of lung CT max-inhale to max-exhale registration. Four fundamentally different algorithms representing main approaches for image registration were compared using clinical images. Each algorithm was assigned to a different person with extensive working knowledge of its usage. Quantitative and qualitative evaluation is performed. Whereas the methods achieve similar results in target registration error, characteristic differences come to show by closer analysis of the displacement fields.
Non-rigid registration of 2D manifolds in 3D Euclidian space
This work describes a non-rigid registration method for open 2D manifold embedded in 3D Euclidian space. The method is based on difference of distance maps and grid based warps interpolated by splines constrained in such a way that the deformation field is diffeomorphic. We then create a dense surface to surface correspondence using angle weighted normals and ray tracing. The implementation using a derivation of the inverse compositional algorithm for optimization of computational speed is described. The results are evaluated as a shape model showing the principal modes of variation.
Model-to-image based 2D-3D registration of angiographic data
Sabine Mollus, Jördis Lübke, Andreas J. Walczuch, et al.
We propose a novel registration method, which combines well-known vessel detection techniques with aspects of model adaptation. The proposed method is tailored to the requirements of 2D-3D-registration of interventional angiographic X-ray data such as acquired during abdominal procedures. As prerequisite, a vessel centerline is extracted out of a rotational angiography (3DRA) data set to build an individual model of the vascular tree. Following the two steps of local vessel detection and model transformation the centerline model is matched to one dynamic subtraction angiography (DSA) target image. Thereby, the in-plane position and the 3D orientation of the centerline is related to the vessel candidates found in the target image minimizing the residual error in least squares manner. In contrast to feature-based methods, no segmentation of the vessel tree in the 2D target image is required. First experiments with synthetic angiographies and clinical data sets indicate that matching with the proposed model-to-image based registration approach is accurate and robust and is characterized by a large capture range.
Robust registration for change detection
Sune Darkner, Dan Witzner Hansen, Rasmus R. Paulsen, et al.
We address the problem of intra-subject registration for change detection. The goal is to separate stationary and changing subsets to be able to robustly perform rigid registration on the stationary subsets and thus improve the subsequent change detection. An iterative approach using a hybrid of parametric and non-parametric statistics is presented. The method uses non-parametric clustering and large scale hypothesis testing with estimation of the empirical null hypothesis. The method is successfully applied to 3D surface scans of human ear impressions containing true changes as well as data with synthesized changes. It is shown that the method improves registration and is capable of reducing the difference between registration using different norms.
Reconstruction and registration of multispectral x-ray images for reliable alignment correction in radiation treatment devices
Boris P. Selby, Georgios Sakas, Stefan Walter, et al.
To align patients in radiation devices in six degrees of freedom (DoF), image-guided approaches perform the task of correction computation for the patient position. Digital radiography (DR) images are compared to projections of a CT series to estimate misalignments. A problem is that digital reconstructed radiographs (DRRs) have to be created from the CT to be registered with the DRs. Depending on the X-ray tube energy, detector sensitivity and body part involved, DRRs and DRs may look very different and often cannot be registered. We present a method that reconstructs multi-spectral DRRs for different X-ray settings, which can be registered to real X-ray images. As short rendering times are crucial, multiple spectra of a DRR are generated in one ray-tracing process. We register our multi-spectral DRR with the DR and add a further DoF to find a best match not only for the translations and in-plane rotation, but also the best fitting spectral planes. The results are used to identify patient misalignments and show that higher reliability can be achieved compared to conventional approaches. Misalignments can be identified even if ineligible X-ray settings have been used. As our approach allows application of lower X-ray energies for DR creation, an additional benefit is the reduction of the delivered dose.
Registration of standardized histological images in feature space
In this paper, we propose three novel and important methods for the registration of histological images for 3D reconstruction. First, possible intensity variations and nonstandardness in images are corrected by an intensity standardization process which maps the image scale into a standard scale where the similar intensities correspond to similar tissues meaning. Second, 2D histological images are mapped into a feature space where continuous variables are used as high confidence image features for accurate registration. Third, we propose an automatic best reference slice selection algorithm that improves reconstruction quality based on both image entropy and mean square error of the registration process. We demonstrate that the choice of reference slice has a significant impact on registration error, standardization, feature space and entropy information. After 2D histological slices are registered through an affine transformation with respect to an automatically chosen reference, the 3D volume is reconstructed by co-registering 2D slices elastically.
A new parametric nonrigid image registration method based on Helmholtz's theorem
Helmholtz's theorem states that, with suitable boundary condition, a vector field is completely determined if both of its divergence and curl are specified everywhere. Based on this, we developed a new parametric non-rigid image registration algorithm. Instead of the displacements of regular control grid points, the curl and divergence at each grid point are employed as the parameters. The closest related work was done by Kybic where the parameters are the Bspline coefficients of the displacement field at each control grid point. However, in Kybic's work, it is very likely to result in grid folding in the final deformation field if the distance between adjacent control grid points (knot spacing) is less than 8. This implies that the high frequency components in the deformation field can not be accurately estimated. Another relevant work is the NiRuDeGG method where by solving a div-curl system, an intermediate vector field is generated and, in turn, a well-regularized deformation field can be obtained. Though the present work does not guarantee the regularity (no mesh folding) of the resulting deformation field, which is also suffered by Kybic's work, it allows for a more efficient optimization scheme over the NiRuDeGG method. Our experimental results showed that the proposed method is less prone to grid folding than Kybic's work and that in many cases, in a multi-resolution fashion; the knot spacing can be reduced down to 1 and thus has the potential to achieve higher registration accuracy. Detailed comparison among the three algorithms is described in the paper.
3-D statistical cancer atlas-based targeting of prostate biopsy using ultrasound image guidance
Prostate cancer is a multifocal disease and lesions are not distributed uniformly within the gland. Several biopsy protocols concerning spatially specific targeting have been reported urology literature. Recently a statistical cancer atlas of the prostate was constructed providing voxelwise probabilities of cancers in the prostate. Additionally an optimized set of biopsy sites was computed with 94 - 96% detection accuracy was reported using only 6-7 needles. Here we discuss the warping of this atlas to prostate segmented side-fire ultrasound images of the patient. A shape model was used to speed up registration. The model was trained from over 38 expert segmented subjects off-line. This training yielded as few as 15-20 degrees of freedom that were optimized to warp the atlas surface to the patient's ultrasound image followed by elastic interpolation of the 3-D atlas. As a result the atlas is completely mapped to the patient's prostate anatomy along with optimal predetermined needle locations for biopsy. These do not preclude the use of additional biopsies if desired. A color overlay of the atlas is also displayed on the ultrasound image showing high cancer zones within the prostate. Finally current biopsy locations are saved in the atlas space and may be used to update the atlas based on the pathology report. In addition to the optimal atlas plan, previous biopsy locations and alternate plans can also be stored in the atlas space and warped to the patient with no additional time overhead.
Optimized GPU implementation of learning-based non-rigid multi-modal registration
Non-rigid multi-modal volume registration is computationally intensive due to its high-dimensional parameter space, where common CPU computation times are several minutes. Medical imaging applications using registration, however, demand ever faster implementations for several purposes: matching the data acquisition speed, providing smooth user interaction and steering for quality control, and performing population registration involving multiple datasets. Current GPUs offer an opportunity to boost the registration speed through high computational power at low cost. In our previous work, we have presented a GPU implementation of a non-rigid multi-modal volume registration that was 6 - 8 times faster than a software implementation. In this paper, we extend this work by describing how new features of the DX10-compatible GPUs and additional optimization strategies can be employed to further improve the algorithm performance. We have compared our optimized version with the previous version on the same GPU, and have observed a speedup factor of 3.6. Compared with the software implementation, we achieve a speedup factor of up to 44.
Efficient 3D rigid-body registration of micro-MR and micro-CT trabecular bone images
C. S. Rajapakse, J. Magland, S. L. Wehrli, et al.
Registration of 3D images acquired from different imaging modalities such as micro-magnetic resonance imaging (µMRI) and micro-computed tomography (µCT) are of interest in a number of medical imaging applications. Most general-purpose multimodality registration algorithms tend to be computationally intensive and do not take advantage of the shape of the imaging volume. Multimodality trabecular bone (TB) images of cylindrical cores, for example, tend to be misaligned along and around the axial direction more than that around other directions. Additionally, TB images acquired by µMRI can differ substantially from those acquired by µCT due to apparent trabecular thickening from magnetic susceptibility boundary effects and non-linear intensity correspondence. However, they share very similar contrast characteristics since the images essentially represent a binary tomographic system. The directional misalignment and the fundamental similarities of the two types of images can be exploited to achieve fast 3D registration. Here we present an intensity cross-correlation based 3D registration algorithm for registering 3D specimen images from cylindrical cores of cadaveric TB acquired by µMRI and µCT in the context of finite-element modeling to assess the bone's mechanical constants. The algorithm achieves the desired registration by first coarsely approximating the three translational and three rotational parameters required to align the µMR images to the µCT scan coordinate frame and fine-tuning the parameters in the neighborhood of the approximate solution. The algorithm described here is suitable for 3D rigid-body image registration applications where through-plane rotations are known to be relatively small. The accuracy of the technique is constrained by the image resolution and in-plane angular increments used.
Histopathology and MR image fusion of the prostate
Hyun Hee Jo, Julip Jung, Yujin Jang, et al.
We propose a method for combining histopathology image with MR image of the prostate by using image correction and nonrigid registration. Our method consists of four steps. First, two or four tissue sections of the prostate in histopathology image are combined to produce a single prostate image by stitching. Second, the intensity of prostate bleeding area on T2-weighted MR image is substituted for that on T1-weighted MR image. Our intensity correction prevents a mistake which a prostate bleeding is considered as a tumor on T2-weighted MR image. Third, rough and fine registration is performed to find the best match for pixel overlap between histopathology and MR images. Then the result of rigid registration is deformed by the TPS warping. Finally, aligned images are visualized by the intensity intermixing. Experimental results show that the prostate tumor lesion can be properly located and clearly visualized within MR images for tissue characterization comparison.
3D-3D alignment using particle swarm optimization
Three-dimensional datasets of complex objects are readily available from the tomographic modalities, and fusion of these data sets leads to new understanding of the data. Automatic alignment of the objects is difficult or time consuming when substantial misalignments are present or point correspondences cannot be established, or the solution space is non-convex. These issues effectively exclude most optimization algorithms used in conventional data alignment. Here, we present the particle swarm optimization (PSO) approach which is not sensitive to initial conditions, local minima or non-convex solution space. Intercommunicating particle swarms are randomly placed in the solution space (representing the parameters of the rigid transformations). Each member of each swarm traverses the solution space, constantly evaluating the objective function at its own position and communicating with other members of the swarm about theirs. In addition, the swarms communicate between themselves. Through this information sharing between swarm members and the swarms, the space is searched completely and efficiently, and as a result all swarms converge near the globally optimal rigid transformation. To evaluate the technique, high-resolution micro-CT data sets of single mouse heads were acquired with large initial misalignments. Using two communicating particle swarms in the same solution space, six distinct mouse head objects were aligned finding the approximate global minima in about 25 iterations or 140 seconds on a standard PC independent of initial conditions. Faster speeds (better accuracy) can be obtained by relaxing (restricting) the convergence criteria. These results indicate that the particle swarm approach may be a valuable tool for stand-alone or hybrid alignments.
Automatic alignment of renal DCE-MRI image series for improvement of quantitative tracer kinetic studies
Darko Zikic, Steven Sourbron, Xinxing Feng, et al.
Tracer kinetic modeling with dynamic contrast enhanced MRI (DCE-MRI) and the quantification of the kinetic parameters are active fields of research which have the potential to improve the measurement of renal function. However, the strong coronal motion of the kidney in the time series inhibits an accurate assessment of the kinetic parameters. Automatic motion correction is challenging due to the large movement of the kidney and the strong intensity changes caused by the injected bolus. In this work, we improve the quantification results by a template matching motion correction method using a gradient-based similarity measure. Thus, a tedious manual motion correction is replaced by an automatic procedure. The only remaining user interaction is reduced to a selection of a reference slice and a coarse manual segmentation of the kidney in this slice. These steps do not present an overhead to the interaction needed for the assessment of the kinetic parameters. In order to achieve reliable and fast results, we constrain the degrees of freedom for the correction method as far as possible. Furthermore, we compare our method to deformable registration using the same similarity measure. In all our tests, the presented template matching correction was superior to the deformable approach in terms of reliability, leading to more accurate parameter quantification. The evaluation on 10 patient data series with 180-230 images each demonstrate that the quantitative analysis by a two-compartment model can be improved by our method.
Consistent detection of mid-sagittal planes for follow-up MR brain studies
The mid-sagittal plane (MSP) is a commonly used anatomic landmark for standardized MR brain acquisition. In addition to the requirement of accurate detection of the MSP geometry, it is also imperative from clinical point of view to consistently prescribe scan planning for evaluation of pathology process in follow-up studies. In this work, an adaptive technique of scan planning has been developed to enforce the consistency among scans acquired at different time points from the same patient by maximizing image similarity in the proximity of MSP. The geometry parameters of the MSP of current study are optimized by simplex algorithm to achieve better similarity to the reference study. Meanwhile different similarity measures are studied and evaluated within the region of the interest of each MSP. The method is successfully tested on self-reference consistency study by manually setting the reference sagittal image. It is also tested with clinical follow-up studies of MR images acquired from 30 patients. By visual inspection, the adaptive consistency method improves the similarity to the reference images in 22 follow-up studies evidently, while the similarity to the reference images in 7 studies improves slightly. This result demonstrates the efficacy of our method on consistent detection of mid-sagittal planes for follow-up MR brain study.
A rapid and robust iterative closest point algorithm for image-guided radiotherapy
Joseph Barbiere, Joseph Hanley
Our work presents a rapid and robust process that can analytically evaluate and correct patient setup error for head and neck radiotherapy by comparing orthogonal megavoltage portal images with digitally reconstructed radiographs. For robust data Photoshop is used to interactively segment images and registering reference contours to the transformed PI. MatLab is used for matrix computations and image analysis. The closest point distance for each PI point to a DRR point forms a set of homologous points. The translation that aligns the PI to the DRR is equal to the difference in centers of mass. The original PI points are transformed and the process repeated with an Iterative Closest Point algorithm until the transformation change becomes negligible. Using a 3.00 GHz processor the calculation of the 2500x1750 CPD matrix takes about 150 sec per iteration. Standard down sampling to about 1000 DRR and 250 PI points significantly reduces that time. We introduce a local neighborhood matrix consisting of a small subset of the DRR points in the vicinity of each PI point to further reduce the CPD matrix size. Our results demonstrate the effects of down sampling on accuracy. For validation, analytical detailed results are displayed as a histogram.
Retinal image mosaicing using the radial distortion correction model
Sangyeol Lee, Michael D. Abràmoff M.D., Joseph M. Reinhardt
Fundus camera imaging can be used to examine the retina to detect disorders. Similar to looking through a small keyhole into a large room, imaging the fundus with an ophthalmologic camera allows only a limited view at a time. Thus, the generation of a retinal montage using multiple images has the potential to increase diagnostic accuracy by providing larger field of view. A method of mosaicing multiple retinal images using the radial distortion correction (RADIC) model is proposed in this paper. Our method determines the inter-image connectivity by detecting feature correspondences. The connectivity information is converted to a tree structure that describes the spatial relationships between the reference and target images for pairwise registration. The montage is generated by cascading pairwise registration scheme starting from the anchor image downward through the connectivity tree hierarchy. The RADIC model corrects the radial distortion that is due to the spherical-to-planar projection during retinal imaging. Therefore, after radial distortion correction, individual images can be properly mapped onto a montage space by a linear geometric transformation, e.g. affine transform. Compared to the most existing montaging methods, our method is unique in that only a single registration per image is required because of the distortion correction property of RADIC model. As a final step, distance-weighted intensity blending is employed to correct the inter-image differences in illumination encountered when forming the montage. Visual inspection of the experimental results using three mosaicing cases shows our method can produce satisfactory montages.
Deformation estimation and analysis for adaptive radiation therapy
Bin Wang, Jianhua Xuan, Jackie Qingrong Wu, et al.
To accommodate the inter- and intra-fractional motion of internal organs in prostate cancer treatment, a large margin (5mm-25mm) has often to be considered during radiation therapy planning. Normally, the inter-fractional motion is more substantial than the intra-fractional counterpart. Therefore, the study of inter-fractional motion pattern is of special interest for adaptive radiation therapy. Existing methods on organ motion analysis mainly focus on the deviation of an organ's shape from its mean shape. The deviation information is helpful in choosing a statistically proper margin, but is of limited use for plan adaptation. In this paper, we propose a new deformation analysis method that can be directly used for plan adaptation. First, deformation estimation is accomplished by a fast deformable registration method, which utilizes a contour based multi-grid strategy to register treatment cone-beam CT (CBCT) images with planning CT images. Second, dominant deformation modes are extracted by a novel deformation analysis approach. To be specific, a cooperative principal component analysis (PCA) method is developed to analyze the deformation field in a coarse-to-fine strategy. The deformation modes are initialized by applying PCA on the organs as a whole and refined by analyzing the individual organs subsequently. The experimental results show that the organ motion can be well characterized by a few dominant deformation modes. Based on the dominant modes, a corresponding set of dominant modal plans could be generated for further optimization. Ultimately, an adaptive plan for each treatment can be obtained on-line while the margin can be effectively reduced to minimize the unnecessary radiation dosage.
Posters: Segmentation
icon_mobile_dropdown
Simultaneous segmentation and motion estimation in 4D-CT data using a variational approach
Spatiotemporal image data sets, like 4D CT or dynamic MRI, open up the possibility to estimate respiratory induced tumor and organ motion and to generate four-dimensional models that describe the temporal change in position and shape of structures of interest. However, two main problems arise: the structures of interest have to be segmented in the 4D data set and and the organ motion has to be estimated in the temporal image sequence. This paper presents a variational approach for simultaneous segmentation and registration applied to temporal image sequences. The proposed method assumes a known segmentation in one frame and then recovers nonlinear registration and segmentation in other frames by minimizing a cost function that combines intensity-based registration, level-set segmentation as well as prior shape and intensity knowledge. The purpose of the presented method is to estimate respiration induced organ motion in spatiotemporal CT image sequences and to segment a structure of interest simultaneously. A validation of the combined registration and segmentation approach is presented using low dose 4D CT data sets of the liver. The results demonstrate that the simultaneous solution of both problems improves the segmentation performance over a sequential application of the registration and segmentation steps.
Recent advances in 3D-CSC based MR brain image segmentation
Frank Schmitt, Lutz Priese
The 3D-CSC is a general segmentation method for voxel images. One of its possible applications is the segmentation of MR images of the human head. We here propose a self-contained method consisting of preprocessing steps which remove common artifacts from the input image, a 3D-CSC segmentation which partitions the input image into gray value similar, spatially connected regions and a final classification of CSC segments into white matter, gray matter and non-brain. We evaluate our method using the brainweb dataset for which a ground truth is available.
Automatic knee cartilage delineation using inheritable segmentation
Sebastian P. M. Dries M.D., Vladimir Pekar, Daniel Bystrov, et al.
We present a fully automatic method for segmentation of knee joint cartilage from fat suppressed MRI. The method first applies 3-D model-based segmentation technology, which allows to reliably segment the femur, patella, and tibia by iterative adaptation of the model according to image gradients. Thin plate spline interpolation is used in the next step to position deformable cartilage models for each of the three bones with reference to the segmented bone models. After initialization, the cartilage models are fine adjusted by automatic iterative adaptation to image data based on gray value gradients. The method has been validated on a collection of 8 (3 left, 5 right) fat suppressed datasets and demonstrated the sensitivity of 83±6% compared to manual segmentation on a per voxel basis as primary endpoint. Gross cartilage volume measurement yielded an average error of 9±7% as secondary endpoint. For cartilage being a thin structure, already small deviations in distance result in large errors on a per voxel basis, rendering the primary endpoint a hard criterion.
Computerized image analysis for acetic acid induced intraepithelial lesions
Wenjing Li, Daron G. Ferris M.D., Rich W. Lieberman
Cervical Intraepithelial Neoplasia (CIN) exhibits certain morphologic features that can be identified during a visual inspection exam. Immature and dysphasic cervical squamous epithelium turns white after application of acetic acid during the exam. The whitening process occurs visually over several minutes and subjectively discriminates between dysphasic and normal tissue. Digital imaging technologies allow us to assist the physician analyzing the acetic acid induced lesions (acetowhite region) in a fully automatic way. This paper reports a study designed to measure multiple parameters of the acetowhitening process from two images captured with a digital colposcope. One image is captured before the acetic acid application, and the other is captured after the acetic acid application. The spatial change of the acetowhitening is extracted using color and texture information in the post acetic acid image; the temporal change is extracted from the intensity and color changes between the post acetic acid and pre acetic acid images with an automatic alignment. The imaging and data analysis system has been evaluated with a total of 99 human subjects and demonstrate its potential to screening underserved women where access to skilled colposcopists is limited.
Improving 3D active appearance model segmentation of the left ventricle with Jacobian tuning
K .Y. E. Leung, M. van Stralen, M. M. Voormolen, et al.
Automated image processing techniques may prove invaluable in the examination of real-time three-dimensional echocardiograms, by providing quantitative and objective measurements of functional parameters such as left ventricular (LV) volume and ejection fraction. In this study, we investigate the use of active appearance models (AAMs) for automatic detection of left ventricular endocardial contours. AAMs are especially useful in segmenting ultrasound images, due to their ability to model the typical LV appearance. However, since only a limited number of images is available for training, the model may be incapable of capturing the large variability in ultrasound image appearance. This may cause standard AAM matching procedures to fail if the model and image are significantly different. Recently, a Jacobian-tuning method for AAM matching was proposed, which allowed the model's training matrix to adapt to the new, unseen image. This may potentially result in a more robust matching. To compare both matching methods, AAMs were built with end-diastolic images from 54 patients. Larger capture ranges and higher accuracy were obtained when the new method was used. In conclusion, this method has great potential for segmentation in echocardiograms and will improve the assessment of LV functional parameters.
Novel method for digital subtraction of tagged stool in virtual colonoscopy
Lutz Guendel, Michael Suehling, Helmut Eckert
Colon cancer is one of the most frequent causes of death. CT colonography is a novel method for the detection of polyps and early cancer. The general principle of CT colonography includes a cathartic bowel preparation. The resulting discomfort for patients leads to limited patient acceptance and therefore to limited cancer detection rates. Reduced bowel preparation, techniques for stool tagging, and electronic cleansing, however, improve the acceptance rates. Hereby, the high density of oral contrast material highlights residual stool and can be digitally removed. Known subtraction methods cause artifacts: additional 3D objects are introduced and small bowel folds are perforated. We propose a new algorithm that is based on the 2nd derivative of the image data using the Hessian matrix and the following principal axis transform to detect tiny folds which shall not be subtracted together with tagged stool found by a thresholding method. Since the stool is usually not homogenously tagged with contrast media a detection algorithm for island-like structures is incorporated. The interfaces of air-stool level and colon wall are detected by a 3-dimensional difference of Gaussian module. A 3-dimensional filter smoothes the transitions between removed stool and colon tissue. We evaluated the efficacy of the new algorithm with 10 patient data sets. The results showed no introduced artificial objects and no perforated folds. The artifacts at the air-stool and colon tissue-stool transitions are considerably reduced compared to those known from the literature.
Airway segmentation by topology-driven local thresholding
We describe a method for segmenting airway trees from greyscale 3D images such as CT (Computed Tomography) scans. Our approach is based on topological analysis of sets obtained by thresholding from thick slices, i.e. sub-images consisting of a small number of consecutive slices. From each thick slice under consideration, we select all sets S obtained from that thick slice by thresholding that have simple enough topological structure. As the selection criterion, we use a simple algebraic condition involving the numbers of connected components in the intersection of the set S with every slice in the thick slice. The condition basically asserts that the intersections of S with each of the slices is small and attempts to limit the number of the branching points of S within the thick slice. The output 3D model of the airway tree is obtained as the largest connected component of the union of all selected sets, extracted from several overlapping thick slices. Experiments with a number of chest CT scans show that the method leads to promising results.
Improving cervical region of interest by eliminating vaginal walls and cotton-swabs for automated image analysis
Image analysis for automated diagnosis of cervical cancer has attained high prominence in the last decade. Automated image analysis at all levels requires a basic segmentation of the region of interest (ROI) within a given image. The precision of the diagnosis is often reflected by the precision in detecting the initial region of interest, especially when some features outside the ROI mimic the ones within the same. Work described here discusses algorithms that are used to improve the cervical region of interest as a part of automated cervical image diagnosis. A vital visual aid in diagnosing cervical cancer is the aceto-whitening of the cervix after the application of acetic acid. Color and texture are used to segment acetowhite regions within the cervical ROI. Vaginal walls along with cottonswabs sometimes mimic these essential features leading to several false positives. Work presented here is focused towards detecting in-focus vaginal wall boundaries and then extrapolating them to exclude vaginal walls from the cervical ROI. In addition, discussed here is a marker-controlled watershed segmentation that is used to detect cottonswabs from the cervical ROI. A dataset comprising 50 high resolution images of the cervix acquired after 60 seconds of acetic acid application were used to test the algorithm. Out of the 50 images, 27 benefited from a new cervical ROI. Significant improvement in overall diagnosis was observed in these images as false positives caused by features outside the actual ROI mimicking acetowhite region were eliminated.
CALM: cascading system with leaking detection mechanism for medical image segmentation
Jiang Liu, Joo Hwee Lim, Huiqi Li
Medical image segmentation is a challenging process due to possible image over-segmentation and under-segmentation (leaking). The CALM medical image segmentation system is constructed with an innovative scheme that cascades threshold level-set and region-growing segmentation algorithms using Union and Intersection set operators. These set operators help to balance the over-segmentation rate and under-segmentation rate of the system respectively. While adjusting the curvature scalar parameter in the threshold level-set algorithm, we observe that the abrupt change in the size of the segmented areas coincides with the occurrences of possible leaking. Instead of randomly choose a value or use the system default curvature scalar values, this observation prompts us to use the following formula in CALM to automatically decide the optimal curvature values γ to prevent the occurrence of leaking : δ2S/δγ2 >= M, where S is the size of the segmented area and M is a large positive number. Motivated for potential applications in organ transplant and analysis, the CALM system is tested on the segmentation of the kidney regions from the Magnetic Resonance images taken from the National University Hospital of Singapore. Due to the nature of MR imaging, low-contrast, weak edges and overlapping regions of adjacent organs at kidney boundaries are frequently seen in the datasets and hence kidney segmentation is prone to leaking. The kidney segmentation accuracy rate achieved by CALM is 22% better compared with those achieved by the component algorithms or the system without leaking detection mechanism. CALM is easy-to-implement and can be applied to many applications besides kidney segmentation.
Validation of automatic landmark identification for atlas-based segmentation for radiation treatment planning of the head-and-neck region
Claudia Leavens, Torbjørn Vik, Heinrich Schulz, et al.
Manual contouring of target volumes and organs at risk in radiation therapy is extremely time-consuming, in particular for treating the head-and-neck area, where a single patient treatment plan can take several hours to contour. As radiation treatment delivery moves towards adaptive treatment, the need for more efficient segmentation techniques will increase. We are developing a method for automatic model-based segmentation of the head and neck. This process can be broken down into three main steps: i) automatic landmark identification in the image dataset of interest, ii) automatic landmark-based initialization of deformable surface models to the patient image dataset, and iii) adaptation of the deformable models to the patient-specific anatomical boundaries of interest. In this paper, we focus on the validation of the first step of this method, quantifying the results of our automatic landmark identification method. We use an image atlas formed by applying thin-plate spline (TPS) interpolation to ten atlas datasets, using 27 manually identified landmarks in each atlas/training dataset. The principal variation modes returned by principal component analysis (PCA) of the landmark positions were used by an automatic registration algorithm, which sought the corresponding landmarks in the clinical dataset of interest using a controlled random search algorithm. Applying a run time of 60 seconds to the random search, a root mean square (rms) distance to the ground-truth landmark position of 9.5 ± 0.6 mm was calculated for the identified landmarks. Automatic segmentation of the brain, mandible and brain stem, using the detected landmarks, is demonstrated.
Segmentation in noisy medical images using PCA model based particle filtering
Wei Qu, Xiaolei Huang, Yuanyuan Jia
Existing common medical image segmentation algorithms such as snake or graph cut usually could not generate satisfying results for noisy medical images such as X-ray angiographical and ultrasound images where the image quality is very poor including substantial background noise, low contrast, clutter, etc. In this paper, we present a robust segmentation method for noisy medical image analysis using Principle Component Analysis (PCA) based particle filtering. It exploits the prior clinical knowledge of desired object's shape through a PCA model. The preliminary results have shown the effectiveness and efficiency of the proposed approach on both synthetic and real clinical data.
Semi-automatic detection of Gd-DTPA-saline filled capsules for colonic transit time assessment in MRI
Christian Harrer, Sonja Kirchhoff M.D., Andreas Keil, et al.
Functional gastrointestinal disorders result in a significant number of consultations in primary care facilities. Chronic constipation and diarrhea are regarded as two of the most common diseases affecting between 2&percent; and 27&percent; of the population in western countries1-3. Defecatory disorders are most commonly due to dysfunction of the pelvic floor or the anal sphincter. Although an exact differentiation of these pathologies is essential for adequate therapy, diagnosis is still only based on a clinical evaluation1. Regarding quantification of constipation only the ingestion of radio-opaque markers or radioactive isotopes and the consecutive assessment of colonic transit time using X-ray or scintigraphy, respectively, has been feasible in clinical settings4-8. However, these approaches have several drawbacks such as involving rather inconvenient, time consuming examinations and exposing the patient to ionizing radiation. Therefore, conventional assessment of colonic transit time has not been widely used. Most recently a new technique for the assessment of colonic transit time using MRI and MR-contrast media filled capsules has been introduced9. However, due to numerous examination dates per patient and corresponding datasets with many images, the evaluation of the image data is relatively time-consuming. The aim of our study was to develop a computer tool to facilitate the detection of the capsules in MRI datasets and thus to shorten the evaluation time. We present a semi-automatic tool which provides an intensity, size10, and shape-based11,12 detection of ingested Gd-DTPA-saline filled capsules. After an automatic pre-classification, radiologists may easily correct the results using the application-specific user interface, therefore decreasing the evaluation time significantly.
A learning-based automatic spinal MRI segmentation
Xiaoqing Liu, Jagath Samarabandu, Greg Garvin, et al.
Image segmentation plays an important role in medical image analysis and visualization since it greatly enhances the clinical diagnosis. Although many algorithms have been proposed, it is still challenging to achieve an automatic clinical segmentation which requires speed and robustness. Automatically segmenting the vertebral column in Magnetic Resonance Imaging (MRI) image is extremely challenging as variations in soft tissue contrast and radio-frequency (RF) in-homogeneities cause image intensity variations. Moveover, little work has been done in this area. We proposed a generic slice-independent, learning-based method to automatically segment the vertebrae in spinal MRI images. A main feature of our contributions is that the proposed method is able to segment multiple images of different slices simultaneously. Our proposed method also has the potential to be imaging modality independent as it is not specific to a particular imaging modality. The proposed method consists of two stages: candidate generation and verification. The candidate generation stage is aimed at obtaining the segmentation through the energy minimization. In this stage, images are first partitioned into a number of image regions. Then, Support Vector Machines (SVM) is applied on those pre-partitioned image regions to obtain the class conditional distributions, which are then fed into an energy function and optimized with the graph-cut algorithm. The verification stage applies domain knowledge to verify the segmented candidates and reject unsuitable ones. Experimental results show that the proposed method is very efficient and robust with respect to image slices.
Reclassification of segmentation boundary base on neighboring function
Jian Chen, Jie Tian
Motivated by the goal of improving the performance of segmentation, a new technique based on Neighboring Function (NF) is presented to reclassify the rough segmentation boundary pixels. The NF is a novel measurement of neighboring relationship, it takes into consideration of both spatial and intensity information and their distribution pattern. With the rough boundary provided by other segmentation algorithms, the value of the NF at each boundary pixel is calculated in a specific neighborhood, then the reclassification will be implemented by comparing these values. In order to obtain the expected boundary, this step is iterated for several times. In our study, the proposed method is applied to synthetic and real medical images, a great improvement of the quality of segmentation boundary has been achieved. The accuracy and reproducibility of this reclassification method has been proven by experimental results. Experiments also show that this method is insensitive to noise.
Effect of various binning methods and ROI sizes on the accuracy of the automatic classification system for differentiation between diffuse infiltrative lung diseases on the basis of texture features at HRCT
To find optimal binning, variable binning size linear binning (LB) and non-linear binning (NLB) methods were tested. In case of small binning size (Q ≤ 10), NLB shows significant better accuracy than the LB. K-means NLB (Q = 26) is statistically significant better than every LB. To find optimal binning method and ROI size of the automatic classification system for differentiation between diffuse infiltrative lung diseases on the basis of textural analysis at HRCT Six-hundred circular regions of interest (ROI) with 10, 20, and 30 pixel diameter, comprising of each 100 ROIs representing six regional disease patterns (normal, NL; ground-glass opacity, GGO; reticular opacity, RO; honeycombing, HC; emphysema, EMPH; and consolidation, CONS) were marked by an experienced radiologist from HRCT images. Histogram (mean) and co-occurrence matrix (mean and SD of angular second moment, contrast, correlation, entropy, and inverse difference momentum) features were employed to test binning and ROI effects. To find optimal binning, variable binning size LB (bin size Q: 4~30, 32, 64, 128, 144, 196, 256, 384) and NLB (Q: 4~30) methods (K-means, and Fuzzy C-means clustering) were tested. For automated classification, a SVM classifier was implemented. To assess cross-validation of the system, a five-folding method was used. Each test was repeatedly performed twenty times. Overall accuracies with every combination of variable ROIs, and binning sizes were statistically compared. In case of small binning size (Q ≤ 10), NLB shows significant better accuracy than the LB. K-means NLB (Q = 26) is statistically significant better than every LB. In case of 30x30 ROI size and most of binning size, the K-means method showed better than other NLB and LB methods. When optimal binning and other parameters were set, overall sensitivity of the classifier was 92.85%. The sensitivity and specificity of the system for each class were as follows: NL, 95%, 97.9%; GGO, 80%, 98.9%; RO 85%, 96.9%; HC, 94.7%, 97%; EMPH, 100%, 100%; and CONS, 100%, 100%, respectively. We determined the optimal binning method and ROI size of the automatic classification system for differentiation between diffuse infiltrative lung diseases on the basis of texture features at HRCT.
Interactive surface correction for 3D shape based segmentation
Tobias Schwarz, Tobias Heimann, Ralf Tetzlaff, et al.
Statistical shape models have become a fast and robust method for segmentation of anatomical structures in medical image volumes. In clinical practice, however, pathological cases and image artifacts can lead to local deviations of the detected contour from the true object boundary. These deviations have to be corrected manually. We present an intuitively applicable solution for surface interaction based on Gaussian deformation kernels. The method is evaluated by two radiological experts on segmentations of the liver in contrast-enhanced CT images and of the left heart ventricle (LV) in MRI data. For both applications, five datasets are segmented automatically using deformable shape models, and the resulting surfaces are corrected manually. The interactive correction step improves the average surface distance against ground truth from 2.43mm to 2.17mm for the liver, and from 2.71mm to 1.34mm for the LV. We expect this method to raise the acceptance of automatic segmentation methods in clinical application.
An approach to segment lung pleura from CT data with high precision
E. Angelats, K. Chaisaowong, A. Knepper, et al.
A new approach to segment pleurae from CT data with high precision is introduced. This approach is developed in the segmentation's framework of an image analysis system to automatically detect pleural thickenings. The new technique to carry out the 3D segmentation of lung pleura is based on supervised range-constrained thresholding and a Gibbs-Markov random field model. An initial segmentation is done using the 3D histogram by supervised range-constrained thresholding. 3D connected component labelling is then applied to find the thorax. In order to detect and remove trachea and bronchi therein, the 3D histogram of connected pulmonary organs is modelled as a finite mixture of Gaussian distributions. Parameters are estimated using the Expectation-Maximization algorithm, which leads to the classification of that pulmonary region. As consequence left and right lungs are separated. Finally we apply a Gibbs-Markov random field model to our initial segmentation in order to achieve a high accuracy segmentation of lung pleura. The Gibbs- Markov random field is combined with maximum a posteriori estimation to estimate optimal pleural contours. With these procedures, a new segmentation strategy is developed in order to improve the reliability and accuracy of the detection of pleural contours and to achieve a better assessment performance of pleural thickenings.
Boundary-precise segmentation of nucleus and plasma of leukocytes
The exact segmentation of nucleus and plasma of a white blood cell (leukocyte) is the basis for the creation of an automatic, image based differential white blood cell count(WBC). In this contribution we present an approach for the according segmentation of leukocytes. For a valid classification of the different cell classes, a precise segmentation is essential. Especially concerning immature cells, which can be distinguished from their mature counterparts only by small differences in some features, a segmentation of nucleus and plasma has to be as precise as possible, to extract those differences. Also the problems with adjacent erythrocyte cells and the usage of a LED illumination are considered. The presented approach can be separated into several steps. After preprocessing by a Kuwahara-filter, the cell is localized by a simple thresholding operation, afterwards a fast-marching method for the localization of a rough cell boundary is defined. To retrieve the cell area a shortest-path-algorithm is applied next. The cell boundary found by the fast-marching approach is finally enhanced by a post-processing step. The concluding segmentation of the cell nucleus is done by a threshold operation. An evaluation of the presented method was done on a representative sample set of 80 images recorded with LED illumination and a 63-fold magnification dry objective. The automatically segmented cell images are compared to a manual segmentation of the same dataset using the Dice-coefficient as well as Hausdorff-distance. The results show that our approach is able to handle the different cell classes and that it improves the segmentation quality significantly.
Fully automated segmentation of carotid and vertebral arteries from contrast-enhanced CTA
Olivier Cuisenaire, Sunny Virmani, Mark E. Olszewski, et al.
We propose a method for segmenting and labeling the main head and neck vessels (common, internal, external carotid, vertebral) from a contrast enhanced computed tomography angiography (CTA) volume. First, an initial centerline of each vessel is extracted. Next, the vessels are segmented using 3D active objects initialized using the first step. Finally, the true centerline is identified by smoothly deforming it away from the segmented mask edges using a spline-snake. We focus particularly on the novel initial centerline extraction technique. It uses a locally adaptive front propagation algorithm that attempts to find the optimal path connecting the ends of the vessel, typically from the lowest image of the scan to the Circle of Willis in the brain. It uses a patient adapted anatomical model of the different vessels both to initialize and constrain this fast marching, thus eliminating the need for manual selection of seed points. The method is evaluated using data from multiple regions (USA, India, China, Israel) including a variety of scanners (10, 16, 40, 64-slice; Brilliance CT, Philips Healthcare, Cleveland, OH, USA), contrast agent dose, and image resolution. It is fully successful in over 90% of patients and only misses a single vessel in most remaining cases. We also demonstrate its robustness to metal and dental artifacts and anatomical variability. Total processing time is approximately two minutes with no user interaction, which dramatically improves the workflow over existing clinical software. It also reduces patient dose exposure by obviating the need to acquire an unenhanced scan for bone suppression as this can be done by applying the segmentation masks.
Simultaneous detection of multiple elastic surfaces with application to tumor segmentation in CT images
Kang Li, Marie-Pierre Jolly
We present a new semi-supervised method for segmenting multiple interrelated object boundaries with spherical topology in volumetric images. The core of our method is a novel graph-theoretic algorithm that simultaneously detects multiple surfaces under smoothness, distance, and elasticity constraints. The algorithm computes the global optimum of an objective function that incorporates boundary, regional and surface elasticity information. A single straight line drawn by the user in a cross-sectional slice is the sole user input, which roughly indicates the extent of the object. We employ a multi-seeded Dijkstra-based range competition algorithm to pre-segment the object on two orthogonal multiplanar reformatted (MPR) planes that pass through the input line. Based on the 2D pre-segmentation results, we estimate the object and background intensity histograms, and employ an adaptive mean-shift mode-seeking process on the object histogram to automatically determine the number of surface layers to be segmented. The final multiple-surface segmentation is performed in an ellipsoidal coordinate frame constructed by an automated ellipsoid fitting procedure. We apply our method to the segmentation of liver lesions with necrosis or calcification, and various other tumors in CT images. For liver tumor segmentation, our method can simultaneously delineate both tumor and necrosis boundaries. This capability is unprecedented and is valuable for cancer diagnosis, treatment planning, and evaluation.
An efficient topology adaptation system for parametric active contour segmentation of 3D images
Jochen Abhau, Otmar Scherzer
Active contour models have already been used succesfully for segmentation of organs from medical images in 3D. In implicit models, the contour is given as the isosurface of a scalar function, and therefore topology adaptations are handled naturally during a contour evolution. Nevertheless, explicit or parametric models are often preferred since user interaction and special geometric constraints are usually easier to incorporate. Although many researchers have studied topology adaptation algorithms in explicit mesh evolutions, no stable algorithm is known for interactive applications. In this paper, we present a topology adaptation system, which consists of two novel ingredients: A spatial hashing technique is used to detect self-colliding triangles of the mesh whose expected running time is linear with respect to the number of mesh vertices. For the topology change procedure, we have developed formulas by homology theory. During a contour evolution, we just have to choose between a few possible mesh retriangulations by local triangle-triangle intersection tests. Our algorithm has several advantages compared to existing ones: Since the new algorithm does not require any global mesh reparametrizations, it is very efficient. Since the topology adaptation system does not require constant sampling density of the mesh vertices nor especially smooth meshes, mesh evolution steps can be performed in a stable way with a rather coarse mesh. We apply our algorithm to 3D ultrasonic data, showing that accurate segmentation is obtained in some seconds.
Time-dependent joint probability speed function for level-set segmentation of rat brain slices
Christoph Palm, Uwe Pietrzyk
Introduction - The segmentation of rat brain slices suffers from illumination inhomogeneities and staining effects. State-of-the-art level-set methods model slice and background with intensity mixture densities defining the speed function as difference between the respective probabilites. Nevertheless, the overlap of these distributions causes an inaccurate stopping at the slice border. In this work, we propose the characterisation of the border area with intensity pairs for inside and outside estimating joint intensity probabilities. Method - In contrast to global object and background models, we focus on the object border characterised by a joint mixture density. This specifies the probability of the occurance of an inside and an outside value in direct adjacency. These values are not known beforehand, because inside and outside depend on the level-set evolution and change during time. Therefore, the speed function is computed time-dependently at the position of the current zero level-set. Along this zero level-set curve, the inside and outside values are derived as mean along the curvature normal directing inside and outside the object. Advantage of the joint probability distribution is to resolve the distribution overlaps, because these are assumed to be not located at the same border position. Results - The novel time-dependent joint probability based speed function is compared expermimentally with single probability based speed functions. Two rat brains with about 40 slices are segmented and the results analysed using manual segmentations and the Tanimoto overlap measure. Improved results are recognised for both data sets.
Multi-phase image segmentation using level sets
A hierarchical multi-phase image segmentation using the original and a modified Chan-Vese 2-phase method is considered. A method of capturing features inside a pre-selected region of interest (ROI) is proposed that effectively restricts the segmentation operation to the ROI. At the first step, a modified image is created by setting the portion of the image outside the ROI to a uniform intensity equal to the mean image intensity inside the ROI. Effectively, this procedure partitions the initial image into two phases, in such a way that the ROI effectively becomes a 'segmented' feature. At the second step, the segmentation procedure is applied to the modified image, partitioning the image in two phases - object and background - inside the ROI. By confining segmentation to the ROI, it is shown, using an artificial image, that objects can be discriminated that could not have been found if segmentation had been performed on the entire image. If necessary, this second step can be repeated to further segment features of interest within the ROI, thereby providing a multi-phase segmentation procedure. ROI placement around features of interest requires prior knowledge, and may be derived from an atlas or manually prescribed by the operator. In this way, segmentation is possible on low-contrast features of interest, while ignoring features irrelevant for a particular application. Examples are provided for segmentation of several 2D/3D images performed both on entire images and inside a ROI.
Bidirectional segmentation of prostate capsule from ultrasound volumes: an improved strategy
Prostate volume is an indirect indicator for several prostate diseases. Volume estimation is a desired requirement during prostate biopsy, therapy and clinical follow up. Image segmentation is thus necessary. Previously, discrete dynamic contour (DDC) was implemented in orthogonal unidirectional on the slice-by-slice basis for prostate boundary estimation. This suffered from the glitch that it needed stopping criteria during the propagation of segmentation procedure from slice-to-slice. To overcome this glitch, axial DDC was implemented and this suffered from the fact that central axis never remains fixed and wobbles during propagation of segmentation from slice-to-slice. The effect of this was a multi-fold reconstructed surface. This paper presents a bidirectional DDC approach, thereby removing the two glitches. Our bidirectional DDC protocol was tested on a clinical dataset on 28 3-D ultrasound image volumes acquired using side fire Philips transrectal ultrasound. We demonstrate the orthogonal bidirectional DDC strategy achieved the most accurate volume estimation compared with previously published orthogonal unidirectional DDC and axial DDC methods. Compared to the ground truth, we show that the mean volume estimation errors were: 18.48%, 9.21% and 7.82% for unidirectional, axial and bidirectional DDC methods, respectively. The segmentation architecture is implemented in Visual C++ in Windows environment.
Robust segmentation using kernel and spatial based fuzzy c-means methods on breast x-ray images
Robust methods for precise segmentation of breast region or volume from breast X-ray images, including mammogram and tomosynthetic image, is crucial for applications of these medical images. However, this task is challenging because the acquired images not only are inherent noisy and inhomogeneous, but there are also connected or overlapped artifacts, or noises on the images as well, due to local volume effect of tissues, parametric resolutions and other physical limitations of the imaging device. This paper proposes and develops robust fuzzy c-means (FCM) segmentation methods for segmentation of breast region on breast x-ray images, including mammography and tomosynthesis, respectively. We develop spatial information- and kernel function- based FCM methods to differentiate breast area or breast volume. Spatial information based FCM method incorporates neighborhood pixels' intensities into segmentation because neighbored pixels on an image are highly correlated. Kernel based FCM algorithm is developed by transforming pixel intensity using kernel functions to better improve segmentation performance. The proposed segmentation methods are implemented on mammograms and tomosynthetic images and compared with conventional FCM results. Experiment results demonstrate the proposed segmentation methods are much better compared with traditional FCM method, and are more robust to noises. The developed kernel and spatial based FCM method will be applied for differentiation of breast density and abnormal regions within the breast region to examine its performance in reducing false positive segmentations.
Hierarchical segmentation of malignant gliomas via integrated contextual filter response
Shishir Dube, Jason J. Corso, Alan Yuille, et al.
We present a novel methodology for the automated segmentation of Glioblastoma Multiforme tumors given only a high-resolution T1 post-contrast enhanced channel, which is routinely done in clinical MR acquisitions. The main contribution of the paper is the integration of contextual filter responses, to obtain a better class separation of abnormal and normal brain tissues, into the multilevel segmentation by weighted aggregation (SWA) algorithm. The SWA algorithm uses neighboring voxel intensities to form an affinity between the respective voxels. The affinities are then recursively computed for all the voxel pairs in the given image and a series of cuts are made to produce segments that contain voxels with similar intensity properties. SWA provides a fast method of partitioning the image, but does not produce segments with meaning. Thus, a contextual filter response component was integrated to label the aggregates as tumor or non-tumor. The contextual filter responses were computed via texture filter responses based on the gray level co-occurrence matrix (GLCM) method. The GLCM results in texture features that are used to quantify the visual appearance of the tumor versus normal tissue. Our results indicate the benefit of incorporating contextual features and applying non-linear classification methods to segment and classify the complex case of grade 4 tumors.
Local control of speed function in level set segmentation using interactive interface for CT images
Automatic segmentation of specific organ from CT image is one of the most important problems for computer aided diagnosis. Therefore, many automatic segmentation methods have been proposed. However, in these conventional methods, it is necessary to set some empirical parameters by users for supporting the automatic segmentation. It takes, however, much time to obtain the optimum parameters for segmentation. Another problem is that although the optimum value of parameters varies spatially, conventional segmentation methods can give only one value of parameters for whole organ. In this paper, we propose a new segmentation method named hybrid segmentation of CT image based on level set segmentation. A local control parameter which can be controlled by the interactive interface used PHANToM omni is introduced for segmentation of pancreas from three dimensional CT image. It is shown that the proposed method can increase accuracy in less time than hand-written segmentation.
Automated retinal layer segmentation in OCT images using spatially variant filtering
We have developed a new method to segment and analyze retinal layers in optical coherence tomography (OCT) images with the intent of monitoring changes in thickness of retinal layers due to disease. OCT is an imaging modality that obtains cross-sectional images of the retina, which makes it possible to measure thickness of individual layers. In this paper we present a method that identifies six key layers in OCT images. OCT images present challenges to conventional edge detection algorithms, including that due to the presence of speckle noise which affects the sharpness of inter-layer boundaries significantly. We use a directional filter bank, which has a wedge shaped passband that helps reduce noise while maintaining edge sharpness, in contrast to previous methods that use Gaussian filter or median filter variants that reduce the edge sharpness resulting in poor edge-detection performance. This filter is utilized in a spatially variant setting which uses additional information from the intersecting scans. The validity of extracted edge cues is determined according to the amount of gray-level transition across the edge, strength, continuity, relative location and polarity. These cues are processed according to the retinal model that we have developed and the processing yields edge contours.
Neuronal nuclei localization in 3D using level set and watershed segmentation from laser scanning microscopy images
Yingxuan Zhu, Eric Olson, Arun Subramanian, et al.
Abnormalities of the number and location of cells are hallmarks of both developmental and degenerative neurological diseases. However, standard stereological methods are impractical for assigning each cell's nucleus position within a large volume of brain tissue. We propose an automated approach for segmentation and localization of the brain cell nuclei in laser scanning microscopy (LSM) embryonic mouse brain images. The nuclei in these images are first segmented by using the level set (LS) and watershed methods in each optical plane. The segmentation results are further refined by application of information from adjacent optical planes and prior knowledge of nuclear shape. Segmentation is then followed with an algorithm for 3D localization of the centroid of nucleus (CN). Each volume of tissue is thus represented by a collection of centroids leading to an approximate 10,000-fold reduction in the data set size, as compared to the original image series. Our method has been tested on LSM images obtained from an embryonic mouse brain, and compared to the segmentation and CN localization performed by an expert. The average Euclidian distance between locations of CNs obtained using our method and those obtained by an expert is 1.58±1.24 µm, a value well within the ~5 µm average radius of each nucleus. We conclude that our approach accurately segments and localizes CNs within cell dense embryonic tissue.
Prostate segmentation on pelvic CT images using a genetic algorithm
Payel Ghosh, Melanie Mitchell
A genetic algorithm (GA) for automating the segmentation of the prostate on pelvic computed tomography (CT) images is presented here. The images consist of slices from three-dimensional CT scans. Segmentation is typically performed manually on these images for treatment planning by an expert physician, who uses the "learned" knowledge of organ shapes, textures and locations to draw a contour around the prostate. Using a GA brings the flexibility to incorporate new "learned" information into the segmentation process without modifying the fitness function that is used to train the GA. Currently the GA uses prior knowledge in the form of texture and shape of the prostate for segmentation. We compare and contrast our algorithm with a level-set based segmentation algorithm, thereby providing justification for using a GA. Each individual of the GA population represents a segmenting contour. Shape variability of the prostate derived from manually segmented images is used to form a shape representation from which an individual of the GA population is randomly generated. The fitness of each individual is evaluated based on the texture of the region it encloses. The segmenting contour that encloses the prostate region is considered more fit than others and is more likely to be selected to produce an offspring over successive generations of the GA run. This process of selection, crossover and mutation is iterated until the desired region is segmented. Results of 2D and 3D segmentation are presented and future work is also discussed here.
Robust segmentation of tubular structures in medical images
Segmentation of blood vessels is a challenging problem due to poor contrast, noise, and specifics of vessels' branching and bending geometry. This paper describes a robust semi-automatic approach to extract the surface between two or more user-supplied end points for tubular- or vessel-like structures. We first use a minimal path technique to extract the shortest path between the user-supplied points. This path is the global minimizer of an active contour model's energy along all possible paths joining the end-points. Subsequently, the surface of interest is extracted using an edge-based level set segmentation approach. To prevent leakage into adjacent tissues, the algorithm uses a diameter constraint that does not allow the moving front to grow wider than the predefined diameter. Points constituting the extracted path(s) are automatically used as initialization seeds for the evolving level set function. To cope with any further leaks that may occur in the case of large variations of the vessel width between the user-supplied end-points, a freezing mechanism is designed to prevent the moving front to leak into undesired areas. The regions to be frozen are determined from few clicks by the user. The potential of the proposed approach is demonstrated on several synthetic and real images.
Segmentation of large periapical lesions toward dental computer-aided diagnosis in cone-beam CT scans
This paper presents an experimental study for assessing the applicability of general-purpose 3D segmentation algorithms for analyzing dental periapical lesions in cone-beam computed tomography (CBCT) scans. In the field of Endodontics, clinical studies have been unable to determine if a periapical granuloma can heal with non-surgical methods. Addressing this issue, Simon et al. recently proposed a diagnostic technique which non-invasively classifies target lesions using CBCT. Manual segmentation exploited in their study, however, is too time consuming and unreliable for real world adoption. On the other hand, many technically advanced algorithms have been proposed to address segmentation problems in various biomedical and non-biomedical contexts, but they have not yet been applied to the field of dentistry. Presented in this paper is a novel application of such segmentation algorithms to the clinically-significant dental problem. This study evaluates three state-of-the-art graph-based algorithms: a normalized cut algorithm based on a generalized eigen-value problem, a graph cut algorithm implementing energy minimization techniques, and a random walks algorithm derived from discrete electrical potential theory. In this paper, we extend the original 2D formulation of the above algorithms to segment 3D images directly and apply the resulting algorithms to the dental CBCT images. We experimentally evaluate quality of the segmentation results for 3D CBCT images, as well as their 2D cross sections. The benefits and pitfalls of each algorithm are highlighted.
Segmentation of sonographic breast lesions: fuzzy cell-competition algorithm and bias field reduction
Chia-Yen Lee, Chi-Chun Hsieh, Chung-Ming Chen
Bias field is a common phenomenon in a breast sonogram. Although artifacts caused by bias filed may carry important information, e.g., shadowing behind a lesion, they are generally disturbing in the process of automatic boundary delineation for sonographic breast lesions. This paper presents a new segmentation algorithm aiming to decompose the region of interest (ROI) into prominent components while estimating the bias field in the ROI. A prominent component is a contiguous region with a visually perceivable boundary, which might be a noise, an artifact, a substructure of a tissue or a part of breast lesion. The prominent components may be used as the basic constructs for a higher level segmentation algorithm to identify the lesion boundary. The bias field in an ROI is modeled as a spatially-variant Gaussian distribution with a constant variance and spatially-variant means, which is a polynomial surface of order n. The true gray levels of the pixels in a prominent component are assumed to be Gaussian-distributed. The proposed algorithm is formulated as an EM-algorithm composed of two major steps. In the E-step, the ROI is decomposed into prominent components using a new fuzzy cell-competition algorithm based on the bias field and model parameters estimated in the previous M-step. In the M-step, the bias field and model parameters are estimated based on the prominent components derived in the E-step using a least squared approach. The results show that the effect of bias field on segmentation has been reduced and better segmentation results have been attained.
Three-dimensional segmentation of bones from CT and MRI using fast level sets
Our task is to segment bones from 3D CT and MRI images. The main application is creation of 3D mesh models for finite element modeling. These surface and volume vector models can be used for further biomechanical processing and analysis. We selected a novel fast level set method because of its high computational efficiency, while preserving all advantages of traditional level set methods. Unlike in traditional level set methods, we are not solving partial differential equations (PDEs). Instead, the contours are represeted by two sets of points, corresponding to the inner and outer edge of the object boundary. We have extended the original implementation in 3D, where the speed advantage over classical level set segmentation are even more pronounced. We can segment a CT image of 512×512×125 in less than 20s by this method. It is approximately two orders of magnitude faster than standard narrow band algorithms. Our experiments with real 3D CT and MRI images presented in this paper showed high ability of the fast level set algorithm to solve complex segmentation problems.
3-D segmentation of articular cartilages by graph cuts using knee MR images from osteoarthritis initiative
Hackjoon Shim, Soochan Lee, Bohyeong Kim, et al.
Knee osteoarthritis is the most common debilitating health condition affecting elderly population. MR imaging of the knee is highly sensitive for diagnosis and evaluation of the extent of knee osteoarthritis. Quantitative analysis of the progression of osteoarthritis is commonly based on segmentation and measurement of articular cartilage from knee MR images. Segmentation of the knee articular cartilage, however, is extremely laborious and technically demanding, because the cartilage is of complex geometry and thin and small in size. To improve precision and efficiency of the segmentation of the cartilage, we have applied a semi-automated segmentation method that is based on an s/t graph cut algorithm. The cost function was defined integrating regional and boundary cues. While regional cues can encode any intensity distributions of two regions, "object" (cartilage) and "background" (the rest), boundary cues are based on the intensity differences between neighboring pixels. For three-dimensional (3-D) segmentation, hard constraints are also specified in 3-D way facilitating user interaction. When our proposed semi-automated method was tested on clinical patients' MR images (160 slices, 0.7 mm slice thickness), a considerable amount of segmentation time was saved with improved efficiency, compared to a manual segmentation approach.
Segmentation and volumetric measurement of renal cysts and parenchyma from MR images of polycystic kidneys using multi-spectral analysis method
K. T. Bae, P. K. Commean, B. S. Brunsden, et al.
For segmentation and volume measurement of renal cysts and parenchyma from kidney MR images in subjects with autosomal dominant polycystic kidney disease (ADPKD), a semi-automated, multi-spectral anaylsis (MSA) method was developed and applied to T1- and T2-weighted MR images. In this method, renal cysts and parenchyma were characterized and segmented for their characteristic T1 and T2 signal intensity differences. The performance of the MSA segmentation method was tested on ADPKD phantoms and patients. Segmented renal cysts and parenchyma volumes were measured and compared with reference standard measurements by fluid displacement method in the phantoms and stereology and region-based thresholding methods in patients, respectively. As results, renal cysts and parenchyma were segmented successfully with the MSA method. The volume measurements obtained with MSA were in good agreement with the measurements by other segmentation methods for both phantoms and subjects. The MSA method, however, was more time-consuming than the other segmentation methods because it required pre-segmentation, image registration and tissue classification-determination steps.
Semi-automated segmentation of the prostate gland boundary in ultrasound images using a machine learning approach
This paper presents a semi-automated algorithm for prostate boundary segmentation from three-dimensional (3D) ultrasound (US) images. The US volume is sampled into 72 slices which go through the center of the prostate gland and are separated at a uniform angular spacing of 2.5 degrees. The approach requires the user to select four points from slices (at 0, 45, 90 and 135 degrees) which are used to initialize a discrete dynamic contour (DDC) algorithm. 4 Support Vector Machines (SVMs) are trained over the output of the DDC and classify the rest of the slices. The output of the SVMs is refined using binary morphological operations and DDC to produce the final result. The algorithm was tested on seven ex vivo 3D US images of prostate glands embedded in an agar mold. Results show good agreement with manual segmentation.
Multiscale hierarchical support vector clustering
Michael Saas Hansen, David Alberg Holm, Karl Sjöstrand, et al.
Clustering is the preferred choice of method in many applications, and support vector clustering (SVC) has proven efficient for clustering noisy and high-dimensional data sets. A method for multiscale support vector clustering is demonstrated, using the recently emerged method for fast calculation of the entire regularization path of the support vector domain description. The method is illustrated on artificially generated examples, and applied for detecting blood vessels from high resolution time series of magnetic resonance imaging data. The obtained results are robust while the need for parameter estimation is reduced, compared to support vector clustering.
Fast approximate surface evolution in arbitrary dimension
James Malcolm, Yogesh Rathi, Anthony Yezzi, et al.
The level set method is a popular technique used in medical image segmentation; however, the numerics involved make its use cumbersome. This paper proposes an approximate level set scheme that removes much of the computational burden while maintaining accuracy. Abandoning a floating point representation for the signed distance function, we use integral values to represent the signed distance function. For the cases of 2D and 3D, we detail rules governing the evolution and maintenance of these three regions. Arbitrary energies can be implemented in the framework. This scheme has several desirable properties: computations are only performed along the zero level set; the approximate distance function requires only a few simple integer comparisons for maintenance; smoothness regularization involves only a few integer calculations and may be handled apart from the energy itself; the zero level set is represented exactly removing the need for interpolation off the interface; and evolutions proceed on the order of milliseconds per iteration on conventional uniprocessor workstations. To highlight its accuracy, flexibility and speed, we demonstrate the technique on intensity-based segmentations under various statistical metrics. Results for 3D imagery show the technique is fast even for image volumes.
An accurate segmentation method for volumetry of brain tumor in 3D MRI
Jiahui Wang, Qiang Li, Toshinori Hirai, et al.
Accurate volumetry of brain tumors in magnetic resonance imaging (MRI) is important for evaluating the interval changes in tumor volumes during and after treatment, and also for planning of radiation therapy. In this study, an automated volumetry method for brain tumors in MRI was developed by use of a new three-dimensional (3-D) image segmentation technique. First, the central location of a tumor was identified by a radiologist, and then a volume of interest (VOI) was determined automatically. To substantially simplify tumor segmentation, we transformed the 3-D image of the tumor into a two-dimensional (2-D) image by use of a "spiral-scanning" technique, in which a radial line originating from the center of the tumor scanned the 3-D image spirally from the "north pole" to the "south pole". The voxels scanned by the radial line provided a transformed 2-D image. We employed dynamic programming to delineate an "optimal" outline of the tumor in the transformed 2-D image. We then transformed the optimal outline back into 3-D image space to determine the volume of the tumor. The volumetry method was trained and evaluated by use of 16 cases with 35 brain tumors. The agreement between tumor volumes provided by computer and a radiologist was employed as a performance metric. Our method provided relatively accurate results with a mean agreement value of 88&percent;.
Automated segmentation of middle hepatic vein in non-contrast x-ray CT images based on an atlas-driven approach
In order to support the diagnosis of hepatic diseases, understanding the anatomical structures of hepatic lobes and hepatic vessels is necessary. Although viewing and understanding the hepatic vessels in contrast media-enhanced CT images is easy, the observation of the hepatic vessels in non-contrast X-ray CT images that are widely used for the screening purpose is difficult. We are developing a computer-aided diagnosis (CAD) system to support the liver diagnosis based on non-contrast X-ray CT images. This paper proposes a new approach to segment the middle hepatic vein (MHV), a key structure (landmark) for separating the liver region into left and right lobes. Extraction and classification of hepatic vessels are difficult in non-contrast X-ray CT images because the contrast between hepatic vessels and other liver tissues is low. Our approach uses an atlas-driven method by the following three stages. (1) Construction of liver atlases of left and right hepatic lobes using a learning datasets. (2) Fully-automated enhancement and extraction of hepatic vessels in liver regions. (3) Extraction of MHV based on the results of (1) and (2). The proposed approach was applied to 22 normal liver cases of non-contrast X-ray CT images. The preliminary results show that the proposed approach achieves the success in 14 cases for MHV extraction.
A deformable model-based minimal path segmentation method for kidney MR images
We developed a new minimal path segmentation method for mouse kidney MR images. We used dynamic programming and a minimal path segmentation approach to detect the optimal path within a weighted graph between two end points. The energy function combines distance and gradient information to guide the marching curve and thus to evaluate the best path and to span a broken edge. An algorithm was developed to automatically place initial end points. Dynamic programming was used to automatically optimize and update end points during the searching procedure. Principle component analysis (PCA) was used to generate a deformable model, which serves as the prior knowledge for the selection of initial end points and for the evaluation of the best path. The method has been tested for kidney MR images acquired from 44 mice. To quantitatively assess the automatic segmentation method, we compared the results with manual segmentation. The mean and standard deviation of the overlap ratios are 95.19%±0.03%. The distance error between the automatic and manual segmentation is 0.82±0.41 pixel. The automatic minimal path segmentation method is fast, accurate, and robust and it can be applied not only for kidney images but also for other organs.
Automated lung tumor detection and quantification for respiratory gated PET/CT images
Purpose: To develop and validate an automatic algorithm for the detection and functional assessment of lung tumors on three-dimensional respiratory gated PET/CT images. Method and Materials: First the algorithm will automatically segment lung regions in CT images, then identify and localize focal increases of activity in lung regions of PET images at each gated bin. Once the tumor voxels have been determined, an integration algorithm will include all the tumor counts collected at different bins within the respiratory cycle into one reference bin. Then the total activity (Bq), concentration (Bq/ml), functional volume (ml) and standard uptake values (SUV) are calculated for each tumor on PET images. Validation of the automatic algorithm was demonstrated by conducting experiments with the computerized 4D NCAT phantom and with a dynamic lung-chest phantom imaged using a GE PET/CT System at Baptist Hospital of Miami. Tumor variables to be controlled were: volume, total number of counts (activity), maximum and average number of counts. These values were the gold standard to which the results of the algorithm were compared. The tumor's motion was also controlled with different respiratory periods and amplitudes. Results: Validation, feasibility and robustness of the algorithm were demonstrated. With the algorithm, the best compromise between short PET scan time and reduced image noise can be achieved, while quantification and clinical analysis become faster and more precise.
Efficient curvature estimations for real-time (25Hz) segmentation of volumetric ultrasound data
Christopher R. Wagner, Douglas P. Perrin
While Moore's law has eliminated the need for algorithm optimization when computing 2D dynamic contours, real-time 3D image analysis remains limited by computational bottlenecks. We are specifically concerned with segmenting 3D volumetric ultrasound streams from echo-cardiograph machines (Phillips Medical Systems, Andover, MA) for analysis of cardiac function. The system uses a 3000 element array that produces 20-25 volumes per second at a resolution of 128x48x204 voxels. This yields a data rate of of 240 Mbits/sec, requiring efficient algorithms and implementations to track moving cardiac tissue in real-time. This paper discusses implementation of active 2D deformable models for real-time volumetric segmentation at the high data rates described above. We demonstrate that using an efficient approximation of local curvature change in the implementation of dynamic contours leads to real-time volumetric segmentation on mid-range off-the-shelf hardware without the use of specialized graphics hardware. Our dynamic contour implementation relies on an optimal estimation of local curvature change based on a Menger curvature calculation. We investigate the role of curvature approximations and smoothness with respect to optimal contour point motions and step size in real-time implementations. This smoothness provides reasonable shape estimates in the absence of appropriate or conflicting external image input. Finally, we present a 3D image segmentation algorithm based on an efficient implementation of 2D dynamic contours, and demonstrate real-time performance with high volumetric data rates.
Semi-automatic segmentation and modeling of the cervical spinal cord for volume quantification in multiple sclerosis patients from magnetic resonance images
Pavlina Sonkova, Iordanis E. Evangelou, Antonio Gallo, et al.
Spinal cord (SC) tissue loss is known to occur in some patients with multiple sclerosis (MS), resulting in SC atrophy. Currently, no measurement tools exist to determine the magnitude of SC atrophy from Magnetic Resonance Images (MRI). We have developed and implemented a novel semi-automatic method for quantifying the cervical SC volume (CSCV) from Magnetic Resonance Images (MRI) based on level sets. The image dataset consisted of SC MRI exams obtained at 1.5 Tesla from 12 MS patients (10 relapsing-remitting and 2 secondary progressive) and 12 age- and gender-matched healthy volunteers (HVs). 3D high resolution image data were acquired using an IR-FSPGR sequence acquired in the sagittal plane. The mid-sagittal slice (MSS) was automatically located based on the entropy calculation for each of the consecutive sagittal slices. The image data were then pre-processed by 3D anisotropic diffusion filtering for noise reduction and edge enhancement before segmentation with a level set formulation which did not require re-initialization. The developed method was tested against manual segmentation (considered ground truth) and intra-observer and inter-observer variability were evaluated.
Integrating local voxel classification and global shape models for medical image segmentation
Segmentation of anatomical structures is a prerequisite for many medical image analysis tasks. We propose a method that integrates local voxel classification and global shape models. The method starts by computing a local feature vector for every voxel and mapping this, via a classifier trained from example segmentations, to a probability that the voxel belongs to the structure to be segmented. Next, this probabilistic output is entered into a global shape model. This shape model is constructed by mapping aligned blurred versions of reference segmentations of the training data into a vector space and applying principal component analysis (PCA). The mapping onto a vector space that is applied guarantees valid results from the PCA. An advantage of using such a shape model is that there is no need to define corresponding landmarks on all training scans, which is a hard task on 3D data. Segmentation of unseen test data is performed by a least squares fit of the results of the voxel classification, after alignment and blurring, into the PCA space. The result of this procedure is for each voxel a probability that it belongs to the structure to be segmented conditioned on both local and global information. We demonstrate the effectiveness of the method on segmentation of lungs containing pathologic abnormalities in 3D CT data.
Lung lobe and segmental lobe extraction from 3D chest CT datasets based on figure decomposition and Voronoi division
Kensaku Mori, Yuichi Nakada, Takayuki Kitasaka, et al.
In this paper, we present a method for segmenting lung lobe regions and segmental lobe regions from 3D CT datasets. In a CAD system for the chest, it is very important to understand structures of the chest by a computer. It is required to develop algorithms that automatically segment organ regions of the chest area from input 3-D chest CT datasets. This paper tries to divide lung regions into lung lobe regions and segmental lobe regions by the figure decomposition process and the Voronoi division process. In this method, we enhance sheet structures on CT images by using eigenvalues of a Hessian matrix. Also, the lung regions are segmented by simple thresholding and morphological filtering. Then, we subtract sheet structures from the lung regions. By applying the figure decomposition process, we obtain each lung lobe region. Segmental lobe regions are obtained by the Voronoi division process using bronchial branch information The proposed method was applied to thirteen cases of 3-D chest CT images. Experimental results showed that the proposed method can extract lung lobe regions and segmental lobe regions even for cases of incomplete fissure or over-extraction of interlobar pleura.
Posters: Shape
icon_mobile_dropdown
A novel approach to fracture-risk-assessment in osteoporosis by ROI-oriented application of the Minkowski-functionals to dual x-ray absorptiometry scans of the hip
Holger F. Boehm, Alexandra Panteleon, Tobias Vogel, et al.
Fractures of the proximal femur represent the worst complication in osteoporosis with a mortality rate of up to 50% during the first post-traumatic year. Bone mineral density (BMD) as obtained from dual energy x-ray absorptiometry (DXA) is a good predictor of fracture risk. However, there is a considerable overlap in the BMD-results between individuals who have fractured and those who have not. As DXA uses highly standardized radiographic projection images to obtain the densitometric information, it can be postulated that these images contain much more information than just mineral density. Lately, geometric dimensions, e.g. hip axis length (HAL) or femoral neck axis length (FNAL), are considered in conjunction with BMD, which may allow to enhance the predictive potential of bone mass measurements. In recent studies we sucessfully introduced a novel methodology for topological analysis of multi-dimensional graylevel datasets, that, for instance, allows to predict the ultimate mechanical strength of femoral bone specimens. The new topolocial parameters are based on the so called Minkowski Functionals (MF), which represent a set of topographical descriptors that can be used universally. Since the DXA-images are multi-graylevel datasets in 2D obtained in a standardized way, they are ideally suited to be processed by the new method. In this study we introduce a novel algorithm to evaluate DXA-scans of the proximal femur using quantitative image analysis procedures based on the MF in 2D. The analysis is conducted in four defined regions of interest in analogy to the standard densitometric evaluation. The objective is to provide a tool to identifiy individuals with critically reduced mechanical competence of the hip. The result of the new method is compared with the evaluation bone mineral density obtained by DXA, which - at present - is the clinical standard of reference.
Reconstructing liver shape and position from MR image slices using an active shape model
Matthias Fenchel, Stefan Thesen, Andreas Schilling
We present an algorithm for fully automatic reconstruction of 3D position, orientation and shape of the human liver from a sparsely covering set of $n$ 2D MR slice images. Reconstructing the shape of an organ from slice images can be used for scan planning, for surgical planning or other purposes where 3D anatomical knowledge has to be inferred from sparse slices. The algorithm is based on adapting an active shape model of the liver surface to a given set of slice images. The active shape model is created from a training set of liver segmentations from a group of volunteers. The training set is set up with semi-manual segmentations of T1-weighted volumetric MR images. Searching for the optimal shape model that best fits to the image data is done by maximizing a similarity measure based on local appearance at the surface. Two different algorithms for the active shape model search are proposed and compared: both algorithms seek to maximize the a-posteriori probability of the grey level appearance around the surface while constraining the surface to the space of valid shapes. The first algorithm works by using grey value profile statistics in normal direction. The second algorithm uses average and variance images to calculate the local surface appearance on the fly. Both algorithms are validated by fitting the active shape model to abdominal 2D slice images and comparing the shapes, which have been reconstructed, to the manual segmentations and to the results of active shape model searches from 3D image data. The results turn out to be promising and competitive to active shape model segmentations from 3D data.
Tracheal stent prediction using statistical deformable models of tubular shapes
R. Pinho, T. Huysmans, W. Vos, et al.
Tracheal stenosis is a narrowing of the trachea that impedes normal breathing. Tracheotomy is one solution, but subjects patients to intubation. An alternative technique employs tracheal stents, which are tubular structures that push the walls of the stenotic areas to their original location. They are implanted with endoscopes, therefore reducing the surgical risk to the patient. Stents can also be used in tracheal reconstruction to aid the recovery of reconstructed areas. Correct preoperative stent length and diameter specification is crucial to successful treatment, otherwise stents might not cover the stenotic area nor push the walls as required. The level of stenosis is usually measured from inside the trachea, either with endoscopes or with image processing techniques that, eg compute the distance from the centre line to the walls of the trachea. These methods are not suited for the prediction of stent sizes because they can not trivially estimate the healthy calibre of the trachea at the stenotic region. We propose an automatic method that enables the estimation of stent dimensions with statistical shape models of the trachea. An average trachea obtained from a training set of CT scans of healthy tracheas is placed in a CT image of a diseased person. The shape deforms according to the statistical model to match the walls of the trachea, except at stenotic areas. Since the deformed shape gives an estimation of the healthy trachea, it is possible to predict the size and diameter of the stent to be implanted in that specific subject.
Vertebral classification using localized pathology-related shape model
R. Zewail, A. Elsafi, N. Durdle
Radiographs of the spine are frequently examined for assessment of vertebral abnormalities. Features like osteophytes (bony growth of vertebra's corners), and disc space narrowing are often used as visual evidence of osteoarthris or degenerative joint disease. These symptoms result in remarkable changes in the shapes of the vertebral body. Statistical analysis of anatomical structure has recently gained increased popularity within the medical imaging community, since they have the potential to enhance the automated diagnosis process. In this paper, we present a novel method for computer-assisted vertebral classification using a localized, pathology-related shape model. The new classification scheme is able to assess the condition of multiple vertebrae simultaneously, hence is possible to directly classify the whole spine anatomy according to the condition of interest (anterior osteophites). At the core of this method is a new localized shape model that uses concepts of sparsity, dimension reduction, and statistical independence to extract sets of localized modes of deformations specific to each of the vertebrae under investigation. By projection of the shapes onto any specific set of deformation modes (or basis), we obtain low-dimensional features that are most directly related to the pathology of the vertebra of interest. These features are then used as input to a support vector machine classifier to classify the vertebra under investigation as normal or upnormal. Experiments are conducted using contours from digital x-ray images of five vertebrae of lumbar spine. The accuracy of the classification scheme is assessed using the ROC curves. An average specifity of 96.8 % is achieved with a sensitivity of 80 %.
Local curvature scale: a new concept of shape description
A new boundary shape description based on the notion of curvature-scale is presented. This shape descriptor performs better than the commonly used Rosenfeld's method of curvature estimation and can be applied directly to digital boundaries without requiring prior approximations. It can extract special points of interest such as convex and concave corners, straight lines, circular segments, and inflection points. The results show that this method produces a complete boundary shape description capable of handling different levels of shape detail. It also has numerous potential applications such as automatic landmark tagging which becomes necessary to build model-based approaches toward the goal of organ modelling and segmentation.
Conditional-mean initialization using neighboring objects in deformable model segmentation
Ja-Yeon Jeong, Joshua V. Stough, J. Steve Marron, et al.
Most model-based segmentation methods find a target object in a new image by constructing an objective function and optimizing it using a standard minimization algorithm. In general, the objective function has two penalty terms: 1) for deforming a template model and 2) for mismatch between the trained image intensities relative to the template model and the observed image intensities relative to the deformed template model in the target image. While it is difficult to establish an objective function with a global minimum at the desired segmentation result, even such an objective function is typically non-convex due to the complexity of the intensity patterns and the many structures surrounding the target object. Thus, it is critical that the optimization starts at a point close to the global minimum of the objective function in deformable model-based segmentation framework. For a segmentation method in maximum a posteriori framework a good objective function can be obtained by learning the probability distributions of the population shape deformations and their associated image intensities because each penalty term can be simplified to a squared function of some distance metric defined in the shape space. The mean shape and intensities of the learned probability distributions also provide a good initialization for segmentation. However, a major concern in estimating the shape prior is the stability of the estimated shape distributions from given training samples because the feature space of a shape model is usually very high dimensional while the number of training samples is limited. A lot of effort in that regard have been made to attain a stable estimation of shape probability distribution. In this paper, we describe our approach to stably estimate a shape probability distribution when good segmentations of objects adjacent to the target object are available. Our approach is to use a conditional shape probability distribution (CSPD) to take into account in the shape distribution the relation of the target object to neighboring objects. In particular, we propose a new method based on principal component regression (PCR) in reflecting in the conditional term of the CSPD the effect of neighboring objects on the target object. The resulting approach is able to give a better and robust initialization with training samples of a few cases. To demonstrate the potential of our approach, we apply it first to training of a simulated data of known deformations and second to male pelvic organs, using the CSPD in m-rep segmentations of the prostate in CT images. Our results show a clear improvement in initializing the prostate by its conditional mean given the bladder and the rectum as neighboring objects, as measured both by volume overlap and average surface distance.
A multi-modal prostate segmentation scheme by combining spectral clustering and active shape models
Robert Toth, Pallavi Tiwari, Mark Rosen, et al.
Segmentation of the prostate boundary on clinical images is useful in a large number of applications including calculating prostate volume during biopsy, tumor estimation, and treatment planning. Manual segmentation of the prostate boundary is, however, time consuming and subject to inter- and intra-reader variability. Magnetic Resonance (MR) imaging (MRI) and MR Spectroscopy (MRS) have recently emerged as promising modalities for detection of prostate cancer in vivo. In this paper we present a novel scheme for accurate and automated prostate segmentation on in vivo 1.5 Tesla multi-modal MRI studies. The segmentation algorithm comprises two steps: (1) A hierarchical unsupervised spectral clustering scheme using MRS data to isolate the region of interest (ROI) corresponding to the prostate, and (2) an Active Shape Model (ASM) segmentation scheme where the ASM is initialized within the ROI obtained in the previous step. The hierarchical MRS clustering scheme in step 1 identifies spectra corresponding to locations within the prostate in an iterative fashion by discriminating between potential prostate and non-prostate spectra in a lower dimensional embedding space. The spatial locations of the prostate spectra so identified are used as the initial ROI for the ASM. The ASM is trained by identifying user-selected landmarks on the prostate boundary on T2 MRI images. Boundary points on the prostate are identified using mutual information (MI) as opposed to the traditional Mahalanobis distance, and the trained ASM is deformed to fit the boundary points so identified. Cross validation on 150 prostate MRI slices yields an average segmentation sensitivity, specificity, overlap, and positive predictive value of 89, 86, 83, and 93&percent; respectively. We demonstrate that the accurate initialization of the ASM via the spectral clustering scheme is necessary for automated boundary extraction. Our method is fully automated, robust to system parameters, and computationally efficient.
Comparison of statistical shape models built on correspondence probabilities and one-to-one correspondences
Heike Hufnagel, Xavier Pennec, Jan Ehrhardt, et al.
In this paper, we present a method to compute a statistical shape model based on shapes which are represented by unstructured point sets with arbitrary point numbers. A fundamental problem when computing statistical shape models is the determination of correspondences between the observations of the associated data set. Often, homologies between points that represent the surfaces are assumed. When working merely with point clouds, this might lead to imprecise mean shape and variability results. To overcome this problem, we propose an approach where exact correspondences are replaced by evolving correspondence probabilities. These are the basis for a novel algorithm that computes a generative statistical shape model. We developed a unified Maximum A Posteriori (MAP) framework to compute the model parameters ('mean shape' and 'modes of variation') and the nuisance parameters which leads to an optimal adaption of the model to the set of observations. The registration of the model on the observations is solved using the Expectation Maximization - Iterative Closest Point algorithm which is based on probabilistic correspondences and proved to be robust and fast. The alternated optimization of the MAP explanation with respect to the observation and the generative model parameters leads to very efficient and closed-form solutions for nearly all parameters. A comparison with a statistical shape model which is built using the Iterative Closest Point (ICP) registration algorithm and a Principal Component Analysis (PCA) shows that our approach leads to better SSM quality measures.
Posters: Texture
icon_mobile_dropdown
Studying the effect of noise on the performance of 2D and 3D texture measures for quantifying the trabecular bone structure as obtained with high resolution MR imaging at 3 tesla
Roberto Monetti, Jan Bauer, Dirk Mueller, et al.
3.0 Tesla MRI devices are becoming popular in clinical applications since they render images with a higher signal-tonoise ratio than the former 1.5 Tesla MRI devices. Here, we investigate if higher signal-to-noise ratio can be beneficial for a quantitative image analysis in the context of bone research. We performed a detailed analysis of the effect of noise on the performance of 2D morphometric linear measures and a 3D nonlinear measure with respect to their correlation with biomechanical properties of the bone expressed by the maximum compressive strength. The performance of both 2D and 3D texture measures was relatively insensitive to superimposed artificial noise. This finding suggests that MR sequences for visualizing bone structures at 3T should rather be optimized to spatial resolution (or scanning time) than to signal-to-noise ratio.
Comparison and combination of scaling index method and Minkowski functionals in the analysis of high resolution magnetic resonance images of the distal radius in vitro
Irina N. Sidorenko, Jan Bauer, Roberto Monetti, et al.
High resolution magnetic resonance (HRMR) imaging can reveal major characteristics of trabecular bone. The quantification of this trabecular micro architecture can be useful for better understanding the progression of osteoporosis and improve its diagnosis. In the present work we applied the scaling index method (SIM) and Minkowski Functionals (MF) for analysing tomographic images of distal radius specimens in vitro. For both methods, the correlation with the maximum compressive strength (MCS) as determined in a biomechanical test and the diagnostic performance with regard to the spine fracture status were calculated. Both local SIM and global MF methods showed significantly better results compared to bone mineral density measured by quantitative computed tomography. The receiver operating characteristic analysis for differentiating fractured and non-fractured subjects revealed area under the curve (AUC) values of 0.716 for BMD, 0.897 for SIM and 0.911 for MF. The correlation coefficients with MCS were 0.6771 for BMD, 0.843 for SIM and 0.772 for MF. We simulated the effect of perturbations, namely noise effects and intensity variations. Overall, MF method was more sensitive to noise than SIM. A combination of SIM and MF methods could, however, increase AUC values from 0.85 to 0.89 and correlation coefficients from 0.71 to 0.82. In conclusion, local SIM and global MF techniques can successfully be applied for analysing HRMR image data. Since these methods are complementary, their combination offers a new possibility of describing MR images of the trabecular bone, especially noisy ones.
Posters: Validation
icon_mobile_dropdown
Semi-synthetic digital phantoms incorporating natural structured noise and boundary inhomogeneities
Sovira Tan, Michael M. Ward
Validating segmentation algorithms remains a difficult problem. Manual segmentation taken as gold standard is timeconsuming and can still be contentious especially in the case of complex 3D objects and in the presence of important partial volume effect (PVE). In contrast digital phantoms have well-defined built-in boundaries even when PVE is simulated. However their degree of realism is questionable. In particular the rich natural structures inside an object that constitute one of the most difficult obstacles to segmentation are to this day too complex to model. A new method for constructing semi-synthetic digital phantoms was recently proposed that incorporates natural structured noise and boundary inhomogeneities. However only one phantom was presented and validation was lacking. In the present work we constructed 5 phantoms of vertebral bodies. Validation of phantoms should test their ability to predict how an algorithm will perform when confronted to real data. Our phantoms were used to compare the performance of two level set based segmentation algorithms and find the parameters that optimize their performances. We validated the phantoms by correlating the results obtained on them with those obtained on 50 real vertebrae. We show that: 1) the phantoms accurately predict which segmentation algorithm will perform better with real clinical data. 2) by combining the results obtained by the 5 different phantoms we can extract useful predictions about the performance of different sets of parameters on real data. Because the phantoms possess the high variability of real data predictions based on only one phantom would fail.
Evaluation of accuracy in partial volume analysis of small objects
Jan Rexilius, Heinz-Otto Peitgen
Accurate and robust assessment of quantitative parameters is a key issue in many fields of medical image analysis, and can have a direct impact on diagnosis and treatment monitoring. Especially for the analysis of small structures such as focal lesions in patients with MS, the finite spatial resolution of imaging devices is often a limiting factor that results in a mixture of different tissue types. We propose a new method that allows an accurate quantification of medical image data, focusing on a dedicated model for partial volume (PV) artifacts. Today, a widely accepted model assumption is that of a uniformly distributed linear mixture of pure tissues. However, several publications have clearly shown that this is not an appropriate choice in many cases. We propose a generalization of current PV models based on the Beta distribution, yielding a more accurate quantification. Furthermore, we present a new classification scheme. Prior knowledge obtained from a set of training data allows a robust initial estimate of the proper model parameters, even in cases of objects with predominant PV artifacts. A maximum likelihood based clustering algorithm is employed, resulting in a robust volume estimate. Experiments are carried out on more than 100 stylized software phantoms as well as on realistic phantom data sets. A comparison with current mixture models shows the capabilities of our approach.
A software assistant for the design of realistic software phantoms
Jan Rexilius, Olaf Konrad, Heinz-Otto Peitgen
Physical and software phantom data sets have become an integral tool during the design, implementation, and utilization of new algorithms. Unfortunately, a common research resource has not been established until now for many applications. We propose a general software assistant for the development of realistic software phantoms. Our aim is an easy to use tool with an intuitive user interface. Furthermore, we provide a publicly available software for researchers including a common basis of reference data, which facilitates a standardized and objective validation of performance and limitations of own developments as well as the comparison of different methods.