Proceedings Volume 5747

Medical Imaging 2005: Image Processing

cover
Proceedings Volume 5747

Medical Imaging 2005: Image Processing

View the digital version of this volume at SPIE Digital Libarary.

Volume Details

Date Published: 29 April 2005
Contents: 16 Sessions, 231 Papers, 0 Presentations
Conference: Medical Imaging 2005
Volume Number: 5747

Table of Contents

icon_mobile_dropdown

Table of Contents

All links to SPIE Proceedings will open in the SPIE Digital Library. external link icon
View Session icon_mobile_dropdown
  • Computer-Aided Diagnosis I: Breast
  • Computer-Aided Diagnosis II
  • Registration I: Rigid Registration
  • Poster Session I
  • Registration II: Non-Rigid Registration
  • Texture
  • Pattern Recognition and Neural Networks
  • Deformable Geometry
  • Tomographic Reconstruction
  • Segmentation I
  • Segmentation II: Vasculature
  • Segmentation III
  • Restoration
  • Statistical Methods
  • Shape and Scale
  • Poster Session I
  • Pattern Recognition and Neural Networks
  • Poster Session I
  • Registration I: Rigid Registration
  • Poster Session I
  • Registration II: Non-Rigid Registration
  • Poster Session I
  • Poster Session II
  • Segmentation II: Vasculature
  • Poster Session II
Computer-Aided Diagnosis I: Breast
icon_mobile_dropdown
A study of several CAD methods for classification of clustered microcalcifications
In this paper we investigate several state-of-the-art machine-learning methods for automated classification of clustered microcalcifications (MCs), aimed to assisting radiologists for more accurate diagnosis of breast cancer in a computer-aided diagnosis (CADx) scheme. The methods we consider include: support vector machine (SVM), kernel Fisher discriminant (KFD), and committee machines (ensemble averaging and AdaBoost), most of which have been developed recently in statistical learning theory. We formulate differentiation of malignant from benign MCs as a supervised learning problem, and apply these learning methods to develop the classification algorithms. As input, these methods use image features automatically extracted from clustered MCs. We test these methods using a database of 697 clinical mammograms from 386 cases, which include a wide spectrum of difficult-to-classify cases. We use receiver operating characteristic (ROC) analysis to evaluate and compare the classification performance by the different methods. In addition, we also investigate how to combine information from multiple-view mammograms of the same case so that the best decision can be made by a classifier. In our experiments, the kernel-based methods (i.e., SVM, KFD) yield the best performance, significantly outperforming a well-established CADx approach based on neural network learning.
Computer-aided detection of breast masses on mammograms: performance improvement using a dual system
We have developed a computer-aided detection (CAD) system for breast masses on mammograms. In this study, our purpose was to improve the performance of our mass detection system by using a new dual system approach which combines a CAD system optimized with "average" masses with another CAD system optimized with subtle masses. The latter system is trained to provide high sensitivity in detecting subtle masses. For an unknown mammogram, the two systems are used in parallel to detect suspicious objects. A feed-forward backpropagation neural network trained to merge the scores of the two linear discriminant analysis (LDA) classifiers from the two systems makes the final decision in differentiation of true masses from normal tissue. A data set of 86 patients containing 172 mammograms with biopsy-proven masses was partitioned into a training set and an independent test set. This data set is referred to as the average data set. A second data set of 214 prior mammograms was used for training the second CAD system for detection of subtle masses. When the single CAD system trained on the average data set was applied to the test set, the Az for false positive (FP) classification was 0.81 and the FP rates were 2.1, 1.5 and 1.3 FPs/image at the case-based sensitivities of 95%, 90% and 85%, respectively. With the dual CAD system, the Az was 0.85 and the FP rates were improved to 1.7, 1.2 and 0.8 FPs/image at the same case-based sensitivities. Our results indicate that the dual CAD system can improve the performance of mass detection on mammograms.
Classification of mammographic lesions into BI-RADS shape categories using the beamlet transform
We present a new algorithm and preliminary results for classifying lesions into BI-RADS shape categories: round, oval, lobulated, or irregular. By classifying masses into one of these categories, computer aided detection (CAD) systems will be able to provide additional information to radiologists. Thus, such a tool could potentially be used in conjunction with a CAD system to enable greater interaction and personalization. For this classification task, we have developed a new set of features using the Beamlet transform, which is a recently developed multi-scale image analysis transform. We trained a k-Nearest Neighbor classifier using images from the Digital Database for Digital Mammography (DDSM). The method was tested on a set of 25 images of each type and we obtained a classification accuracy of 78% for classifying masses as oval or round and an accuracy of 72% for classifying masses as lobulated or round.
Evidence based detection of spiculated masses and architectural distortions
Mehul P. Sampat, Gary J. Whitman, Mia K. Markey, et al.
Mass detection algorithms generally consist of two stages. The aim of the first stage is to detect all potential masses. In the second stage, the aim is to reduce the false-positives by classifying the detected objects as masses or normal tissue. In this paper, we present a new evidence based, stage-one algorithm for the detection of spiculated masses and architectural distortions. By evidence based, we mean that we use the statistics of the physical characteristics of these abnormalities to determine the parameters of the detection algorithm. Our stage-one algorithm consists of two steps, an enhancement step followed by a filtering step. In the first step, we propose a new technique for the enhancement of spiculations in which a linear filter is applied to the Radon transform of the image. In the second step, we filter the enhanced images with a new class of linear image filters called Radial Spiculation Filters. We have invented these filters specifically for detecting spiculated masses and architectural distortions that are marked by converging lines or spiculations. These filters are highly specific narrowband filters, which are designed to match the expected structures of these abnormalities and form a new class of wavelet-type filterbanks derived from optimal theories of filtering. A key aspect of this work is that each parameter of the filter has been incorporated to capture the variation in physical characteristics of spiculated masses and architectural distortions and that the parameters of the stage-one detection algorithm are determined by the physical measurements.
DNA: directional neighborhood analysis for detection of breast masses in screening mammograms
We introduce a computer-assisted detection (CAD) system for the automated detection of breast masses in screening mammograms. The system targets the directional behavior of the neighborhood pixels surrounding a reference image pixel. The underlying hypothesis is that in the presence of a mass the directional properties of the breast tissue surrounding the mass should be altered. The hypothesis was tested using a database of 1,337 mammographic regions of interest (ROIs) extracted from DDSM mammograms. There were 681 ROIs containing a biopsy-proven mass centered in the ROI (340 malignant, 341 benign) and 656 ROIs depicting normal breast parenchyma. Initially, eight main directional propagations were identified and modeled given the center of the ROI as the reference pixel. Subsequently, eight novel morphological features were extracted for each direction. The features were designed to characterize the disturbance occurring in normal breast parenchyma due to the presence of a mass. Finally, the extracted features were merged using a back propagation neural network (BPANN). The network served as a non linear classifier trained to determine the presence of a mass centered at the reference image pixel. The BPANN was trained and tested using a leave-one-out sampling scheme. Its performance was evaluated with Receiver Operating Characteristics (ROC) analysis. Our CAD system showed an ROC area index of Az=0.88±0.01 for discriminating mass vs. normal ROIs. Detection performance was robust for both malignant (Az=0.88±0.01) and benign masses (Az=0.87±0.01). Thus, the proposed directional neighborhood analysis (DNA) can be applied effectively to identify suspicious masses in screening mammograms.
New approach to breast cancer CAD using partial least squares and kernel-partial least squares
Walker H. Land Jr., John Heine, Mark Embrechts, et al.
Breast cancer is second only to lung cancer as a tumor-related cause of death in women. Currently, the method of choice for the early detection of breast cancer is mammography. While sensitive to the detection of breast cancer, its positive predictive value (PPV) is low, resulting in biopsies that are only 15-34% likely to reveal malignancy. This paper explores the use of two novel approaches called Partial Least Squares (PLS) and Kernel-PLS (K-PLS) to the diagnosis of breast cancer. The approach is based on optimization for the partial least squares (PLS) algorithm for linear regression and the K-PLS algorithm for non-linear regression. Preliminary results show that both the PLS and K-PLS paradigms achieved comparable results with three separate support vector learning machines (SVLMs), where these SVLMs were known to have been trained to a global minimum. That is, the average performance of the three separate SVLMs were Az = 0.9167927, with an average partial Az (Az90) = 0.5684283. These results compare favorably with the K-PLS paradigm, which obtained an Az = 0.907 and partial Az = 0.6123. The PLS paradigm provided comparable results. Secondly, both the K-PLS and PLS paradigms out performed the ANN in that the Az index improved by about 14% (Az ≈ 0.907 compared to the ANN Az of ≈ 0.8). The "Press R squared" value for the PLS and K-PLS machine learning algorithms were 0.89 and 0.9, respectively, which is in good agreement with the other MOP values.
Computer-Aided Diagnosis II
icon_mobile_dropdown
Automated prostate cancer diagnosis and Gleason grading of tissue microarrays
Ali Tabesh, Vinay P. Kumar, Ho-Yuen Pang, et al.
We present the results on the development of an automated system for prostate cancer diagnosis and Gleason grading. Images of representative areas of the original Hematoxylin-and-Eosin (H&E)-stained tissue retrieved from each patient, either from a tissue microarray (TMA) core or whole section, were captured and analyzed. The image sets consisted of 367 and 268 color images for the diagnosis and Gleason grading problems, respectively. In diagnosis, the goal is to classify a tissue image into tumor versus non-tumor classes. In Gleason grading, which characterizes tumor aggressiveness, the objective is to classify a tissue image as being from either a low- or high-grade tumor. Several feature sets were computed from the image. The feature sets considered were: (i) color channel histograms, (ii) fractal dimension features, (iii) fractal code features, (iv) wavelet features, and (v) color, shape and texture features computed using Aureon Biosciences' MAGIC system. The linear and quadratic Gaussian classifiers together with a greedy search feature selection algorithm were used. For cancer diagnosis, a classification accuracy of 94.5% was obtained on an independent test set. For Gleason grading, the achieved accuracy of classification into low- and high-grade classes of an independent test set was 77.6%.
Computer-aided-diagnosis (CAD) for colposcopy
Holger Lange, Daron G. Ferris M.D.
Uterine cervical cancer is the second most common cancer among women worldwide. Colposcopy is a diagnostic method, whereby a physician (colposcopist) visually inspects the lower genital tract (cervix, vulva and vagina), with special emphasis on the subjective appearance of metaplastic epithelium comprising the transformation zone on the cervix. Cervical cancer precursor lesions and invasive cancer exhibit certain distinctly abnormal morphologic features. Lesion characteristics such as margin; color or opacity; blood vessel caliber, intercapillary spacing and distribution; and contour are considered by colposcopists to derive a clinical diagnosis. Clinicians and academia have suggested and shown proof of concept that automated image analysis of cervical imagery can be used for cervical cancer screening and diagnosis, having the potential to have a direct impact on improving women’s health care and reducing associated costs. STI Medical Systems is developing a Computer-Aided-Diagnosis (CAD) system for colposcopy -- ColpoCAD. At the heart of ColpoCAD is a complex multi-sensor, multi-data and multi-feature image analysis system. A functional description is presented of the envisioned ColpoCAD system, broken down into: Modality Data Management System, Image Enhancement, Feature Extraction, Reference Database, and Diagnosis and directed Biopsies. The system design and development process of the image analysis system is outlined. The system design provides a modular and open architecture built on feature based processing. The core feature set includes the visual features used by colposcopists. This feature set can be extended to include new features introduced by new instrument technologies, like fluorescence and impedance, and any other plausible feature that can be extracted from the cervical data. Preliminary results of our research on detecting the three most important features: blood vessel structures, acetowhite regions and lesion margins are shown. As this is a new and very complex field in medical image processing, the hope is that this paper can provide a framework and basis to encourage and facilitate collaboration and discussion between industry, academia, and medical practitioners.
A complete CAD system for pulmonary nodule detection in high resolution CT images
Xiangwei Zhang, Geoffrey McLennan, Eric A. Hoffman, et al.
The purpose of this study is to develop a computer-aided diagnosis (CAD) system to detect small-sized (from 2mm to 10mm) pulmonary nodules in high resolution helical CT scans. A new CAD system is proposed to locate both juxtapleural nodules and non-pleural nodules. Isotropic resampling and lung segmentation are performed as preprocessing steps. Morphological closing was utilized to smooth the lung contours to include the indented possible juxtapleural locations, thresholding and 3D component analysis were used to obtain 3D volumetric nodule candidates; furthermore, gray level and geometric features were extracted, and analyzed using linear discriminant analysis (LDA) classifier. Leave one case out method was used to evaluate the LDA. To deal with non-pleural nodules, a discrete-time cellular neural network (DTCNN) based on local shape features was developed. This scheme employed the local shape property to perform voxel classification. The shape index feature successfully captured the local shape difference between nodules and non-nodules, especially vessels. To tailor it for lung nodule detection, this DTCNN was trained using genetic algorithms (GAs) to derive the shape index variation pattern of nodules. Nonoverlapping training and testing sets were utilized in the non-pleural nodule detection. 19 clinical thoracic CT cases involving a total of 4838 sectional images were used in this work. The juxtapleural nodule detection method was able to obtain sensitivity 81.25% with an average of 8.29 FPs per case. The non-pleural nodule finding scheme attained sensitivity of 83.9% with an average 3.47 FPs/case. Combining the two subsystems together, an overall performance of 82.98% sensitivity with 11.76 FPs/case can be obtained.
Effect of massive training artificial neural networks for rib suppression on reduction of false positives in computerized detection of nodules on chest radiographs
Kenji Suzuki, Junji Shiraishi, Feng Li, et al.
A major challenge in computer-aided diagnostic (CAD) schemes for nodule detection on chest radiographs is the detection of nodules that overlap with ribs. Our purpose was to develop a technique for false-positive reduction in a CAD scheme using a rib-suppression technique based on massive training artificial neural networks (MTANNs). We developed a multiple MTANN (multi-MTANN) consisting of eight MTANNs for removing eight types of false positives. For further removal of false positives caused by ribs, we developed a rib-suppression technique using a multi-resolution MTANN consisting of three different resolution MTANNs. To suppress the contrast of ribs, the multi-resolution MTANN was trained with input chest radiographs and the teaching soft-tissue images obtained by using a dual-energy subtraction technique. Our database consisted of 91 nodules in 91 chest radiographs. With our original CAD scheme based on a difference image technique with linear discriminant analysis, a sensitivity of 82.4% (75/91 nodules) with 4.5 (410/91) false positives per image was achieved. The trained multi-MTANN was able to remove 62.7% (257/410) of false positives with a loss of one true positive. With the rib-suppression technique, the contrast of ribs in chest radiographs was suppressed substantially. Due to the effect of rib-suppression, 41.2% (63/153) of the remaining false positives were removed without a loss of any true positives. Thus, the false-positive rate of our CAD scheme was improved substantially, while a high sensitivity was maintained.
Computerized nodule detection in thin-slice CT using selective enhancement filter and automated rule-based classifier
Qiang Li, Feng Li, Kunio Doi
We have been developing computer-aided diagnostic (CAD) scheme to assist radiologists detect lung nodules in thoracic CT images. In order to improve the sensitivity for nodule detection, we developed a selective nodule enhancement filter for nodule which can simultaneously enhance nodules and suppress other normal anatomic structures such as blood vessels and airway walls. Therefore, as preprocessing steps, this filter is useful for improving the sensitivity of nodule detection and for reducing the number of false positives. Another new technique we employed in this study is an automated rule-based classifier. It can significantly reduce the extent of the disadvantages of existing rule-based classifiers, including manual design, poor reproducibility, poor evaluation methods such as re-substitution, and a large overtraining effect. Experimental results performed with Monte Carlo simulation and a real lung nodule CT dataset demonstrated that the automated method can completely eliminate overtraining effect in the procedure of cutoff threshold selection, and thus can minimize overall overtraining effect in the rule-based classifier.
A computer-based method of segmenting ground glass nodules in pulmonary CT images: comparison to expert radiologists’ interpretations
Li Zhang, Tiantian Zhang, Carol L. Novak, et al.
Ground glass nodules (GGNs) have proved especially problematic in lung cancer diagnosis, as despite frequently being malignant they have extremely slow growth rates. In this work, the GGN segmentation results of a computer-based method were compared with manual segmentation performed by two dedicated chest radiologists. CT volumes of 8 patients were acquired by multi-slice CT. 21 pure or mixed GGNs were identified and independently segmented by the computer-based method and by two readers. The computer-based method is initialized by a click point, and uses a Markov random field (MRF) model for segmentation. While the intensity distribution varies for different GGNs, the intensity model used in MRF is adapted for each nodule based on initial estimates. This method was run three times for each nodule using different click points to evaluate consistency. In this work, consistency was defined by the overlap ratio (overlap volume/mean volume). The consistency of the computer-based method with different initial points, with a mean overlap ratio of 0.96±0.02 (95% confidence interval on mean), was significantly higher than the inter-observer consistency between the two radiologists, indicated by a mean overlap ratio of 0.73±0.04. The computer consistency was also significantly higher than the intra-observer consistency of two measurements from the same radiologist, indicated by an overlap ratio of 0.69±0.05 (p-value < 1E-05). The concordance of the computer with the expert interpretation demonstrated a mean overlap ratio of 0.69±0.05. As shown by our data, the consistency provided by the computer-based method is significantly higher than between observers, and the accuracy of the method is no worse than that of one physician’s accuracy with respect to another, allowing more reproducible assessment of nodule growth.
A novel technique for assessing the case-specific reliability of decisions made by CAD tools
We present a novel technique that provides a case-specific confidence measure for artificial neural network (ANN) based computer-assisted diagnostic (CAD) decisions. The technique relies on the analysis of the feature space neighborhood for each query case and dynamically creates a validation set that allows estimation of a local accuracy of the decisions made by the network. Then a case-specific reliability measure is assigned to each system's response, which can be used to stratify network's predictions according to the acceptable validation error value. The study was performed using a database containing 1,337 mammographic regions of interest (ROIs) with biopsy-proven diagnosis (681 with masses, 656 with normal parenchyma). Two types of neural networks (1) a feed forward network with error back propagation (BPNN) and (2) a generalized regression neural network with RBF nodes (GRNN) were developed to detect masses based on 8 morphological features automatically extracted from each ROI. The performance of the networks was evaluated with Receiver Operating Characteristics (ROC) analysis. The study shows that as the threshold on the acceptable validation error declines, the technique rejects more CAD decisions as not reliable enough. However, the ROC performance for the reliable results steadily improves (from Az = 0.88 to Az = 0.98 for BPNN, from Az = 0.86 to Az = 0.97 for GRNN). The proposed technique provides a stratification strategy for predictions made by CAD tools and can be applied to any type of decision algorithms.
Registration I: Rigid Registration
icon_mobile_dropdown
Measuring image similarity in the presence of noise
Gustavo Kunde Rohde, Carlos A. Berenstein, Dennis M. Healy Jr.
Measuring the similarity between discretely sampled intensity values of different images as a function of geometric transformations is necessary for performing automatic image registration. Arbitrary spatial transformations require a continuous model for the intensity values of the discrete images. Because of computation cost most researchers choose to use low order basis functions, such as the linear hat function or low order B-splines, to model the discrete images. Using the theory of random processes we show that low order interpolators cause undesirable local optima artifacts in similarity measures based on the L2 norm, linear correlation coefficient, and mutual information. We show how these artifacts can be significantly reduced, and at times completely eliminated, by using sinc approximating kernels.
2D/3D registration based on volume gradients
Wolfgang Wein, Barbara Roeper, Nassir Navab
We present a set of new methods for efficient and precise registration of any X-Ray modality (fluoroscopy, portal imaging or regular X-Ray imaging) to a CT data set. These methods require neither feature extraction nor 2D or 3D segmentation. Our main contribution is to directly perform the computations on the gradient vector volume of the CT data, which has several advantages. It can increase the precision of the registration as it assesses mainly the alignment of intensity edges in both CT and X-Ray images. By using only significant areas of the gradient vector volume, the amount of information needed in each registration step can be reduced up to a factor of 10. This both speeds up the registration process and allows for using the CT data with full precision, e.g. 5123 voxels. We introduce a Volume Gradient Rendering (VGR) as well as a Volume Gradient Correlation (VGC) method, where the latter one can be used directly for computing the image similarity without DBR generation.
Unbiased rigid registration using transfer functions
Dieter A. Hahn, Joachim Hornegger, Werner Bautz, et al.
The evaluation of tumor growth as regression under therapy is an important clinical issue. Rigid registration of sequentially acquired 3D-images has proven its value for this purpose. Existing approaches to rigid image registration use the whole volume for the estimation of the rigid transform. Non-rigid soft tissue deformation, however, will imply a bias to the registration result, because local deformations cannot be modeled by rigid transforms. Anatomical substructures, like bones or teeth, are not affected by these deformations, but follow a rigid transform. This important observation is incorporated in the proposed registration algorithm. The selection of anatomical substructure is done by manual interaction of medical experts adjusting the transfer function of the volume rendering software. The parameters of the transfer function are used to identify the voxels that are considered for registration. A rigid transform is estimated by a quaternion gradient descent algorithm based on the intensity values of the specified tissue classes. Commonly used voxel intensity measures are adjusted to the modified registration algorithm. The contribution describes the mathematical framework of the proposed registration method and its implementation in a commercial software package. The experimental evaluation includes the discussion of different similarity measures, the comparison of the proposed method to established rigid registration techniques and the evaluation of the efficiency of the new method. We conclude with the discussion of potential medical applications of the proposed registration algorithm.
Optimal parameter choice for automatic fast rigid multimodal registration
Ulrich Mueller, Juergen Hesser, Reinhard Maenner
The issue of this paper is about real-time or interactive 2D-2D resp. 3D-3D matching. Based on Viola's sample-based stochastic Mutual Information (MI) gradient matching we developed a technique that allows to optimally set all necessary parameters in a short preprocessing step using typical images. In this paper we concentrate on finding an optimal parameter set for Rprop, the underlying stochastic optimizer. The relevant parameters are the start and the minimum learning rate given a pair of aligned images. Rprop is modelled by a Markov chain whose transition kernel is estimated by a mean gradient. We introduce a truncated recursion to simulate Rprop and obtain an expectation for the number of iterations for each parameter combination. This way near optimal parameters are found within 20-50 seconds, depending on the data. Using automatically set parameters for Rprop and the sample size, matching requires 0.3-1.3 s for 2D-2D and 0.6-2.1 s for 3D-3D on our test data using an Athlon 800 MHz processor. Altogether we get a real-time registration algorithm that optimizes its control parameters for the given data within less than a minute.
Poster Session I
icon_mobile_dropdown
Fast interpolation operations in non-rigid image registration
Much literature on image registration has worked with purely geometric image deformation models. For such models, interpolation/resampling operations are often the computationally intensive steps when iteratively minimizing the deformation cost function. This article discusses some techniques for efficiently implementing and accelerating these operations. To simplify presentation, we discuss our ideas in the context of 2D imaging. However, the concepts readily generalize to 3D. Our central technique is a table-lookup scheme that makes somewhat liberal use of RAM, but should not strain the resources of modern processors if certain design parameters are appropriately selected. The technique works by pre-interpolating and tabulating the grid values of the reference image onto a finer grid along one of the axes of the image. The lookup table can be rapidly constructed using FFTs. Our results show that this technique reduces iterative computation by an order of magnitude. When a minimization algorithm employing coordinate block alternation is used, one can obtain still faster computation by storing certain intermediate quantities as state variables. We refer to this technique as state variable hold-over. When combined with table-lookup, state variable hold-over reduces CPU time by about a factor two, as compared to table-lookup alone.
Registration II: Non-Rigid Registration
icon_mobile_dropdown
Comparison of gradient approximation techniques for optimisation of mutual information in nonrigid registration
Nonrigid registration of medical images by maximisation of their mutual information, in combination with a deformation field parameterised by cubic B-splines, has been shown to be robust and accurate in many applications. However, the high computation time is a big disadvantage. This work focusses on the optimisation procedure. Many implementations follow a gradient-descent like approach. The time needed for computing the derivative of the mutual information with respect to the B-spline parameters is the bottleneck in this process. We investigate the influence of several gradient approximation techniques on the number of iterations needed and the computation time per iteration. Three methods are studied: a simple finite difference strategy, the so-called simultaneous perturbation method, and a more analytic computation of the gradient based on a continuous, and differentiable representation of the joint histogram. In addition, the effect of decreasing the number of image samples, used for computing the gradient in each iteration, is investigated. Two types of experiments are performed. Firstly, the registration of an image to itself, after application of a known, randomly generated deformation, is considered. Secondly, experiments are performed with 3D ultrasound brain scans, and 3D CT follow-up scans of the chest. The experiments show that the method using an analytic gradient computation outperforms the other two. Furthermore, the computation time per iteration can be extremely decreased, without affecting the rate of convergence and final accuracy, by using very few samples of the image (randomly chosen every iteration) to compute the derivative. With this approach, large data sets (2563) can be registered within 5 minutes on a moderate PC.
Hybrid point-and-intensity-based deformable registration for abdominal CT images
In this paper, we examine the problem of non-rigid, image-to-image registration for CT images of the abdomen. This problem has been previously addressed in many different contexts (e.g., visualization using different imaging modalities, modelling of organ deformation after surgical resection). The particular application in which we are interested is modelling of respiratory motion of abdominal organs, so that we may achieve a more accurate representation of the dose distribution in both targeted structures and non-targeted areas during radiosurgical treatment. Our goal is to register two CT images, each acquired at different positions in the respiratory cycle. We use a transformation model based on B-splines, and take advantage of B-splines' "locality" or "compact support" property to ensure computational efficiency and robust convergence to a satisfactory result. We demonstrate that, although a purely intensity-based registration metric performs well in matching the deformation of the lungs during the respiratory cycle, the movement of other organs (e.g., liver and kidney) is poorly represented due to the poor contrast within and between these organs in the CT images. We introduce a registration metric that is a weighted combination of intensity difference and distance between corresponding points that are manually identified in the two images being registered; we show how the influence of these points can be elegantly added to the metric so that it remains differentiable with respect to the spline control points. Visual inspection reveals that resulting registrations appear to be superior to the intensity-only ones in terms of representation of visceral organ deformation and movement.
Nonrigid registration with adaptive content-based filtering of the deformation field
In present-day medical practice it is often necessary to nonrigidly align image data, either intra- or inter-patient. Current registration algorithms usually do not take different tissue types into account. A problem that might occur with these algorithms is that rigid tissue, like bone, also deforms elastically. We propose a method to correct a deformation field, that is calculated with a nonrigid registration algorithm. The correction is based on a second feature image, which represents the tissue stiffness. The amount of smoothing of the deformation field is related to this stiffness coefficient. By filtering the deformation field on rigid tissue, the deformation field will represent a local rigid transformation. Other parts of the image containing nonrigid tissue are smoothed less, which leaves the original elastic deformation (almost) untouched. It is shown on a synthetic example and on inspiration-expiration CT data of the thorax, that a filtering of the deformation field based on tissue type indeed keeps rigid tissue rigid, thus improving the registration results.
Comparison of 3D dense deformable registration methods for breath-hold reproducibility study in radiotherapy
Vlad Boldea, David Sarrut, Christian Carrie
Breath holding (BH) allows to immobilize organs during radiotherapy treatment of lung cancer. Deformable registration methods applied on 3D Computerized Tomography (CT) scans acquired in BH can be used to evaluate the breath holding reproducibility. Resulting 3D vector fields could then be used to adapt internal margins for each patient. In this work we compare three non-rigid registration schemes with Gaussian, linear-elastic and Nagel-Enckelmann based regularizations. As we do not dispose of gold standard, we analyze vector fields by several operators (transitivity, symmetry, volume dilatation, Jacobian). Experiments were based on clinical data sets of two patients: one with normal lung behavior and second with lung discrepancies which lead to bad BH reproducibility. Results show that none of operators allows to clearly highlight the superiority of a method, except for convergence rapidity and Jacobian.
Texture
icon_mobile_dropdown
Structural analysis of human proximal femur for the prediction of biomechanical strength in vitro: the locally adapted scaling vector method
We introduce an image structure analysis technique suitable in cases where anisotropy plays an important role. The so-called Locally Adapted Scaling Vector Method (LSVM) comprises two steps. First, a procedure to estimate the local main orientation at every point of the image is applied. These orientations are then incorporated in a structure characterization procedure. We apply this methodology to High Resolution Magnetic Resonance Images (HRMRI) of human proximal femoral specimens IN VITRO. We extract a 3D local texture measure to establish correlations with the biomechanical properties of bone specimens quantified via the bone maximum compressive strength. The purpose is to compare our results with the prediction of bone strength using similar isotropic texture measures, bone mineral density, and standard 2D morphometric parameters. Our findings suggest that anisotropic texture measures are superior in cases where directional properties are relevant.
Improving the textural characterization of trabecular bone structure to quantify its changes: the locally adapted scaling vector method
We extend the recently introduced scaling vector method (SVM) to improve the textural characterization of oriented trabecular bone structures in the context of osteoporosis. Using the concept of scaling vectors one obtains non-linear structural information from data sets, which can account for global anisotropies. In this work we present a method which allows us to determine the local directionalities in images by using scaling vectors. Thus it becomes possible to better account for local anisotropies and to implement this knowledge in the calculation of the scaling properties of the image. By applying this adaptive technique, a refined quantification of the image structure is possible: we test and evaluate our new method using realistic two-dimensional simulations of bone structures, which model the effect of osteoblasts and osteoclasts on the local change of relative bone density. The partial differential equations involved in the model are solved numerically using cellular automata (CA). Different realizations with slightly varying control parameters are considered. Our results show that even small changes in the trabecular structures, which are induced by variation of a control parameters of the system, become discernible by applying the locally adapted scaling vector method. The results are superior to those obtained by isotropic and/or bulk measures. These findings may be especially important for monitoring the treatment of patients, where the early recognition of (drug-induced) changes in the trabecular structure is crucial.
Performance of linear and nonlinear texture measures in 2D and 3D for monitoring architectural changes in osteoporosis using computer-generated models of trabecular bone
Holger F. Boehm, Thomas M. Link, Roberto A. Monetti, et al.
Osteoporosis is a metabolic bone disease leading to de-mineralization and increased risk of fracture. The two major factors that determine the biomechanical competence of bone are the degree of mineralization and the micro-architectural integrity. Today, modern imaging modalities (high resolution MRI, micro-CT) are capable of depicting structural details of trabecular bone tissue. From the image data, structural properties obtained by quantitative measures are analysed with respect to the presence of osteoporotic fractures of the spine (in-vivo) or correlated with biomechanical strength as derived from destructive testing (in-vitro). Fairly well established are linear structural measures in 2D that are originally adopted from standard histo-morphometry. Recently, non-linear techniques in 2D and 3D based on the scaling index method (SIM), the standard Hough transform (SHT), and the Minkowski Functionals (MF) have been introduced, which show excellent performance in predicting bone strength and fracture risk. However, little is known about the performance of the various parameters with respect to monitoring structural changes due to progression of osteoporosis or as a result of medical treatment. In this contribution, we generate models of trabecular bone with pre-defined structural properties which are exposed to simulated osteoclastic activity. We apply linear and non-linear texture measures to the models and analyse their performance with respect to detecting architectural changes. This study demonstrates, that the texture measures are capable of monitoring structural changes of complex model data. The diagnostic potential varies for the different parameters and is found to depend on the topological composition of the model and initial “bone density”. In our models, non-linear texture measures tend to react more sensitively to small structural changes than linear measures. Best performance is observed for the 3rd and 4th Minkowski Functionals and for the scaling index method.
Pattern Recognition and Neural Networks
icon_mobile_dropdown
Comparison of decision tree classifiers with neural network and linear discriminant analysis classifiers for computer-aided diagnosis: a Monte Carlo simulation study
The goal of this study was to compare the performance of decision tree (DT) classifiers with artificial neural network (ANN) and linear discriminant analysis (LDA) classifiers under different conditions for the class distributions, feature space dimensionality, and training sample size using a Monte Carlo simulation study. We also investigated a bagging technique for improving the accuracy of the DT. The resubstitution (training) and test accuracies of the classifiers were compared by using the area Az under the ROC curve as the performance measure. Three types of feature space distributions were studied: the Gaussian feature space, a mixture of Gaussians, and a mixture of uniform distributions. The feature space dimensionality was varied between 2 and 12. In a given experiment, 1000 cases were randomly sampled from each distribution, Nt trainers per class was used for classifier design, and the remaining cases were used to test the classifier. The effect of the training sample size was investigated by varying Nt between 30 and 500. Performance measures from 100 experiments were averaged. Our results indicated that, in the Gaussian feature space, the LDA outperformed the other two classifiers, especially when the number of trainers was low. For the mixture of uniform distributions, the Az value of the DT was in general higher than that of the ANN and the LDA. For the mixture of Gaussians, the performances of the DT and ANN classifiers were comparable. Our results indicate that a DT can be a viable alternative to ANN and LDA classifiers in certain feature spaces.
Recognition of micro-array protein crystals images using multi-scale representations
Ya Wang, David H. Kim, Elsa D. Angelini, et al.
Micro-array protein crystal images are now routinely acquired automatically by CCD cameras. High-throughput automatic classification of protein crystals requires to alleviation of the time-consuming task of manual visual inspection. We propose a classification framework combined with a multi-scale image processing method for recognizing protein crystals and precipitates versus clear drops. The main two points of the processing method are the multi-scale Laplacian pyramid filters and histogram analysis techniques to find an effective feature vector. The processing steps include: 1. Tray well cropping using Radon Transform; 2. Droplet cropping using an ellipsoid Hough Transform; 3. Multi-scale image separation with Laplacian pyramidal filters; 4. Feature vector extraction from the histogram of the multi-scale boundary images. The feature vector combines geometric and texture features of each image and provides input to a feed forward binomial neural network classifier. Using human (expert crystallographers) classified images as ground truth, the current experimental results gave 86% true positive and 94% true negative rates (average true percentage is 90%) using an image database which contained over 2,000 images. To enable NESG collaborators to carry our crystal classification, a web-based Matlab server was also developed. Users at other locations on the internet can input micro-array crystal image folders and parameters for training and testing processes through a friendly web interface. Recognition results are shown on the client side website and may be downloaded by a remote user as an Excel spreadsheet file.
Stomach, intestine, and colon tissue discriminators for wireless capsule endoscopy images
Jeff Berens, Michal Mackiewicz, Duncan Bell
Wireless Capsule Endoscopy (WCE) is a new colour imaging technology that enables close examination of the interior of the entire small intestine. Typically, the WCE operates for ~8 hours and captures ~40,000 useful images. The images are viewed as a video sequence, which generally takes a doctor over an hour to analyse. In order to activate certain key features of the software provided with the capsule, it is necessary to locate and annotate the boundaries between certain gastrointestinal (GI) tract regions (stomach, intestine and colon) in the footage. In this paper we propose a method of automatically discriminating stomach, intestine and colon tissue in order to significantly reduce the video assessment time. We use hue saturation chromaticity histograms which are compressed using a hybrid transform, incorporating the Discrete Cosine Transform (DCT) and Principal Component Analysis (PCA). The performance of two classifiers is compared: k-nearest neighbour (kNN) and Support Vector Classifier (SVC). After training the classifier, we applied a narrowing step algorithm to converge to the points in the video where the capsule firstly passes through the pylorus (the valve between the stomach and the intestine) and later the ileocaecal valve (IV, the valve between the intestine and colon). We present experimental results that demonstrate the effectiveness of this method.
Informative-frame filtering in endoscopy videos
Yong Hwan An, Sae Hwang, JungHwan Oh, et al.
Advances in video technology are being incorporated into today’s healthcare practice. For example, colonoscopy is an important screening tool for colorectal cancer. Colonoscopy allows for the inspection of the entire colon and provides the ability to perform a number of therapeutic operations during a single procedure. During a colonoscopic procedure, a tiny video camera at the tip of the endoscope generates a video signal of the internal mucosa of the colon. The video data are displayed on a monitor for real-time analysis by the endoscopist. Other endoscopic procedures include upper gastrointestinal endoscopy, enteroscopy, bronchoscopy, cystoscopy, and laparoscopy. However, a significant number of out-of-focus frames are included in this type of videos since current endoscopes are equipped with a single, wide-angle lens that cannot be focused. The out-of-focus frames do not hold any useful information. To reduce the burdens of the further processes such as computer-aided image processing or human expert’s examinations, these frames need to be removed. We call an out-of-focus frame as non-informative frame and an in-focus frame as informative frame. We propose a new technique to classify the video frames into two classes, informative and non-informative frames using a combination of Discrete Fourier Transform (DFT), Texture Analysis, and K-Means Clustering. The proposed technique can evaluate the frames without any reference image, and does not need any predefined threshold value. Our experimental studies indicate that it achieves over 96% of four different performance metrics (i.e. precision, sensitivity, specificity, and accuracy).
Deformable Geometry
icon_mobile_dropdown
A novel framework for automated 3D PDM construction using deformable models
Zheen Zhao, Eam Khwang Teoh
This paper describes a novel framework to build 3D Point Distribution Model (PDM) from a set of segmented volumetric images. This method is based on a deformable model algorithm. Each training sample deforms to approximate all other training shapes. The training sample with best approximation results is then chosen as the template. Finally, the poor approximation results from this template are improved by the "bridge over" scheme, which deforms the template to approximate intermediate training shapes and then deforms the approximations to outliers. The method is applied to construct a 3D PDM of 20 human brain ventricles. The results show that the algorithm leads to more accurate representation than traditional framework. Also, the performance of the PDM of soft tissue is comparable with the PDM of bone structures by a previous method. The traditional framework of deformable model based approach selects the template arbitrarily and deforms the template to approximate training shapes directly. The limitation of the traditional framework is that the representation accuracy of the PDM entirely depends on the direct approximation. Moreover, the arbitrary template selection deteriorates the accuracy of the approximation. Our framework that features template selection and indirect approximation solves the shortcomings and improves the PDM representation accuracy. Furthermore, the "bridge over" framework could be used with any deformable model algorithm. In this sense, the method is a generic framework open to future investigation.
A marker-induced vector field for reduced sensitivity to initialization for parametric and geometric deformable models
Deformable models are powerful approaches to medical image analysis, particularly segmentation. However, the outcome of applying a parametric or geometric deformable model is often significantly dependent on its initialization. This is an obstacle to robust automatic segmentation. Based on theoretical analyses of the watershed transform, we propose a novel approach to reducing this sensitivity to initialization by deriving a vector field from topographic and Euclidean distance transforms. This vector field is aimed to extend the influence of the gradients at the boundary of the segmentation target over the entire image in a consistent fashion, while ignoring any irrelevant gradients in the original image. Initiated by one or more segmentation seeds, the vector field is first computed using an efficient numerical method, and subsequently participates in the model's evolution process. Integration of the vector field has so far been performed with a two-dimensional (2D) parametric deformable model and with a three-dimensional (3D) geodesic active contour level set model. We believe that our approach will enable a higher degree of automation for deformable-model-based segmentation, particularly in situations where the seeds can be placed automatically based on, for example, a priori knowledge regarding the anatomy and the intensity differentiation between the target and the background. Experiments on segmenting organs and tumors from CT and MR images using the integrated models have shown that this is a promising approach.
Physics-based deformable organisms for medical image analysis
Previously, "Deformable organisms" were introduced as a novel paradigm for medical image analysis that uses artificial life modelling concepts. Deformable organisms were designed to complement the classical bottom-up deformable models methodologies (geometrical and physical layers), with top-down intelligent deformation control mechanisms (behavioral and cognitive layers). However, a true physical layer was absent and in order to complete medical image segmentation tasks, deformable organisms relied on pure geometry-based shape deformations guided by sensory data, prior structural knowledge, and expert-generated schedules of behaviors. In this paper we introduce the use of physics-based shape deformations within the deformable organisms framework yielding additional robustness by allowing intuitive real-time user guidance and interaction when necessary. We present the results of applying our physics-based deformable organisms, with an underlying dynamic spring-mass mesh model, to segmenting and labelling the corpus callosum in 2D midsagittal magnetic resonance images.
Bi-temporal 3D active appearance models with applications to unsupervised ejection fraction estimation
Mikkel B. Stegmann, Dorthe Pedersen
Rapid and unsupervised quantitative analysis is of utmost importance to ensure clinical acceptance of many examinations using cardiac magnetic resonance imaging (MRI). We present a framework that aims at fulfilling these goals for the application of left ventricular ejection fraction estimation in four-dimensional MRI. The theoretical foundation of our work is the generative two-dimensional Active Appearance Models by Cootes et al., here extended to bi-temporal, three-dimensional models. Further issues treated include correction of respiratory induced slice displacements, systole detection, and a texture model pruning strategy. Cross-validation carried out on clinical-quality scans of twelve volunteers indicates that ejection fraction and cardiac blood pool volumes can be estimated automatically and rapidly with accuracy on par with typical inter-observer variability.
Accurate left atrium segmentation in multislice CT images using a shape model
Segmentation and labelling of the left atrium from pre-operative images could be a valuable source of information for the planning of electrophysiology procedures to cure atrial fibrillation. A method is presented that uses multi-slice computed tomography (MSCT) images for this purpose that were initially acquired for coronary assessment. The method combines the power of active shape models (robustness by use of prior anatomical knowledge) with the advantages of solely data driven segmentation methods (accuracy). A triangular shape model was built for the human left atrium and its pulmonary vein trunks. It was automatically adapted to the MSCT images, labelling these structures and segmenting them coarsely. In addition, a segmentation of the blood pool by a Hounsfield threshold was applied to the images. The enclosed volumes were triangulated to get a fine surface representation yet still including many distracting objects (the artery tree, coronaries, adjacent chambers, and bones). A correspondence between surface triangles of the coarse, but anatomically labelled model surface and those of the fine iso-surface was established by a similarity criterion on position and orientation. This allows for the refinement of the model-based segmentation showing more anatomical details by selection of corresponding parts of the iso-surface. Vice versa, the correspondence could be used to assign anatomical labels to each iso-surface patch.
Cardiac deformation recovery via incompressible transformation decomposition
This paper presents a method for automated deformation recovery of the left and right ventricular wall from a time sequence of anatomical images of the heart. The deformation is recovered within the heart wall, i.e. it is not limited only to the epicardium and endocardium. Most of the suggested methods either ignore or approximately model incompressibility of the heart wall. This physical property of the cardiac muscle is mathematically guaranteed to be satisfied by the proposed method. A scheme for decomposition of a complex incompressible geometric transformation into simpler components and its application to cardiac deformation recovery is presented. A general case as well as an application specific solution is discussed. Furthermore, the manipulation of the constructed incompressible transformations, including the computation of the inverse transformation, is computationally inexpensive. The presented method is mathematically guaranteed to generate incompressible transformations which are experimentally shown to be a very good approximation of actual cardiac deformations. The transformation representation has a relatively small number of parameters which leads to a fast deformation recovery. The approach was tested on six sequences of two-dimensional short-axis cardiac MR images. The cardiac deformation was recovered with an average error of 1.1 pixel. The method is directly extendable to three dimensions and to the entire heart.
Tomographic Reconstruction
icon_mobile_dropdown
Reconstruction-plane-dependent weighted FDK algorithm for cone beam volumetric CT
The original FDK algorithm has been extensively employed in medical and industrial imaging applications. With an increased cone angle, cone beam (CB) artifacts in images reconstructed by the original FDK algorithm deteriorate, since the circular trajectory does not satisfy the so-called data sufficiency condition (DSC). A few “circular plus” trajectories have been proposed in the past to reduce CB artifacts by meeting the DSC. However, the circular trajectory has distinct advantages over other scanning trajectories in practical CT imaging, such as cardiac, vascular and perfusion applications. In addition to looking into the DSC, another insight into the CB artifacts of the original FDK algorithm is the inconsistency between conjugate rays that are 180° apart in view angle. The inconsistence between conjugate rays is pixel dependent, i.e., it varies dramatically over pixels within the image plane to be reconstructed. However, the original FDK algorithm treats all conjugate rays equally, resulting in CB artifacts that can be avoided if appropriate view weighting strategy is exercised. In this paper, a modified FDK algorithm is proposed, along with an experimental evaluation and verification, in which the helical body phantom and a humanoid head phantom scanned by a volumetric CT (64 x 0.625 mm) are utilized. Without extra trajectories supplemental to the circular trajectory, the modified FDK algorithm applies reconstruction-plane-dependent view weighting on projection data before 3D backprojection, which reduces the inconsistency between conjugate rays by suppressing the contribution of one of the conjugate rays with a larger cone angle. Both computer-simulated and real phantom studies show that, up to a moderate cone angle, the CB artifacts can be substantially suppressed by the modified FDK algorithm, while advantages of the original FDK algorithm, such as the filtered backprojection algorithm structure, 1D ramp filtering, and data manipulation efficiency, can be maintained.
Estimating 0th and 1st moments in C-arm CT data for extrapolating truncated projections
C-Arm CT systems suffer from artifacts due to truncated projections caused by a finite detector size. One method used to mitigate the truncation artifacts is projection extrapolation without a priori knowledge. This work focuses on estimating the 0th and 1st moments of an image, which can be used to extrapolate a set of truncated projections. If some projections are not truncated, then accurate estimation of the moments can be achieved using only those projections. The more difficult case arises when all projections are truncated by some amount. For this case we make simplifying assumptions and fit the truncated projections with elliptical profiles. From this fit, we estimate the 0th and 1st moments of the original image. These estimated moments are then used to perform an extrapolation of the truncated projections, where each projection meets a constraint based on the 0th and 1st moments (moment extrapolation). This work focuses on how accurate moment estimates must be for moment extrapolation to be effective. The algorithm was tested on simulated and real data for the head, thorax, and abdomen, and those results were compared to symmetric mirroring by Ohnesorge et al., another extrapolation technique that requires no a priori knowledge. Overall, moment estimation and mass extrapolation alleviates a large amount of image artifact, and can improve on other extrapolation techniques. For the real CT head and abdominal data, the average reconstruction error for mass extrapolation was 48% less than the reconstruction error for symmetric mirroring.
Exact and efficient cone-beam reconstruction algorithm for a short-scan circle combined with various lines
X-ray 3D rotational angiography based on C-arm systems has become a versatile and established tomographic imaging modality for high contrast objects in interventional environment. Improvements in data acquisition, e.g. by use of flat panel detectors, will enable C-arm systems to resolve even low-contrast details. However, further progress will be limited by the incompleteness of data acquisition on the conventional short-scan circular source trajectories. Cone artifacts, which result from that incompleteness, significantly degrade image quality by severe smearing and shading. To assure data completeness a combination of a partial circle with one or several line segments is investigated. A new and efficient reconstruction algorithm is deduced from a general inversion formula based on 3D Radon theory. The method is theoretically exact, possesses shift-invariant filtered backprojection (FBP) structure, and solves the long object problem. The algorithm is flexible in dealing with various circle and line configurations. The reconstruction method requires nothing more than the theoretically minimum length of scan trajectory. It consists of a conventional short-scan circle and a line segment approximately twice as long as the height of the region-of-interest. Geometrical deviations from the ideal source trajectory are considered in the implementation in order to handle data of real C-arm systems. Reconstruction results show excellent image quality free of cone artifacts. The proposed scan trajectory and reconstruction algorithm assure excellent image quality and allow low-contrast tomographic imaging with C-arm based cone-beam systems. The method can be implemented without any hardware modifications on systems commercially available today.
Reconstruction of kinetic parameter images directly from dynamic PET sinograms
Mustafa E. Kamasak, Charles A. Bouman, Evan D. Morris, et al.
Recently, there has been interest in estimating kinetic model parameters for each voxel in a PET image. To do this, the activity images are first reconstructed from PET sinogram frames at each measurement time, and then the kinetic parameters are estimated by fitting a model to the reconstructed time-activity response of each voxel. However, this indirect approach to kinetic parameter estimation tends to reduce signal-to-noise ratio (SNR) because of the requirement that the sinogram data be divided into individual time frames. In 1985, Carson and Lange proposed, but did not implement, a method based on the EM algorithm for direct parametric reconstruction. More recently, researchers have developed semi-direct methods which use spline-based reconstruction, or direct methods for estimation of kinetic parameters from image regions. However, direct voxel-wise parametric reconstruction has remained a challenge due to the unsolved complexities of inversion and required spatial regularization. In this work, we demonstrate an efficient method for direct voxel-wise reconstruction of kinetic parameters (as a parametric image) from all frames of the PET data. The direct parametric image reconstruction is formulated in a Bayesian framework, and uses the parametric iterative coordinate descent (PICD) algorithm to solve the resulting optimization problem. This PICD algorithm is computationally efficient and allows the physiologically important kinetic parameters to be spatially regularized. Our experimental simulations demonstrate that direct parametric reconstruction can substantially reduce estimation error of kinetic parameters as compared to indirect methods.
Region of interest reconstruction from truncated data in circular cone-beam CT
In many applications of circular cone-beam CT, it is not uncommon that the size of the field of view (FOV) is smaller than that of the imaging object, thus leading to transverse truncation in projection data. Exact reconstruction in any region is not possible from such truncated data using conventional algorithms. Recently, an exact algorithm for image reconstruction on PI-line segments in helical cone-beam CT has been proposed. This algorithm, which we refer to as the backprojection-filtration (BPF) algorithm, can naturally address the problem of exact region of interest (ROI) reconstruction from such truncated data. In this work, we modified this algorithm to reconstructing images in circular cone-beam scan. The unique property of this modified algorithm is that it can reconstruct exact ROIs in midplane and approximate ROIs in other planes from transversely truncated data. We have performed computer-simulation studies to validate the theoretical assertions. Preliminary results demonstrate that the proposed algorithm provides a solution to the truncation problems caused by limited FOV size.
Segmentation I
icon_mobile_dropdown
Shape regularized active contour based on dynamic programming for anatomical structure segmentation
We present a method to incorporate nonlinear shape prior constraints into segmenting different anatomical structures in medical images. Kernel space density estimation (KSDE) is used to derive the nonlinear shape statistics and enable building a single model for a class of objects with nonlinearly varying shapes. The object contour is coerced by image-based energy into the correct shape sub-distribution (e.g., left or right lung), without the need for model selection. In contrast to an earlier algorithm that uses a local gradient-descent search (susceptible to local minima), we propose an algorithm that iterates between dynamic programming (DP) and shape regularization. DP is capable of finding an optimal contour in the search space that maximizes a cost function related to the difference between the interior and exterior of the object. To enforce the nonlinear shape prior, we propose two shape regularization methods, global and local regularization. Global regularization is applied after each DP search to move the entire shape vector in the shape space in a gradient descent fashion to the position of probable shapes learned from training. The regularized shape is used as the starting shape for the next iteration. Local regularization is accomplished through modifying the search space of the DP. The modified search space only allows a certain amount of deformation of the local shape from the starting shape. Both regularization methods ensure the consistency between the resulted shape with the training shapes, while still preserving DP’s ability to search over a large range and avoid local minima. Our algorithm was applied to two different segmentation tasks for radiographic images: lung field and clavicle segmentation. Both applications have shown that our method is effective and versatile in segmenting various anatomical structures under prior shape constraints; and it is robust to noise and local minima caused by clutter (e.g., blood vessels) and other similar structures (e.g., ribs). We believe that the proposed algorithm represents a major step in the paradigm shift to object segmentation under nonlinear shape constraints.
Inducing node specification in active shape models for accurate lung-field segmentation
Amit Singhal, Mark Bolin, Hui Luo, et al.
We have developed an active shape model (ASM)-based segmentation scheme that uses the original Cootes et al. formulation for the underlying mechanics of the ASM but improves the model by fixating selected nodes at specific structural boundaries called transitional landmarks. Transitional landmarks identify the change from one boundary type (such as lung-field/heart) to another (lung-field/diaphragm). This results in a multi-segmented lung-field boundary where each segment correlates to a specific boundary type (lung-field/heart, lung-field/aorta, lung-field/rib-cage, etc.). The node-specified ASM is built using a fixed set of equally spaced feature nodes for each boundary segment. This allows the nodes to learn local appearance models for a specific boundary type, rather than generalizing over multiple boundary types, which results in a marked improvement in boundary accuracy. In contrast, existing lung-field segmentation algorithms based only on ASM simply space the nodes equally along the entire boundary without specification. We have performed extensive experiments using multiple datasets (public and private) and compared the performance of the proposed scheme with other contour-based methods. Overall, the improved accuracy is 3-5 &percent; over the standard ASM and, more importantly, it corresponds to increased alignment with salient anatomical structures. Furthermore, the automatically generated lung-field masks lead to the same fROC for lung-nodule detection as hand-drawn lung-field masks. The accurate landmarks can be easily used for detecting other structures in the lung field. Based on the related landmarks (mediastinum-heart transition, heart-diaphragm transition), we have extended the work to heart segmentation.
Active appearance model-based segmentation of hip radiographs
Nabil Boukala, Eric Favier, Bernard Laget
Despite the advantages that 3D medical image analysis methods offer and the fast introduction of CT and MRI, to date most hospitals use radiographs to perform preoperative planning of hip surgeries and automatic analysis of hip radiographs is still of interest. In this paper, we present a novel method for segmentation of bone structures in anterior-posterior (AP) radiographs based on Active Appearance Models. The pelvis shape is decomposed in circular regions which reflect convex local arrangement of shape points. A priori global knowledge of the geometric structure of this region representation is captured by a statistical deformable template integrating a set of admissible deformations. The texture of each region is modeled separately, and we build a local Active Appearance Model for each region. A leave-one-out test was used to evaluate the performance of the proposed method and to compare it with conventional Active Appearance Model. The results demonstrate that the method is precise and very robust to large-scale noise present in radiographs, and that it can be useful in the context of preoperative planning of hip surgery.
Deformable associate net approach for chest CT image segmentation
We propose a new deformable model Deformable Associate Net (DAN). It is represented by a set of nodes which are associated by deformation constrains such as topology association, inter-part association, intra-part association, and geometry to atlas association. Each node in the model is given a priority, and hence DAN is a hierarchical model in which each layer is decided by nodes with same priority. Directional edges and dynamic generated local atlases are used in energy function to incorporate knowledge about tissue and image acquisition. A fast digital topology based method is designed to check whether topology of the model is changed under deformation. The deformation procedure hierarchically combines global and local deformations. Layers with high priority deform first. Once a higher layer is deformed to its target position in an image, the nodes in this layer are fixed, and then used as reference to help lower layers deform to their initial positions. At a particular layer, the model is first deformed by using global affine transformation to fit the image roughly, and then is warped by using a local deformation to fit the image better. The proposed method has been used to segment chest CT images for thoracic surgical planning, and it is also promising for other medical applications, such as model based image registration, and model-based 3D modeling.
Automated quantification of cardiac short-axis multi-slice CT images for assessment of left ventricular global function
Mikhail G. Danilouchkine, Faiza Admiraal-Behloul, Rob J. van der Geest, et al.
This paper describes a method for automatic contour detection in reformatted short-axis (SA) cardiac computed tomography (CT) using a virtual exploring robot. The robot is a tricycle with a steering front wheel. Its motion obeys a set of kinematic equations and is subject to the non-holonomic constraints. The robot is designed to navigate in the binary representation of a cardiac image, consisting of the allowed navigational and obstacle spaces. It is initially positioned inside the allowed navigational space. Avoiding obstacles, the robot autonomously cruises through the navigational space and collects information about the location of the left ventricular (LV) boundaries. Consequently, the obtained information is used to reconstruct the endocardial and epicardial contours. Validation of the method was performed on in-vivo multislice multiphase short-axis cardiac CT images of ten subjects. Results showed good correlation between the quantitative parameters, computed from manual and automatic segmentation: for end-diastolic volume (EDV) r=0.99, for end-systolic volume (ESV) r=0.98, ejection fraction (EF) r=0.83, and LV mass (LVM) r=0.95.
Stopping rules for active contour segmentation of ultrasound cardiac images
The presence of speckle (a spatial stochastic process in an ultrasound image) makes ultrasound segmentation difficult. Speckle introduces local minima in the MAP energy function of an active contour, and when evolving under gradient descent, the contour gets trapped in a spurious local minimum. In this paper, we propose an alternate technique for evolving a MAP active contour. The technique has two parts: a deterministic evolution strategy called tunneling descent which escapes from spurious local minima, and a stopping rule for terminating the evolution. The combination gives an algorithm that is robust and gives good segmentations. The algorithm also benefits from having only a few free parameters which do not require tweaking. We present the conceptual framework of the algorithm in this paper, and study the impact of different stopping rules on the performance of the algorithm. The algorithm is used to segment the endocardium in cardiac ultrasound images. We present segmentation results in this paper and an experimental evaluation of different stopping rules on the performance of the algorithm. Although the algorithm is presented as an ultrasound segmentation technique, in fact, it can be used to segment any first-order texture boundary.
Robust fast automatic skull stripping of MRI-T2 data
The efficacy of image processing and analysis on skull stripped MR images vis-a-vis the original images is well established. Additionally, compliance with the Health Insurance Portability and Accountability Act (HIPAA) requires neuroimage repositories to anonymise the images before sharing them. This makes the non-trivial skull stripping process all the more significant. While a number of optimal approaches exist to strip the skull from T1-weighted MR images to the best of our knowledge, there is no simple, robust, fast, parameter free and fully automatic technique to perform the same on T2-weighted images. This paper presents a strategy to fill this gap. It employs a fast parameterization of the T2 image intensity onto a standardized T1 intensity scale. The parametric "T1-like" image obtained via the transformation, which takes only a few seconds to compute, is subsequently processed by any of the many T1-based brain extraction techniques to derive the brain mask. Masking the original T2 image with this brain mask strips the skull. By standardizing the intensity of the parametric image, preset algorithm-specific parameters (if any) could be used across multiple datasets. The proposed scheme has been used in a number of phantom and clinical T2 brain datasets to successfully strip the skull.
Segmentation II: Vasculature
icon_mobile_dropdown
Multidimensional segmentation of coronary intravascular ultrasound images using knowledge-based methods
Mark Eric Olszewski, Andreas Wahle, Sarah C. Vigmostad, et al.
In vivo studies of the relationships that exist among vascular geometry, plaque morphology, and hemodynamics have recently been made possible through the development of a system that accurately reconstructs coronary arteries imaged by x-ray angiography and intravascular ultrasound (IVUS) in three dimensions. Currently, the bottleneck of the system is the segmentation of the IVUS images. It is well known that IVUS images contain numerous artifacts from various sources. Previous attempts to create automated IVUS segmentation systems have suffered from either a cost function that does not include enough information, or from a non-optimal segmentation algorithm. The approach presented in this paper seeks to strengthen both of those weaknesses -- first by building a robust, knowledge-based cost function, and then by using a fully optimal, three-dimensional segmentation algorithm. The cost function contains three categories of information: a compendium of learned border patterns, information theoretic and statistical properties related to the imaging physics, and local image features. By combining these criteria in an optimal way, weaknesses associated with cost functions that only try to optimize a single criterion are minimized. This cost function is then used as the input to a fully optimal, three-dimensional, graph search-based segmentation algorithm. The resulting system has been validated against a set of manually traced IVUS image sets. Results did not show any bias, with a mean unsigned luminal border positioning error of 0.180 ± 0.027 mm and an adventitial border positioning error of 0.200 ± 0.069 mm.
Brain aneurysm segmentation in CTA and 3DRA using geodesic active regions based on second order prototype features and nonparametric density estimation
Coupling the geodesic active contours model with statistical information based on regions introduces robustness in the segmentation of images with weak or inhomogeneous gradients. In the estimation of the probability density function for each region take part the definition of the features that describe the image inside the different regions and the method of density estimation itself. A Gaussian Mixture Model is frequently proposed for density estimation. This approach is based on the assumption that the intensity distribution of the image is the most discriminant feature in a region. However, the use of second order features provides a better discrimination of the different regions, as these features represent more accurately the local properties of the image manifold. Due to the high dimensionality of the problem, the use of non parametric density estimation methods becomes necessary. In this article, we present a novel method of introducing the second order information of an image for non parametric estimation of the probability density functions of the different tissues that are present in medical images. The novelty of the method stems on the use of the response of the image under an orthogonal harmonic operator set projected onto a prototype space for feature generation. The technique described here is applied to the segmentation of brain aneurysms in Computed Tomography Angiography (CTA) and 3D Rotational Angiography (3DRA) showing a qualitative improvement from the Gaussian Mixture Model approach.
Complexity analysis of angiogenesis vasculature
Vijay Mahadevan, James Alexander Tyrell, Ricky T. Tong, et al.
Tumor vasculature has a high degree of irregularity as compared to normal vasculature. The quantification of the morphometric complexity in tumor images can be useful in diagnosis. Also, it is desirable in several other medical applications to have an automated complexity analysis to aid in diagnosis and prognosis under treatment. e.g. in diabetic retinopathy and in arteriosclerosis. In addition, prior efforts at segmentation of the tumor vasculature using matched filtering, template matching and splines have been hampered by the irregularity of these vessels. We try to solve both problems by introducing a novel technique for vessel detection, followed by a tracing-independent complexity analysis based on a combination of ideas. First, the vessel cross-sectional profile is modeled using a continuous and everywhere differentiable family of super-Gaussian curves. This family generates rectangular profiles that can accurately localize the vessel boundaries in microvasculature images. Second, a robust non-linear regression algorithm based on M-estimators is used to estimate the parameters that optimally characterize the vessel’s shape. A framework for the quantitative analysis of the complexity of the vasculature based on the vessel detection is presented. A set of measures that quantify the complexity are proposed viz. Squared Error, Entropy-based and Minimum Description Length-based Shape Complexities. They are completely automatic and can deal with complexities of the entire vessel unlike existing tortuousity measures which deal only with vessel centerlines. The results are validated using carefully constructed phantom and real image data with ground truth information from an expert observer.
A novel node-structural map for angiogenesis analysis
Yizhi Xiong, Tong Zhao, Dodanim Talavera, et al.
We introduce a novel node-structural map method for digital image-based analysis of angiogenesis, and apply it to quantitate the effect of pro-angiogenic (e.g. VEGF) and anti-angiogenic (e.g. anti-VEGF) treatments on blood vessel patterns in the quail chorioallantoic membrane (CAM) in vivo model. After vessel segmentation and skeletonization, a node-structural map is derived to represent the structural property of each pixel in the skeletonized image, such as end nodes (root and leaf), branch and furcation/junction nodes. By combining the node-structural map with vascular thickness map, our proposed method provides more detailed morphometric and structural measurement of vascular angiogenesis, such as average diameter, average branch length, branch thickness distribution, density of branch and furcation nodes, etc. Furthermore, the concept of vascular generations will be proposed. Accordingly, more detailed measurements related to the generations of the vascular tree can be easily evaluated. From the quantitative analysis results, it correctly shows the fact that VEGF/anti-VEGF can modify the blood vessel pattern and increase/decrease the vessel density in CAM significantly, compared to the phosphate-buffered saline treated controls.
Segmentation III
icon_mobile_dropdown
Segmentation of medical images combining local, regional, global, and hierarchical distances into a bottom-up region merging scheme
Thomas Martin Lehmann, Daniel Beier, Christian Thies, et al.
Segmentation of medical images is fundamental for many high-level applications. Unsupervised techniques such as region growing or merging allow automated processing of large data amounts. The regions are usually described by a mean feature vector, and the merging decisions are based on the Euclidean distance. This kind of similarity model is strictly local, since the feature vector of each region is calculated without evaluating the region's surrounding. Therefore, region merging often fails to extract visually comprehensible and anatomically relevant regions. In our approach, the local model is extended. Regional similarity is calculated for a pair of adjacent regions, e.g. considering the contrast along their common border. Global similarity components are obtained by analyzing the entire image partitioning before and after a hypothetical merge. Hierarchical similarities are derived from the iteration history. Local, regional, global, and hierarchical components are combined task-specifically guiding the iterative region merging process. Starting with an initial watershed segmentation, the process terminates when the entire image is represented as a single region. A complete segmentation takes only a few seconds. Our approach is evaluated contextually on plain radiographs that display human hands acquired for bone age determination. Region merging based on a local model fails to detect most bones, while a correct localization and delineation is obtained with the combined model. A gold standard is computed from ten manual segmentations of each radiograph to evaluate the quality of delineation. The relative error of labeled pixels is 15.7%, which is slightly more than the mean error of the ten manual references to the gold standard (12%). The flexible and powerful similarity model can be adopted to many other segmentation tasks in medical imaging.
Automatic lung lobe segmentation in x-ray CT images by 3D watershed transform using anatomic information from the segmented airway tree
The human lungs are divided into five distinct anatomic compartments called lobes. The physical boundaries between the lobes are called the lobar fissures. Detection of lobar fissure positions in pulmonary X-ray CT images is of increasing interest for the diagnosis of lung disease. We have developed an automatic method for segmentation of all five lung lobes simultaneously using a 3D watershed transform on the distance transform of a previously generated vessel mask, linearly combined with the original data. Due to the anatomically separate airway sub-trees for individual lobes, we can accurately and automatically place seed points for the watershed segmentation based on the airway tree anatomical description, due to the fact that lower generation airway and vascular tree segments are located near each other. This, along with seed point placement using information on the spatial location of the lobes, can give a close approximation to the actual lobar fissures. The accuracy of the lobar borders is assessed by comparing the automatic segmentation to manually traced lobar boundaries. Averaged over all volumes, the RMS distance errors for the left oblique fissure, right oblique fissure and right horizontal fissure are 3.720 mm, 0.713 mm and 1.109 mm respectively.
Mid-sagittal plane and mid-sagittal surface optimization in brain MRI using a local symmetry measure
Mikkel B. Stegmann, Karl Skoglund, Charlotte Ryberg
This paper describes methods for automatic localization of the mid-sagittal plane (MSP) and mid-sagittal surface (MSS). The data used is a subset of the Leukoaraiosis And DISability (LADIS) study consisting of three-dimensional magnetic resonance brain data from 62 elderly subjects (age 66 to 84 years). Traditionally, the mid-sagittal plane is localized by global measures. However, this approach fails when the partitioning plane between the brain hemispheres does not coincide with the symmetry plane of the head. We instead propose to use a sparse set of profiles in the plane normal direction and maximize the local symmetry around these using a general-purpose optimizer. The plane is parameterized by azimuth and elevation angles along with the distance to the origin in the normal direction. This approach leads to solutions confirmed as the optimal MSP in 98 percent of the subjects. Despite the name, the mid-sagittal plane is not always planar, but a curved surface resulting in poor partitioning of the brain hemispheres. To account for this, this paper also investigates an optimization strategy which fits a thin-plate spline surface to the brain data using a robust least median of squares estimator. Albeit computationally more expensive, mid-sagittal surface fitting demonstrated convincingly better partitioning of curved brains into cerebral hemispheres.
Super-resolved multi-channel fuzzy segmentation of MR brain images
Ying Bai, Xiao Han, Dzung L. Pham, et al.
We propose a new fuzzy segmentation framework that incorporates the idea of super-resolution image reconstruction. The new framework is designed to segment data sets comprised of orthogonally acquired magnetic resonance (MR) images by taking into account their different system point spread functions. Formulating the reconstruction within the segmentation framework improves its robustness and stability, and makes it possible to incorporate multispectral scans that possess different contrast properties into the super-resolution reconstruction process. Our method has been tested on both simulated and real 3D MR brain data.
A level set segmentation for computer-aided dental x-ray analysis
Shuo Li, Thomas Fevens, Adam Krzyzak, et al.
A level-set-based segmentation framework for Computer Aided Dental X-rays Analysis (CADXA) is proposed. In this framework, we first employ level set methods to segment the dental X-ray image into three regions: Normal Region (NR), Potential Abnormal Region (PAR), Abnormal and Background Region (ABR). The segmentation results are then used to build uncertainty maps based on a proposed uncertainty measurement method and an analysis scheme is applied. The level set segmentation method consists of two stages: a training stage and a segmentation stage. During the training stage, manually chosen representative images are segmented using hierarchical level set region detection. The segmentation results are used to train a support vector machine (SVM) classifier. During the segmentation stage, a dental X-ray image is first classified by the trained SVM. The classifier provides an initial contour which is close to the correct boundary for the coupled level set method which is then used to further segment the image. Different dental X-ray images are used to test the framework. Experimental results show that the proposed framework achieves faster level set segmentation and provides more detailed information and indications of possible problems to the dentist. To our best knowledge, this is one of the first results on CADXA using level set methods.
Pull-push level sets: a new term to encode prior knowledge for the segmentation of teeth images
Rodrigo de Luis Garcia, Raul San Jose Estepar, Carlos Alberola-Lopez
This paper presents a novel level set method for contour detection in multiple-object scenarios applied to the segmentation of teeth images. Teeth segmentation from 2D images of dental plaster cast models is a difficult problem because it is necessary to independently segment several objects that have very badly defined borders between them. Current methods for contour detection which only employ image information cannot successfully segment such structures. Being therefore necessary to use prior knowledge about the problem domain, current approaches in the literature are limited to the extraction of shape information of individual objects, whereas the key factor in such a problem are the relative positions of the different objects composing the anatomical structure. Therefore, we propose a novel method for introducing such information into a level set framework. This results in a new energy term which can be explained as a regional term that takes into account the relative positions of the different objects, and consequently creates an attraction or repulsion force that favors a determined configuration. The proposed method is compared with balloon and GVF snakes, as well as with the Geodesic Active Regions model, showing accurate results.
Morphological filtering based on the Minkowski functionals in 3D for segmentation of macromolecular structures in intact eukaryotic cells depicted by cryo-electron tomography
Holger F. Boehm, Ferdinand Jamitzki, Roberto A. Monetti, et al.
In this contribution, we propose a novel approach to the segmentation of tomographic image data considering topological properties of binarized image components expressed in terms of the Minkowski Functionals in 3D. Electron tomography is a non-invasive method for three-dimensional (3D) reconstruction of cellular sub-structures from a series of projection images (i.e. from a tilt series) recorded with a transmission electron microscope. Data obtained by electron tomography provide a rich source of quantitative information concerning the structural composition and organization of cellular components. It allows to obtain 3D information on structural cellular arrangements at a significantly higher resolution than any other of the currently available imaging modalities. A major challenge, in this context, is the segmentation of the image data with respect to the identification macro-molecular structures such as the actin-cytoskeleton or cell organelles. We introduce a morphological filtering algorithm based on the Minkowski Functionals in 3D for segmentation of macromolecular structures in intact eukaryotic cells depicted by cryo-electron tomography. In mathematical topology, multi-dimensional convex objects can be characterized with respect to shape, structure, and the connectivity of their components using a set of morphological descriptors known as the Minkowski functionals. In a 3D-Euclidian space, these correspond to volume, surface area, mean integral curvature, and the Euler-Poincare characteristic. The morphological filtering procedure is applied to a 3D image data of an intact, ice-embedded Dictyostelium cell obtained by low dose transmission electron microscopy using a tilt series of -50° to +41.5° with an increment of 1.5°. Our method allows to separate cellular components with predefined textural properties, e.g. filamentary or globular structures, from the image data, which may then be studied and interpreted further.
Restoration
icon_mobile_dropdown
Resolution enhancement in dual-energy x-ray imaging
Pierre Gravel, Philippe Despres, Gilles Beaudoin, et al.
We have developed a method that uses a large amount of a priori information to generate super resolution radiographs. We measured and modeled analytically the point spread function of a low-dose gas microstrip x-ray detector at several beam energies. We measured the relationship between the local image intensity and the noise variance in the radiographs. The soft-tissue signal in the images was modeled using a minimum-curvature filtering technique. These results were then combined into an image deconvolution procedure using wavelet filtering to reduce restoration noise while keeping the enhanced small-scale features. The method was applied to a resolution grid image to measure its effects on the detector’s modulation transfer function. The restored images of a radiological human-torso phantom revealed small-scale details on the bones that were not seen before, and this, with improved SNR and image contrast. Dual-energy imaging was integrated to the restoration process in order to generate separate high-resolution images of the bones, the soft tissues, and the mean atomic number. This information could be used to detect bone micro-fractures in athletes and to assess bone demineralization in seniors due to osteoporosis. Super resolution radiographs are easier to segment due to their enhanced contrasts and uniform backgrounds; the boundaries of the features of interest can be delimited with a sub-pixel accuracy. This is highly relevant to the morphometric analysis of complex bone structures like individual vertebrae. The restoration method can be automated for a clinical environment use.
Partial volume correction using reverse diffusion
Olivier Salvado, Claudia Hillenbrand, David L. Wilson
Many medical images suffer from the partial volume effect where a boundary between two structures of interest falls in the middle of a voxel giving a signal value that is a mixture of the two. We propose a method to restore the ideal boundary by splitting a voxel into sub-voxels and reapportioning the signal into the sub-voxels. We designed this method to correct MRI 2D slice images where partial volume can be a considerable limitation. Each voxel is divided into four (or more) sub-voxels by nearest neighbor interpolation. The gray level of each sub-voxel is considered as “materials” able to move between sub-voxels but not between voxels. A partial differential equation is written to allow the material to flow towards the highest gradient direction, creating a “reverse” diffusion process. Flow is subject to constraints that tend to create step edges. Material is conserved in the process thereby conserving MR signal. The method proceeds until the flow decreases to a low value. To test the method, synthetic images were down-sampled to simulate the partial volume artifact and restored. Corrected images were remarkably closer both visually and quantitatively to the original images than those obtained from common interpolation methods: on simulated data mean square errors were 0.35, 1.09, and 1.24 for the proposed method, bicubic, and bilinear interpolation respectively. The method was relatively insensitive to noise. On MRI physical phantom and brain images, restored images processed with the new method were visually much closer to high-resolution counter-parts than those obtained with common interpolation methods.
Statistically based spatially adaptive noise reduction of planar nuclear studies
A. Hans Vija, Timothy R. Gosnell, Amos Yahil, et al.
The data-driven Pixon noise-reduction method is applied to nuclear studies. By using the local information content, it preserves all statistically justifiable image features without generating artifacts. Statistical measures provide the user a feedback to judge if the processing parameters are optimal. The present work focuses on planar nuclear images with known Poisson noise characteristics. Its ultimate goals are to: (a) increase sensitivity for detection of lesions of small size and/or of small activity-to-background ratio, (b) reduce data acquisition time, and (c) reduce patient dose. Data are acquired using Data Spectrum’s cylinder phantom in two configurations: (a) with hot and cold rod inserts at varying total counts and (b) with hot sphere inserts at varying activity-to-background ratios. We show that the method adapts automatically to both hot and cold lesions, concentration ratios, and different noise levels and structure dimensions. In clinical applications, slight adjustment of the parameters may be needed to adapt to the specific clinical protocols and physician preference. Visually, the processed images are comparable to raw images with ~16 times as many counts, and quantitatively the reduced noise equals that obtained with ~50 times as many counts. We also show that the Pixon method allows for identification of spheres at low concentration ratios, where raw planar imaging fails and matched filtering underperforms. Conclusion: The Pixon method significantly improves the image quality of data at either reduced count levels, or low target-to-background ratios. An analysis of clinical studies is now warranted to assess the clinical impact of the method.
MRI denoising via phase error estimation
If the phase error at each pixel in a complex-valued MRI image is known the noise in the image can be reduced resulting in improved detection of medically significant details. However, given a complex-valued MRI image, estimating the phase error at each pixel is a difficult problem. Several approaches have previously been suggested including non-linear least squares fitting and smoothing filters. We propose a new scheme based on iteratively applying a series of non-linear filters, each used to modify the estimate into greater agreement with one piece of knowledge about the problem, until the output converges to a stable estimate. We compare our results with other phase estimation and MRI denoising schemes using synthetic data.
Statistical Methods
icon_mobile_dropdown
Estimating intensity variance due to noise in registered images
Gustavo Kunde Rohde, Alan S. Barnett, Peter J. Basser, et al.
Image registration refers to the process of finding the spatial correspondence between two or more images. This is usually done by applying a spatial transformation, computed automatic or manually, to a given image using a continuous image model computed either with interpolation or approximation methods. We show that noise induced signal variance in interpolated images differs significantly from the signal variance of the original images in native space. We describe a simple approach to compute the signal variance in registered images based on the signal variance and covariance of the original images, the spatial transformations computed by the registration procedure, and the interpolation or approximation kernel chosen. Our approach is applied to diffusion tensor (DT) MRI data. We show that incorrect noise variance estimates in registered diffusion weighted images can affect the estimated DT parameters, their estimated uncertainty, as well as indices of goodness of fit such as chi-square maps. In addition to DT-MRI, we believe that this methodology would be useful any time parameter extraction methods are applied to registered or interpolated data.
The use and benefit of stereology in choosing a CT scanning protocol for the lung
When a patient is examined at different times using different protocols, how can we know whether the observed differences in the area or volume estimate are due to the patient, the protocol, or both? Specifically, we ask what is the smallest difference in lung volume that can be computed reliably when two sets of CT data are acquired by varying the number and thickness of the slices, but while holding constant the in-plane resolution. The accuracy and precision of the total lung volume estimates are calculated based on the principles of stereology using uniform design sampling. Comparisons of the lung volume estimate based on fewer slices using stereological principles are employed. A formal test made of the hypothesis that the use of fewer slices can yield satisfactory precision of the lung estimate. It is known that estimation of lung volume based on CT images is sensitive to the acquisition parameters used during scanning: dose, scan time, number of cross-sectional slices, and slice collimation. Those parameters are very different depending on the lung examination required: routine studies or high-resolution detailed studies. Thus, if different protocols are to be used confidently for volume estimation, it is important to understand the factors that influence volume estimate accuracy and to provide the associated confidence intervals for the measurements.
Proposal and validation of a method to construct confidence intervals for clinical outcomes around FROC curves for mammography CAD systems
Hans Bornefalk
This paper introduces a method for constructing confidence intervals for possible clinical outcomes around the FROC curve of a mammography CAD system. Given the architecture of a CAD classifying machine, there is one and only one system threshold that will yield a desired sensitivity on a certain population. The limited training sample size leads to a sampling error and an uncertainty in determining the optimal system threshold. This leads to an uncertainty in the operating point in the direction along the FROC curve which can be captured by a Bayesian approach where the distribution of possible thresholds is estimated. This uncertainty contributes to a large and spread-out confidence interval which is important to consider when one is intending to make comparisons between CAD algorithms trained on different data sets. The method is validated using a Monte Carlo method designed to capture the effect of correctly determining the system threshold.
Unsupervised spatio-temporal detection of brain functional activation based on hidden Markov multiple event sequence models
Sylvain Faisan, Laurent Thoraval, Jean-Paul Armspach, et al.
This paper presents a novel, completely unsupervised fMRI brain mapping approach that addresses the three problems of hemodynamic response function (HRF) shape variability, neural event timing, and fMRI response linearity. To make it robust, the method takes into account spatial and temporal information directly into the core of the activation detection process. In practice, activation detection is formulated in terms of temporal alignment between the sequence of hemodynamic response onsets (HROs) detected in the fMRI signal at υ and in the spatial neighbourhood of υ, and the sequence of "off-on" transitions observed in the input blocked stimulation paradigm (when considering epoch-related fMRI data), or the sequence of stimuli of the event-based paradigm (when considering event-related fMRI data). This multiple event sequence alignment problem, which comes under multisensor data fusion, is solved within the probabilistic framework of hidden Markov multiple event sequence models (HMMESMs), a special class of hidden Markov models. Results obtained on real and synthetic data compete with those obtained with the popular statistical parametric mapping (SPM) approach, but without necessitating any prior definition of the expected activation patterns, the HMMESM mapping approach being completely unsupervised.
Improved quantitation for PET/CT image reconstruction with system modeling and anatomical priors
Adam M. Alessio, Paul M. Kinahan, Thomas Lewellen
Accurate quantitation of PET tracer uptake levels in small tumors remains a challenge. This work uses an improved reconstruction algorithm to reduce the quantitative errors due to limited system resolution and due to necessary image noise reduction. We propose a method for finding and using the detection system response in the projection matrix of a statistical reconstruction algorithm. In addition we use aligned anatomical information, available in PET/CT scanners, to govern the penalty term applied during each image update. These improvements are combined with FORE rebinning in a clinically feasible algorithm for reconstructing fully 3D PET data. Simulated results show improved tumor bias and variance characteristics with the new algorithm.
A probabilistic model for predicting diameters of lung airways
Spencer Yuen, Matthew Brown, Sumit Shah, et al.
The accurate characterization of pulmonary airways on CT is potentially very useful for diagnosis and evaluation of lung diseases. The task is challenging due to their small size and variable orientation. We propose a probabilistic modeling technique and a set of measurement tools to quantitate airway morphology. We extract the airway tree structure from high resolution CT scans with a seeded region growing algorithm. Individual airway branches are identified by reducing the airway tree to a set of central axes. Properties such as lumen diameter and branch angle are measured from these central axes. The structure of the Bayesian model is inferred from a set of equations representing the parent-daughter relationships between branches, such as equations of air flow ratio and flow conservation. The CT measurements are used to instantiate the conditional probability tables of the Bayesian model. To evaluate the model, it was used to predict the airway diameter for the 2nd, 3rd, 4th, 5th, and 6th generations of the airway tree. We show that the model can reasonably predict the diameter of a particular airway branch, given information of its parent.
Shape and Scale
icon_mobile_dropdown
Fetal head detection and measurement in ultrasound images by a direct inverse randomized Hough transform
Wei Lu, Jinglu Tan, Randall C. Floyd
Object detection in ultrasound fetal images is a challenging task for the relatively low resolution and low signal-to-noise ratio. A direct inverse randomized Hough transform (DIRHT) is developed for filtering and detecting incomplete curves in images with strong noise. The DIRHT combines the advantages of both the inverse and the randomized Hough transforms. In the reverse image, curves are highlighted while a large number of unrelated pixels are removed, demonstrating a “curve-pass filtering” effect. Curves are detected by iteratively applying the DIRHT to the filtered image. The DIRHT was applied to head detection and measurement of the biparietal diameter (BPD) and head circumference (HC). No user input or geometric properties of the head were required for the detection. The detection and measurement took 2 seconds for each image on a PC. The inter-run variations and the differences between the automatic measurements and sonographers’ manual measurements were small compared with published inter-observer variations. The results demonstrated that the automatic measurements were consistent and accurate. This method provides a valuable tool for fetal examinations.
Shape regression for vertebra fracture quantification
Michael Tillge Lund, Marleen de Bruijne, Laszlo B. Tanko M.D., et al.
Accurate and reliable identification and quantification of vertebral fractures constitute a challenge both in clinical trials and in diagnosis of osteoporosis. Various efforts have been made to develop reliable, objective, and reproducible methods for assessing vertebral fractures, but at present there is no consensus concerning a universally accepted diagnostic definition of vertebral fractures. In this project we want to investigate whether or not it is possible to accurately reconstruct the shape of a normal vertebra, using a neighbouring vertebra as prior information. The reconstructed shape can then be used to develop a novel vertebra fracture measure, by comparing the segmented vertebra shape with its reconstructed normal shape. The vertebrae in lateral x-rays of the lumbar spine were manually annotated by a medical expert. With this dataset we built a shape model, with equidistant point distribution between the four corner points. Based on the shape model, a multiple linear regression model of a normal vertebra shape was developed for each dataset using leave-one-out cross-validation. The reconstructed shape was calculated for each dataset using these regression models. The average prediction error for the annotated shape was on average 3%.
Generalized scale-based image filtering
In medical imaging, low signal-to-noise ratio (SNR) and/or contrast-to-noise ratio (CNR) often cause many image processing algorithms to perform poorly. Postacquisition image filtering is an important off-line image processing approach widely employed to enhance the SNR and CNR. A major drawback of many filtering techniques is image degradation by diffusing/blurring edges and/or fine structures. In this paper, we introduce a scale-based filtering method that employs scale-dependent diffusion conductance to perform filtering. This approach utilizes novel object scale information via a concept called generalized ball scale, which imposes no shape, size, or anisotropic constraints unlike previously published ball scale-based filtering strategies. The object scale allows us to better control the filtering process by constraining smoothing in regions with fine details and in the vicinity of boundaries while permitting effective smoothing in the interior of homogeneous regions. Quantitative evaluations based on the Brainweb data sets show superior performance of generalized ball scale-based diffusive filtering over two existing methods, namely, ball scale-based and nonlinear complex diffusion process. Qualitative experiments based on both phantom and patient magnetic resonance images demonstrate that the generalized ball scale-based approach leads to better preserved fine details and edges.
Characterization of cerebral aneurysms using 3D moment invariants
Raul Daniel Millan, Monica Hernandez, Daniel Gallardo, et al.
The rupture mechanism of intracranial aneurysms is still not fully understood. Although the size of the aneurysm is the shape index most commonly used to predict rupture, some controversy still exists about its adequateness as an aneurysm rupture predictor. In this work, an automatic method to geometrically characterize the shape of cerebral saccular aneurysms using 3D moment invariants is proposed. Geometric moments are efficiently computed via application of the Divergence Theorem over the aneurysm surface using a non-structured mesh. 3D models of the aneurysm and its connected parent vessels have been reconstructed from segmentations of both 3DRA and CTA images. Two alternative approaches have been used for segmentation, the first one based on isosurface deformable models, and the second one based on the level set method. Several experiments were also conducted to both assess the influence of pre-processing steps in the stability of the aneurysm shape descriptors, and to know the robustness of the proposed method. Moment invariants have proved to be a robust technique while providing a reliable way to discriminate between ruptured and unruptured aneurysms (Sensitivity=0.83, Specificity=0.74) on a data set containing 55 aneurysms. Further investigation over larger databases is necessary to establish their adequateness as reliable predictors of rupture risk.
3D pulmonary airway color image reconstruction via shape from shading and virtual bronchoscopy imaging techniques
The dependence on macro-optical imaging of the human body in the assessment of possible disease is rapidly increasing concurrent with, and as a direct result of, advancements made in medical imaging technologies. Assessing the pulmonary airways through bronchoscopy is performed extensively in clinical practice however remains highly subjective due to limited visualization techniques and the lack of quantitative analyses. The representation of 3D structures in 2D visualization modes, although providing an insight to the structural content of the scene, may in fact skew the perception of the structural form. We have developed two methods for visualizing the optically derived airway mucosal features whilst preserving the structural scene integrity. Shape from shading (SFS) techniques can be used to extract 3D structural information from 2D optical images. The SFS technique presented addresses many limitations previously encountered in conventional techniques resulting in high-resolution 3D color images. The second method presented to combine both color and structural information relies on combined CT and bronchoscopy imaging modalities. External imaging techniques such as CT provide a means of determining the gross structural anatomy of the pulmonary airways, however lack the important optically derived mucosal color. Virtual bronchoscopy is used to provide a direct link between the CT derived structural anatomy and the macro-optically derived mucosal color. Through utilization of a virtual and true bronchoscopy matching technique we are able to directly extract combined structurally sound 3D color segments of the pulmonary airways. Various pulmonary airway diseases are assessed and the resulting combined color and texture results are presented demonstrating the effectiveness of the presented techniques.
Poster Session I
icon_mobile_dropdown
Computer-aided diagnosis algorithm for lung cancer using retrospective CT images
This paper presents a method for detecting suspicious nodules based on successive low-dose helical CT images. The method uses both initial and follow-up images to improve nodule detection performance. The basic idea of the detection is to register nodule images measured at different time and to assess the changes in size, shape, and density of the nodule. Since there are several variations of nodule changes, such as stable, shrinking, expansion in size, disappearance, appearance, and separation, a coarse-to-fine registration technique was adopted to deal with large nodule deformation. Especially, the fine registration is performed by excluding nodule regions and using nodule surroundings to avoid effects of nodule deformations in alignment task. In a preliminary experiment, the method was applied to ten cases with successive scans. From visual inspection, the corresponding results between initial and follow-up images were acceptable in clinical use. More researches using a large data set will be required. Still, we believe that the method has the potential of detecting suspicious nodules for use in a computer-aide diagnosis system.
Liver cancer detection by using transition features obtained from multi-phase CT images
Shigeto Watanabe, Yoshito Mekada, Junichi Hasegawa, et al.
This paper presents a method for automated detection of liver cancer regions based on transition of density at each point obtained from multi-phase X-ray CT images. For describing transition of density, two kinds of feature vectors named Density Transition (DT) and Density Change Transition (DCT) are introduced. DCT is used for extraction of cancer candidates and DT is used for suppression of false candidates. In the experiments using 14 real abdominal CT images with cancer, it was shown that the detection rate was 100% and the number of false-positives was 0.71 regions per case.
False-positive reduction using Hessian features in computer-aided detection of pulmonary nodules on thoracic CT images
We are developing a computer-aided detection (CAD) system for lung nodules in thoracic CT volumes. During false positive (FP) reduction, the image structures around the identified nodule candidates play an important role in differentiating nodules from vessels. In our previous work, we exploited shape and first-order derivative information of the images by extracting ellipsoid and gradient field features. The purpose of this study was to explore the object shape information using second-order derivatives and the Hessian matrix to further improve the performance of our detection system. Eight features related to the eigenvalues of the Hessian matrix were extracted from a volume of interest containing the object, and were combined with ellipsoid and gradient field features to discriminate nodules from FPs. A data set of 82 CT scans from 56 patients was used to evaluate the usefulness of the FP reduction technique. The classification accuracy was assessed using the area Az under the receiving operating characteristic curve and the number of FPs per section at 80% sensitivity. In the combined feature space, we obtained a test Az of 0.97 ± 0.01, and 0.27 FPs/section at 80% sensitivity. Our results indicate that combining the Hessian, ellipsoid and gradient field features can significantly improve the performance of our FP reduction stage.
A novel quantification method for determining previously undetected silent infarcts on MR-perfusion in patients following carotid endarterectomy
The purpose of this paper is to evaluate the post-operative Magnetic Resonance Perfusion (MRP) scans of patients undergoing carotid endarterectomy (CEA), using a novel image-analysis algorithm, to determine if post-operative neurocognitive decline is associated with cerebral blood flow changes. CEA procedure reduces the risk of stroke in appropriately selected patients with significant carotid artery stenosis. However, 25% of patients experience subtle cognitive deficits after CEA compared to a control group. It was hypothesized that abnormalities in cerebral blood flow (CBF) are responsible for these cognitive deficits. A novel algorithm for analyzing MR-perfusion (MRP) scans to identify and quantify the amount of CBF asymmetry in each hemisphere was developed and to quantify the degree of relative difference between three corresponding vascular regions in the ipsilateral and contralateral hemispheres, the Relative Difference Map (RDM). Patients undergoing CEA and spine surgery (controls) were examined preoperatively, and one day postoperatively with a battery of neuropsychometric (NPM) tests, and labeled “injured” patients with significant cognitive deficits, and “normal” if they demonstrated no decline in neurocognitive function. There are apparently significant RDM differences with MRP scans between the two hemispheres in patients with cognitive deficits which can be used to guide expert reviews of the imagery. The proposed methodology aids in the analysis of MRP parameters in patients with cognitive impairment.
Computer-aided detection of microcalcification clusters on full-field digital mammograms: multiscale pyramid enhancement and false positive reduction using an artificial neural network
We are developing a computer-aided detection (CAD) system to detect microcalcification clusters automatically on full field digital mammograms (FFDMs). The CAD system includes five stages: preprocessing, image enhancement and/or box-rim filtering, segmentation of microcalcification candidates, false positive (FP) reduction, and clustering. In this study, we investigated the performance of a nonlinear multiscale Laplacian pyramid enhancement method in comparison with a box-rim filter at the image enhancement stage and the use of a new error metric to improve the efficiency and robustness of the training of a convolution neural network (CNN) at the FP reduction stage of our CAD system. A data set of 96 cases with 200 images was collected at the University of Michigan. This data set contained 215 microcalcification clusters, of which 64 clusters were proven by biopsy to be malignant and 151 were proven to be benign. The data set was separated into two independent data sets. One data set was used to train and validate the CNN in our CAD system. The other data set was used to evaluate the detection performance. For this data set, Laplacian pyramid multiscale enhancement did not improve the performance of the microcalcification detection system in comparison with our box-rim filter previously optimized for digitized screen-film mammograms. With the new error metric, the training of CNN could be accelerated and the classification performance in validation was improved from an Az value of 0.94 to 0.97 on average. The CNN in combination with rule-based classifiers could reduce FPs with a small tradeoff in sensitivity. By using the free-response receiver operating characteristic (FROC) methodology, it was found that our CAD system can achieve a cluster-based sensitivity of 70%, 80%, and 88% at 0.23, 0.39, and 0.71 FP marks/image, respectively. For case-based performance evaluation, a sensitivity of 80%, 90%, and 98% can be achieved at 0.17, 0.27, and 0.51 FP marks/image, respectively.
Computer-aided system for measuring the mandibular cortical width on panoramic radiographs in osteoporosis diagnosis
Agus Zainal Arifin, Akira Asano, Akira Taguchi D.D.S., et al.
Osteoporotic fractures are associated with substantial morbidity, increased medical cost and high mortality risk. Several equipments of bone assessment have been developed to identify individuals, especially postmenopausal women, with high risk of osteoporotic fracture; however, a large segment of women with low skeletal bone mineral density (BMD), namely women with high risk of osteoporotic fractures, cannot be identified sufficiently because osteoporosis is asymptomatic. Recent studies have been demonstrating that mandibular inferior cortical width manually measured on panoramic radiographs may be useful for the identification of women with low BMD. Automatic measurement of cortical width may enable us to identify a large number of asymptomatic women with low BMD. The purpose of this study was to develop a computer-aided system for measuring the mandibular cortical width on panoramic radiographs. Initially, oral radiologists determined the region of interest based on the position of mental foramen. Some enhancing image techniques were applied so as to measure the cortical width at the best point. Panoramic radiographs of 100 women who had BMD assessments of the lumbar spine and femoral neck were used to confirm the efficacy of our new system. Cortical width measured with our system was compared with skeletal BMD. There were significant correlation between cortical width measured with our system and skeletal BMD. These correlations were similar with those between cortical width manually measured by the dentist and skeletal BMD. Our results suggest that our new system may be useful for mass screening of osteoporosis.
A classification method of liver tumors based on temporal change of Hounsfield unit in CT images
Masaki Ishiguro, Ichiro Murase, Noriyuki Moriyama, et al.
We present an automatic diagnosis method of liver cancer by using sequential images with contrast material of dynamic CT. Our method identifies and classifies liver tumors by extracting temporal change of CT values [Hounsfield Unit(HU)] of tumors from four kinds of CT images (i.e. plain CT, early phase, portal phase, late phase of dynamic CT images) in addition to morphological features of tumors. Automatic diagnosis of liver tumors is very difficult, because contrast of liver tumors is very small compared with liver background, shapes of tumors are diverse, and extraction of temporal change of CT values is very difficult due to morphological and contrast complexity of temporal change of tumor segments. Our method extracts temporal change of CT values of objects by mapping segments of same objects in different CT phase based on overlap ratio and position adjustment. We also implemented a graphical user interface for searching such images from an image database that include tumors similar to an image given as a search condition with respect to features of morphorogical and temporal change of contrast.
A relevance vector machine technique for the automatic detection of clustered microcalcifications
Microcalcification (MC) clusters in mammograms can be important early signs of breast cancer in women. Accurate detection of MC clusters is an important but challenging problem. In this paper, we propose the use of a recently developed machine learning technique -- relevance vector machine (RVM) -- for automatic detection of MCs in digitized mammograms. RVM is based on Bayesian estimation theory, and as a feature it can yield a decision function that depends on only a very small number of so-called relevance vectors. We formulate MC detection as a supervised-learning problem, and use RVM to classify if an MC object is present or not at each location in a mammogram image. MC clusters are then identified by grouping the detected MC objects. The proposed method is tested using a database of 141 clinical mammograms, and compared with a support vector machine (SVM) classifier which we developed previously. The detection performance is evaluated using the free-response receiver operating characteristic (FROC) curves. It is demonstrated that the RVM classifier matches closely with the SVM classifier in detection performance, and does so with a much sparser kernel representation than the SVM classifier. Consequently, the RVM classifier greatly reduces the computational complexity, making it more suitable for real-time processing of MC clusters in mammograms.
Tree-structured grading of pathological images of prostate
Reza Farjam, Hamid Soltanian-Zadeh, Reza Aghaizadeh Zoroofi, et al.
This paper presents a new algorithm for Gleason grading of pathological images of prostate. Structural features of the glands are extracted and used in a tree-structured (TS) algorithm to classify the images into five Gleason grades of 1 to 5. In this algorithm the image is first segmented to locate the glandular regions using texture features and a K-means clustering algorithm. The glands are then labeled from the glandular regions. In each stage of the proposed TS algorithm, shape and intensity-based features of the glands are extracted and used in a linear classifier to classify the image into two groups. Despite some proposed methods in the literature which use only texture features, this technique uses the features like roundness and shape distribution, which are related to the structure of the glands in each grade and are independent of the magnification. The proposed method is therefore robust to illumination and magnification variations. To evaluate the performance of the proposed method, we use two datasets. Data set 1 contains 91 images with similar magnifications and illuminations. Data set 2 contains 199 images with different magnifications and illuminations. Using leave-one-out technique, we achieve 95% and 85% accuracy for dataset 1 and 2, respectively.
Computerized pectoral muscle identification on MLO-view mammograms for CAD applications
Chuan Zhou, Lubomir M. Hadjiiski, Chintana Paramagul, et al.
Automatic identification of the pectoral muscle on MLO view is an essential step for computerized analysis of mammograms. It can reduce the bias of mammographic density estimation, will enable region-specific processing in lesion detection programs, and also may be used as a reference in image registration algorithms. We are developing a computerized method for the identification of pectoral muscle on mammograms. The upper portion of the pectoral edges was first detected to estimate the direction of the pectoral muscle boundary. A gradient-based directional (GD) filter was used to enhance the linear texture structures, and then a gradient-based texture analysis was designed to extract a texture orientation image that represented the dominant texture orientation at each pixel. The texture orientation image was enhanced by a second GD filter. An edge flow propagation method was developed to extract edges around the pectoral boundary using geometric features and anatomic constraints. The pectoral boundary was finally generated by a second-order curve fitting. 118 MLO view mammograms were used in this study. The pectoral muscle boundary identified on each image by an experienced radiologist was used as the gold standard. The accuracy of pectoral boundary detection was evaluated by two performance metrics. One is the overlap percentage between the computer-identified area and the gold standard, and the other is the root-mean-square (RMS) distance between the computer and manually identified pectoral boundary. For 118 MLO view mammograms, 99.2% (117/118) of the pectoral muscles could be identified. The average of the overlap percentage is 94.8% with a standard deviation of 20.9%, and the average of the RMS distance is 4.3 mm with a standard deviation of 5.9 mm. These results indicate that the pectoral muscle on mammograms can be detected accurately by our automated method.
Issues in assessing multi-institutional performance of BI-RADS-based CAD systems
The purpose of this study was to investigate factors that impact the generalization of breast cancer computer-aided diagnosis (CAD) systems that utilize the Breast Imaging Reporting and Data System (BI-RADS). Data sets from four institutions were analyzed: Duke University Medical Center, University of Pennsylvania Medical Center, Massachusetts General Hospital, and Wake Forest University. The latter two data sets are subsets of the Digital Database for Screening Mammography. Each data set consisted of descriptions of mammographic lesions according to the BI-RADS lexicon, patient age, and pathology status (benign/malignant). Models were developed to predict pathology status from the BI-RADS descriptors and the patient age. Comparisons between the models built on data from the different institutions were made in terms of empirical (non-parametric) receiver operating characteristic (ROC) curves. Results suggest that BI-RADS-based CAD systems focused on specific classes of lesions may be more generally applicable than models that cover several lesion types. However, better generalization was seen in terms of the area under the ROC curve than in the partial area index (>90% sensitivity). Previous studies have illustrated the challenges in translating a BI-RADS-based CAD system from one institution to another. This study provides new insights into possible approaches to improve the generalization of BI-RADS-based CAD systems.
Automated segmentation of mesothelioma volume on CT scan
Binsheng Zhao, Lawrence Schwartz, Raja Flores, et al.
In mesothelioma, response is usually assessed by computed tomography (CT). In current clinical practice the Response Evaluation Criteria in Solid Tumors (RECIST) or WHO, i.e., the uni-dimensional or the bi-dimensional measurements, is applied to the assessment of therapy response. However, the shape of the mesothelioma volume is very irregular and its longest dimension is almost never in the axial plane. Furthermore, the sections and the sites where radiologists measure the tumor are rather subjective, resulting in poor reproducibility of tumor size measurements. We are developing an objective three-dimensional (3D) computer algorithm to automatically identify and quantify tumor volumes that are associated with malignant pleural mesothelioma to assess therapy response. The algorithm first extracts the lung pleural surface from the volumetric CT images by interpolating the chest ribs over a number of adjacent slices and then forming a volume that includes the thorax. This volume allows a separation of mesothelioma from the chest wall. Subsequently, the structures inside the extracted pleural lung surface, including the mediastinal area, lung parenchyma, and pleural mesothelioma, can be identified using a multiple thresholding technique and morphological operations. Preliminary results have shown the potential of utilizing this algorithm to automatically detect and quantify tumor volumes on CT scans and thus to assess therapy response for malignant pleural mesothelioma.
A computerized approach for estimating pulmonary nodule growth rates in three-dimensional thoracic CT images based on CT density histogram
Yoshiki Kawata, Noboru Niki, Hironobu Ohmatsu, et al.
In research and development of computer-aided differential diagnosis, there is now a widespread interest in the use of nodule doubling time for measuring the volumetric changes of pulmonary nodule. To assess nodule status requires not only the measurement of volume changes but also one of nodule density variations. This paper proposes a computerized approach to measure nodule density variation inside small pulmonary nodule using CT images. The approach consists of five steps: (1) nodule segmentation, (2) computation of CT density histogram, (3) nodule classification based on CT density histogram and size, (4) computation of doubling time based on CT density histogram, and (5) classification between benign and malignant. Our approach was applied to follow-up scans of lung nodules. The preliminary experimental result demonstrated that our approach has a highly potential usefulness to assess the nodule evolution using high-resolution CT images.
The effects of 3D region-based compression on the performance of an automatic lung nodule detection system
Benjamin L Odry, Karthik Krishnan, Mariappan Nadar, et al.
There is growing interest in computer aided diagnosis applications including automatic detection of lung nodules from multislice computed tomography (CT). However the increase in the number and size of CT datasets introduces high costs for data storage and transmission, and becomes an obstacle to routine clinical exam as well as hindering widespread utilization of computerized applications. We investigated the effects of 3D lossy region-based JPEG2000 standard compression on the results of an automatic lung nodule detection system. As the algorithm detects the lungs within the datasets, we used this lung segmentation to define a region of interest (ROI) where the compression should be of higher fidelity. We tested 4 methods of 3D compression: 1) default compression of the whole image, 2) default compression of segmented lungs with masking out all non-lung regions, 3) ROI-based compression as specified in the JPEG2000 standard and 4) compression where voxels in the ROI are weighted to be given emphasis in the encoding. We tested 7 compression ratios per method: 1, 4, 6, 8, 10, 20, and 30 to 1. We then evaluated our experimental CAD algorithm on 10 patients with 67 documented nodules initially identified on the decompressed data. Sensitivities and false positive rates were compared for the various compression methods and ratios. We found that region-based compression generally performs better than default compression. The sensitivity with default compression decreased from 85% at no compression to 61% at 30:1 compression, a decrease of 25%, whereas the masked compression method saw a decreased in sensitivity on only 13.5% at maximum compression. At compression levels up to 10:1, all 3 region-based compression methods had decreases in sensitivity of 7.5% or less. Detection of small nodules (< 4mm in diameter) was more affected by compression than detection of large nodules; sensitivity to calcified nodules was less affected by compression than to non-calcified nodules.
Applying knowledge engineering and representation methods to improve support vector machine and multivariate probabilistic neural network CAD performance
Walker H. Land Jr., Frances Anderson, Tom Smith, et al.
Achieving consistent and correct database cases is crucial to the correct evaluation of any computer-assisted diagnostic (CAD) paradigm. This paper describes the application of artificial intelligence (AI), knowledge engineering (KE) and knowledge representation (KR) to a data set of ≈2500 cases from six separate hospitals, with the objective of removing/reducing inconsistent outlier data. Several support vector machine (SVM) kernels were used to measure diagnostic performance of the original and a “cleaned” data set. Specifically, KE and ER principles were applied to the two data sets which were re-examined with respect to the environment and agents. One data set was found to contain 25 non-characterizable sets. The other data set contained 180 non-characterizable sets. CAD system performance was measured with both the original and “cleaned” data sets using two SVM kernels as well as a multivariate probabilistic neural network (PNN). Results demonstrated: (i) a 10% average improvement in overall Az and (ii) approximately a 50% average improvement in partial Az.
Detection of circumscribed masses in mammograms using morphological segmentation
We present a method for detecting circumscribed masses in digital mammograms. Morphological hierarchical watersheds are used in the segmentation process. Oversegmentation is prevented by employing a reconstructive open/close alternating sequential filter to simplify the image. The advantage of this method of simplification is that the object shapes and edges are preserved. The regional maxima of the simplified input image are then extracted and used as internal markers for the hierarchical watershed transform, which is applied to the gradient image of the simplified input image. An image-based classification technique is applied to reduce the number of false positives. The method is applied to 18 mammograms from the MIAS database, containing 20 circumscribed masses in background tissue of varying density. We obtain a high true detection rate using combined with a low number of false positives per image.
Computer-aided diagnosis for detection of cardiomegaly in digital chest radiographs
Takayuki Ishida, Shigehiko Katsuragawa, Koichi Chida, et al.
The cardio-thoracic ratio (CTR) is commonly measured manually for the evaluation of cardiomegaly. To determine the CTR automatically, we have developed a computerized scheme based on gray-level histogram analysis and an edge detection technique with feature analysis. The database used in this study consisted of 392 chest radiographs, which included 304 normals and 88 abnormals with cardiomegaly. The pixel size and the quantization level of the image were 0.175 mm and 1024, respectively. We performed a nonlinear density correction to maintain consistency in the density and contrast of the image. Initial heart edge detection was performed by selection of a certain range of pixel values in the histogram of a rectangular area at the center of a low-resolution image. Feature analysis with use of an edge gradient and with the orientation obtained by a Sobel operator was applied for accurate identification of the heart edges, which tend to have large edge gradients in a certain range of orientations. In addition, to determine the CTR, we detected the ribcage edges automatically by using image profile analysis. In 94.9% of all of the cases, the heart edges were detected accurately by use of this scheme. The area under the ROC curve (Az value) in distinguishing between normals and abnormals with cardiomegaly based on the CTR was 0.912. Because the CTR is measured automatically and quickly (in less than 1 sec.), radiologists could save reading time. The computerized scheme will be useful for the assessment of cardiomegaly on chest radiographs.
Local noise reduction for emphysema scoring in low-dose CT images
Arnold Schilham, Mathias Prokop, Hester Gietema, et al.
Computed Tomography (CT) has become the new reference standard for quantification of emphysema. The most popular measure for emphysema derived from CT is the Pixel Index (PI), which expresses the fraction of the lung volume with abnormally low intensity values. As PI is calculated from a single, fixed threshold on intensity, this measure is strongly influenced by noise. This effect shows up clearly when comparing the PI score for a high-dose scan to the PI score for a low-dose (i.e. noisy) scan of the same subject. This paper presents a class of noise filters that make use of a local noise estimate to specify the filtering strength: Local Noise Variance Weighted Averaging (LNVWA). The performance of the filter is assessed by comparing high-dose and low-dose PI scores for 11 subjects. LNVWA improves the reproducibility of high-dose PI scores: For an emphysema threshold of -910 HU, the root-mean-square difference in PI score drops from 10% of the lung volume to 3.3% of the lung volume if LNVWA is used.
Detection of architectural distortion in mammograms using fractal analysis
Several studies have demonstrated the fractal properties of screening mammograms. The purpose of this study was to investigate fractal texture analysis for the automated detection of architectural distortion (AD) in screening mammograms. The study was based on the Digital Database for Screening Mammography (DDSM). Initially, a database of 708 mammographic regions with confirmed pathology was created. They were all 512x512 pixel regions of interest (ROIs). The ROI size was determined empirically. Fifty-two regions were extracted around biopsy-proven architectural distortion. The remaining 656 ROIs depicted normal breast parenchyma. Fractal analysis was performed on each ROI at multiple resolutions (512x512, 256x256, 128x128, and 64x64). The fractal dimension of each ROI was calculated using the circular average power spectrum technique. Overall, the average fractal dimension (FD) estimate of the normal ROIs was statistically significantly higher than the average FD of the ROIs with AD. This result was consistent across all resolutions. However, best detection performance was achieved when the fractal dimension was estimated on ROIs subsampled with a factor of 2 (ROC area index Az=0.89±0.02). Specifically, there was perfect performance in fatty breasts (Az=1.0), Az=0.95±0.02 in fibroglandular breasts, Az=0.84±0.05 in heterogeneous breasts, and Az=0.66±0.10 in dense breasts. Overall, the present study demonstrates that the presence of AD disrupts the normal parenchymal structure, thus resulting in a lower fractal dimension. Consequently, fractal texture analysis could play an important role in the development of computer-assisted detection tools tailored towards architectural distortion.
Computer-aided diagnosis of leukoencephalopathy in children treated for acute lymphoblastic leukemia
John O. Glass, Chin-Shang Li, Kathleen J. Helton, et al.
The purpose of this study was to use objective quantitative MR imaging methods to develop a computer-aided diagnosis tool to differentiate white matter (WM) hyperintensities as either leukoencephalopathy (LE) or normal maturational processes in children treated for acute lymphoblastic leukemia with intravenous high dose methotrexate. A combined imaging set consisting of T1, T2, PD, and FLAIR MR images and WM, gray matter, and cerebrospinal fluid a priori maps from a spatially normalized atlas were analyzed with a neural network segmentation based on a Kohonen Self-Organizing Map. Segmented regions were manually classified to identify the most hyperintense WM region and the normal appearing genu region. Signal intensity differences normalized to the genu within each examination were generated for two time points in 203 children. An unsupervised hierarchical clustering algorithm with the agglomeration method of McQuitty was used to divide data from the first examination into normal appearing or LE groups. A C-support vector machine (C-SVM) was then trained on the first examination data and used to classify the data from the second examination. The overall accuracy of the computer-aided detection tool was 83.5% (299/358) with sensitivity to normal WM of 86.9% (199/229) and specificity to LE of 77.5% (100/129) when compared to the readings of two expert observers. These results suggest that subtle therapy-induced leukoencephalopathy can be objectively and reproducibly detected in children treated for cancer using this computer-aided detection approach based on relative differences in quantitative signal intensity measures normalized within each examination.
Automated detection of mammographic masses: preliminary assessment of an information-theoretic CAD scheme for reduction of false positives
The purpose of this work was to evaluate an information-theoretic computer-aided detection (CAD) scheme for improving the specificity of mass detection in screening mammograms. The study was based on images from the Lumisys set of the Digital Database for Screening Mammography (DDSM). Initially, the craniocaudal views of 49 DDSM mammograms were analyzed using an automated detection algorithm developed to prescreen mammograms. The prescreening algorithm followed a morphological concentric layer analysis and resulted in 319 false positive detections at 92% sensitivity. These 319 suspicious yet normal regions were extracted for further analysis with our information-theoretic CAD scheme. Our scheme follows a knowledge-based decision strategy. The strategy relies on information theoretic principles for similarity assessment between a query case and a knowledge databank of cases with known ground truth. Receiver Operating Characteristic (ROC) analysis was performed to determine how well the CAD scheme can discriminate the false positive regions from 681 true masses. The overall ROC area index of the information-theoretic CAD system was 0.75±0.02. At 97%, 95%, and 90% sensitivity, the system eliminated safely 20%, 30%, and 42% of the previously identified false positives respectively. Thus, information-theoretic CAD analysis can yield a significant reduction in false-positive detections while maintaining reasonable sensitivity.
Automated bone fracture detection
Fractures of bone are a common affliction. In most developed countries the number of fractures associated with age-related bone loss is increasing rapidly. Each year many fractures are missed during x-ray diagnosis, resulting in ineffective patient management and expensive litigation. From both an orthopaedic and radiologic point of view, the fully automatic detection and classification of fractures in long-bones is an important but difficult problem. In this paper, a fully automated method of detecting fractures in the diaphysis of a long-bone is described. X-rays are very difficult to process automatically, so to extract the required information a non-linear anisotropic diffusion method, the Affine Morphological Scale Space, was implemented to smooth the image without losing information about the location of boundaries within the image. Next, an iterative peak detection algorithm is used to accurately locate the bone centreline and articular surfaces. A method based on orthogonal projections calculated from a modified Hough transform is used to automatically locate the long-bone diaphysis. At this point, our algorithm accurately localises the area of the fracture, and would allow further image registration if necessary. Finally, a gradient-based algorithm is used to detect fractures present in the region of interest. The magnitude and direction of the gradient are combined to produce a measure of the likelyhood of the presence of a fracture. A library of long-bone fracture images was created. Experimental tests performed on a series of x-ray images show that the method is capable of accurately segmenting the diaphysis from the epiphyses, and is also able to detect many mid-shaft fractures of long-bones.
CAD scheme for detection of intracranial aneurysms in MRA based on 3D analysis of vessel skeletons and enhanced aneurysms
Hidetaka Arimura, Qiang Li, Yukunori Korogi, et al.
We have developed a computer-aided diagnostic (CAD) scheme for detection of unruptured intracranial aneurysms in magnetic resonance angiography (MRA) based on findings of short branches in vessel skeletons, and a three-dimensional (3D) selective enhancement filter for dots (aneurysms). Fifty-three cases with 61 unruptured aneurysms and 62 non-aneurysm cases were tested in this study. The isotropic 3D MRA images with 400 x 400 x 128 voxels (a voxel size of 0.5 mm) were processed by use of the dot enhancement filter. The initial candidates were identified not only on the dot-enhanced images by use of a multiple gray-level thresholding technique, but also on the vessel skeletons by finding short branches on parent skeletons, which can indicate a high likelihood of small aneurysms. All candidates were classified into four categories of candidates according to effective diameter and local structure of the vessel skeleton. In each category, a number of false positives were removed by use of two rule-based schemes and by linear discriminant analysis on localized image features related to gray level and morphology. Our CAD scheme achieved a sensitivity of 97% with 5.0 false positives per patient by use of a leave-one-out-by-patient test method. This CAD system may be useful in assisting radiologists in the detection of small intracranial aneurysms as well as medium-size aneurysms in MRA.
Regrouping initial CAD mass detections to facilitate classification of suspicious regions in mammography
There is a lot of interest in developing computer-aided detection (CAD) techniques for mammography that use multiple view information. During the development of such techniques we have noticed that they are hampered by the phenomena that mass lesions are sometimes detected by multiple regions. This has encouraged us to develop a technique to regroup initial CAD detections to facilitate the final classification of suspicious regions. The regrouping technique searches for detections that belong to the same structure. Therefore, it takes into account the distance between the detections and the image structure along a path between the detections. When correspondence is found, the two detections are replaced by a new detection in between the initial detections. Our regrouping technique correctly regrouped the detections in 48 percent of the masses initially detected by multiple regions. Of the false positive detections two percent were combined, and the percentage of true positive - false positive combinations was one. Incorporation of the algorithm into our CAD scheme resulted in a slight increase in detection performance. In addition, in our multiple view scheme it also resulted in a decrease in the number of incorrectly linked regions in corresponding mammographic views.
CAD system for lung cancer screening using low dose thick-slice CT images
Our group developed the computer aided diagnosis (CAD) system for lung cancer in 1996, and has been used in clinical field since 1997. From this CAD system (conventional system), we discovered problem and we attempted to solve the problem by using our proposed algorithm. The proposed algorithm succeeded in the improvement of the following three problems of the conventional system. (1) Weak extraction algorithm of region of interest (ROI) with noise, (2) Poor knowledge of chest structure, and (3) diagnostic processing for nodule of limited size. In this paper, the algorithm that solves problem (2) and (3) is described. We evaluated the proposed algorithm, which was applied to the following four databases. (A) Lung cancer database, (B) detailed examination database, (C) a large-scale screening database by 10mm-thickness images reconstructed from single-slice CT scan, and (D) a large-scale screening database by 10mm-thickness images reconstructed from multi-slice CT scan. The proposed method obtained the following successful results: Lung cancer database 95.7% TP and detailed examination 94.8% TP. For the large-scale screening database, we evaluated each examination process from physicians’ reading to cancer decision. The extraction rate of proposed algorithm improved as the examinations proceed. Two false positive results were obtained. False positive 1 (6.8-9.2 shadows/case) needed for a detailed examination and the object of false positive 2 (2.6-4.0 shadows/case) was an abnormal shadow.
A probabilistic approach to segmentation and classification of neoplasia in uterine cervix images using color and geometric features
Yeshwanth Srinivasan, Dana Hernes, Bhakti Tulpule, et al.
Automated segmentation and classification of diagnostic markers in medical imagery are challenging tasks. Numerous algorithms for segmentation and classification based on statistical approaches of varying complexity are found in the literature. However, the design of an efficient and automated algorithm for precise classification of desired diagnostic markers is extremely image-specific. The National Library of Medicine (NLM), in collaboration with the National Cancer Institute (NCI), is creating an archive of 60,000 digitized color images of the uterine cervix. NLM is developing tools for the analysis and dissemination of these images over the Web for the study of visual features correlated with precancerous neoplasia and cancer. To enable indexing of images of the cervix, it is essential to develop algorithms for the segmentation of regions of interest, such as acetowhitened regions, and automatic identification and classification of regions exhibiting mosaicism and punctation. Success of such algorithms depends, primarily, on the selection of relevant features representing the region of interest. We present color and geometric features based statistical classification and segmentation algorithms yielding excellent identification of the regions of interest. The distinct classification of the mosaic regions from the non-mosaic ones has been obtained by clustering multiple geometric and color features of the segmented sections using various morphological and statistical approaches. Such automated classification methodologies will facilitate content-based image retrieval from the digital archive of uterine cervix and have the potential of developing an image based screening tool for cervical cancer.
Automatic detection of multi-level acetowhite regions in RGB color images of the uterine cervix
Holger Lange
Uterine cervical cancer is the second most common cancer among women worldwide. Colposcopy is a diagnostic method used to detect cancer precursors and cancer of the uterine cervix, whereby a physician (colposcopist) visually inspects the metaplastic epithelium on the cervix for certain distinctly abnormal morphologic features. A contrast agent, a 3-5% acetic acid solution, is used, causing abnormal and metaplastic epithelia to turn white. The colposcopist considers diagnostic features such as the acetowhite, blood vessel structure, and lesion margin to derive a clinical diagnosis. STI Medical Systems is developing a Computer-Aided-Diagnosis (CAD) system for colposcopy -- ColpoCAD, a complex image analysis system that at its core assesses the same visual features as used by colposcopists. The acetowhite feature has been identified as one of the most important individual predictors of lesion severity. Here, we present the details and preliminary results of a multi-level acetowhite region detection algorithm for RGB color images of the cervix, including the detection of the anatomic features: cervix, os and columnar region, which are used for the acetowhite region detection. The RGB images are assumed to be glare free, either obtained by cross-polarized image acquisition or glare removal pre-processing. The basic approach of the algorithm is to extract a feature image from the RGB image that provides a good acetowhite to cervix background ratio, to segment the feature image using novel pixel grouping and multi-stage region-growing algorithms that provide region segmentations with different levels of detail, to extract the acetowhite regions from the region segmentations using a novel region selection algorithm, and then finally to extract the multi-levels from the acetowhite regions using multiple thresholds. The performance of the algorithm is demonstrated using human subject data.
Compression of digital mammograms with region-of-interest coding evaluated on a CAD system
Kjersti Engan, Martin Ruoff Lillo, Thor Ole Gulsrud
Screening programs produce large amount of mammographic data, and good compression schemes would be beneficial for both storage and transmission purposes. In medical data it is crucial that diagnostic important information is preserved. In this work we have implemented two different region-of-interest (ROI) coding methods together with a Set Partitioning in Hierarchical Trees (SPIHT) scheme to be used for compression of mammograms. Region-of-interest coding allows a region of the image to be compressed with higher fidelity than the rest of the image. This is useful in medical data to be able to compress a region containing a possibly cancer area with very high fidelity, but still manage an overall good compression ratio. Both the ROI methods, the basic SPIHT method as well as JPEG compression standard, the latter two without possibility of ROI coding, are evaluated by studying the results from a Computer Aided Detection (CAD) system for microcalcifications tested on the original and the compressed mammograms. In addition a visual inspection is performed as well as Peak Signal-to-Noise-Ratio (PSNR) calculations. Mammograms from the MIAS database is used. We show that mammograms can be compressed to less than 0.5 (0.3) bpp without any visual degradation and without significantly influence on the performance of the CAD system.
Four-dimensional compression of fMRI using JPEG2000
Hariharan G. Lalgudi, Ali Bilgin, Michael W Marcellin, et al.
Many medical imaging techniques available today generate 4D data sets. One such technique is functional magnetic resonance imaging (fMRI) which aims to determine regions of the brain that are activated due to various cognitive and/or motor functions or sensory stimuli. These data sets often require substantial resources for storage and transmission and hence call for efficient compression algorithms. fMRI data can be seen as a time-series of 3D images of the brain. Many different strategies can be employed for compressing such data. One possibility is to treat each 2D slice independently. Alternatively, it is also possible to compress each 3D image independently. Such methods do not fully exploit the redundancy present in 4D data. In this work, methods using 4D wavelet transforms are proposed. They are compared to different 2D and 3D methods. The proposed schemes are based on JPEG2000, which is included in the DICOM standard as a transfer syntax. Methodologies to test the effects of lossy compression on the end result of fMRI analysis are introduced and used to compare different compression algorithms.
Seed image reconstruction using a template matching technique
One of the problems in fluoroscopy based 3D seed reconstruction for prostate brachytherapy is the clustering of seeds in the fluoroscopic images. A template matching based method is proposed in this study to reconstruct the orientations and locations of individual seed images in the cluster. An idealized projection image of implanted seeds was used as a template to reconstruct the cluster, and different optimization strategies were implemented to find the best orientation and location for individual seeds. The four search methods compared were: 1) Down hill simplex method; 2) Powell method; 3) Multi-resolution based method with exhaustive initial search; and 4) Multi-resolution based method without exhaustive initial search. These methods were applied to 10 test images. Five of the ten images had only 2-seed clusters and five had 3-seed clusters. The results demonstrate that the first two methods didn’t perform well, and that the results were dependent on the initial guesses used to start the optimization process. The third method successfully found the best configurations for all of the 10 images while the fourth method succeeded in 9 of the 10 cases. We conclude therefore that multi-resolution approaches are appropriate for the seed image reconstruction problem. Since the possible configurations of the template are pre-computed, an additional advantage is that less execution time was needed for the multi-resolution methods. When applied to the seed image reconstruction process, this method will potentially significantly improve the accuracy of the 3D reconstruction of implanted seeds from fluoroscopic images used in prostate brachytherapy.
Wavelet-based multiscale anisotropic diffusion for MR imaging
Junmei Zhong, Bernard Dardzinski, Scott Holland, et al.
Magnetic resonance (MR) images acquired with a high temporal resolution or high spatial resolution are usually with a penalty of low signal to noise ratio (SNR). It is necessary to remove the noise artifacts with important image features such as edges preserved. In this paper, we propose to use the improved wavelet-based multiscale anisotropic diffusion algorithm for MR imaging. Experimental results demonstrate that this denoising algorithm can significantly improve the SNR for MR images while preserving edges with good visual quality. The denoising results indicate that in MR imaging applications, we can almost doubly improve the temporal resolution or improve the spatial resolution while achieving high SNR and acceptable image quality.
Noise reduction in digital radiography by wavelet packet using nonlinear correlation threshold
Takahide Okamoto, Shigeru Furui, Hiroshi Ichiji, et al.
The influence of quantum mottle appears as degradation of graininess with reduction of the amount of incidence X-rays. The results of a Wiener spectrum study showed that graininess increased as the dose was reduced, and noise affected all frequencies. However, in clinical images, these effects are seen only in the high-frequency domain above 0.3 cycle/mm. Moreover, the effects of a grid are restricted to a parallel component or a perpendicular component based on its structure. And the influence appears in the decomposition wavelet image of H or V. From these result, the decomposed wavelet coefficients at complete binary tree are divided into seven frequency-coefficient bands. About the noise processing method, we tried to reduce noise by applying modification of Wavelet Transform Modules Maxima method proposed by Mallet, et al. Then we tried to the adaptive nonlinear threshold based on wiener spectrum study and power spectrum study. Based on the above considerations, evaluation was performed using clinical radiographs obtained at a standard dose and reduced dose with the noise reduction processing applied. The results showed that noise caused by quantum mottle and the grid can be reduced by this method without the need for threshold processing based on clinical experience.
An artifact-free structure-saving noise reduction using the correlation between two images for threshold determination in the wavelet domain
A new method of noise reduction based on shrinkage in the wavelet domain has been created for the application in projection radiography. The method is based on comparing two similar or quasi-identical images of the same object. Using an appropriate measure of similarity, these images are compared with each other in order to produce the weighting matrices. The weighting factors for the wavelet coefficients are chosen to be proportional to the elements of the weighting matrices. One image of the pair is then reconstructed from the weighted wavelet coefficients. The effect of this kind of de-noising is a suppression of those structures in the image which don’t correlate with the structures in the other image of the pair. Normally the suppressed structures are quantum or scatter noise, while the correlated structures which are not affected at all, are the real anatomical structures.
Coarse-to-fine markerless gait analysis based on PCA and Gauss-Laguerre decomposition
Michela Goffredo, Maurizio Schmid, Silvia Conforto, et al.
Human movement analysis is generally performed through the utilization of marker-based systems, which allow reconstructing, with high levels of accuracy, the trajectories of markers allocated on specific points of the human body. Marker based systems, however, show some drawbacks that can be overcome by the use of video systems applying markerless techniques. In this paper, a specifically designed computer vision technique for the detection and tracking of relevant body points is presented. It is based on the Gauss-Laguerre Decomposition, and a Principal Component Analysis Technique (PCA) is used to circumscribe the region of interest. Results obtained on both synthetic and experimental tests provide significant reduction of the computational costs, with no significant reduction of the tracking accuracy.
Improvement of ultrasound image based on wavelet transform: speckle reduction and edge enhancement
For 2-dimensional B-mode ultrasound images, we propose an image enhancement algorithm based on a multi-resolution approach. In the proposed algorithm, we perform the directional filtering and noise reducing procedures from the coarse to fine resolution images that are obtained from the wavelet-transformed data. For directional filtering, the structural feature at each pixel is examined through the eigen-analysis. Then, if the pixel belongs to the edge region, we perform two-step directional filtering, namely, directional smoothing along the tangential direction of the edge to improve its continuity, and directional sharpening along the normal direction to enhance the contrast. Meanwhile, speckle noise is alleviated by reducing the wavelet coefficients corresponding to the homogeneous region. The reducing rate of the wavelet coefficients is determined by considering the frequency characteristics of speckle. Thereby, the algorithm reduces speckle noise efficiently without affecting the edge sharpness and enhances edges regardless their size. Note that the proposed speckle reduction scheme is based on the structural information rather than the statistics of the magnitude of wavelet coefficients as in the existing methods. The proposed algorithm is compared to the algorithm based on nonlinear anisotropic diffusion filtering and the one based on the wavelet shrinkage scheme. The experimental results show that the proposed algorithm considerably improves the subjective image quality without providing any noticeable artifact.
Pattern Recognition and Neural Networks
icon_mobile_dropdown
Combining contour detection algorithms for the automatic extraction of the preparation line from a dental 3D measurement
Volker Ahlers, Paul Weigl, Hartmut Schachtzabel
Due to the increasing demand for high-quality ceramic crowns and bridges, the CAD/CAM-based production of dental restorations has been a subject of intensive research during the last fifteen years. A prerequisite for the efficient processing of the 3D measurement of prepared teeth with a minimal amount of user interaction is the automatic determination of the preparation line, which defines the sealing margin between the restoration and the prepared tooth. Current dental CAD/CAM systems mostly require the interactive definition of the preparation line by the user, at least by means of giving a number of start points. Previous approaches to the automatic extraction of the preparation line rely on single contour detection algorithms. In contrast, we use a combination of different contour detection algorithms to find several independent potential preparation lines from a height profile of the measured data. The different algorithms (gradient-based, contour-based, and region-based) show their strengths and weaknesses in different clinical situations. A classifier consisting of three stages (range check, decision tree, support vector machine), which is trained by human experts with real-world data, finally decides which is the correct preparation line. In a test with 101 clinical preparations, a success rate of 92.0% has been achieved. Thus the combination of different contour detection algorithms yields a reliable method for the automatic extraction of the preparation line, which enables the setup of a turn-key dental CAD/CAM process chain with a minimal amount of interactive screen work.
Poster Session I
icon_mobile_dropdown
Feature extraction in digital mammograms based on optimal and morphological filtering
The aim of this study is to provide feature images of digital mammograms in which regions corresponding to masses are enhanced. Subsequently, the feature images can be segmented and classified into two classes; masses and normal tissue. Our proposed feature extraction method is based on a local energy measure as texture feature. The local energy measure is extracted using a filter optimized with respect to the relative distance between the average feature values. In order to increase the sensitivity of the texture feature extraction scheme each mammogram is preprocessed using wavelet transformation, adaptive histogram equalization, and a morphology based enhancement technique. Initial experiments indicate that our scheme is able to provide useful feature images of digital mammograms. In order to quantify the system performance the feature images of 38 mammograms from the MIAS database -- 19 containing circumscribed masses, and 19 containing spiculated masses -- were segmented using simple gray level thresholding. For the circumscribed masses a true positive (TP) rate of 89% with a corresponding 2.3 false detections (false positives, FPs) per image was achieved. For the spiculated masses the performance was somewhat lower, yielding an overall TP rate of 84% with a corresponding 2.6 FPs per image.
Comparative assessment of retinal vasculature using topological and geometric measures
Denis Fan, Abhir Bhalerao, Roland Wilson
We present a quantitative method for the comparison of vascular topology and geometry measured from retinal fundus photographs. The measure compares the difference between distributions taken from a graph representation of the vasculature, which is derived by image segmentation. The measure uses the Kullback-Leibler distance between statistical measures on the reference and test segmentations which can be geometrical, like the distribution of vessel widths, or topological, like local connectivity or a combination of the two. The user is free to build any meaningful description and here we illustrate two local topology measures graphically. Using this assessment method, we also show that our model based segmentation method has better geometrical accuracy than a technique based on matched filtering. We have tested out the measures on a set of 20 images from the STARE project data.
Gene expression detection and expression visualization in in situ hybridized cross-sectional images of the mouse brain
Sayan Pathak, Ed Lein, Simon Smith, et al.
Understanding gene expression in the mouse brain should provide a better understanding of the underlying topology of the mammalian brain, thereby opening previously unexplored avenues in neuroscience and brain informatics. An important step in this direction is to develop robust algorithms to quantify gene expression in in-situ hybridization (ISH) data. ISH methodology involves the use of labeled nucleic acid probes that bind to specific mRNA transcripts in tissue sections. The bound probe is detected using colorimetric methods and the resulting stained tissue sections are imaged at high resolution. The goals of the present study are to first identify a staining method that produces maximum signal to noise ratio (SNR), and second to develop a method for gene expression detection for a wide range of ISH images spanning different intensity and expression patterns. A simple k-means based clustering method is used to separate foreground labeling from background and non-expressing tissues in a variety of images stained with different staining/counter staining techniques. We found NBT/BCIP with no counterstain produces the best signal to background separation. The foreground cluster detected using the k-means algorithm was further modeled using a normal distribution. A novel one sided Mahalanobis distance based metric with majority partial ordered voting method was then developed to generate a fuzzy segmentation of the gene expression in each ISH image. This algorithm is fully automatic and facilitates high throughput analysis of large amount of image data. Using this methodology a cluster of 10 PCs was able to process approximately 10% of the mouse genome (17 TBytes of JPEG2000 lossless compressed images) over a period of 2 weeks. The results may be visualized at the Allen Institute for Brain Science web site www.brain-map.org.
A novel decision-tree based classification of white blood cells
Automated medical image processing and analysis offer a powerful tool for medical diagnosis. In this work, a decision-tree based white blood cell (WBC) classification scheme for peripheral blood images is developed. Based on the sufficient analysis on the characteristics of white blood cells, 10 efficient features are extracted, including size, shape, intensity and color, and a classification scheme based on decision-tree is designed to classify 6 different types of normal white blood cells. Especially, an efficient approach to separate two types of neutrophil is presented. The presented scheme is tested on 59 WBCs coming from 3 sets of blood images, which are obtained under different dying and imaging conditions. Results show classification accuracy above 96%.
Character recognition and image manipulation for the clinical translation of CAD for breast ultrasound
To be clinically viable, computer-aided diagnosis (CAD) systems must be as automated and user-friendly as possible. CAD systems for breast ultrasound are still preliminary and are not adapted for use in a standard clinical environment. For example, computer detection and classification schemes need the pixel size of each image to operate correctly, and while the DICOM standard allows pixel size to be encoded in the image file, some equipment manufacturers neglect to utilize the encoding. As a result, the pixel size is calculated from user input. In order to increase clinical efficiency and reduce the likelihood of error due to incorrect image specifications, automating this input process is a highly desirable asset. We developed and applied a character recognition algorithm to the annotation region of each ultrasound image in our database. A set of numerical masks, which corresponded to the characters used in the annotation information, enabled the filtering of each image. Numerical masks yielding the maximum output from the comparison operation between image data and mask were output to obtain the annotation information. Each image was then automatically cropped to remove the annotation banner and leave only the image data. The cropped image matrix dimensions and character recognition output were used to determine the corresponding pixel size. The algorithm was tested on 1110 images with various pixel sizes. In every case, the value output by the algorithm corresponded exactly to the true value. Our recognition algorithm now allows for the clinical translation of our fully-automated breast ultrasound CAD system.
An automatic method for estimating noise-induced signal variance in magnitude-reconstructed magnetic resonance images
Signal intensity in magnetic resonance images (MRIs) is affected by random noise. Assessing noise-induced signal variance is important for controlling image quality. Knowledge of signal variance is required for correctly computing the chi-square value, a measure of goodness of fit, when fitting signal data to estimate quantitative parameters such as T1 and T2 relaxation times or diffusion tensor elements. Signal variance can be estimated from measurements of the noise variance in an object- and ghost-free region of the image background. However, identifying a large homogeneous region automatically is problematic. In this paper, a novel, fully automated approach for estimating the noise-induced signal variance in magnitude-reconstructed MRIs is proposed. This approach is based on the histogram analysis of the image signal intensity, explicitly by extracting the peak of the underlining Rayleigh distribution that would characterize the distribution of the background noise. The peak is extracted using a nonparametric univariate density estimation like the Parzen window density estimation; the corresponding peak position is shown here to be the expected signal variance in the object. The proposed method does not depend on prior foreground segmentation, and only one image with a small amount of background is required when the signal-to-noise ratio (SNR) is greater than three. This method is applicable to magnitude-reconstructed MRIs, though diffusion tensor (DT)-MRI is used here to demonstrate the approach.
New methods of MR image intensity standardization via generalized scale
Image intensity standardization is a post-acquisition processing operation designed for correcting acquisition-to-acquisition signal intensity variations (non-standardness) inherent in Magnetic Resonance (MR) images. While existing standardization methods based on histogram landmarks have been shown to produce a significant gain in the similarity of resulting image intensities, their weakness is that, in some instances the same histogram-based landmark may represent one tissue, while in other cases it may represent different tissues. This is often true for diseased or abnormal patient studies in which significant changes in the image intensity characteristics may occur. In an attempt to overcome this problem, in this paper, we present two new intensity standardization methods based on the concept of generalized scale. In reference 1 we introduced the concept of generalized scale (g-scale) to overcome the shape, topological, and anisotropic constraints imposed by other local morphometric scale models. Roughly speaking, the g-scale of a voxel in a scene was defined as the largest set of voxels connected to the voxel that satisfy some homogeneity criterion. We subsequently formulated a variant of the generalized scale notion, referred to as generalized ball scale (gB-scale), which, in addition to having the advantages of g-scale, also has superior noise resistance properties. These scale concepts are utilized in this paper to accurately determine principal tissue regions within MR images, and landmarks derived from these regions are used to perform intensity standardization. The new methods were qualitatively and quantitatively evaluated on a total of 67 clinical 3D MR images corresponding to four different protocols and to normal, Multiple Sclerosis (MS), and brain tumor patient studies. The generalized scale-based methods were found to be better than the existing methods, with a significant improvement observed for severely diseased and abnormal patient studies.
Registration I: Rigid Registration
icon_mobile_dropdown
Feature based registration of fluorescent LSCM imagery using region centroids
We present a novel semi-automated registration technique for 3D volume reconstruction from fluorescent laser scanning confocal microscope (LSCM) imagery. The developed registration procedure consists of (1) highlighting segmented regions as salient feature candidates, (2) defining two region correspondences by a user, (3) computing a pair of region centroids, as control points for registration, and (4) transforming images according to estimated transformation parameters determined by solving a set of linear equations with input control points. The presented semi-automated method is designed based on our observations that (a) an accurate point selection is much harder for a human than an accurate region (segment) selection, (b) a centroid selection of any region is less accurate by a human than by a computer, and (c) registration based on structural shape of a region rather than on intensity-defined point is more robust to noise and to morphological deformation of features across stacks. We applied the method to image mosaicking and image alignment registration steps and evaluated its performance with 20 human subjects on LSCM images with stained blood vessels. Our experimental evaluation showed significant benefits of automation for 3D volume reconstruction in terms of achieved accuracy, consistency of results and performance time. In addition, the results indicate that the differences between registration accuracy obtained by experts and by novices disappear with an advanced automation while the absolute registration accuracy increases.
Poster Session I
icon_mobile_dropdown
Qualifying CT for wrist arthroplasty: extending techniques for total hip arthroplasty to total wrist arthroplasty
Yvonne Alcala, Henrik Olivecrona, Lotta Olivecrona, et al.
The purpose of this study was to extend previous work to detect migration of total wrist arthroplasty non-invasively, and with greater accuracy. Two human cadaverous arms, each with a cemented total wrist implant, were used in this study. In one of the arms, 1 mm tantalum balls were implanted, six in the carpal bones and five in the radius. Five CT scans of each arm were acquired, changing the position of the arm each time to mimic different positions patients might take on repeated examinations. Registration of CT volume data sets was performed using an extensively validated, 3D semi-automatic volume fusion tool in which co-homologous point pairs (landmarks) are chosen on each volume to be registered. Three sets of ten cases each were obtained by placing landmarks on 1) bone only (using only arm one), 2) tantalum implants only, and 3) bone and tantalum implants (both using only arm two). The accuracy of the match was assessed visually in 2D and 3D, and numerically by calculating the distance difference between the actual position of the transformed landmarks and their ideal position (i.e., the reference landmark positions). All cases were matched visually within one width of cortical bone and numerically within one half CT voxel (0.32 mm, p = 0.05). This method matched only the bone/arm and not the prosthetic component per se, thus making it possible to detect prosthetic movement and wear. This method was clinically used for one patient with pain. Loosening of the carpal prosthetic component was accurately detected and this was confirmed at surgery.
Localization of perfusion abnormalities in brain SPECT imaging
Elise Nguyen, Jean Meunier, Jean-Paul Soucy, et al.
SPECT (Single Photon Emission Computed Tomography) imagery has become widely available and is particularly useful for regional cerebral blood flow (rCBF) studies. Distribution of rCBF is still essentially studied by visual observation, searching for abnormalities, and comparing with other studies. In order to facilitate the localization of these abnormalities, we propose a simple, automatic and direct method to register a SPECT rCBF study with a commonly used atlas in the neurological community, the Talairach Atlas. The Talairach atlas still gives today the most extensive information of regions of interests, coupled with a coordinate system. The proposed method will therefore allow a physician to precisely navigate in a SPECT image by interpreting the abnormalities coordinates. The registration of these two volumes is carried out in two steps, a rough alignment followed by an elastic registration. The rough alignment step consists in computing the mass centroid of each volume and in scaling the volumes accordingly if necessary. A simple threshold method (30% of the maximum intensity of the SPECT image) is used to determine the volume of the brain being studied. In order to facilitate the fine registration, the Talairach atlas was previously segmented in three classes: cerebrospinal fluid (CSF), white and gray matters. Then, an automatic intensity transformation as well as a low-pass filtering is performed to closely resemble the spatial resolution and intensities of the SPECT volume. This intensity transformation is a simple method which combines the use of a joint 2D histogram of the segmented atlas and the individual volume as well as a clustering algorithm. The fine registration is then computed with an optical flow methodology. The effectiveness of this scheme was tested on a database of virtual patients, simulated from a database of 45 healthy and diseased brains. The rate of pixels misclassification in each class within a one pixel neighborhood (CSF 0.5%; white matter 1.37%, gray matter 2.80%) indicates that this proposed method will be useful for the nuclear physician in helping localize abnormalities.
Motion correction for synthesis and analysis of respiratory-gated lung SPECT image
Hidenori Ue, Hideaki Haneishi, Hideyuki Iwanaga, et al.
A conventional SPECT image of lung is obtained by accumulating the detected count of gamma rays over long acquisition time that contains many respiratory cycles. The lung motion due to respiration during the acquisition makes reconstructed image blurred and may lead to a misdiagnosis. If a respiratory-gated SPECT is used, reconstructed images at various phase of respiration are obtained and the blur in an image can be avoided. However, the respiratory-gated SPECT requires long time to accumulate sufficient number of counts at each phase. If the acquisition time is not long enough, the detected count becomes inadequately small and hence the reconstructed image becomes noisy. We propose a method for correcting the motion between different phase images obtained with the respiratory-gated SPECT. In this method, an objective function consisting of both the degree of similarity between a reference and a deformed image and the smoothness of deformation is defined and optimized. The expansion ratio defined as a ratio of the change of the local volume due to the deformation is introduced to preserve the total activity during the motion correction process. By summing each phase images corrected by this method, a less noisy and less blurred SPECT image can be obtained. Furthermore, this method allows us to analyze the local movement of lung. This method was applied to the computer phantom, the real phantom and some clinical data and the motion correction and visualization of local movements between inspiration and expiration phase images were successfully achieved.
Registration II: Non-Rigid Registration
icon_mobile_dropdown
Articulated registration: elastic registration based on a wire-model
Miguel A. Martin-Fernandez, Emma Munyoz-Moreno, Marcos Martin-Fernandez, et al.
In this paper we propose a new method of elastic registration of anatomical structures that bears an inner skeleton, such as the knee, hand or spine. Such a method has to deal with great degrees of variability, specially for the case of inter-subject registration; but even for the intra-subject case the degree of variability of images will be large since the structures we bear in mind are articulated. Rigid registration methods are clearly inappropriate for this problem, and well-known elastic methods do not usually incorporate the restriction of maintaining long skeletal structures straight. A new method is therefore needed to deal with such a situation; we call this new method "articulated registration". The inner bone skeleton is modeled with a wire model, where wires are drawn by connecting landmarks located in the main joints of the skeletal structure to be registered (long bones). The main feature of our registration method is that within the bone axis (specifically, where the wires are) an exact registration is guaranteed, while for the remaining image points an elastic registration is carried out based on a distance transform (with respect to the model wires); this causes the registration on long bones to be affine to all practical purposes, while the registration of soft tissue -- far from the bones -- is elastic. As a proof-of-concept of this method we describe the registration of hands on radiographs.
Poster Session I
icon_mobile_dropdown
Segmentation guided registration for medical images
Registration and segmentation are two most important problems in the field of medical image analysis. Traditionally, they were treated as separate problems. In this paper, we introduce a unified variational framework for simultaneously carrying out image segmentation and registration. Segmentation information is integrated into the process of registration in leading to a more stable and noise-tolerant shape evolution, while a diffusion model is used to infer the volumetric deformation across the image. One of the major advantages of our model is its robustness against image noise. We present several 2D examples on synthetic and real data.
Combining global and local parallel optimization for medical image registration
Optimization is an important component in linear and nonlinear medical image registration. While common non-derivative approaches such as Powell's method are accurate and efficient, they cannot easily be adapted for parallel hardware. In this paper, new optimization strategies are proposed for parallel, shared-memory (SM) architectures. The Dividing Rectangles (DIRECT) global method is combined with the local Generalized Pattern Search (GPS) and Multidirectional Search (MDS) and to improve efficiency on multiprocessor systems. These methods require no derivatives, and can be used with all similarity metrics. In a multiresolution framework, DIRECT is performed with relaxed convergence criteria, followed by local refinement with MDS or GPS. In 3D-3D MRI rigid registration of simulated MS lesion volumes to normal brains with varying noise levels, DIRECT/MDS had the highest success rate, followed by DIRECT/GPS. DIRECT/GPS was the most efficient (5-10 seconds with 8 CPUs, and 10-20 seconds with 4 CPUs). DIRECT followed by MDS or GPS greatly increased efficiency while maintaining accuracy. Powell's method generally required more than 30 seconds (1 CPU) with a low success rate (0.3 or lower). This work indicates that parallel optimization on shared memory systems can markedly improve registration speed and accuracy, particularly for large initial misorientations.
Adaptive reduction of intensity levels in 3D images for mutual information-based registration
Mutual information is currently one of the most widely used image similarity measures for multimodality image registration. An important step in the calculation of the mutual information of two images is the estimation of their joint histogram. Most algorithms use lateral joint histogram sizes that are smaller than the actual number of intensity levels present in the images being registered. Using a reduced joint histogram size is especially useful when registering small portions of the images to obtain local deformations in nonrigid registration algorithms, and when implementing hardware solutions for acceleration of mutual information calculation. The most commonly used method for reducing the size of the joint histogram is to perform a linear rescaling of intensity values. The main problem with this method is that image regions with similar intensity values but corresponding to distinct tissue types tend to merge, thus compromising the accuracy of registration. We present new algorithms for reducing the number of gray levels present in 3D medical images, and compare their performance with previously reported ones. The tested algorithms are classified in three categories: histogram shape preserving algorithms, entropy maximization algorithms and quantization error minimization algorithms. Results show that in CT and MRI registration the best accuracy is achieved using entropy maximization algorithms, while in PET and MRI registration the best accuracy is achieved using histogram shape preservation algorithms.
Evaluation of sub-voxel registration accuracy between MRI and 3D MR spectroscopy of the brain
Francois Rousseau, Andrew Maudsley, Andreas Ebel, et al.
The implementation of Magnetic Resonance Spectroscopic Imaging (MRSI) for diagnostic imaging benefits from close integration of the lower-spatial resolution MRSI information with information from high-resolution structural MRI. Since patients can commonly move between acquisitions, it is necessary to account for possible mis-registration between the datasets arising from differences in patient positioning. In this paper we evaluate the use of 4 common multi-modality registration criteria to recover alignment between high resolution structural MRI and 3D MRSI data of the brain with sub-voxel accuracy. We explore the use of alternative MRSI water reference images to provide different types of structural information for the alignment process. The alignment accuracy was evaluated using both synthetically created MRSI and MRI data and a set of carefully collected subject image data with known ground truth spatial transformation between image volumes. The final accuracy and precision of estimates were assessed using multiple random starts of the registration algorithm. Sub voxel accuracy was found by all four similarity criteria with normalized mutual information providing the lowest target registration error for the 7 subject images. This effort supports the ongoing development of a database of brain metabolite distributions in normal subjects, which will be used in the evaluation of metabolic changes in neurological diseases.
Modular toolbox for derivative-based medical image registration
Astrid Franz, Ingwer C. Carlsen, Sven Kabus, et al.
Registration of medical images, i.e. the integration of two or more images into a common geometrical system of reference so that corresponding image structures correctly align, is an active field of current research. Registration algorithms in general are composed of three main building blocks: a geometrical transformation is applied in order to transform the images into the geometrical system of reference, a similarity measure puts the comparison of the images into quantifiable terms, and an optimization algorithm searches for that transformation that leads to optimal similarity between the images. Whereas in the literature fixed configurations of registration algorithms are investigated, here we present a modular toolbox containing several similarity measures, transformation classes and optimization strategies. Derivative-free optimization is applicable for any similarity measure, but is not fast enough in clinical practice. Hence we consider much faster derivative-based Gauss-Newton and Levenberg-Marquardt optimization algorithms that can be used in conjunction with frequently needed similarity measures for which derivatives can be easily obtained. The implemented similarity measures, geometrical transformations and optimization methods can be freely combined in order to configure a registration algorithm matching the requirements of a particular clinical application. Test examples show that particular algorithm configurations out of this toolbox allow e.g. for an improved lesion identification and localization in PET-CT or MR registration applications.
Automatic registration of ICG images using mutual information and perfusion analysis
Introduction: Indocyanin green fundus angiographic images (ICGA) of the eyes is useful method in detecting and characterizing the choroidal neovascularization (CNV), which is the major cause of the blindness over 65 years of age. To investigate the quantitative analysis of the blood flow on ICGA, systematic approach for automatic registration of using mutual information and a quantitative analysis was developed. Methods: Intermittent sequential images of indocyanin green angiography were acquired by Heidelberg retinal angiography that uses the laser scanning system for the image acquisition. Misalignment of each image generated by the minute eye movement of the patients was corrected by the mutual information method because the distribution of the contrast media on image is changing throughout the time sequences. Several region of interest (ROI) were selected by a physician and the intensities of the selected region were plotted according to the time sequences. Results: The registration of ICGA time sequential images is required not only translate transform but also rotational transform. Signal intensities showed variation based on gamma-variate function depending on ROIs and capillary vessels show more variance of signal intensity than major vessels. CNV showed intermediate variance of signal intensity and prolonged transit time. Conclusion: The resulting registered images can be used not only for quantitative analysis, but also for perfusion analysis. Various investigative approached on CNV using this method will be helpful in the characterization of the lesion and follow-up.
Comparison and evaluation of joint histogram estimation methods for mutual information based image registration
Joint histogram is the only quantity required to calculate the mutual information (MI) between two images. For MI based image registration, joint histograms are often estimated through linear interpolation or partial volume interpolation (PVI). It has been pointed out that both methods may result in a phenomenon known as interpolation induced artifacts. In this paper, we implemented a wide range of interpolation/approximation kernels for joint histogram estimation. Some kernels are nonnegative. In this case, these kernels are applied in two ways as the linear kernel is applied in linear interpolation and PVI. In addition, we implemented two other joint histogram estimation methods devised to overcome the interpolation artifact problem. They are nearest neighbor interpolation with jittered sampling with/without histogram blurring and data resampling. We used the clinical data obtained from Vanderbilt University for all of the experiments. The objective of this study is to perform a comprehensive comparison and evaluation of different joint histogram estimation methods for MI based image registration in terms of artifacts reduction and registration accuracy.
Robust and staining-invariant elastic registration of a series of images from histologic slices
Stefan Wirtz, Nils Papenberg, Bernd Fischer, et al.
In image registration of medical data a common and challenging problem is handling intensity-inhomogeneities. These inhomogeneities appear for instance in images of serially sectioned brains caused by the histological staining process or in medical imaging with contrast agents. Beneath this, natural outliers (for instance cells or vessels) produced by the underlying material itself may be mistaken as noise. Both image registration applications have in common that the well known sum of squared differences (SSD) measure would detect false differences. To deal with these kinds of problems, we supplement the common SSD-measure with image derivatives of higher order. Additionally we introduce a non-quadratic penalizer function to the distance measure leading to robust energy. The concepts are well known in optical flow. Overall, we present a variational model which combines all of these properties. This formulation leads to a fast and efficient algorithm. We demonstrate its applicability at the problems described above.
Deformable registration for integration of MRI/MRSI information in TRUS-guided prostate biopsy
Wei Shao, Ruoyun Wu, Choon Hua Thng, et al.
Prostate cancer has been ranked as the second leading cause of cancer death in men. The existence of cancer in prostate is usually examined by a biopsy procedure under the transrectal ultrasound (TRUS) guidance. Development of a prostate biopsy robotics can alleviate urologists' labor and guarantee accuracy. However, it is usually impossible to identify cancer region in the noisy ultrasound images, thus leading to a random biopsy protocol for prostate. It is being recognized that Magnetic Resonance Imaging (MRI) and Magnetic Resonance Spectroscopy Imaging (MRSI) techniques are potential to diagnose cancer distribution in prostate. So navigating the biopsy needle towards those cancer-suspected sites could improve the cancer detection rate and reduce the possibility of false negative diagnosis results. As the prostate usually deforms under the different rectal filling of probes and change of patient postures, a deformable registration scheme is implemented for the integration of the pre-operative MRI/MRSI information with the intra-operative TRUS images. A framework including a global rigid alignment and a sequent non-rigid transformation was described in this paper to match the cross-modal prostate surfaces and thereafter their volumes. For validation, an elastic prostate phantom that simulated the human condition was built up, with fiducial markers implanted inside the phantom prostate as the "ground truth". It shows that our method can achieve at least 30% improvement in accuracy compared with an affine transformation. Preliminary study was also conducted on patient data but with visual assessments presented only due to the current lack of "ground truth".
Multimodal image registration based on compound mutual information
Zhaohui Sun, Lawrence A. Ray
In this paper, we study image description based on compound mutual information (CMI) and its application to multimodal image registration. CMI is an aggregate information measure derived from the multiple marginal densities of image distributions. It extends histogram-based mutual information to that based on various marginal densities, which encode spatial intensity distribution and other image characteristics. Therefore, CMI can overcome the difficulties (such as the lack of spatial information) inherent in the color histogram, enrich the vocabulary of image description, and help improve registration accuracy and robustness. CMI is not sensitive to illumination and absolute appearance, and it is particularly suited for multimodal applications.
Assessment of artery dilation by using image registration based on spatial features
The use of affine image registration based on normalized mutual information (NMI) has recently been proposed by Frangi et al. as an automatic method for assessing brachial artery flow mediated dilation (FMD) for the characterization of endothelial function. Even though this method solves many problems of previous approaches, there are still some situations that can lead to misregistration between frames, such as the presence of adjacent vessels due to probe movement, muscle fibres or poor image quality. Despite its widespread use as a registration metric and its promising results, MI is not the panacea and can occasionally fail. Previous work has attempted to include spatial information into the image similarity metric. Among these methods the direct estimation of α-MI through Minimum Euclidean Graphs allows to include spatial information and it seems suitable to tackle the registration problem in vascular images, where well oriented structures corresponding to vessel walls and muscle fibres are present. The purpose of this work is twofold. Firstly, we aim to evaluate the effect of including spatial information in the performance of the method suggested by Frangi et al. by using α-MI of spatial features as similarity metric. Secondly, the application of image registration to long image sequences in which both rigid motion and deformation are present will be used as a benchmark to prove the value of α-MI as a similarity metric, and will also allow us to make a comparative study with respect to NMI.
Constructing and assessing brain templates from Chinese pediatric MRI data using SPM
Qingjie Yin, Qing Ye, Li Yao, et al.
Spatial normalization is a very important step in the processing of magnetic resonance imaging (MRI) data. So the quality of brain templates is crucial for the accuracy of MRI analysis. In this paper, using the classical protocol and the optimized protocol plus nonlinear deformation, we constructed the T1 whole brain templates and apriori brain tissue data from 69 Chinese pediatric MRI data (age 7-16 years). Then we proposed a new assessment method to evaluate our templates. 10 pediatric subjects were chosen to do the assessment as the following steps. First, the cerebellum region, the region of interest (ROI), was located on both the pediatric volume and the template volume by an experienced neuroanatomist. Second, the pediatric whole brain was mapped to the template with affine and nonlinear deformation. Third, the parameter, derived from the second step, was used to only normalize the ROI of the child to the ROI of the template. Last, the overlapping ratio, which described the overlapping rate between the ROI of the template and the normalized ROI of the child, was calculated. The mean of overlapping ratio normalized to the classical template was 0.9687, and the mean normalized to the optimized template was 0.9713. The results show that the two Chinese pediatric brain templates are comparable and their accuracy is adequate to our studies.
Maximization of feature potential mutual information in multimodality image registration using particle swarm optimization
Xuan Yang, Jihong Pei, Weixin Xie
Standard Mutual Information function contains local maxima, which make against to convergence of registration transformation parameters for automated multimodality image registration problems. We proposed Feature Potential Mutual Information (FPMI) to increases the smoothness of the registration measure function and use Particle Swarm Optimization to search the optimal registration transformation parameter in this paper. At first, Edges of images are detected. Next, edge feature potential is defined by expanding edges to the neighborhood region using potential function. Each edge point influences the whole potential field, just like the particle of physics in the gravitation field space. FPMI is computed on the edge feature potential of two images. It substitutes the edge feature potential values for gray values in images. It can avoid great change of joint probability distribution and has less local maxima. The registration accuracy of FPMI is analyzed under different edge detection cases. It is shown that the registration accuracy of FPMI is more accurate and more robust than that of MI. Maximization of FPMI is done by PSO. PSO combines local search methods with global search methods, attempting to balance exploration and exploitation. Its complex behavior follows from a few simple rules and has less computational complexity. Multimodal medical images are used to compare the response of FPMI and MI to translation and rotation. Experiments show that FPMI is smoother and has less local fluctuations than that of MI. Registration results show that PSO do it better than Powell’s method to search the optimal registration parameters.
Implicit function-based phantoms for evaluation of registration algorithms
Girish Gopalakrishnan, Timothy Poston, Nithin Nagaraj, et al.
Medical image fusion is increasingly enhancing diagnostic accuracy by synergizing information from multiple images, obtained by the same modality at different times or from complementary modalities such as structural information from CT and functional from PET. An active, crucial research topic in fusion is validation of the registration (point-to-point correspondence) used. Phantoms and other simulated studies are useful in the absence of, or as a preliminary to, definitive clinical tests. Software phantoms in specific have the added advantage of robustness, repeatability and reproducibility. Our virtual-lung-phantom-based scheme can test the accuracy of any registration algorithm and is flexible enough for added levels of complexity (addition of blur/anti-alias, rotate/warp, and modality-associated noise) to help evaluate the robustness of an image registration/fusion methodology. Such a framework extends easily to different anatomies. The feature of adding software-based fiducials both within and outside simulated anatomies prove more beneficial when compared to experiments using data from external fiducials on a patient. It would help the diagnosing clinician make a prudent choice of registration algorithm.
A new look at Markov random field (MRF) model-based MR image analysis
Pixel intensities of MR images reconstructed by Fourier Transform and Projection methods have been proved to be spatially asymptotically independent (S.A.I.) and to have exponential correlation coefficient (E.C.C.). Based on S.A.I. and E.C.C., the MR image has been proved to be embedded in an MRF with respect to a proper neighborhood system. Further, the MR image is proved to be modeled by a Finite Normal Mixture (FNM) with an MRF as its prior. A unified Expectation-Maximization (EM) algorithm is presented for performing image segmentation. S.A.I., E.C.C. and Markovianity provide means for selecting the order of neighborhood systems and the values of clique potentials. The use of the 3rd-order neighborhood system and the correlation coefficient-based assignments of clique potentials strike an optimal trade-off between good accuracy and sufficient simplicity in image segmentation.
Automatic landmarking of magnetic resonance brain images
Camille Izard, Bruno M. Jedynak, Craig E. L. Stark
Landmarking MR images is crucial in registering brain structures from different images. It consists in locating the voxel in the image that corresponds to a well-defined point in the anatomy, called the landmark. Example of landmarks are the apex of the head (HoH) of Hippocampus, the tail and the tip of the splenium of the corpus collosum (SCC). Hand landmarking is tedious and time-consuming. It requires an adequate training. Experimental studies show that the results are dependent on the landmarker and drifting with time. We propose a generic algorithm performing automated detection of landmarks. The first part consists in learning from a training set of landmarked images the parameters of a probabilistic model, using the EM algorithm. The second part inputs the estimated parameters and a new image, and outputs a voxel as a predicted location for the landmark. The algorithm is demonstrated on the HoH and the SCC. In contrast with competing approaches, the algorithm is generic: it can be used to detect any landmark, given a hand-labeled training set of images.
Quantizing calcification in the lumbar aorta on 2-D lateral x-ray images
Lars Arne Conrad-Hansen, Francois Lauze, Laszlo B. Tanko M.D., et al.
In this paper we seek to improve upon the standard method of assessing the degree of calcification in the lumbar aorta, which is commonly used on lateral 2-D x-rays. The necessity for improvement arises from the fact that the existing method can not measure subtle progressions in the plaque development; neither is it possible to express the density of individual plaques. Both of these qualities would be desireable to assess, since they are the key for making progression studies as well as for testing the effect of drugs in longitudinal studies. Our approach is based on inpainting, a technique used in image restoration as well as postprocessing of film. In this study we discuss the potential implications of total variation inpainting for characterizing aortic calcification.
Texture analysis of high-resolution FLAIR images for TLE
This paper presents a study of the texture information of high-resolution FLAIR images of the brain with the aim of determining the abnormality and consequently the candidacy of the hippocampus for temporal lobe epilepsy (TLE) surgery. Intensity and volume features of the hippocampus from FLAIR images of the brain have been previously shown to be useful in detecting the abnormal hippocampus in TLE. However, the small size of the hippocampus may limit the texture information. High-resolution FLAIR images show more details of the abnormal intensity variations of the hippocampi and therefore are more suitable for texture analysis. We study and compare the low and high-resolution FLAIR images of six epileptic patients. The hippocampi are segmented manually by an expert from T1-weighted MR images. Then the segmented regions are mapped on the corresponding FLAIR images for texture analysis. The 2-D wavelet transforms of the hippocampi are employed for feature extraction. We compare the ability of the texture features from regular and high-resolution FLAIR images to distinguish normal and abnormal hippocampi. Intracranial EEG results as well as surgery outcome are used as gold standard. The results show that the intensity variations of the hippocampus are related to the abnormalities in the TLE.
Modelization of three-dimensional bone micro-architecture using Markov random fields with a multi-level clique system
Thomas Lamotte, Jean-Marc Dinten, Francoise Peyrin
The goal of this work is to find a new parametric model of three-dimensional (3D) bone micro-architecture which is both representative of some bone characteristics and can be used as a prior model in tomographic reconstruction or image segmentation. For this purpose, Markov Random Fields (MRFs) are adapted. Bone micro-architecture is typically organized as a pseudo-periodic arrangement of parallel plates with a given thickness (≈ 100 μm), and a given spacing (≈ 700 μm). As standard MRFs could not represent the complexity of bone, we investigated appropriated energy functions, parameters and a design of the neighborhood system including local and distant interactions. Then, parameters of the new model were related to the standard morphometric parameters of the bone microarchitecture using an usual bone quantification tool.
Textural content in 3T MR: an image-based marker for Alzheimer's disease
In this paper, we propose a study, which investigates the first-order and second-order distributions of T2 images from a magnetic resonance (MR) scan for an age-matched data set of 24 Alzheimer's disease and 17 normal patients. The study is motivated by the desire to analyze the brain iron uptake in the hippocampus of Alzheimer's patients, which is captured by low T2 values. Since, excess iron deposition occurs locally in certain regions of the brain, we are motivated to investigate the spatial distribution of T2, which is captured by higher-order statistics. Based on the first-order and second-order distributions (involving gray level co-occurrence matrix) of T2, we show that the second-order statistics provide features with sensitivity >90% (at 80% specificity), which in turn capture the textural content in T2 data. Hence, we argue that different texture characteristics of T2 in the hippocampus for Alzheimer's and normal patients could be used as an early indicator of Alzheimer's disease.
Poster Session II
icon_mobile_dropdown
Automatic landmark selection for active shape models
The first step in Active Shape Model (ASM) based image segmentation and processing is to create a point distribution model (PDM) during the training phase. Correct point (landmark) correspondences across each of the training shapes must be determined for a successful and effective statistical model building process. Effective and automatic solutions for this problem are needed for the practical use of ASM methods. In this paper, we provide a solution for this problem which consists of: (i) a process of generating a mean shape without requiring landmarks, (ii) a process of automatic landmark selection for the mean shape, and (iii) a process of propagating landmarks on to each training shape for defining landmarks in them. This paper describes the method of generating the mean shape, and the landmark selection and correspondence process. Although the method is generally applicable to spaces of any dimensionality, our first implementation and evaluation has been carried out for 2D shapes. The method is evaluated on 20 MRI foot data sets, the object of interest being the talus bone. The results indicate that, for the same given number of points, better compactness (number of parameters) of the ASM by using our method can be achieved than by using the commonly used equi-spaced point selection method.
Functional analysis of cardiac MR images using SPHARM modeling
Heng Huang, Li Shen, James Ford, et al.
Visualization and quantification of cardiac function can provide direct and reliable indicators of cardiac health. The heart's operation occurs in three dimensions, and is dependent on three dimensional forces and ventricular geometry, making the observation of its shape important. Many approaches have been presented to extract cardiac shape and do functional analysis from a variety of imaging modalities. We apply a spherical harmonics (SPHARM) model to cardiac function analysis using magnetic resonance (MR) images. Our three dimensional SPHARM approach increases measurement accuracy over two dimensional approaches and also simplifies the management and indexing of clinical data by providing access to many important functional measures directly from the SPHARM representation.
On the adequacy of principal factor analysis for the study of shape variability
Miguel Angel Gonzalez Ballester, Marius George Linguraru, Mauricio Reyes Aguirre, et al.
The analysis of shape variability of anatomical structures is of key importance in a number of clinical disciplines, as abnormality in shape can be related to certain diseases. Statistical shape analysis techniques commonly employed in the medical imaging community, such as Active Shape Models or Active Appearance Models rely on Principal Component Analysis (PCA) to decompose shape variability into a reduced set of interpretable components. In this paper we propose Principal Factor Analysis (PFA) as an alternative to PCA and argue that PFA is a better suited technique for medical imaging applications. PFA provides a decomposition into modes of variation that are more easily interpretable, while still being a linear, efficient technique that performs dimensionality reduction (as opposed to Independent Component Analysis, ICA). Both PCA and PFA are described. Examples are provided for 2D landmark data of corpora callosa outlines, as well as vector-valued 3D deformation fields resulting from non-rigid registration of ventricles in MRI. The results show that PFA is a more descriptive tool for shape analysis, at a small cost in size (as in theory more components may be necessary to explain a given percentage of total variance in the data). In conclusion, we argue that it is important to study the potential of factor analysis techniques other than PCA for the application of shape analysis, and defend PFA as a good alternative.
Extraction of the cerebral cortical boundaries from MRI for measurement of cortical thickness
Simon Fristed Eskildsen, Mark Uldahl, Lasse Riis Ostergaard
Several neurodegenerative diseases, such as Alzheimer's disease, cause atrophy of the cerebral cortex. Measurements of cerebral cortical thickness and volume are used in the quantification and localization of atrophy. It is possible to measure the thickness of the cerebral cortex manually from magnetic resonance imaging, but partial volume effects, orthogonality problems, large amounts of manual labor and operator bias makes it difficult to conduct measurements on large patient populations. Automatic quantification and localization of atrophy is a highly desirable goal, as it facilitates the study of early anatomical changes and track disease progression on large populations. The first step in achieving this goal is to develop robust and accurate methods for measuring cortical thickness and volume automatically. We have developed a new method, capable of both extracting surface representations of the cortical boundaries from magnetic resonance imaging and measuring the cortical thickness. Experiments show that the developed method is robust and performs well on datasets of both healthy subjects and subjects suffering from Alzheimer's disease.
Statistical shape model generation using nonrigid deformation of a template mesh
Active shape models (ASMs) have been studied extensively for the statistical analysis of three-dimensional shapes. These models can be used as prior information for segmentation and other image analysis tasks. In order to create an ASM, correspondence between surface points on the training shapes must be provided. Various groups have previously investigated methods that attempted to provide correspondences between points on pre-segmented shapes. This requires a time-consuming segmentation stage before the statistical analysis can be performed. This paper presents a method of ASM generation that requires as input only a single segmented template shape obtained from a mean grayscale image across the training set. The triangulated mesh representing this template shape is then propagated to the other shapes in the training set by a nonrigid transformation. The appropriate transformation is determined by intensity-based nonrigid registration of the corresponding grayscale images. Following the transformation of the template, the mesh is treated as an active surface, and evolves towards the image edges while preserving certain curvature constraints. This process results in automatic segmentation of each shape, but more importantly also provides an automatic correspondence between the points on each shape. The resulting meshes are aligned using Procrustes analysis, and a principal component analysis is performed to produce the statistical model. For demonstration, a model of the lower cervical vertebrae (C6 and C7) was created. The resulting model is evaluated for accuracy, compactness, and generalization ability.
Reconstruction of rotary DSA vessel axis based on the matching of multiple projections
A novel method of the matching and reconstruction of DSA vessel axis is proposed based on the redundant information from multiple two-dimensional (2-D) projections of the object. First, a correlation scheme on pixels gray level is used to extract the vessel structure on every projection. Through this way, we acquire a binary image. Secondly, Hit-Miss Transform (HMT) is applied on the binary image to produce a vessel axis image. Thirdly, an arbitrary projection is chosen as base image, and others are chosen as reference images. For each key point in the base image, the matching points are found from the key points of the reference images according to epipolar geometry and the topological linking relations of vessel branches. The matching segments are determined from matched key points. If one segment of vessel axis in the base image is confirmed to be matching with one in the reference image, an interpolation process may be used to find out the corresponding relationship of points on these two segments. Then we can get a continuous 3-D segment of DSA vessel axis according to reconstruction the matched points between two views. The whole process is executed repeatedly when another image is selected as base image. The experiment result shows that this method may provide satisfactory reconstruction even the vessel extraction result is not very accurate.
3D reconstruction of an IVUS transducer trajectory with a single view in cineangiography
Melissa Jourdain, Jean Meunier, Rosaire Mongrain, et al.
During an Intravascular Ultrasound (IVUS) intervention, a catheter with an ultrasound transducer is introduced in the body through a blood vessel and then pulled back to image a sequence of vessel cross-sections. Unfortunately, there is no 3D information about the position and orientation of these cross-section planes. To position the IVUS images in space, some researchers have proposed complex stereoscopic procedures relying on biplane angiography to get two X-ray image sequences of the IVUS transducer trajectory along the catheter. We have elaborated a much simpler algorithm to recover the transducer 3D trajectory with only a single view X-ray image sequence. The known pullback distance of the transducer during the IVUS intervention is used as an a priori to perform this task. Considering that biplane system are difficult to operate and rather expensive and uncommon in hospitals; this simple pose estimation algorithm could lead to an affordable and useful tool to better assess the 3D shape of vessels investigated with IVUS.
3D segmentation of non-isolated pulmonary nodules in high resolution CT images
Xiangwei Zhang, Geoffrey McLennan, Eric A. Hoffman, et al.
The purpose of this study is to develop a computer-aided diagnosis (CAD) system to segment small size non-isolated pulmonary nodules in high resolution helical CT scans. A new automated method of segmenting juxtapleural nodules was proposed, in which a quadric surface fitting procedure was used to create a boundary between a juxtapleural nodule and its neighboring pleural surface. Experiments on some real CT nodule data showed that this method was able to yield results that reflect the local shape of the pleural surface. Additionally, a scheme based on parametrically deformable geometric model was developed to deal with the problem of segmenting nodules attached to vessels. A vessel segment connected to a nodule was modeled using superquadrics with parametric deformations. The boundary between a vascularized nodule and the attached vessels can be recovered by finding the deformed superquadrics which approximates the vessels. Gradient descent scheme was utilized to optimize the parameters of the superquadrics. Simple experiments on synthetic data showed this scheme is promising.
Curved planar reformation of CT spine data
Tomaz Vrtovec, Bostjan Likar, Franjo Pernus
Traditional techniques for visualizing anatomical structures are based on planar sections from volume images, like images obtained by computed tomography (CT) or magnetic resonance imaging (MRI). However, planar slices taken in the coordinate system of the 3D image often do not provide sufficient or qualitative enough diagnostic information. The reason is that because planar slices do not follow curved anatomical structures (e.g. arteries, colon, spine, etc.), not all important details can be shown simultaneously. For better visualization of curved structures, reformatted images in the coordinate system of a structure must be created (an operation called curved planar reformation). In this paper we focus on automated curved planar reformation (CPR). The obtained spine-based 3D coordinate system is determined by the natural curvature of the spine, described by a curve that is parameterized by a polynomial model. The model is optimized to fit the curvature of the spine basing on the values of a pre-calculated distance map. The first coordinate is defined by the resulting spine curve, while the other two coordinates are defined by the natural rotations of the vertebrae around the spine curve. The proposed approach benefits from reduced structural complexity in favor of improved feature perception of the spine, and is not only important for extracting diagnostically important images, but also for easier navigation, manipulation and orientation in 3D space, which is helpful for morphometric analysis, automated image analysis (e.g. segmentation), and normalization of spine images.
Semi-automatic border detection method for left ventricular volume estimation in 4D ultrasound data
Marijn van Stralen, Johan G. Bosch, Marco M. Voormolen, et al.
We propose a semi-automatic endocardial border detection method for LV volume estimation in 3D time series of cardiac ultrasound data. It is based on pattern matching and dynamic programming techniques and operates on 2D slices of the 4D data requiring minimal user-interaction. We evaluated on data acquired with the Fast Rotating Ultrasound (FRU) transducer: a linear phased array transducer rotated at high speed around its image axis, generating high quality 2D images of the heart. We automatically select a subset of 2D images at typically 10 rotation angles and 16 cardiac phases. From four manually drawn contours a 4D shape model and a 4D edge pattern model is derived. For the selected images, contour shape and edge patterns are estimated using the models. Pattern matching and dynamic programming is applied to detect the contours automatically. The method allows easy corrections in the detected 2D contours, to iteratively achieve more accurate models and improved detections. An evaluation of this method on FRU data against MRI was done for full cycle LV volumes on 10 patients. Good correlations were found against MRI volumes [r=0.94, y=0.72x + 30.3, difference of 9.6 +/- 17.4 ml (Av +/- SD)] and a low interobserver variability for US (r=0.94, y=1.11x - 16.8, difference of 1.4 +/- 14.2 ml). On average only 2.8 corrections per patient were needed (in a total of 160 images). Although the method shows good correlations with MRI without corrections, applying these corrections can make significant improvements.
Fast intersection checking for parametric deformable models
Douglas P. Perrin, Andrew M. Ladd, Lydia E. Kavraki, et al.
Parametric active deformable models for image-based segmentation offer a distinct advantage over level sets: speed. This paper presents an extension to active deformable models that makes real-time volume segmentation possible on mid-range off-the-shelf hardware and without the use of specialized graphics hardware. The proposed method uses region-based parametric deformable models. A region-based parametric model, represented by a polygon, must remain non-self intersecting (simple) while undergoing deformation. The simplicity constraint can be enforced by allowing topological changes or by restricting motions of the curve. In either case, intersections of curve segments must be detected otherwise catastrophic divergence results. Good performance relies on the efficiency of the intersection check operation. This paper presents a parameter-free and efficient technique for on-line simplicity checking of polygons undergoing motion. We present timing results validating our approach; in particular, we segment 3-D ultrasound data at 20 volumes per second.
Model based LV-reconstruction in bi-plane x-ray angiography
Werner Backfrieder, Martin Carpella, Roland Swoboda, et al.
Interventional x-ray angiography is state of the art in diagnosis and therapy of severe diseases of the cardiovascular system. Diagnosis is based on contrast enhanced dynamic projection images of the left ventricle. A new model based algorithm for three dimensional reconstruction of the left ventricle from bi-planar angiograms was developed. Parametric super ellipses are deformed until their projection profiles optimally fit measured ventricular projections. Deformation is controlled by a simplex optimization procedure. A resulting optimized parameter set builds the initial guess for neighboring slices. A three dimensional surface model of the ventricle is built from stacked contours. The accuracy of the algorithm has been tested with mathematical phantom data and clinical data. Results show conformance with provided projection data and high convergence speed makes the algorithm useful for clinical application. Fully three dimensional reconstruction of the left ventricle has a high potential for improvements of clinical findings in interventional cardiology.
Locating articular cartilage in MR images
Jenny Folkesson, Erik Dam, Paola Pettersen M.D., et al.
Accurate computation of the thickness of the articular cartilage is of great importance when diagnosing and monitoring the progress of joint diseases such as osteoarthritis. A fully automated cartilage assessment method is preferable compared to methods using manual interaction in order to avoid inter- and intra-observer variability. As a first step in the cartilage assessment, we present an automatic method for locating articular cartilage in knee MRI using supervised learning. The next step will be to fit a variable shape model to the cartilage, initiated at the location found using the method presented in this paper. From the model, disease markers will be extracted for the quantitative evaluation of the cartilage. The cartilage is located using an ANN-classifier, where every voxel is classified as cartilage or non-cartilage based on prior knowledge of the cartilage structure. The classifier is tested using leave-one-out-evaluation, and we found the average sensitivity and specificity to be 91.0% and 99.4%, respectively. The center of mass calculated from voxels classified as cartilage are similar to the corresponding values calculated from manual segmentations, which confirms that this method can find a good initial position for a shape model.
Growing deformable surface patches for topology-adaptive object detection in MR images
Xujian Chen, Eam Khwang Teoh
A novel deformable model for 3D surface extraction, called growing deformable surface patches, is presented in this work. In the proposed method, a growing mechanism is introduced to 3D deformable model. With the growing framework, the proposed deformable model could achieve topologically adaptable surface extraction by connecting new surface patches with active patches and automated triangulating the square patch in particular situations. A number of experiments demonstrate that the proposed algorithm can extract the surface of complex anatomic structures effectively. Compared with the existing topologically adaptable deformable surfaces, computational cost is reduced because no splitting or merging judgement is carried out among all the vertices in each deformation step. Topologically adaptable detection is achieved by analyzing the possibility of patch connection among the "active" surface patches only.
Enhanced 3-D reconstruction in NMR Fresnel diffractive imaging technique
Satoshi Ito, Yoshifumi Yamada
Enhanced 3-D reconstruction in NMR diffractive imaging technique which is a new approach to MR angiography, has been investigated. The expression of NMR signals obtained in the NMR diffractive imaging technique is similar to the equation for Fresnel diffraction in light waves or sound waves. Therefore, it is possible to reconstruct images focusing on optional plane in the depth direction using the data scanned two-dimensionally. However blurred image components out of focal-plane superimpose on the focal-plane image and the spatial resolution in the longitudinal direction is degraded. We have developed a new algorithm by which blurred image components are effectively removed. These studies demonstrate the possibility of the proposed method as a fast imaging technique for MR angiography.
Generalized ball-scale: theory, algorithms, and application to image inhomogeneity correction
The concept of generalized scale (g-scale) was introduced previously to overcome the shape, topological, and anisotropic constraints imposed by previous local morphometric scale models. Roughly speaking, the g-scale of a voxel in a scene was defined as the largest set of all voxels associated with it, that satisfy some homogeneity criterion. g-scale was shown to have interesting theoretical properties, and its superiority to an existing image background inhomogeneity correction method was demonstrated. In this paper, we present a variant of g-scale that we refer to as gB-scale. The difference between g- and gB-scale is that, while for g-scale, individual voxels are included into the g-scale set one at a time, the gB-scale set is grown by including hyperballs, the hyperball corresponding to the local ball scale at every voxel (which, briefly, is the radius of the largest hyperball of homogeneous intensity centered at the voxel). The gB-scale model was found to be more resistant to severe levels of inhomogeneity and noise compared to g-scale. A methodology to perform image background inhomogeneity correction based on the idea of gB-scale was qualitatively and quantitatively compared on nearly 250 clinical and phantom datasets, with a ball-scale- and a g-scale-based correction methodology. For scenes containing inhomogeneity but no noise, the g-, and gB-scale methods performed comparably and were superior to the ball-scale method. For scenes containing both noise and inhomogeneity, the gB-scale-based method outperformed both the g- and ball-scale correction methods.
Thickness correction of mammographic images by anisotropic filtering and interpolation of dense tissue
Peter R. Snoeren, Nico Karssemeijer
Without image processing, the dynamic range of display systems is too small to optimally display both the interior and the peripheral zone of a compressed breast. To overcome manual adjustment of contrast, we propose an algorithm for peripheral enhancement of digital or digitized mammograms. This is done by virtually adding homogeneous tissue at the peripheral zone, where the breast comes loose from the compression paddle. The gradual signal increase due to a smaller breast thickness near the breast edge is estimated by the solution of the anisotropic diffusion equation. The conductivity is set small in the direction perpendicular to the breast edge, and large in the parallel direction. By this, large conductivities (much blurring) can be applied, while undesirable artifacts that would be caused by isotropic filtering are minimized. This measure is not always enough to prevent some reduction of relevant contrast. Therefore, dense tissue along parallel curves to the breast edge is interpolated before smoothing the image. Anisotropic diffusion filtering and dense tissue interpolation are both new techniques to improve peripheral enhancement. Comparison with some other methods showed that our approach performs at least as good as other methods.
Deformable model for 3D intramodal nonrigid breast image registration with fiducial skin markers
Mehmet Z. Unlu, Andrzej Krol, Ioana L.. Coman, et al.
We implemented a new approach to intramodal non-rigid 3D breast image registration. Our method uses fiducial skin markers (FSM) placed on the breast surface. After determining the displacements of FSM, finite element method (FEM) is used to distribute the markers’ displacements linearly over the entire breast volume using the analogy between the orthogonal components of the displacement field and a steady state heat transfer (SSHT). It is valid because the displacement field in x, y and z direction and a SSHT problem can both be modeled using LaPlace’s equation and the displacements are analogous to temperature differences in SSHT. It can be solved via standard heat conduction FEM software with arbitrary conductivity of surface elements significantly higher than that of volume elements. After determining the displacements of the mesh nodes over the entire breast volume, moving breast volume is registered to target breast volume using an image warping algorithm. Very good quality of the registration was obtained. Following similarity measurements were estimated: Normalized Mutual Information (NMI), Normalized Correlation Coefficient (NCC) and Sum of Absolute Valued Differences (SAVD). We also compared our method with rigid registration technique.
High resolution retinal image restoration with wavefront sensing and self-extracted filtering
Shuyu Yang, Gavin Erry, Sheila Nemeth, et al.
Diagnosis and treatment of retinal diseases such as diabetic retinopathy commonly rely on a clear view of the retina. High quality retinal images are essential in early detection and more accurate diagnosis of many retinal diseases. Conventional fundus cameras usually lack the ability to provide high resolution details required for diagnostic accuracy. Major factors contributing to the degradation of retinal image quality are the aberrations from the eye and the imaging device. The challenge in obtaining high quality retinal image lies in the design of the imaging system that can reduce the strong aberrations of the human eye. Since the amplitudes of human eye aberrations decrease rapidly as the aberration order goes up, it is more cost-effective to correct low order aberrations with adaptive optical devices while process high order aberrations through image processing. A cost effective fundus imaging device that can capture high quality retinal images with 2-5 times higher resolution than conventional retinal images has been designed. This imager improves image quality by attaching complementary adaptive optical components to a conventional fundus camera. However, images obtained with the high resolution camera are still blurred due to some uncorrected aberrations as well as defocusing resulting from non-isoplanatic effect. Therefore, advanced image restoration algorithms have been employed for further improvement in image quality. In this paper, we use wavefront-based and self-extracted blind deconvolution techniques to restore images captured by the high resolution fundus camera. We demonstrate that through such techniques, pathologies that are critical to retinal disease diagnosis but not clear or not observable in the original image can be observed clearly in the restored images. Image quality evaluation is also used to finalize the development of a cost-effective, fast, and automated diagnostic system that can be used clinically.
A new non-parametric method for image intensity inhomogeneity correction using a non-uniform gradient filter and path integrals
A new local gradient-based non-parametric inhomogeneity correction method is developed that is independent of image acquisition modality and protocol. Image intensity inhomogeneity, mostly caused by imperfections in imaging devices and underlying processes, results in different image intensity values at different regions of an object. This phenomenon often poses a major challenge to different post processing applications (e.g., segmentation, quantitative analysis), especially in medical imaging. Prospective intensity correction approaches in MRI are time consuming and costly, and they fail to resolve patient-specific magnetic susceptibility and RF coil attenuation. Most of the image post-processing methods for intensity inhomogeneity correction require segmentation of iso-tissue regions (therefore, prone to segmentation related errors and unreliability) and other methods do not apply direct analyses on spatial image domain (therefore, causes different artifacts and less reliable). Here, we present a new approach that directly works on spatial domain, requires no pre-segmentation or parametric model, and uses only local image gradient -- a low level information. The key idea is to distinguish the two components of the gradient at a point -- (1) slow background intensity variations, and (2) intensity variations due to “true edges.” It allows computing intensity inhomogeneity along a path by integrating slow intensity variations which subsequently, yields intensity inhomogeneity surface. The method requires no expert’s intervention and may easily be amended to an imaging system. The new method is applied on image slices taken from MR data of different body regions.
MR image quality improvement by zero-filling Fresnel transform imaging technique
Bin Rong Wu, Satoshi Ito, Yoshitsugu Kamimura, et al.
The quality of MR image is an important factor to control the accuracy of the diagnosis and treatment, so the image quality improvement methods, which can remove the noise and do not deteriorate the spatial resolution of images, are demanded in medical imaging field. There is an MR image reconstruction method, which applies the Fourier transform after a quadratic phase modulation in the Fresnel transform imaging technique. This method has a great property that can diffuse the noise contained in NMR signal by the quadratic phase modulation. If we prepare suitable area for diffusion of the noise by using zero-filling technique before reconstructing an image with this method, the spectrum of noise will evenly distribute over the zero filled area, and on the other hand, the image spectrum appear as an aspect of multi-resolution type image which locally distribute. After the noise was removed by a threshold filter in this aspect space, we return the data to the beginning signal space by doing a series of inverse processes, and reconstruct the denoised signal by using usual Fresnel transform image reconstruction technique. An MR image, which the noise is greatly improved, whereas the deterioration in spatial resolution is hardly caused, can be obtained. In this work, we present a new MR image quality improvement method that uses the property of noise diffusing in the Fresnel transform imaging technique; and we describe the effectiveness from simulations that are evaluated on SNR improvement and extent of deterioration in spatial resolution, and compare them with standard Wiener filtering and Wavelet Wiener filtering. Finally, we verify the powerfulness of the proposed method and it is applicable to phase-scrambled Fourier transform imaging technique through experiment.
Automatic cropping of breast regions for registration in MR mammography
Matthias Koenig, Sven Kohle, Heinz-Otto Peitgen
In MR mammography, usually the complete upper part of the body is recorded although for most diagnostic examinations and therapeutic planning only the regions around the breasts are important. This can lead to some disadvantages for automatic processing of images, e.g. higher time consumption and lesser accuracy in image registration. In this paper, we present a straight-forward method for automatic cropping of breast regions for registration in MR mammography. The method starts with processing two-dimensional slices. The result of this first step is statistically analyzed and a cropping-region for three-dimensional volumes is calculated. For each two-dimensional slice, the boundary between breast and air is identified by applying a threshold operator. This boundary describes a bimodal curve; the peaks for the right and left breast and the breastbone are found by searching among the curve's extremal points. An heuristic method analyzes this curve and further yields an estimation of the chest wall boundary to the inner body. For image registration purpose, we compare our proposed automatic cropping method with a simple cropping mechanism that defines the regions of the left and right breasts by reducing the volume to 60 percent from the margin of each side. It is shown that the proposed method works reliably and gives advantages regarding the time-quality tradeoff for automatic image registration in MR mammography.
Segmentation of hand radiographs by using multi-level connected active appearance models
Joost A. Kauffman, Cornelis H. Slump, Hein J. Bernelot Moens
Robust and accurate segmentation methods are important for the computerized evaluation of medical images. For treatment of rheumatoid arthritis, joint damage assessment in radiographs of hands is frequently used for monitoring disease progression. Current clinical scoring methods are based on visual measurements that are time-consuming and subject to intra and inter-reader variance. A solution may be found in the development of partially automated assessment procedures. This requires reliable segmentation algorithms. Our work demonstrates a segmentation method based on multiple connected active appearance models (AAM) with multiple search steps using different quality levels. The quality level can be regulated by setting the image resolution and the number of landmarks in the AAMs. We performed experiments using two models of different quality levels for shape and texture information. Both models included AAMs for the carpal region, the metacarpals, and all phalanges. By starting an iterative search with the faster, low-quality model, we were able to determine the initial parameters of the second, high-quality model. After the second search, the results showed successful segmentation for 22 of 30 test images. For these images, 70% of the landmarks were found within 1.3 mm difference from manual placement by an expert. The multi-level search approach resulted in a reduction of 50% in calculation time compared to a search using a single model. Results are expected to improve when the model is refined by increasing the number of training examples and the resolution of the models.
Cerebella segmentation on MR images of pediatric patients with medulloblastoma
In this study, an automated method has been developed to identify the cerebellum from T1-weighted MR brain images of patients with medulloblastoma. A new objective function that is similar to Gibbs free energy in classic physics was defined; and the brain structure delineation was viewed as a process of minimizing Gibbs free energy. We used a rigid-body registration and an active contour (snake) method to minimize the Gibbs free energy in this study. The method was applied to 20 patient data sets to generate cerebellum images and volumetric results. The generated cerebellum images were compared with two manually drawn results. Strong correlations were found between the automatically and manually generated volumetric results, the correlation coefficients with each of manual results were 0.971 and 0.974, respectively. The average Jaccard similarities with each of two manual results were 0.89 and 0.88, respectively. The average Kappa indexes with each of two manual results were 0.94 and 0.93, respectively. These results showed this method was both robust and accurate for cerebellum segmentation. The method may be applied to various research and clinical investigation in which cerebellum segmentation and quantitative MR measurement of cerebellum are needed.
Color lesion boundary detection using live wire
Artur Chodorowski, Ulf Mattsson, Morgan Langille, et al.
The boundaries of oral lesions in color images were detected using a live-wire method and compared to expert delineations. Multiple cost terms were analyzed for their inclusion in the final total cost function including color gradient magnitude, color gradient direction, Canny edge detection, and Laplacian zero crossing. The gradient magnitude and direction cost terms were implemented so that they acted directly on the three components of the color image, instead of using a single derived color band. The live-wire program was shown to be considerably more accurate and faster compared to manual segmentations by untrained users.
3D live-wire-based semi-automatic segmentation of medical images
Ghassan Hamarneh, Johnson Yang, Chris McIntosh, et al.
Segmenting anatomical structures from medical images is usually one of the most important initial steps in many applications, including visualization, computer-aided diagnosis, and morphometric analysis. Manual 2D segmentation suffers from operator variability and is tedious and time-consuming. These disadvantages are accentuated in 3D applications and, the additional requirement of producing intuitive displays to integrate 3D information for the user, makes manual segmentation even less approachable in 3D. Robust, automatic medical image segmentation in 2D to 3D remains an open problem caused particularly by sensitivity to low-level parameters of segmentation algorithms. Semi-automatic techniques present possible balanced solution where automation focuses on low-level computing-intensive tasks that can be hidden from the user, while manual inter- vention captures high-level expert knowledge nontrivial to capture algorithmically. In this paper we present a 3D extension to the 2D semi-automatic live-wire technique. Live-wire based contours generated semi-automatically on a selected set of slices are used as seed points on new unseen slices in different orientations. The seed points are calculated from intersections of user-based live-wire techniques with new slices. Our algorithm includes a step for ordering the live-wire seed points in the new slices, which is essential for subsequent multi-stage optimal path calculation. We present results of automatically detecting contours in new slices in 3D volumes from a variety of medical images.
Optic nerve head segmentation in multimodal retinal images
Radim Chrastek, Heinrich Niemann, Libor Kubecka, et al.
An established method for glaucoma diagnosis is the morphological analysis of the optic nerve head (ONH) by the scanning-laser-tomography (SLT). This analysis depends on prior manual outlining of the ONH. The first automated segmentation method that we developed is limited in its reliability by noise, non-uniform illumination and presence of blood vessels. Inspired by recent medical research we developed a new algorithm improving our previous method by segmenting in registered multimodal retinal images. The multimodal approach combines SLT-images with color fundus photographs (CFP). The first step of the algorithm, the registration, is based on gradient-image mutual information maximization using controlled random search as the optimization procedure. The kernel of the segmentation module consists in the anchored active contours. The initial contour is obtained from the CFP. The points the initial curve should be attracted to, the anchors, are constrained by the Hough transform applied to a morphologically processed SLT-image. The false anchors are eliminated by masking out blood vessels that are extracted in the CFP. The method was tested on 174 multimodal image pairs. The overall performance of the system yielded 89% correctly segmented ONH, qualitatively evaluated comparing the automated contours with manual ones drawn by an experienced ophthalmologist. This represents an appreciable improvement in reliability (from 74% to 89%) compared to monomodal approach. The developed method is the basis for a promising tool for glaucoma screening.
Automatic breast border extraction
In computer aided mammography algorithms there are several processing steps, which must be performed. The basic segmentation procedure involves extracting the principal feature on a mammogram; the breast border. This is performed by segmenting the breast and the non-breast into distinct regions. In this paper, a method for extracting the breast border is proposed. The method has performance similar to established techniques but with higher degrees of automatization and robustness. It iteratively adapts a model of the background to ensure a robust object detection yielding a smooth outline of the breast. The main idea is to identify the "knee" in the cumulative intensity histogram of the image. The intensity value at the knee is thereafter used to automatically define a region, to be modelled by a two-dimensional polynomial surface of degree two. The modelled background is then subtracted from the original image. The procedure described is iteratively performed until the degree of non-uniformity of the grey-scale background is smaller then a certain value. Thereafter the difference image is post-processed by a flood-filling algorithm, a new threshold is estimated as above and applied to yield a binary image. Lastly morphological operations are performed to smoothen the breast border. In conclusion, the strength in the proposed method, compared to similar methods, is that it makes use of an iterative approach to reduce the effects of the background, it produces smooth edges and automatically finds thresholds. It is also evaluated on the entire MIAS database (322 images) with a performance of 94%.
Bounding-object segmentation
Marcus Vetter, Jochen Neuhaus, Ingmar Wegner, et al.
The paper presents a new method for the semiautomatic segmentation of anatomical or pathological structures in MRI, CT or ultrasound images. The concept of bounding-object segmentation is based on an efficient combination of a new interactive approach with well known automatic segmentation algorithms. The efficiency of this new method is based on the transparent interaction between a 3D scene as well arbitrary 2D views of the scene. Bounding-object segmentation can also be described as a combination of interactive 3D segmentation with region-based, level-set-based, and/or texture based 3D-segmentation algorithms.
Assessment of similarity indices to quantify segmentation accuracy of scaffold images for tissue engineering
Existing similarity metrics to compare the accuracy of n-Dimensional image segmentation with the corresponding ground truth is restricted to a limited set of volume fractions which, by themselves, lack robustness. This paper introduces a comprehensive list of linear and non-linear similarity measures widely used in such diverse fields as ecology, toxicology and patent trending. These metrics based on the binary "absence/presence" data were computed for assessing the delineation of tissue engineering scaffold images into porous and polymeric space using a wide variety of thresholding techniques.
Prostate ultrasound image segmentation using level set-based region flow with shape guidance
Lixin Gong, Lydia Ng, Sayan Dev Pathak, et al.
Prostate segmentation in ultrasound images is a clinically important and technically challenging task. Despite several research attempts, few effective methods are available. One problem is the limited algorithmic robustness to common artifacts in clinical data sets. To improve the robustness, we have developed a hybrid level set method, which incorporates shape constraints into a region-based curve evolution process. The online segmentation method alternates between two steps, namely, shape model estimation (ME) and curve evolution (CE). The prior shape information is encoded in an implicit parametric model derived offline from manually outlined training data. Utilizing this prior shape information, the ME step tries to compute the maximum a posteriori estimate of the model parameters. The estimated shape is then used to guide the CE step, which in turn provides a new model initialization for the ME step. The process stops automatically when the curve locks onto the specific prostate shape. The ME and the CE steps complement each other to capture both global and local shape details. With shape guidance, this algorithm is less sensitive to initial contour placement and more robust even in the presence of large boundary gaps and strong clutter. Promising results are demonstrated on both synthetic and real prostate ultrasound images.
Semiautomatic segmentation of textured laser range scans for use in image-guided procedures
Logan W. Clements, David Marshall Cash, Tuhin K. Sinha, et al.
Laser range scanners produce high resolution surface data of anatomic structures, which facilitates the determination of intraoperative soft tissue deformation and the performance of surface based image-to-physical space registration. Segmentation of the range scans is required for the data to be effectively incorporated into current image-guided procedures. Due to time constraints in the operating room, manual segmentation methods are not feasible. We propose a novel segmentation algorithm based on the level set method that uses information from the texture map and curvature of the acquired point cloud to provide an accurate edge map for computation of the speed image. Specifically, the edge image is created by combining the curvature values, computed from a surface fitted to the acquired point cloud using radial basis functions, and gradients of the RGB intensities in the texture map. Preliminary results, obtained from comparing the semiautomatic segmentations of intraoperatively acquire liver LRS data with manual gold standard segmentations, shows the method to be a significant first step towards the implementation of semiautomatic LRS segmentation routine during image-guided surgery.
Estimation of three-dimensional knee joint movement using bi-plane x-ray fluoroscopy and 3D-CT
Hideaki Haneishi, Satoshi Fujita, Takahiro Kohno, et al.
Acquisition of exact information of three-dimensional knee joint movement is desired in plastic surgery. Conventional X-ray fluoroscopy provides dynamic but just two-dimensional projected image. On the other hand, three-dimensional CT provides three-dimensional but just static image. In this paper, a method for acquiring three-dimensional knee joint movement using both bi-plane, dynamic X-ray fluoroscopy and static three-dimensional CT is proposed. Basic idea is use of 2D/3D registration using digitally reconstructed radiograph (DRR) or virtual projection of CT data. Original ideal is not new but the application of bi-plane fluoroscopy to natural bones of knee is reported for the first time. The technique was applied to two volunteers and successful results were obtained. Accuracy evaluation through computer simulation and phantom experiment with a knee joint of a pig were also conducted.
3D live-wires on pre-segmented volume data
Sebastian Koenig, Juergen Hesser
This article discusses our approach for extending interactive Live-Wire segmentation on 3D for pre-segmented MRI volume data. We generate a mosaic image volume by region growing. Surfaces between two neighboring regions (called facets) are considered as nodes and the borders where two or more surfaces meet are defined as edges thus converting the mosaic image into a graph onto which the 3D-live-wire operates. Due to the pre-segmentation the graph is concentrated to the areas of strong gradients which are the regions of interest for finding object outlines. The cost optimal path between two graph nodes can be calculated by applying a standard Live-Wire algorithm to this graph data. To get a surface patch without holes from three user defined seeds, an initial frame is calculated in a first step. This frame consists of the three paths connecting the seeds. By calculating the paths between a few additional facets taken from the initial frame we obtain supporting paths which increase the quality of segmentation. The remaining gaps in the resulting grating are closed in three steps. Firstly, we detect object and background regions of the grating. Next, these regions are expanded by adding similar unlabeled regions. And finally, we add facets enclosed by object and background regions to the grating until all gaps are closed. The preprocessing is done in about 188 seconds and the interactive segmentation in our examples takes about 8.4 seconds. We achieve a mean deviation from the correct object boundary extracted from available model data of about 0.2 voxels.
Segmentation II: Vasculature
icon_mobile_dropdown
Coronary vessel cores from 3D imagery: a topological approach
Andrzej Szymczak, Allen Tannenbaum, Konstantin Mischaikow
We propose a simple method for reconstructing thin, low-contrast blood vessels from three-dimensional greyscale images. Our algorithm first extracts persistent maxima of the intensity on all axis-aligned two-dimensional slices through the input volume. Those maxima tend to concentrate along one-dimensional intensity ridges, in particular along blood vessels. Persistence (which can be viewed as a measure of robustness of a local maximum with respect to perturbations of the data) allows to filter out the `unimportant' maxima due to noise or inaccuracy in the input volume. We then build a minimum forest based on the persistent maxima that uses edges of length smaller than a certain threshold. Because of the distribution of the robust maxima, the structure of this forest already reflects the structure of the blood vessels. We apply three simple geometric filters to the forest in order to improve its quality. The first filter removes short branches from the forest's trees. The second filter adds edges, longer than the edge length threshold used earlier, that join what appears (based on geometric criteria) to be pieces of the same blood vessel to the forest. Such disconnected pieces often result from non-uniformity of contrast along a blood vessel. Finally, we let the user select the tree of interest by clicking near its root (point from which blood would flow out into the tree). We compute the blood flow direction assuming that the tree is of the correct structure and cut it in places where the vessel's geometry would force the blood flow direction to change abruptly. Experiments on clinical CT scans show that our technique can be a useful tool for segmentation of thin and low contrast blood vessels. In particular, we successfully applied it to extract coronary arteries from heart CT scans. Volumetric 3D models of blood vessels can be obtained from the graph described above by adaptive thresholding.
Poster Session II
icon_mobile_dropdown
A comparison of the tissue classification and the segmentation propagation techniques in MRI brain image segmentation
Jinsong Ren, Beatrix Sneller, Daniel Rueckert, et al.
Tissue classifications of the MRI brain images can either be obtained by segmenting the images or propagating the segmentations of the atlas to the target image. This paper compares the classification results of the direct segmentation method using FAST with those of the segmentation propagation method using nreg and the MNI Brainweb phantom images. The direct segmentation is carried out by extracting the brain and classifying the tissues by FAST. The segmentation propagation is carried out by registering the Brainweb atlas image to the target images by affine registration, followed by non-rigid registration at different control spacing, then transforming the PVE (partial volume effect) fuzzy membership images of cerebrospinal fluid (CSF), grey matter (GM) and white matter (WM) of the atlas image into the target space respectively. We have compared the running time, reproducibility, global and local differences between the two methods. Direct segmentation is much faster. There is no significant difference in reproducibility between the two techniques. There are significant global volume differences on some tissue types between them. Visual inspection was used to localize these differences. This study had no gold standard segmentations with which to compare the automatic segmentation solutions, but the global and local volume differences suggest that the most appropriate algorithm is likely to be application dependent.
Interactive constraints for 3D-simplex meshes
Thomas Boettger, Tobias Kunert, Hans-Peter Meinzer, et al.
Medical image segmentation is still a very time consuming task and therefore not often integrated into clinical routine. Various 3D segmentation approaches promise to facilitate the work. But they are rarely used in clinical setups due to complex intialization and parametrization of such models. Clinical users need interactive tools, intuitive and easy to handle. They do not want to play around with a set of parameters which will differ from dataset to dataset and often have a non-intuitive meaning. In this work new interactive constraints for deformable three-dimensional 2-simplex meshes are presented. The user can define attracting points in the original image data. These attractors are considered during model deformation and the new forces guarantee that the surface model will pass through these interactively set points. By using the constraints the model parameterization is simplified. Segmentation is started with a spherical surface model which is placed inside the structure of interest and then adapts to the boundaries. The user can directly influence the evolution of the deformable model and gets direct feedback during the segmentation process. The model deformation algorithm was implemented and integrated in ITK (Insight Segmentation and Registration Toolkit). The newly developed segmentation tool was tested on cardiac image data and MRI lung images, but is suitable for any kind of 3D and 3D+t medical image data. It has been shown that the model is less sensitive to preprocessing of the input data as well as model initialization.
Expectation maximization approach to vessel enhancement in thoracic CT scans
Vessel enhancement in volumetric data is a necessary prerequisite in various medical imaging applications. In the context of automated lung nodule detection in thoracic CT scans, segmented blood vessels can be used to resolve local ambiguities based on global considerations and so improve the performance of lung nodule detection algorithms. Segmenting the data correctly is a difficult problem with direct consequences for subsequent processing steps. Voxels belonging to vessels and nodules in thoracic CT scans are both characterized by high contrast with respect to a local neighborhood. Thus in order to enhance vessels while suppressing nodules, additional characteristics should be used. In this paper we propose a novel vessel enhancement filter that is capable of enhancing vessels and junctions in thoracic CT scans while suppressing nodules. The proposed filters are based on a Gaussian mixture model which is optimized through expectation maximization. The proposed filters are based on first order differential quantities and so are less sensitive to noise compared with known Hessian-based vessel enhancement filters. Moreover, the proposed filters utilize an adaptive window and so avoid the common need for multiple scale analysis. The proposed filters are evaluated and compared to known techniques qualitatively and quantitatively on both synthetic and actual clinical data and it is shown that the proposed filters perform better.
Semiautomatic segmentation of the heart from CT images based on intensity and morphological features
The incidence of certain types of cardiac arrhythmias is increasing. Effective, minimally invasive treatment has remained elusive. Pharmacologic treatment has been limited by drug intolerance and recurrence of disease. Catheter based ablation has been moderately successful in treating certain types of cardiac arrhythmias, including typical atrial flutter and fibrillation, but there remains a relatively high rate of recurrence. Additional side effects associated with cardiac ablation procedures include stroke, perivascular lung damage, and skin burns caused by x-ray fluoroscopy. Access to patient specific 3-D cardiac images has potential to significantly improve the process of cardiac ablation by providing the physician with a volume visualization of the heart. This would facilitate more effective guidance of the catheter, increase the accuracy of the ablative process, and eliminate or minimize the damage to surrounding tissue. In this study, a semiautomatic method for faithful cardiac segmentation was investigated using Analyze - a comprehensive processing software package developed at the Biomedical Imaging Resource, Mayo Clinic. This method included use of interactive segmentation based on math morphology and separation of the chambers based on morphological connections. The external surfaces of the hearts were readily segmented, while accurate separation of individual chambers was a challenge. Nonetheless, a skilled operator could manage the task in a few minutes. Useful improvements suggested in this paper would give this method a promising future.
Cerebrovascular plaque segmentation using object class uncertainty snake in MR images
Bipul Das, Punam Kumar Saha, Ronald Wolf, et al.
Atherosclerotic cerebrovascular disease leads to formation of lipid-laden plaques that can form emboli when ruptured causing blockage to cerebral vessels. The clinical manifestation of this event sequence is stroke; a leading cause of disability and death. In vivo MR imaging provides detailed image of vascular architecture for the carotid artery making it suitable for analysis of morphological features. Assessing the status of carotid arteries that supplies blood to the brain is of primary interest to such investigations. Reproducible quantification of carotid artery dimensions in MR images is essential for plaque analysis. Manual segmentation being the only method presently makes it time consuming and sensitive to inter and intra observer variability. This paper presents a deformable model for lumen and vessel wall segmentation of carotid artery from MR images. The major challenges of carotid artery segmentation are (a) low signal-to-noise ratio, (b) background intensity inhomogeneity and (c) indistinct inner and/or outer vessel wall. We propose a new, effective object-class uncertainty based deformable model with additional features tailored toward this specific application. Object-class uncertainty optimally utilizes MR intensity characteristics of various anatomic entities that enable the snake to avert leakage through fuzzy boundaries. To strengthen the deformable model for this application, some other properties are attributed to it in the form of (1) fully arc-based deformation using a Gaussian model to maximally exploit vessel wall smoothness, (2) construction of a forbidden region for outer-wall segmentation to reduce interferences by prominent lumen features and (3) arc-based landmark for efficient user interaction. The algorithm has been tested upon T1- and PD-weighted images. Measures of lumen area and vessel wall area are computed from segmented data of 10 patient MR images and their accuracy and reproducibility are examined. These results correspond exceptionally well with manual segmentation completed by radiology experts. Reproducibility of the proposed method is estimated for both intra- and inter-operator studies.
Evaluation of ischemic stroke hybrid segmentation in a rat model of temporary middle cerebral artery occlusion using ground truth from histologic and MR data
A segmentation method that quantifies cerebral infarct using rat data with ischemic stroke is evaluated using ground truth from histologic and MR data. To demonstrate alternative approach to rapid quantification of cerebral infarct volumes using histologic stained slices that requires scarifying animal life, a study with MR acquire volumetric rat data is proposed where ground truth is obtained by manual delineations by experts and automated segmentation is assessed for accuracy. A framework for evaluation of segmentation is used that provides more detailed accuracy measurements than mere cerebral infarct volume. Our preliminary experiment shows that ground truth derived from MRI data is at least as good as the one obtained from the histologic slices for evaluating segmentation algorithms for accuracy. Therefore we can develop and evaluate automated segmentation methods for rapid quantification of stroke without the necessitating animal sacrifice.
A software framework for preprocessing and level set segmentation of medical image data
Karl David Fritscher, Rainer Schubert M.D.
In this work a software platform for semiautomatic segmentation of medical images based on geometric deformable models will be presented. Including filters for image preprocessing, image segmentation and 3D visualization this toolkit offers the possibility of creating highly effective segmentation pipelines by combining classic segmentation techniques like seeded region growing and manual segmentation with modern level set segmentation algorithms. By individually combining input and output of different segmentation methods, specific and at the same time easy to use segmentation pipelines can be created. Using open source libraries for the implementation of a number of frequently used preprocessing and segmentation algorithms allowed effective programming by at the same time providing stable and highly effective algorithms. The usage of modern programming standards and developing cross-platform algorithm classes guarantees extensibility and flexible implementation in different hard- and software settings. Segmentation results, created in different research projects will be presented and the efficient usage of this framework will be demonstrated. The implementation of parts of the framework in a clinical setting is in progress and currently we are working on the embedding of statistical models and prior knowledge in the segmentation framework.
Fast algorithm for probabilistic bone edge detection (FAPBED)
Danilo Scepanovic, Joshua Kirshtein, Ameet Kumar Jain, et al.
The registration of preoperative CT to intra-operative reality systems is a crucial step in Computer Assisted Orthopedic Surgery (CAOS). The intra-operative sensors include 3D digitizers, fiducials, X-rays and Ultrasound (US). FAPBED is designed to process CT volumes for registration to tracked US data. Tracked US is advantageous because it is real time, noninvasive, and non-ionizing, but it is also known to have inherent inaccuracies which create the need to develop a framework that is robust to various uncertainties, and can be useful in US-CT registration. Furthermore, conventional registration methods depend on accurate and absolute segmentation. Our proposed probabilistic framework addresses the segmentation-registration duality, wherein exact segmentation is not a prerequisite to achieve accurate registration. In this paper, we develop a method for fast and automatic probabilistic bone surface (edge) detection in CT images. Various features that influence the likelihood of the surface at each spatial coordinate are combined using a simple probabilistic framework, which strikes a fair balance between a high-level understanding of features in an image and the low-level number crunching of standard image processing techniques. The algorithm evaluates different features for detecting the probability of a bone surface at each voxel, and compounds the results of these methods to yield a final, low-noise, probability map of bone surfaces in the volume. Such a probability map can then be used in conjunction with a similar map from tracked intra-operative US to achieve accurate registration. Eight sample pelvic CT scans were used to extract feature parameters and validate the final probability maps. An un-optimized fully automatic Matlab code runs in five minutes per CT volume on average, and was validated by comparison against hand-segmented gold standards. The mean probability assigned to nonzero surface points was 0.8, while nonzero non-surface points had a mean value of 0.38 indicating clear identification of surface points on average. The segmentation was also sufficiently crisp, with a full width at half maximum (FWHM) value of 1.51 voxels.
Fast and robust diaphragm detection and tracking in cardiac x-ray projection images
Alexandru Condurache, Til Aach, Kai Eck, et al.
A number of image analysis tasks of the heart region have to cope with both the problem of respiration and heart contraction. While the heart contraction status can be estimated based on the ECG, respiration status estimation must be based on the images themselves, unless additional devices for respiration measurements are used. Since diaphragm motion is closely linked to respiration, we describe a method to detect and track the diaphragm in x-ray projections. We model the diaphragm boundary as being approximately circular. Diaphragm detection is then based on edge detection followed by a Hough transform for circles. To avoid that the detection algorithm is misled by high frequency image content, we first apply a morphological multi-scale top hat operator. A Canny edge detector is then applied to the top hat filtered images. In the edge images, the circle corresponding to the diaphragm boundary is found by the Hough transform. To restrict the search in the 3D Hough parameter space (parameters are circle center coordinates and radius), prior anatomical knowledge about position and size of the diaphragm for the given image acquisition geometry is taken into account. In subsequent frames, diaphragm position and size are predicted from previous detection and tracking results. For each detection result, a confidence measure is computed by analyzing the Hough parameter space with respect to the goodness of the peak giving the circle parameters and by analyzing the coefficient of variation of the pixel that form the circle described by the maximum in Hough parameter space. If the confidence is not sufficiently high -- indicating a poor fit between the Hough circle and true diaphragm boundary -- the detection result is optionally refined by an active contour algorithm.
Automated lung segmentation in magnetic resonance images
Segmentation of the lungs within magnetic resonance (MR) scans is a necessary preprocessing step in the computerized analysis of thoracic MR images. This task is complicated by potentially significant cardiac and pulmonary motion artifacts, partial volume effect, and morphological deformation from disease. We have developed an automated segmentation method to account for these complications. First, the thorax is segmented using a threshold obtained from analysis of the cumulative gray-level histogram constructed along a diagonal line through the center of the image. Next two separate lung-thresholded images are created. The first lung-thresholded image is created using histogram-based gray-level thresholding techniques applied to the segmented thorax. To include lung areas that may be adversely affected by artifact or disease, a second lung-thresholded image is created by applying a grayscale erosion operator to the first lung-thresholded image. After a rolling ball filter is applied to the lung contour to eliminate non-lung pixels from the thresholded lung regions, a logical OR operation is used to combine the two lung-thresholded images into the final segmented lung regions. Modifications to this approach were required to properly segment sections in the lung bases. In a preliminary evaluation, the automated method was applied to 10 MR scans, an observer evaluated the segmented lung regions using a five-point scale (“highly accurate segmentation” to “highly inaccurate segmentation”). Eighty-five percent of the segmented lung regions were rated as highly or moderately accurate.
Automatic segmentation and registration for the lung nodule matching in temporal chest CT scans
To investigate changes of pulmonary nodules in temporal chest CT scans, we propose a novel technique for segmentation and registration of lungs. Our method is composed of the following steps. First, automatic segmentation is used to identify lungs in chest CT scans. Second, optimal cube registration is performed to correct gross translational mismatch of lungs. This initial registration does not require any anatomical landmarks. Third, a 3D distance map is generated by the narrow-band distance propagation, which drives fast and robust convergence to the optimum value. Fourth, the distance measure between surface boundary points is evaluated repeatedly by the selective distance measure (SDM). Then the final geometrical transformations are applied to ten pairs of successive chest CT scans. Fifth, nodule correspondences are established by the pairs with the smallest Euclidean distances. The performance of our method was evaluated with the aspects of visual inspection and accuracy. The positional differences between lungs of initial and follow-up CT scans were much reduced by the optimal cube registration. Then this initial alignment was refined by the subsequent iterative surface registration. For accuracy assessment, we have evaluated a root-mean-square (RMS) error between corresponding nodules on a per-center basis. The reduction of RMS error was obtained with the optimal cube registration, subsequent iterative surface registration and nodule registration. Experimental results show that our segmentation and registration method extracts accurate lungs and aligns them much faster than the conventional ones using a distance measure. Accurate and fast result of our method would be more useful for the radiologist’s evaluation of pulmonary nodules on chest CT scans.
Model-based 3D segmentation of the bones of joints in medical images
There are several medical application areas that require the segmentation and separation of the component bones of joints in a sequence of acquired images of the joint under various loading conditions, our own target area being joint motion analysis. This is a challenging problem due to the proximity of bones at the joint, partial volume effects, and other imaging modality-specific factors that confound boundary contrast. A model-based strategy is proposed in this paper wherein a rigid model of the bone is generated from a segmentation of the bone in the image corresponding to one position of the joint by using the live wire method. In other images of the joint, this model is used to search for the same bone by minimizing an energy functional that utilizes both boundary- and region-based information. An evaluation of the method by utilizing a total of 60 data sets on MR and CT images of the ankle complex and cervical spine indicates that the segmentations agree very closely with the live wire segmentations yielding true positive and false positive volume fractions in the range 89-97% and 0.2-0.7%. The method requires 1-2 minutes of operator time and 6-7 minutes of computer time, which makes it significantly more efficient than live wire -- the only method currently available for the task.
Atlas based automatic identification of abdominal organs
Due to intensity inhomogeneities, partial volume effects, as well as organ shape variations, automatic segmentation of abdominal organs has always been a high challenging task. To conquer these difficulties, we employ a pre-labeled atlas (VIP-Man) to supplement anatomical knowledge to the segmentation process. First, an atlas-subject registration is carried out to establish the proper correspondence between the atlas and the subject. The registration consists of two steps. In the global registration step, a similarity transformation is found to eliminate the stature differences. In the organ registration step, organs of interest are registered respectively to achieve better alignments. Second, we utilize the fuzzy connectedness framework to segment abdominal organs of interest from the subject image. Under the guidance of the registered atlas, the seeds and intensity parameters of organs are specified in an auto-adaptive way. Further more, the anatomical knowledge contained in the atlas is blended into the frame work, to make the segmentation result more reliable. To remove possible jags on boundary, a level set smooth method which implements fuzzy connectedness as external speed forces, is utilized on the segmentation result. Our purpose is to accomplish the segmentation task like how anatomy experts do. So far, this approach has been applied to segment organs, including liver, spleen and kidneys, in the female MRI T1 data set from the VHP. All organs of interest are identified correctly, and delineated with considerable precision.
Mass segmentation of dense breasts on digitized mammograms: analysis of a probability-based function
Lisa M. Kinnard, Shih-Chung Benedict Lo, Eva Duckett, et al.
In this study, a segmentation algorithm based on the steepest changes of a probabilistic cost function was tested on non-processed and pre-processed dense breast images in an attempt to determine the efficacy of pre-processing for dense breast masses. Also, the inter-observer variability between expert radiologists was studied. Background trend correction was used as the pre-processing method. The algorithm, based on searching the steepest changes on a probabilistic cost function, was tested on 107 cancerous masses and 98 benign masses with density ratings of 3 or 4 according to the American College of Radiology's density rating scale. The computer-segmented results were validated using the following statistics: overlap, accuracy, sensitivity, specificity, Dice similarity index, and kappa. The mean accuracy statistic value ranged from 0.71 to 0.84 for cancer cases and 0.81 to 0.86 for benign cases. For nearly all statistics there were statistically significant differences between the expert radiologists.
Free software tools for atlas-based volumetric neuroimage analysis
Pierre-Louis Bazin, Dzung L. Pham, William Gandler, et al.
We describe new and freely available software tools for measuring volumes in subregions of the brain. The method is fast, flexible, and employs well-studied techniques based on the Talairach-Tournoux atlas. The software tools are released as plug-ins for MIPAV, a freely available and user-friendly image analysis software package developed by the National Institutes of Health. Our software tools include a digital Talairach atlas that consists of labels for 148 different substructures of the brain at various scales.
Evaluation of a nonlinear diffusion process for segmentation and quantification of lesions in optical coherence tomography images
Delia Cabrera Fernandez, Harry M. Salinas
We evaluate the ability of a nonlinear anisotropic diffusion process to enhance the contrast for structural boundary regions and to reduce the speckle noise in optical coherence tomography (OCT) images. We also investigate the suitability of various image features, such as gradient magnitude and intensity or gradient profiles, for boundary localization. The results suggest that the nonlinear anisotropic diffusion method has potential in assisting segmentation and quantification of fluid-filled regions in clinical OCT images.
Automatic segmentation of the left ventricle and computation of diagnostic parameters using regiongrowing and a statistical model
The manual segmentation and analysis of high-resolution multi-slice cardiac CT datasets is both labor intensive and time consuming. Therefore it is necessary to supply the cardiologist with powerful software tools to segment the myocardium and compute the relevant diagnostic parameters. In this work we present a semi-automatic cardiac segmentation approach with minimal user interaction. It is based on a combination of an adaptive slice-based regiongrowing and a modified Active Shape Model (ASM). Starting with a single manual click point in the ascending aorta, the aorta, the left atrium and the left ventricle get segmented with the slice-based adaptive regiongrowing. The approximate position of the aortic and mitral valve as well as the principal axes of the left ventricle (LV) are determined. To prevent the regiongrowing from draining into neighboring anatomical structures via CT artifacts, we implemented a draining control by examining a cubic region around the currently processed voxel. Additionally, we use moment-based parameters to integrate simple anatomical knowledge into the regiongrowing process. Using the results of the preceding regiongrowing process, a ventricle-centric and normalized coordinate system is established which is used to adapt a previously trained ASM to the image, using an iterative multi-resolution approach. After fitting the ASM to the image, we can use the generated model-points to create an exact surface model of the left ventricular myocardium for visualization and for computing the diagnostically relevant parameters, like the ventricular blood volume and the myocardial wall thickness.
Shape-based 3D vascular tree extraction for perforator flaps
Perforator flaps have been increasingly used in the past few years for trauma and reconstructive surgical cases. With the thinned perforated flaps, greater survivability and decrease in donor site morbidity have been reported. Knowledge of the 3D vascular tree will provide insight information about the dissection region, vascular territory, and fascia levels. This paper presents a scheme of shape-based 3D vascular tree reconstruction of perforator flaps for plastic surgery planning, which overcomes the deficiencies of current existing shape-based interpolation methods by applying rotation and 3D repairing. The scheme has the ability to restore the broken parts of the perforator vascular tree by using a probability-based adaptive connection point search (PACPS) algorithm with minimum human intervention. The experimental results evaluated by both synthetic and 39 harvested cadaver perforator flaps show the promise and potential of proposed scheme for plastic surgery planning.
Propagating labels of the human brain based on non-rigid MR image registration: an evaluation
Rolf A. Heckemann, Joseph V. Hajnal, Daniel Rueckert, et al.
Background: In order to perform statistical analysis of cohorts based on images, reliable methods for automated anatomical segmentation are required. Label propagation (LP) from manually segmented atlases onto newly acquired images is a particularly promising approach. Methods: We investigated LP on a set of 6 three-dimensional T1-weighted magnetic resonance data sets of the brains of normal individuals. For each image, a manually prepared segmentation of 67 structures was available. Each subject image was used in turn as an atlas and registered non-rigidly to each other subject's image. The resulting transformations were applied to the label sets, yielding five different generated segmentations for each subject, which we compared with the native manual segmentations using an overlap measure (similarity index, SI). We then reviewed the LP results for five structures with varied anatomical and label characteristics visually to determine how the registration procedure had affected the delineation of their boundaries. Results: The majority of structures propagated well as measured by SI (SI > 70 in 80% of measurements). Boundaries that were marked in the atlas image by definite intensity differences were congruent, with good agreement between the manual and the generated segmentations. Some boundaries in the manual segmentation were defined as planes marked by landmarks; such boundaries showed greater mismatch. In some cases, the proximity of structures with similar intensity distorted the LP results: e.g., parts of the parahippocampal gyrus were labeled as hippocampus in two cases. Conclusion: The size and shape of anatomical structures can be determined reliably using label propagation, especially where boundaries are defined by distinct differences in grey scale image intensity. These results will inform further work to evaluate potential clinical uses of information extracted from images in this way.
Analysis of blood and bone marrow smears using multispectral imaging analysis techniques
Qiongshui Wu, Libo Zeng, Hengyu Ke, et al.
Counting of different classes of white blood cells in bone marrow smears can give pathologists valuable information regarding various cancers. But it is tedious to manually locate, identify, and count these classes of cells, even by skilled hands. This paper presents a novel approach for automatic detection of White Blood Cells in bone marrow microscopic images. Different from traditional color imaging method, we use multispectral imaging techniques for image acquisition. The combination of conventional digital imaging with spectroscopy can provide us with additional useful spectral information in common pathological samples. With our spectral calibration method, device-independent images can be acquired, which is almost impossible in conventional color imaging method. A novel segmentation algorithm using spectral operation is presented in this paper. Experiments show that the segmentation is robust, precise, with low computational cost and insensitive to smear staining and illumination condition. Once the nuclei and cytoplasm have been segmented, more than a hundred of features are extracted under the direction of a pathologist, including shape features, textural features and spectral ratio features. In pattern recognition, a maximum likelihood classifier (MLC) is implemented in a hierarchical tree. The classification results are also discussed. This paper is focused on image acquisition and segmentation.
Tracking neural stem cells in time-lapse microscopy image sequences
This paper describes an algorithm for tracking neural stem/progenitor cells in a time-lapse microscopy image sequence. The cells were segmented in a semiautomatic way using dynamic programming. Since the interesting cells were identified by fluorescent staining at the end of the sequence, the tracking was performed backwards. The number of detected cells varied throughout the sequence: cells could appear or disappear at the image boundaries or at cell clusters, some cells split, and the segmentation was not always correct. To solve this asymmetric assignment problem, a modified version of the auction algorithm by Bertsekas was used. The assignment weights were calculated based on distance, correlation and size between possible matching cells. Cell splits are of special interest, therefore tracks without a matching cell were divided into two groups: 1. Merging cells (splitting cells, moving forward in time) and 2. Non-merging cells. These groups were separated based on difference in size of the involved cells, and difference in image intensity of the contour and interior of the possibly merged cell. The tracking algorithm was evaluated using a sequence consisting of 57 images, each image containing approximately 50 cells. The evaluation showed that 99% of the cell-to-cell associations were correct. In most cases, only one association per track was incorrect so in total 55 out of 78 different tracks in the sequence were tracked correctly. Further improvements will be to apply interleaved segmentation and tracking to produce a more reliable segmentation as well as better tracking results.
Digital staining of pathological tissue specimens using spectral transmittance
Staining of tissue specimens is a classical procedure in pathological diagnosis to enhance the contrast between tissue components such that identification and classification of these components can be easily performed. In this paper, a framework for digital staining of pathological specimens using the information derived from the L-band spectral transmittance of various pathological tissue components is introduced, particularly the transformation of a Hematoxylin and Eosin (HE) stained specimen to its Masson-Trichrome (MT) stained counterpart. The digital staining framework involves the classification of tissue components, which are highlighted when the specimen is actually stained with MT stain, e.g. fibrosis, from the HE-stained image; and the linear mapping between specific sets of HE and MT stained transmittance spectra through pseudo-inverse procedure to produce the LxL transformation matrices that will be used to transform the HE stained transmittance to its equivalent MT stained transmittance configuration. To generate the digitally stained image, the decisions of multiple quadratic classifiers are pooled to form the weighting factors for the transformation matrices. Initial results of our experiments on liver specimens show the viability of multispectral imaging (MSI) for the implementation of digital staining in the pathological context.
Geometric flows for vascular segmentation
Yongqiang Zhao, Minglu Li
Accurate description of vasculature structure plays an important role in many clinical applications. The purpose of this paper is to provide a new method based on geometric flows for vessel extraction from magnetic resonance angiography (MRA) images. This method is based on recent surface evolution work which models the object boundary as a manifold that evolves to maximize the rate of increase of flux of an appropriate vector field, and the geodesic active contour is used for regularization. In addition, to improve the insufficient of geometrical description of blood vessel structure, the centerlines of the vascular structure are regarded as space curves, and a tube of small radius around the centerline can be regarded as a distance function. Furthermore, the method uses the level set method to represent the surface evolution as it is intrinsic and topologically flexible. Results on cases demonstrate the effectiveness and accuracy of the approach comparing with the maximum intensity projection (MIP) and other methods.
On the use of lossless integer wavelet transforms in medical image segmentation
Recent trends in medical image processing involve computationally intensive processing techniques on large data sets, especially for 3D applications such as segmentation, registration, volume rendering etc. Multi-resolution image processing techniques have been used in order to speed-up these methods. However, all well-known techniques currently used in multi-resolution medical image processing rely on using Gaussain-based or other equivalent floating point representations that are lossy and irreversible. In this paper, we study the use of Integer Wavelet Transforms (IWT) to address the issue of lossless representation and reversible reconstruction for such medical image processing applications while still retaining all the benefits which floating-point transforms offer such as high speed and efficient memory usage. In particular, we consider three low-complexity reversible wavelet transforms namely the - Lazy-wavelet, the Haar wavelet or (1,1) and the S+P transform as against the Gaussian filter for multi-resolution speed-up of an automatic bone removal algorithm for abdomen CT Angiography. Perfect-reconstruction integer wavelet filters have the ability to perfectly recover the original data set at any step in the application. An additional advantage with the reversible wavelet representation is that it is suitable for lossless compression for purposes of storage, archiving and fast retrieval. Given the fact that even a slight loss of information in medical image processing can be detrimental to diagnostic accuracy, IWTs seem to be the ideal choice for multi-resolution based medical image segmentation algorithms. These could also be useful for other medical image processing methods.
Computing the thickness of the ventricular heart wall from 3D MRI images
A method for measuring the thickness of the ventricular heart wall from 3D MRI images is presented. The quantification of thickness could be useful clinically to measure the health of the heart muscle. The method involves extending a Laplace-equation-based definition of thickness between two surfaces to the ventricular heart wall geometry. Based on the functional organization of the heart, it is proposed that the heart be segmented into two volumes, the left ventricular wall which completely encloses the left ventricle and the right ventricular wall which attaches to the left ventricular wall to enclose the right ventricle, and that the thickness of these two volumes be calculated separately. An algorithm for performing this segmentation automatically is presented. The results of the automatic segmentation algorithm were compared to the results of manual segmentations of both normal and failing hearts and an average of 99.28% of ventricular wall voxels were assigned the same label in both the automatic and the manual segmentations. The thickness of eleven hearts, seven normal and four failing was measured.
Vessel addition using fuzzy technique in CT angiography
CT angiography (CTA) is increasingly used for vascular disease assessment because of its non-invasive characteristics. In order to get a comprehensive overview of the vascular anatomy, the bone has to be removed since it occludes the cranial vessels. One of the commonly used algorithms is bone subtraction, which obtains vessel images by subtracting pre-contrast images from post-contrast images. The current problems are that it removes too much vessel and sometimes pieces of bone still exist near vessel. The purpose of this study is to provide radiologist with a fuzzy technique tool to add back parts of missing vessel. A seed point is put on part of the vessel that was not well preserved and an area is selected to restrict the vessel growing. Vessel extraction is based on fuzzy-connectedness technique proposed by Udupa in 1996. Incorporating intensity information from both pre-contrast and post-contrast images creates the membership images. The value of each voxel in the membership images represents strength of fuzzy connectedness. The bigger the strength value, the more likely the voxel belongs to the classified vessel. After choosing a threshold for the strength, the vessel is extracted and added back. This method may also apply to the whole images to segment out the bone and the vessel. The study will improve the current vessel extraction and bone removal algorithms and provide a good tool for aiding radiologist to diagnose vascular diseases.
Classification and calculation of breast fibroglandular tissue volume on SPGR fat suppressed MRI
Jianhua Yao, Jo Anne Zujewski, Jennifer Orzano, et al.
This paper presents an automatic method to classify and quantify breast fibroglandular tissues on T1 weighted spoiled gradient-echo (SPGR) fat suppressed MRI. The breast region is segmented from the image using mathematical morphology, region growing, and active contour models. The breast-air and breast-chest wall boundaries are located using smooth and continuous curves. Three tissue types are defined: fatty tissue, fibroglandular tissue, and skin. We then employ a fuzzy C-means (FCM) method for tissue classification. For each pixel inside the breast region, the normalized pixel intensity and normalized distance to the breast-air boundary are computed. These two values form a two-dimensional feature space. A fuzzy class is defined for each tissue type. The initial centroid for each class is obtained from training images. The pixel membership values indicate the possibility of a pixel belonging to a certain tissue class. Pixels with highest membership in the fibroglandular class are then classified as fibroglandular tissue. We have tested our method on 29 patients. We automatically segmented the breasts and computed the volume percentage of fibroglandular tissue for both left and right breasts. We then compared the calculated tissue classification with manually generated tissue classification by two experienced radiologists. The two results agreed on 94.95% of breast segmentation, and the average fibroglandular percentage difference is about 3%. This method is useful in research studies assessing breast cancer risk.
Solitary pulmonary nodule characterization on CT by use of contrast enhancement maps
Sumit K. Shah, Michael F. McNitt-Gray, Iva Petkovska, et al.
Studies have shown that vascular structure of a solitary pulmonary nodule (SPN) can give insight into the diagnosis of the nodule. The purpose of this study is to investigate the utility of texture analysis as a quantitative measure of the vascular structure of a nodule. A contrast CT study was conducted for 29 patients with an indeterminate SPN. For each patient, the post-contrast series at maximum enhancement was volumetrically registered to the pre-contrast series. The two registered series were subtracted to form difference images of the nodule and each voxel was color-coded into 7 bins. Initially, a representative image of each nodule was subjectively rated on a five-point by a radiologist as to the magnitude, extent, and heterogeneity of the enhancement. From the initial analysis the heterogeneity of the nodule was found to be significantly different for benign versus malignant nodules (p<0.01), while the other two ratings were found not to be significant. We then attempted to quantify this subjective rating of heterogeneity by calculating 14 textural features based on co-occurrence matrices. These features included various measures of contrast, entropy, energy, etc. Dimension reduction techniques such as principal component and factor analysis were applied to the features to reduce the 14 variables to one factor. The mean of this factor was significantly different for malignant versus benign nodules (p=0.010). Texture analysis of contrast enhancement maps appears to be useful tool to characterize SPNs.
Quantitative analysis of packed and compacted granular systems by x-ray microtomography
Xiaowei Fu, Georgina E. Milroy, Meenakshi Dutt, et al.
The packing and compaction of powders are general processes in pharmaceutical, food, ceramic and powder metallurgy industries. Understanding how particles pack in a confined space and how powders behave during compaction is crucial for producing high quality products. This paper outlines a new technique, based on modern desktop X-ray tomography and image processing, to quantitatively investigate the packing of particles in the process of powder compaction and provide great insights on how powder densify during powder compaction, which relate in terms of materials properties and processing conditions to tablet manufacture by compaction. A variety of powder systems were considered, which include glass, sugar, NaCl, with a typical particle size of 200-300 μm and binary mixtures of NaCl-Glass Spheres. The results are new and have been validated by SEM observation and numerical simulations using discrete element methods (DEM). The research demonstrates that XMT technique has the potential in further investigating of pharmaceutical processing and even verifying other physical models on complex packing.
Optimizing connected component labeling algorithms
Kesheng Wu, Ekow Otoo, Arie Shoshani
This paper presents two new strategies that can be used to greatly improve the speed of connected component labeling algorithms. To assign a label to a new object, most connected component labeling algorithms use a scanning step that examines some of its neighbors. The first strategy exploits the dependencies among them to reduce the number of neighbors examined. When considering 8-connected components in a 2D image, this can reduce the number of neighbors examined from four to one in many cases. The second strategy uses an array to store the equivalence information among the labels. This replaces the pointer based rooted trees used to store the same equivalence information. It reduces the memory required and also produces consecutive final labels. Using an array instead of the pointer based rooted trees speeds up the connected component labeling algorithms by a factor of 5 ~ 100 in our tests on random binary images.
Effect of PDE-based noise removal on GVF-based deformation model on lesion detection in breast phantom x-ray images from Fischer’s fused FFDM and ultrasound (FFDMUS) imaging system
Jasjit Suri, Yujun Guo, Tim Danielson, et al.
It has been recently established that fusion of multi-modalities has led to better diagnostic capability and increased sensitivity and specificity. Fischer has been developing fused full-field digital mammography and ultrasound (FFDMUS) system. In FFDMUS, two sets of acquisitions are performed: 2-D X-ray and 3-D ultrasound. The segmentation of acquired lesions in phantom images is important: (1) to assess the image quality of X-ray and ultrasound images; (2) to register multi-modality images, and (3) to establish an automatic lesion detection methodology to assist the radiologist. In this paper, we studied the effect of PDE-based smoother on the gradient vector flow (GVF)-based active contour model for breast lesion detection. CIRS X-ray phantom images were acquired using FFDMUS, and region of interest (ROI) samples were extracted. PDE-based smoother was implemented to generate noise free images. The GVF-based strategy was then implemented on these noise free samples. Initial contours were set as default, and GVF snake then converged to extract lesion topology. The performance index was calculated by computing the difference between estimated lesion area and ideal lesion area. Our performance index with GVF (without PDE smoothing) yielded an average percentage error of 10.32%, while GVF with PDE yielded an average error of 9.61%, an improvement of 7%. We also optimized our PDE smoother for least GVF error estimation, and to our observation, we found the optimal number of iteration was 140. We also tested our program written in C++ on synthetic datasets.
JESS: Java extensible snakes system
Tim McInerney, M. Reza Akhavan Sharif, Nasrin Pashotanizadeh
Snakes (Active Contour Models) are powerful model-based image segmentation tools. Although researchers have proven them especially useful in medical image analysis over the past decade, Snakes have remained primarily in the academic world and they have not become widely used in clinical practice or widely available in commercial packages. A number of confusing and specialized variants exist and there has been no standard open-source implementation available. To address this problem, we present a Java Extensible Snakes System (JESS) that is general, portable, and extensible. The system uses Java Swing classes to allow for the rapid development of custom graphical user interfaces (GUI's). It also incorporates the Java Advanced Imaging (JAI) class library, which provide custom image preprocessing, image display and general image I/O. The Snakes algorithm itself is written in a hierarchical fashion, consisting of a general Snake class and several subclasses that span the main variants of Snakes including a new, powerful, robust subdivision-curve Snake. These subclasses can be easily and quickly extended and customized for any specific segmentation and analysis task. We demonstrate the utility of these classes for segmenting various anatomical structures from 2D medical images. We also demonstrate the effectiveness of JESS by using it to rapidly build a prototype semi-automatic sperm analysis system. The JESS software will be made publicly available in early 2005.
Skeleton-based region competition for automated gray matter and white matter segmentation of human brain MR images
Yong Chu, Ya-Fang Chen, Min-Ying Su, et al.
Image segmentation is an essential process for quantitative analysis. Segmentation of brain tissues in magnetic resonance (MR) images is very important for understanding the structural-functional relationship for various pathological conditions, such as dementia vs. normal brain aging. Different brain regions are responsible for certain functions and may have specific implication for diagnosis. Segmentation may facilitate the analysis of different brain regions to aid in early diagnosis. Region competition has been recently proposed as an effective method for image segmentation by minimizing a generalized Bayes/MDL criterion. However, it is sensitive to initial conditions -- the "seeds", therefore an optimal choice of “seeds” is necessary for accurate segmentation. In this paper, we present a new skeleton-based region competition algorithm for automated gray and white matter segmentation. Skeletons can be considered as good "seed regions" since they provide the morphological a priori information, thus guarantee a correct initial condition. Intensity gradient information is also added to the global energy function to achieve a precise boundary localization. This algorithm was applied to perform gray and white matter segmentation using simulated MRI images from a realistic digital brain phantom. Nine different brain regions were manually outlined for evaluation of the performance in these separate regions. The results were compared to the gold-standard measure to calculate the true positive and true negative percentages. In general, this method worked well with a 96% accuracy, although the performance varied in different regions. We conclude that the skeleton-based region competition is an effective method for gray and white matter segmentation.
Identify the centerline of tubular structure in medical images
Finding the centerline of the tubular structure helps to segment or analyze the organs such as the vessels or neuron fibers in medical images. This paper described a semi-automatic method using the minimum cost path finding and Hessian matrix analysis in scale space to calculate the centerline of tubular structure organs. Unlike previous approaches, exhaustive search for line-like shapes in every scale is prevent. Centerline pixels candidates and the width of the vessel are extracted by analyzing the intensity profile along the gradient vectors in the image. A verification procedure using Hassian matrix analysis with the scale obtained from the gradient analysis is applied to those candidates. Results obtained from the Hessian matrix analysis are used to construct a weighted graph. Finding the minimum cost path in the graph gives the centerline of the tubular structure. The method is applied to find the centerline of the vessels in the 2D angiogram and the neuron fibers in the 3D confocal microscopic images.
Automatic brain tumor detection in MRI: methodology and statistical validation
Khan M. Iftekharuddin, Mohammad A. Islam, Jahangheer Shaik, et al.
Automated brain tumor segmentation and detection are immensely important in medical diagnostics because it provides information associated to anatomical structures as well as potential abnormal tissue necessary to delineate appropriate surgical planning. In this work, we propose a novel automated brain tumor segmentation technique based on multiresolution texture information that combines fractal Brownian motion (fBm) and wavelet multiresolution analysis. Our wavelet-fractal technique combines the excellent multiresolution localization property of wavelets to texture extraction of fractal. We prove the efficacy of our technique by successfully segmenting pediatric brain MR images (MRIs) from St. Jude Children’s Research Hospital. We use self-organizing map (SOM) as our clustering tool wherein we exploit both pixel intensity and multiresolution texture features to obtain segmented tumor. Our test results show that our technique successfully segments abnormal brain tissues in a set of T1 images. In the next step, we design a classifier using Feed-Forward (FF) neural network to statistically validate the presence of tumor in MRI using both the multiresolution texture and the pixel intensity features. We estimate the corresponding receiver operating curve (ROC) based on the findings of true positive fractions and false positive fractions estimated from our classifier at different threshold values. An ROC, which can be considered as a gold standard to prove the competence of a classifier, is obtained to ascertain the sensitivity and specificity of our classifier. We observe that at threshold 0.4 we achieve true positive value of 1.0 (100%) sacrificing only 0.16 (16%) false positive value for the set of 50 T1 MRI analyzed in this experiment.
Quadratic regularization design for fan beam transmission tomography
Statistical methods for tomographic image reconstruction have shown considerable potential for improving image quality in X-ray CT. Penalized-likelihood (PL) image reconstruction methods require maximizing an objective function that is based on the log-likelihood of the sinogram measurements and on a roughness penalty function to control noise. In transmission tomography, PL methods (and MAP methods) based on conventional quadratic regularization functions lead to nonuniform and anisotropic spatial resolution, even for idealized shift-invariant imaging systems. We have previously addressed this problem for parallel-beam emission tomography by designing data-dependent, shift-variant regularizers that improve resolution uniformity. This paper extends those methods to the fan-beam geometry used in X-ray CT imaging. Simulation results demonstrate that the new method for regularization design requires very modest computation and leads to nearly uniform and isotropic spatial resolution in the fan-beam geometry when using quadratic regularization.
Helical cone beam reconstruction in volumetric CT: maintaining a large field-of-view (FOV) at very high pitches
In volumetric CT, the suppression of cone beam (CB) and helical artifacts is very challenging. At very high pitches larger than 1.5, not only artifact suppression but also maintenance of a large field of view (FOV) become even more challenging. At large helical pitches, the data redundancy corresponding to each pixel within a reconstruction FOV varies dramatically. An inappropriate dealing with the fast-varying data redundancy over pixels to be reconstructed can result in severe shading/glaring artifacts in reconstructed images and lead a clinical diagnosis impossible. One existing approach to combat the non-uniform data redundancy and extend the FOV at very high pitches is to reconstruct images on tilted planes. However, tilting a reconstruction plane can significantly impacts data flow of filtered backprojection -- the most attractive one practiced by all major CT manufacturers -- and slows down image generation speed considerably. Along with an experimental evaluation, a helical CB filtered backprojection reconstruction algorithm using three-dimensional view weighting (namely 3D view weighted CB-FBP algorithm) is proposed here. The novelty of the proposed algorithm is the 3D-nature of the weighting function, which enables the proposed algorithm to reach the optimal image qualities and provides the freedom in controlling image quality over accuracy, noise characteristics, spatial resolution and temporal resolution. Particularly, it is the 3D view weighting that enables the proposed algorithm to maintain an FOV at helical pitch larger than 1.5, which is as large as that at helical pitches below 1.0. It is believed that the proposed 3D view weighted CB-FBP algorithm is applicable to and robust over all CT imaging applications, including diagnostic imaging and dynamic imaging for functional evaluation. It is also believed that the proposed algorithm will work well for larger cone angles when more sophisticated 3D view weighting strategies are employed.
A consistency condition for cone-beam CT with general source trajectories
Recent algorithm development for image reconstruction for cone-beam CT has tackled exact image reconstruction for very general scanning configurations. The heart of the new algorithms is the concept of reconstruction on the chordn of a general source trajectory. Volume ROI reconstruction becomes possible by concatenating the chords on which the image has been obtained. For some scanning trajectories there maybe points in the image space where the image can theoretically be obtained exactly, yet no chord intersects these points. This article provides a consistency condition, based on the ideas of John's equation, that may be used to rebin cone-beam data so that all points satisfying Tuy's condition are reconstructible by a chord algorithm.
Exact image reconstruction from n-PI acquisition data in helical cone-beam CT
In helical cone-beam computerized tomography, the data within the Tam-Danielsson window is sufficient for exact reconstruction of the images. In some practical situation, the projections outside the Tam-Danielsson window are available that indicate the data redundancy. In this work, we investigate image reconstruction on n-PI lines from data within n-PI window. We performed a preliminary numerical study, and the results in these studies shown that our algorithm can exactly reconstruct images from n-PI data.
Utilization of redundancy in ROI reconstruction with backprojection filtration from fan-beam truncated data
In fan-beam computed tomography (CT), one may be interested in image reconstruction in a region of interest (ROI) from truncated data acquired over an angular range less than half-scan data. We developed recently a backprojection filtration (BPF) algorithm to reconstruct an ROI image from reduced scan data containing data truncations. In a reduced scan, the truncated data may still contain redundancy. In this work, we describe a new algorithm that can exploit data redundancy in truncated data for potentially suppressing the aliasing and noise artifacts in reconstructed images. We have performed numerical studies to demonstrate the BPF algorithm.
Sinogram noise reduction for low-dose CT by statistics-based nonlinear filters
Low-dose CT (computed tomography) sinogram data have been shown to be signal-dependent with an analytical relationship between the sample mean and sample variance. Spatially-invariant low-pass linear filters, such as the Butterworth and Hanning filters, could not adequately handle the data noise and statistics-based nonlinear filters may be an alternative choice, in addition to other choices of minimizing cost functions on the noisy data. Anisotropic diffusion filter and nonlinear Gaussian filters chain (NLGC) are two well-known classes of nonlinear filters based on local statistics for the purpose of edge-preserving noise reduction. These two filters can utilize the noise properties of the low-dose CT sinogram for adaptive noise reduction, but can not incorporate signal correlative information for an optimal regularized solution. Our previously-developed Karhunen-Loeve (KL) domain PWLS (penalized weighted least square) minimization considers the signal correlation via the KL strategy and seeks the PWLS cost function minimization for an optimal regularized solution for each KL component, i.e., adaptive to the KL components. This work compared the nonlinear filters with the KL-PWLS framework for low-dose CT application. Furthermore, we investigated the nonlinear filters for post KL-PWLS noise treatment in the sinogram space, where the filters were applied after ramp operation on the KL-PWLS treated sinogram data prior to backprojection operation (for image reconstruction). By both computer simulation and experimental low-dose CT data, the nonlinear filters could not outperform the KL-PWLS framework. The gain of post KL-PWLS edge-preserving noise filtering in the sinogram space is not significant, even the noise has been modulated by the ramp operation.
Conjugate gradient Mojette reconstruction
Myriam Servieres, Jerome Idier, Niccolas Normand, et al.
Iterative methods are now recognized as powerful tools to solve inverse problems such as tomographic reconstruction. In this paper, the main goal is to present a new reconstruction algorithm made from two components. An iterative algorithm, namely the Conjugate Gradient (CG) method, is used to solve the tomographic problem in the least square (LS) sense for our specific discrete Mojette geometry. The results are compared (with the same geometry) to the corresponding Mojette Filtered Back Projection (FBP) method. In the fist part of the paper, we recall the discrete geometry used to define the projection M and backprojection M* operators. In the second part, the CG algorithm is presented within the context of the Mojette geometry. Noise is then added onto these Mojette projections with respect to the sampling and reconstructions are performed. Finally the Toeplitz block Toeplitz (TBT) character of M*M is demonstrated.
A radial adaptive filter for metal artifact reduction
Matthieu Bal, Hasan Celik, Krishna Subramanyan, et al.
High-density objects, such as metal prostheses or surgical clips, generate streak-like artifacts in CT images. We designed a radial adaptive filter, which directly operates on the corrupted reconstructed image, to effectively and efficiently reduce such artifacts. The filter adapts to the severity of local artifacts to preserve spatial resolution as much as possible. The widths and direction of the filter are derived from the local structure tensor. Visual inspection shows that this novel radial adaptive filter is superior with respect to existing methods in the case of mildly distorted images. In the presence of strong artifacts we propose a hybrid approach. An image corrected with a standard method, which performs well on images with regions of severe artifacts, is fused with an adaptively filtered clone to combine the strengths of both methods.
A new reconstruction algorithm for energy-resolved coherent scatter computed tomography
Udo van Stevendaal, Jens-Peter Schlomka, Axel Thran, et al.
For the first time, a reconstruction technique based on filtered back-projection (FBP) using curved 3D back-projection lines is applied to energy-resolved coherent-scatter projection data. Coherent-scatter computed tomography (CSCT) yields information about the molecular structure of an object. It has been shown that the relatively poor spectral resolution due to the application of a polychromatic X-ray source can be overcome, when energy-resolved detection is used. So far, the energy-resolved projection data, acquired with a CSCT scanner, are reconstructed with the help of algebraic reconstruction techniques (ART). Due to the computational complexity of iterative reconstruction, these methods lead to relatively long reconstruction times. In this contribution, a reconstruction algorithm based on 3D FBP is introduced and applied to projection data acquired with a demonstrator setup similar to a multi-line CT scanner geometry using an energy-resolving CdTe-detector. Within a fraction of the computation time of algebraic reconstruction methods, an image of comparable quality is generated when using FBP reconstruction. In addition, the FBP approach has the advantage, that sub-field-of-view reconstruction becomes feasible. This allows a selective reconstruction of the scatter function for a region of interest. The method is based on a high-pass filtering of the scatter data in fan-beam direction applied to all energy channels. The 3D back-projection is performed along curved lines through a volume defined by the in-plane spatial coordinates and the wave-vector transfer.
Image reconstruction on virtual PI-lines using filtered-backprojection in circular cone-beam CT
Recently, a 3D filtered-backprojection (FBP)-based algorithm for image reconstruction on PI-line segments in a helical cone-beam CT scan has been developed (Zou and Pan, 2004). In the present work, we derive new reconstruction algorithms for circular cone-beam scans based upon this algorithm and a concept of virtual PI-line. We prove that, in the case of conventional full- and short-scan, the newly derived algorithms are mathematically identical to existing algorithms. More importantly, in the case of reduced-scans in which the scanning angle range is less than that in a short-scan, the new algorithms can yield exact region of interest (ROI) reconstruction in mid-plane and approximate ROI reconstruction in off-mid-planes. We have performed a preliminary numerical study that verifies our theoretical assertions.
Noise behavior of spline Mojette FBP reconstruction
Myriam Servieres, Niccolas Normand, Yves Bizais, et al.
The goal of this paper is to characterize the noise properties of a spline Filtered BackProjection (denoted as FBP) reconstruction scheme. More specifically, the paper focuses on angular and radial sampling of projection data and on assumed local properties of the function to be reconstructed. This new method is visually and quantitatively compared to standard sampling used for FBP scheme. In the second section, we recall the sampling geometry adapted to the discrete geometry of the reconstructed image. Properties of the discrete zero order Spline Ramp filter for classic angles and discrete angles generated from Farey’s series reconstruction are used to generate their equivalent representations for first order Spline filters. Digital phantoms are used to assess the results and the correctness of the linearity and shift-invariantness assumption for the discrete reconstructions. The filter gain has been studied in the Mojette case since the number of projections can be very different from one angle to another. In the third section, we describe the Spline filter implementation and the continuous/discrete correspondence. In section 4, Poisson noise is added to noise-free onto the projections. The reconstructions between classic angle distribution and Mojette acquisition geometry are compared. Even if the number of bins per projections is fixed for classic FBP while it varies for the Mojette geometry (leading to very different noise behavior per bin) the results of both algorithms are very close. The discussion allows for a general comparison between classic FBP reconstruction and Mojette FBP. The very encouraging results obtained for the Mojette case conclude for the developments of future acquisition devices modeled with the Mojette geometry.
Artifact reduction in truncated CT using sinogram completion
Truncation of projection data in CT produces significant artifacts in the reconstruction process due to non-locality of the Radon transform. In this paper, we present a method for reducing these truncation artifacts by estimating features that lie outside the region of interest (ROI) and using these features to complete the truncated sinogram. Projection images of an object are obtained. A sinogram is obtained by stacking profile data from all projection angles. A simulated truncated sinogram is generated by setting pixel values outside an ROI to zero. The truncated sinogram is then transformed into a (radius, phase) image, with pixel values in what we term as the Polar representation (PR) image corresponding to the minimum value along sine curves given by x = r*cos(projection angle + phase). The PR image contains data for radii greater than the ROI radius. Pixel values outside the ROI in the completed sinogram are determined as follows. For each pixel in the PR image, a sine curve is generated in the completed sinogram image outside the ROI, having the same pixel value as that of the PR image for that radius and phase. Successive sine curves are laid and the values of each are summed. The intensity outside is then equalized to the intensity inside the ROI. The completed sinogram is then reconstructed, to obtain completed reconstruction. The percentage error in the difference image between the full FOV reconstruction and the corresponding completed reconstruction and the extrapolated-average reconstruction are 1.1% and 3.3% respectively. This indicates that the completed reconstruction is closer to full FOV reconstruction. Thus, the sinogram completion can be used to improve reconstructions from truncated data.
A new algorithm for determining 3D biplane imaging geometry: theory and implementation
Vikas Singh, Jinhui Xu, Kenneth R. Hoffmann, et al.
Biplane imaging is a primary method for visual and quantitative assessment of the vasculature. A key problem (called Imaging Geometry Determination problem or IGD for short) in this method is to determine the rotation-matrix (R) and the translation vector (t) which relate the two coordinate systems. In this paper, we propose a new approach, called IG-Sieving, to calculate R and t using corresponding points in the two images. Our technique first generates an initial estimate of R and t from the gantry angles of the imaging system, and then optimizes them by solving an optimal-cell-search problem in a 6-D parametric space (three variables defining R plus the three variables of t). To efficiently find the optimal imaging geometry (IG) in 6-D, our approach divides the high dimensional search domain into a set of lower-dimensional regions, (holding two variables constant at each optimization step), thereby reducing the optimal-cell-search problem to a set of optimization problems in 3D sub-spaces (one other variable is correlated). For each such sub-space, our approach first applies efficient computational geometry techniques to identify "possibly-feasible" IG’s, and then uses a criterion we call fall-in-number to sieve out good IGs. We show that in a bounded number of optimization steps, a (possibly infinite) set of near optimal IGs (which are equally good) can be determined. Simulation results indicate that our method can reconstruct 3D points with average 3D center-of-mass errors of about 0.8cm for input image-data errors as high as 0.1cm, which is comparable to existing techniques. More importantly, our algorithm provides a novel insight into the geometric structure of the solution space, which could be exploited to significantly improve the accuracy of other biplane algorithms.
Implementation of sensitivity and resolution modeling for SPECT with cone-beam collimator
Andrzej Krol, Vikram R. Kunniyur, Wei Lee, et al.
We implemented a fully-3D ordered-subsets expectation-maximization (OSEM) algorithm with attenuation compensation, distance-dependent blurring (DDB), and sensitivity modeling for SPECT performed with a cone-beam collimator (CBC). The experimentally obtained detector response to point sources across FOV was fitted to a two-dimensional Gaussian function with its width (FWHM) varying linearly with the source-to-detector distance and with very weak sensitivity dependence on the emission angle. We obtained CBC SPECT scans of a physical point-source phantom, a Defrise phantom, and a female patient, and we investigated performance of our algorithm. To correctly simulate DDB and sensitivity, a blurring kernel with a radius of up to 10 elements had to be used for a 128x128 acquisition matrix, and volumetric ray tracing rather than line-element-based ray tracing has to be implemented. In the point-source phantom reconstruction we evaluated the uniformity of FWHM for the radial, tangential and longitudinal directions, and sensitivity vs. distance. An isotropic and stationary resolution was obtained at any location by OSEM with DDB and sensitivity modeling, only when volumetric ray tracing was used. We analyzed axial and transaxial profiles obtained for the Defrise phantom and evaluated the reconstructed breast SPECT patient images. The proposed fully-3D OSEM reconstruction algorithm with DBB and sensitivity modeling, and attenuation compensation with volumetric rays tracing is efficient and effective with significant resolution and sensitivity recovery.
Coronary artery motion compensation in spiral CT
Because of motion artifacts, spiral Computer Tomography (CT) is not presently a widely useful diagnostic tool in cardiac imaging. There are two time scales in CT data acquisition. The first is the time-scale of a single projection which is roughly a half millisecond. The second time-scale is that of a single rotation of the x-ray source which is about 400 ms. For diagnostic purposes, the fastest components of the heart cycle are on the order of 5 ms. Those of general interest are in the range of about 100 ms. Full-rotation CT acquisition is slower than required for freezing the cardiac motion of interest. This mismatch in speed causes motion artifacts in the reconstructed images. We focus our attention on accurately measuring lumen dimensions and on visualizing lesion architecture. This analysis requires local imaging; it does not require global motion compensation. The methods we use relate to one-dimensional motion tracking and motion compensation applied to the projection data. Dynamic programming is used for the tracking. Following the projection processing, the CT reconstruction algorithm then acts on the motion-corrected projection data to produce the reconstructed image. In addition to applying motion compensation to the projections, we also use region-of-interest (ROI) CT reconstruction algorithms in order to utilize the locally motion-corrected projection data without generating artifacts. Those artifacts come from missing projection data that should come from outside ROI. In our case the ROI is the cross-section of the vessel being imaged.
Streak artifact reduction in cardiac cone beam CT
Gilad Shechter, Galit Naveh, Jonathan Lessick, et al.
Cone beam reconstructed cardiac CT images suffer from characteristic streak artifacts that affect the quality of coronary artery imaging. These artifacts arise from inhomogeneous distribution of noise. While in non-tagged reconstruction inhomogeneity of noise distribution is mainly due to anisotropy of the attenuation of the scanned object (e.g. shoulders), in cardiac imaging it is largely influenced by the non-uniform distribution of the acquired data used for reconstructing the heart at a given phase. We use a cardiac adaptive filter to reduce these streaks. In difference to previous methods of adaptive filtering that locally smooth data points on the basis of their attenuation values, our filter is applied as a function of the noise distribution of the data as it is used in the phase selective reconstruction. We have reconstructed trans-axial images without adaptive filtering, with a regular adaptive filter and with the cardiac adaptive filter. With the cardiac adaptive filter significant reduction of streaks is achieved, and thus image quality is improved. The coronary vessel is much more pronounced in the cardiac adaptive filtered images, in slab MIP the main coronary artery branches are more visible, and non-calcified plaque is better differentiated from vessel wall. This improvement is accomplished without altering significantly the border definition of calcified plaques.
MLEM algorithm adaptation for improved SPECT scintimammography
Andrzej Krol, David H. Feiglin, Wei Lee, et al.
Standard MLEM and OSEM algorithms used in SPECT Tc-99m sestamibi scintimammography produce hot-spot artifacts (HSA) at the image support peripheries. We investigated a suitable adaptation of MLEM and OSEM algorithms needed to reduce HSA. Patients with suspicious breast lesions were administered 10 mCi of Tc-99m sestamibi and SPECT scans were acquired for patients in prone position with uncompressed breasts. In addition, to simulate breast lesions, some patients were imaged with a number of breast skin markers each containing 1 μCi of Tc-99m. In order to reduce HSA in reconstruction, we removed from the backprojection step the rays that traverse the periphery of the support region on the way to a detector bin, when their path length through this region was shorter than some critical length. Such very short paths result in a very low projection counts contributed to the detector bin, and consequently to overestimation of the activity in the peripheral voxels in the backprojection step -- thus creating HSA. We analyzed the breast-lesion contrast and suppression of HSA in the images reconstructed using standard and modified MLEM and OSEM algorithms vs. critical path length (CPL). For CPL ≥ 0.01 pixel size, we observed improved breast-lesion contrast and lower noise in the reconstructed images, and a very significant reduction of HSA in the maximum intensity projection (MIP) images.
Limited-angle reconstruction algorithms in computed-tomographic microscopic imaging
Ravil Chamgoulov, Michael Tsiroulnikov, Pierre Lane, et al.
We present our results on development of algorithms for image reconstruction in optical computed-tomography microscopy. The optical computed-tomography microscope with light modulation, recently developed by our group at the BC Cancer Research Centre is a novel imaging device for three-dimensional visualization and quantitative analysis of absorption-stained biological samples. The angles for projection in the system are limited to the range of 0 ≤ φ ≤ 135° by the numerical aperture of the illumination objective. For the limited-angle tomography problem we have developed several reconstruction algorithms. One algorithm is based on the Radon transformation and assumes parallel ray projections. In order to compensate for limited data, several reconstructions are generated from several sets of projections of a specimen, acquired at different orientations of parallel-ray light scanning. The reconstructions are combined together using a vote criteria to create a final volume. Another reconstruction algorithm developed by the group employs both transform-based and iterative methods to address the limited-angle reconstruction problem. In this algorithm the transform-based method is used as an initial starting point for the following iterative reconstruction. A feedback correction of the reconstruction image is made on each iteration step. The method enables to incorporate previously known information about the object into the reconstruction process. The algorithm improves reconstruction accuracy at a reasonable computational cost and programming commitment. Three-dimensional microscopic images of quantitatively absorption-stained cells have been reconstructed with the resolution better than 6 microns.
Parallelizable 3D statistical reconstruction for C-arm tomosynthesis system
Beilei Wang, Kenneth Barner, Denny Lee
Clinical diagnosis and security detection tasks increasingly require 3D information which is difficult or impossible to obtain from 2D (two dimensional) radiographs. As a 3D (three dimensional) radiographic and non-destructive imaging technique, digital tomosynthesis is especially fit for cases where 3D information is required while a complete projection data is not available. Nowadays, FBP (filtered back projection) is extensively used in industry for its fast speed and simplicity. However, it is hard to deal with situations where only a limited number of projections from constrained directions are available, or the SNR (signal to noises ratio) of the projections is low. In order to deal with noise and take into account a priori information of the object, a statistical image reconstruction method is described based on the acquisition model of X-ray projections. We formulate a ML (maximum likelihood) function for this model and develop an ordered-subsets iterative algorithm to estimate the unknown attenuation of the object. Simulations show that satisfied results can be obtained after 1 to 2 iterations, and after that there is no significant improvement of the image quality. An adaptive wiener filter is also applied to the reconstructed image to remove its noise. Some approximations to speed up the reconstruction computation are also considered. Applying this method to computer generated projections of a revised Shepp phantom and true projections from diagnostic radiographs of a patient’s hand and mammography images yields reconstructions with impressive quality. Parallel programming is also implemented and tested. The quality of the reconstructed object is conserved, while the computation time is considerably reduced by almost the number of threads used.
Automatic glare removal in reflectance imagery of the uterine cervix
Holger Lange
Colposcopy is a diagnostic method used to detect cancer precursors and cancer of the uterine cervix. Computer-Aided-Diagnosis (CAD) for colposcopy is a new field in medical image processing. Colposcopists analyze glare (glint or specular reflection) patterns on the cervix to assess the surface contour (3D topology) of lesions, an important feature used to evaluate lesion severity. However, glare in the imagery presents major problems for automated image analysis systems. Glare eliminates all information in affected pixels and can introduce artifacts in feature extraction algorithms, such as acetowhite region detection. Although cross-polarization filters can be used to eliminate glare, the reality is that we have to deal with glare when we want to use existing cervical image databases or use an instrument that does not provide cross-polarized imagery. Here, we present the details and preliminary results of a glare removal algorithm for RGB color images of the cervix that can be used as a pre-processing step in CAD systems. The algorithm can be extended to multispectral and hyperspectral imagery. The basic approach of the algorithm is to extract a feature image from the RGB image that provides a good glare to background ratio, to detect the glare regions in the feature image, to extend the glare regions to cover all pixels that have been affected by the glare, and to remove the glare in the affected regions by filling in an estimate of the underlying image features. In our current implementation we use the green (G) image component as the feature image, given its high glare to background ratio and simplicity of calculation. Glare regions are either detected as saturated regions or small high contrasted bright regions. Saturated regions are detected using an adaptive thresholding method. Small high contrasted bright regions are detected using morphological top hat filters with different sizes and thresholds. The full extent of the glare regions is estimated by using a morphological constraint watershed segmentation to find the contour of the glare regions and adding a constant dilatation. The image features are estimated by interpolating the R,G,B color components individually from the surrounding regions based on Laplace’s equation and modifying the intensity (I) component of the HSI color space transformed image. As the glare pattern is important to the physician, we embed it as a just visible intensity texture that does not affect the image processing. The performance of the algorithm is demonstrated using human subject data.