Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 16 de 16
Filter
1.
Comput Biol Med ; 170: 107955, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38215618

ABSTRACT

Multi-organ segmentation is vital for clinical diagnosis and treatment. Although CNN and its extensions are popular in organ segmentation, they suffer from the local receptive field. In contrast, MultiLayer-Perceptron-based models (e.g., MLP-Mixer) have a global receptive field. However, these MLP-based models employ fully connected layers with many parameters and tend to overfit on sample-deficient medical image datasets. Therefore, we propose a Cascaded Spatial Shift Network, CSSNet, for multi-organ segmentation. Specifically, we design a novel cascaded spatial shift block to reduce the number of model parameters and aggregate feature segments in a cascaded way for efficient and effective feature extraction. Then, we propose a feature refinement network to aggregate multi-scale features with location information, and enhance the multi-scale features along the channel and spatial axis to obtain a high-quality feature map. Finally, we employ a self-attention-based fusion strategy to focus on the discriminative feature information for better multi-organ segmentation performance. Experimental results on the Synapse (multiply organs) and LiTS (liver & tumor) datasets demonstrate that our CSSNet achieves promising segmentation performance compared with CNN, MLP, and Transformer models. The source code will be available at https://github.com/zkyseu/CSSNet.


Subject(s)
Liver Neoplasms , Humans , Neural Networks, Computer , Software , Image Processing, Computer-Assisted
2.
Comput Biol Med ; 166: 107567, 2023 Oct 13.
Article in English | MEDLINE | ID: mdl-37852109

ABSTRACT

Medical image segmentation is crucial for accurate diagnosis and treatment in the medical field. In recent years, convolutional neural networks (CNNs) and Transformers have been frequently adopted as network architectures in medical image segmentation. The convolution operation is limited in modeling long-range dependencies because it can only extract local information through the limited receptive field. In comparison, Transformers demonstrate excellent capability in modeling long-range dependencies but are less effective in capturing local information. Hence, effectively modeling long-range dependencies while preserving local information is essential for accurate medical image segmentation. In this paper, we propose a four-axis fusion framework called FAFuse, which can exploit the advantages of CNN and Transformer. As the core component of our FAFuse, a Four-Axis Fusion module (FAF) is proposed to efficiently fuse global and local information. FAF combines Four-Axis attention (height, width, main diagonal, and counter diagonal axial attention), a multi-scale convolution, and a residual structure with a depth-separable convolution and a Hadamard product. Furthermore, we also introduce deep supervision to enhance gradient flow and improve overall performance. Our approach achieves state-of-the-art segmentation accuracy on three publicly available medical image segmentation datasets. The code is available at https://github.com/cczu-xiao/FAFuse.

3.
PeerJ Comput Sci ; 9: e1460, 2023.
Article in English | MEDLINE | ID: mdl-37547396

ABSTRACT

Purpose: To compare the diagnostic efficiencies of deep learning single-modal and multi-modal for the classification of benign and malignant breast mass lesions. Methods: We retrospectively collected data from 203 patients (207 lesions, 101 benign and 106 malignant) with breast tumors who underwent breast magnetic resonance imaging (MRI) before surgery or biopsy between January 2014 and October 2020. Mass segmentation was performed based on the three dimensions-region of interest (3D-ROI) minimum bounding cube at the edge of the lesion. We established single-modal models based on a convolutional neural network (CNN) including T2WI and non-fs T1WI, the dynamic contrast-enhanced (DCE-MRI) first phase was pre-contrast T1WI (d1), and Phases 2, 4, and 6 were post-contrast T1WI (d2, d4, d6); and Multi-modal fusion models with a Sobel operator (four_mods:T2WI, non-fs-T1WI, d1, d2). Training set (n = 145), validation set (n = 22), and test set (n = 40). Five-fold cross validation was performed. Accuracy, sensitivity, specificity, negative predictive value, positive predictive value, and area under the ROC curve (AUC) were used as evaluation indicators. Delong's test compared the diagnostic performance of the multi-modal and single-modal models. Results: All models showed good performance, and the AUC values were all greater than 0.750. Among the single-modal models, T2WI, non-fs-T1WI, d1, and d2 had specificities of 77.1%, 77.2%, 80.2%, and 78.2%, respectively. d2 had the highest accuracy of 78.5% and showed the best diagnostic performance with an AUC of 0.827. The multi-modal model with the Sobel operator performed better than single-modal models, with an AUC of 0.887, sensitivity of 79.8%, specificity of 86.1%, and positive prediction value of 85.6%. Delong's test showed that the diagnostic performance of the multi-modal fusion models was higher than that of the six single-modal models (T2WI, non-fs-T1WI, d1, d2, d4, d6); the difference was statistically significant (p = 0.043, 0.017, 0.006, 0.017, 0.020, 0.004, all were greater than 0.05). Conclusions: Multi-modal fusion deep learning models with a Sobel operator had excellent diagnostic value in the classification of breast masses, and further increase the efficiency of diagnosis.

4.
Front Neurosci ; 15: 761610, 2021.
Article in English | MEDLINE | ID: mdl-34803593

ABSTRACT

Accurately predicting the quality of depth-image-based-rendering (DIBR) synthesized images is of great significance in promoting DIBR techniques. Recently, many DIBR-synthesized image quality assessment (IQA) algorithms have been proposed to quantify the distortion that existed in texture images. However, these methods ignore the damage of DIBR algorithms on the depth structure of DIBR-synthesized images and thus fail to accurately evaluate the visual quality of DIBR-synthesized images. To this end, this paper presents a DIBR-synthesized image quality assessment metric with Texture and Depth Information, dubbed as TDI. TDI predicts the quality of DIBR-synthesized images by jointly measuring the synthesized image's colorfulness, texture structure, and depth structure. The design principle of our TDI includes two points: (1) DIBR technologies bring color deviation to DIBR-synthesized images, and so measuring colorfulness can effectively predict the quality of DIBR-synthesized images. (2) In the hole-filling process, DIBR technologies introduce the local geometric distortion, which destroys the texture structure of DIBR-synthesized images and affects the relationship between the foreground and background of DIBR-synthesized images. Thus, we can accurately evaluate DIBR-synthesized image quality through a joint representation of texture and depth structures. Experiments show that our TDI outperforms the competing state-of-the-art algorithms in predicting the visual quality of DIBR-synthesized images.

5.
IEEE Trans Med Imaging ; 39(6): 2151-2162, 2020 06.
Article in English | MEDLINE | ID: mdl-31940526

ABSTRACT

Sufficient data with complete annotation is essential for training deep models to perform automatic and accurate segmentation of CT male pelvic organs, especially when such data is with great challenges such as low contrast and large shape variation. However, manual annotation is expensive in terms of both finance and human effort, which usually results in insufficient completely annotated data in real applications. To this end, we propose a novel deep framework to segment male pelvic organs in CT images with incomplete annotation delineated in a very user-friendly manner. Specifically, we design a hybrid loss network derived from both voxel classification and boundary regression, to jointly improve the organ segmentation performance in an iterative way. Moreover, we introduce a label completion strategy to complete the labels of the rich unannotated voxels and then embed them into the training data to enhance the model capability. To reduce the computation complexity and improve segmentation performance, we locate the pelvic region based on salient bone structures to focus on the candidate segmentation organs. Experimental results on a large planning CT pelvic organ dataset show that our proposed method with incomplete annotation achieves comparable segmentation performance to the state-of-the-art methods with complete annotation. Moreover, our proposed method requires much less effort of manual contouring from medical professionals such that an institutional specific model can be more easily established.


Subject(s)
Image Processing, Computer-Assisted , Tomography, X-Ray Computed , Humans , Male , Pelvis/diagnostic imaging
6.
IEEE Trans Biomed Eng ; 67(10): 2710-2720, 2020 10.
Article in English | MEDLINE | ID: mdl-31995472

ABSTRACT

Obtaining accurate segmentation of the prostate and nearby organs at risk (e.g., bladder and rectum) in CT images is critical for radiotherapy of prostate cancer. Currently, the leading automatic segmentation algorithms are based on Fully Convolutional Networks (FCNs), which achieve remarkable performance but usually need large-scale datasets with high-quality voxel-wise annotations for full supervision of the training. Unfortunately, such annotations are difficult to acquire, which becomes a bottleneck to build accurate segmentation models in real clinical applications. In this paper, we propose a novel weakly supervised segmentation approach that only needs 3D bounding box annotations covering the organs of interest to start the training. Obviously, the bounding box includes many non-organ voxels that carry noisy labels to mislead the segmentation model. To this end, we propose the label denoising module and embed it into the iterative training scheme of the label denoising network (LDnet) for segmentation. The labels of the training voxels are predicted by the tentative LDnet, while the label denoising module identifies the voxels with unreliable labels. As only the good training voxels are preserved, the iteratively re-trained LDnet can refine its segmentation capability gradually. Our results are remarkable, i.e., reaching  âˆ¼ 94% (prostate),  âˆ¼ 91% (bladder), and  âˆ¼ 86% (rectum) of the Dice Similarity Coefficients (DSCs), compared to the case of fully supervised learning upon high-quality voxel-wise annotations and also superior to several state-of-the-art approaches. To our best knowledge, this is the first work to achieve voxel-wise segmentation in CT images from simple 3D bounding box annotations, which can greatly reduce many labeling efforts and meet the demands of the practical clinical applications.


Subject(s)
Prostatic Neoplasms , Tomography, X-Ray Computed , Algorithms , Humans , Male , Prostatic Neoplasms/diagnostic imaging , Urinary Bladder
7.
Comput Intell Neurosci ; 2019: 8214975, 2019.
Article in English | MEDLINE | ID: mdl-30863436

ABSTRACT

Zebrafish embryo fluorescent vessel analysis, which aims to automatically investigate the pathogenesis of diseases, has attracted much attention in medical imaging. Zebrafish vessel segmentation is a fairly challenging task, which requires distinguishing foreground and background vessels from the 3D projection images. Recently, there has been a trend to introduce domain knowledge to deep learning algorithms for handling complex environment segmentation problems with accurate achievements. In this paper, a novel dual deep learning framework called Dual ResUNet is developed to conduct zebrafish embryo fluorescent vessel segmentation. To avoid the loss of spatial and identity information, the U-Net model is extended to a dual model with a new residual unit. To achieve stable and robust segmentation performance, our proposed approach merges domain knowledge with a novel contour term and shape constraint. We compare our method qualitatively and quantitatively with several standard segmentation models. Our experimental results show that the proposed method achieves better results than the state-of-art segmentation methods. By investigating the quality of the vessel segmentation, we come to the conclusion that our Dual ResUNet model can learn the characteristic features in those cases where fluorescent protein is deficient or blood vessels are overlapped and achieves robust performance in complicated environments.


Subject(s)
Blood Vessels/anatomy & histology , Blood Vessels/diagnostic imaging , Neural Networks, Computer , Signal Processing, Computer-Assisted , Algorithms , Animals , Blood Vessels/embryology , Embryo, Nonmammalian , Models, Anatomic , Zebrafish
8.
Med Phys ; 45(5): 2063-2075, 2018 May.
Article in English | MEDLINE | ID: mdl-29480928

ABSTRACT

PURPOSE: Accurate 3D image segmentation is a crucial step in radiation therapy planning of head and neck tumors. These segmentation results are currently obtained by manual outlining of tissues, which is a tedious and time-consuming procedure. Automatic segmentation provides an alternative solution, which, however, is often difficult for small tissues (i.e., chiasm and optic nerves in head and neck CT images) because of their small volumes and highly diverse appearance/shape information. In this work, we propose to interleave multiple 3D Convolutional Neural Networks (3D-CNNs) to attain automatic segmentation of small tissues in head and neck CT images. METHOD: A 3D-CNN was designed to segment each structure of interest. To make full use of the image appearance information, multiscale patches are extracted to describe the center voxel under consideration and then input to the CNN architecture. Next, as neighboring tissues are often highly related in the physiological and anatomical perspectives, we interleave the CNNs designated for the individual tissues. In this way, the tentative segmentation result of a specific tissue can contribute to refine the segmentations of other neighboring tissues. Finally, as more CNNs are interleaved and cascaded, a complex network of CNNs can be derived, such that all tissues can be jointly segmented and iteratively refined. RESULT: Our method was validated on a set of 48 CT images, obtained from the Medical Image Computing and Computer Assisted Intervention (MICCAI) Challenge 2015. The Dice coefficient (DC) and the 95% Hausdorff Distance (95HD) are computed to measure the accuracy of the segmentation results. The proposed method achieves higher segmentation accuracy (with the average DC: 0.58 ± 0.17 for optic chiasm, and 0.71 ± 0.08 for optic nerve; 95HD: 2.81 ± 1.56 mm for optic chiasm, and 2.23 ± 0.90 mm for optic nerve) than the MICCAI challenge winner (with the average DC: 0.38 for optic chiasm, and 0.68 for optic nerve; 95HD: 3.48 for optic chiasm, and 2.48 for optic nerve). CONCLUSION: An accurate and automatic segmentation method has been proposed for small tissues in head and neck CT images, which is important for the planning of radiotherapy.


Subject(s)
Head and Neck Neoplasms/diagnostic imaging , Imaging, Three-Dimensional/methods , Joints/diagnostic imaging , Neural Networks, Computer , Tomography, X-Ray Computed , Humans
9.
Pattern Recognit ; 63: 531-541, 2017 Mar.
Article in English | MEDLINE | ID: mdl-29062159

ABSTRACT

It is fundamentally important to fuse the brain atlas from magnetic resonance (MR) images for many imaging-based studies. Most existing works focus on fusing the atlases from high-quality MR images. However, for low-quality diagnostic images (i.e., with high inter-slice thickness), the problem of atlas fusion has not been addressed yet. In this paper, we intend to fuse the brain atlas from the high-thickness diagnostic MR images that are prevalent for clinical routines. The main idea of our works is to extend the conventional groupwise registration by incorporating a novel super-resolution strategy. The contribution of the proposed super-resolution framework is two-fold. First, each high-thickness subject image is reconstructed to be isotropic by the patch-based sparsity learning. Then, the reconstructed isotropic image is enhanced for better quality through the random-forest-based regression model. In this way, the images obtained by the super-resolution strategy can be fused together by applying the groupwise registration method to construct the required atlas. Our experiments have shown that the proposed framework can effectively solve the problem of atlas fusion from the low-quality brain MR images.

10.
Med Phys ; 44(11): 5916-5929, 2017 Nov.
Article in English | MEDLINE | ID: mdl-28875551

ABSTRACT

PURPOSE: Lung field segmentation for chest radiography is critical to pulmonary disease diagnosis. In this paper, we propose a new deformable model using weighted sparse shape composition with robust initialization to achieve robust and accurate lung field segmentation. METHODS: Our method consists of three steps: initialization, deformation and regularization. The steps of deformation and regularization are iteratively employed until convergence. First, since a deformable model is sensitive to the initial shape, a robust initialization is obtained by using a novel voting strategy, which allows the reliable patches on the image to vote for each landmark of the initial shape. Then, each point of the initial shape independently deforms to the lung boundary under the guidance of the appearance model, which can distinguish lung tissues from nonlung tissues near the boundary. Finally, the deformed shape is regularized by weighted sparse shape composition (SSC) model, which is constrained by both boundary information and the correlations between each point of the deformed shape. RESULTS: Our method has been evaluated on 247 chest radiographs from well-known dataset Japanese Society of Radiological Technology (JSRT) and achieved high overlap scores (0.955 ± 0.021). CONCLUSIONS: The experimental results show that the proposed deformable segmentation model is more robust and accurate than the traditional appearance and shape model on the JSRT database. Our method also shows higher accuracy than most state-of-the-art methods.


Subject(s)
Image Processing, Computer-Assisted/methods , Lung/diagnostic imaging , Humans , Machine Learning
11.
Med Phys ; 43(12): 6569, 2016 Dec.
Article in English | MEDLINE | ID: mdl-27908182

ABSTRACT

PURPOSE: Prostate brachytherapy is an important treatment technique for patients with localized prostate cancer. An inflatable rectal ultrasound probe cover is frequently utilized during the procedure to adjust for unfavorable prostate position relative to the implant grid. However, the inflated cover causes prostate deformation, which is not accounted for during dosimetric planning. Most of the therapeutic dose is delivered after the procedure when the prostate and surrounding organs-at-risk are less deformed. The aim of this study is to quantify the potential dosimetry changes between the initial plan (prostate deformed) and the more realistic dosimetry when the prostate is less deformed without the cover. METHODS: The authors prospectively collected the ultrasound images of the prostate immediately preceding and just after inflation of the rectal probe cover from thirty-four consecutive patients undergoing real-time planning of I-125 permanent seed implant. Manual segmentations of the deformed and undeformed images from each case were used as the input for model training to generate the initial transformation of a testing patient. During registration, the pixel-to-pixel transformation was further optimized to maximize the mutual information between the transferred deformed image and the undeformed images. The accuracy of image registration was evaluated by comparing the displacement of the urethra and calcification landmarks and by determining the Dice index between the registered and manual prostate contours. After registration, using the optimized transformation, the implanted seeds were mapped from the deformed prostate onto the undeformed prostate. The dose distribution of the undeformed anatomy, calculated using the VariSeed treatment planning system, was then analyzed and compared with that of the deformed prostate. RESULTS: The accuracy of image registration was 1.5 ± 1.0 mm when evaluated by the displacement of calcification landmarks, 1.9 ± 1.1 mm when characterized by the displacement of the centroid of the urethra, and 0.86 ± 0.05 from the determination of the Dice index of prostate contours. The magnitude of dosimetric changes was associated with the degree of prostate deformation. The prostate coverage V100% dropped from 96.6 ± 1.7% on prostate-deformed plans to 92.6 ± 3.8% (p < 0.01) on undeformed plans, and the rectum V100% decreased from 0.48 ± 0.39 to 0.06 ± 0.14 cm3 (p < 0.01). The dose to the urethra increased, with the V150% increasing from 0.02 ± 0.06 to 0.11 ± 0.10 cm3 (p < 0.01) and D1% changing from 203.5 ± 22.7 to 239.5 ± 25.6 Gy (p < 0.01). CONCLUSIONS: Prostate deformation from the inflation of an ultrasound rectal probe cover can significantly alter brachytherapy dosimetry. The authors have developed a deformable image registration method that allows for the characterization of dose with the undeformed anatomy. This may be used to more accurately reflect the dosimetry when the prostate is not deformed by the probe cover.


Subject(s)
Brachytherapy/instrumentation , Prostate/diagnostic imaging , Prostheses and Implants , Rectum , Aged , Humans , Image Processing, Computer-Assisted , Male , Middle Aged , Prostate/radiation effects , Prostatic Neoplasms/diagnostic imaging , Prostatic Neoplasms/radiotherapy , Radiometry , Ultrasonography
12.
IEEE Trans Med Imaging ; 35(6): 1532-43, 2016 06.
Article in English | MEDLINE | ID: mdl-26800531

ABSTRACT

Segmenting male pelvic organs from CT images is a prerequisite for prostate cancer radiotherapy. The efficacy of radiation treatment highly depends on segmentation accuracy. However, accurate segmentation of male pelvic organs is challenging due to low tissue contrast of CT images, as well as large variations of shape and appearance of the pelvic organs. Among existing segmentation methods, deformable models are the most popular, as shape prior can be easily incorporated to regularize the segmentation. Nonetheless, the sensitivity to initialization often limits their performance, especially for segmenting organs with large shape variations. In this paper, we propose a novel approach to guide deformable models, thus making them robust against arbitrary initializations. Specifically, we learn a displacement regressor, which predicts 3D displacement from any image voxel to the target organ boundary based on the local patch appearance. This regressor provides a non-local external force for each vertex of deformable model, thus overcoming the initialization problem suffered by the traditional deformable models. To learn a reliable displacement regressor, two strategies are particularly proposed. 1) A multi-task random forest is proposed to learn the displacement regressor jointly with the organ classifier; 2) an auto-context model is used to iteratively enforce structural information during voxel-wise prediction. Extensive experiments on 313 planning CT scans of 313 patients show that our method achieves better results than alternative classification or regression based methods, and also several other existing methods in CT pelvic organ segmentation.


Subject(s)
Image Interpretation, Computer-Assisted/methods , Machine Learning , Pelvis/diagnostic imaging , Prostate/diagnostic imaging , Tomography, X-Ray Computed/methods , Decision Trees , Humans , Male , Neural Networks, Computer , Prostatic Neoplasms/diagnostic imaging , Rectum/diagnostic imaging , Urinary Bladder/diagnostic imaging
13.
Med Image Anal ; 26(1): 345-56, 2015 Dec.
Article in English | MEDLINE | ID: mdl-26439938

ABSTRACT

Automatic and accurate segmentation of the prostate and rectum in planning CT images is a challenging task due to low image contrast, unpredictable organ (relative) position, and uncertain existence of bowel gas across different patients. Recently, regression forest was adopted for organ deformable segmentation on 2D medical images by training one landmark detector for each point on the shape model. However, it seems impractical for regression forest to guide 3D deformable segmentation as a landmark detector, due to large number of vertices in the 3D shape model as well as the difficulty in building accurate 3D vertex correspondence for each landmark detector. In this paper, we propose a novel boundary detection method by exploiting the power of regression forest for prostate and rectum segmentation. The contributions of this paper are as follows: (1) we introduce regression forest as a local boundary regressor to vote the entire boundary of a target organ, which avoids training a large number of landmark detectors and building an accurate 3D vertex correspondence for each landmark detector; (2) an auto-context model is integrated with regression forest to improve the accuracy of the boundary regression; (3) we further combine a deformable segmentation method with the proposed local boundary regressor for the final organ segmentation by integrating organ shape priors. Our method is evaluated on a planning CT image dataset with 70 images from 70 different patients. The experimental results show that our proposed boundary regression method outperforms the conventional boundary classification method in guiding the deformable model for prostate and rectum segmentations. Compared with other state-of-the-art methods, our method also shows a competitive performance.


Subject(s)
Anatomic Landmarks/diagnostic imaging , Prostate/diagnostic imaging , Prostatic Neoplasms/diagnostic imaging , Radiotherapy, Image-Guided/methods , Rectum/diagnostic imaging , Tomography, X-Ray Computed/methods , Humans , Imaging, Three-Dimensional/methods , Male , Pattern Recognition, Automated/methods , Prostatic Neoplasms/radiotherapy , Radiographic Image Enhancement/methods , Radiographic Image Interpretation, Computer-Assisted/methods , Radiotherapy Planning, Computer-Assisted/methods , Regression Analysis , Reproducibility of Results , Sensitivity and Specificity , Subtraction Technique
14.
IEEE Trans Med Imaging ; 33(9): 1761-80, 2014 Sep.
Article in English | MEDLINE | ID: mdl-25181734

ABSTRACT

Lung field segmentation in the posterior-anterior (PA) chest radiograph is important for pulmonary disease diagnosis and hemodialysis treatment. Due to high shape variation and boundary ambiguity, accurate lung field segmentation from chest radiograph is still a challenging task. To tackle these challenges, we propose a joint shape and appearance sparse learning method for robust and accurate lung field segmentation. The main contributions of this paper are: 1) a robust shape initialization method is designed to achieve an initial shape that is close to the lung boundary under segmentation; 2) a set of local sparse shape composition models are built based on local lung shape segments to overcome the high shape variations; 3) a set of local appearance models are similarly adopted by using sparse representation to capture the appearance characteristics in local lung boundary segments, thus effectively dealing with the lung boundary ambiguity; 4) a hierarchical deformable segmentation framework is proposed to integrate the scale-dependent shape and appearance information together for robust and accurate segmentation. Our method is evaluated on 247 PA chest radiographs in a public dataset. The experimental results show that the proposed local shape and appearance models outperform the conventional shape and appearance models. Compared with most of the state-of-the-art lung field segmentation methods under comparison, our method also shows a higher accuracy, which is comparable to the inter-observer annotation variation.


Subject(s)
Algorithms , Image Processing, Computer-Assisted/methods , Lung/diagnostic imaging , Radiography, Thoracic/methods , Databases, Factual , Humans
15.
Med Phys ; 41(7): 072303, 2014 Jul.
Article in English | MEDLINE | ID: mdl-24989402

ABSTRACT

PURPOSE: Automatic prostate segmentation from MR images is an important task in various clinical applications such as prostate cancer staging and MR-guided radiotherapy planning. However, the large appearance and shape variations of the prostate in MR images make the segmentation problem difficult to solve. Traditional Active Shape/Appearance Model (ASM/AAM) has limited accuracy on this problem, since its basic assumption, i.e., both shape and appearance of the targeted organ follow Gaussian distributions, is invalid in prostate MR images. To this end, the authors propose a sparse dictionary learning method to model the image appearance in a nonparametric fashion and further integrate the appearance model into a deformable segmentation framework for prostate MR segmentation. METHODS: To drive the deformable model for prostate segmentation, the authors propose nonparametric appearance and shape models. The nonparametric appearance model is based on a novel dictionary learning method, namely distributed discriminative dictionary (DDD) learning, which is able to capture fine distinctions in image appearance. To increase the differential power of traditional dictionary-based classification methods, the authors' DDD learning approach takes three strategies. First, two dictionaries for prostate and nonprostate tissues are built, respectively, using the discriminative features obtained from minimum redundancy maximum relevance feature selection. Second, linear discriminant analysis is employed as a linear classifier to boost the optimal separation between prostate and nonprostate tissues, based on the representation residuals from sparse representation. Third, to enhance the robustness of the authors' classification method, multiple local dictionaries are learned for local regions along the prostate boundary (each with small appearance variations), instead of learning one global classifier for the entire prostate. These discriminative dictionaries are located on different patches of the prostate surface and trained to adaptively capture the appearance in different prostate zones, thus achieving better local tissue differentiation. For each local region, multiple classifiers are trained based on the randomly selected samples and finally assembled by a specific fusion method. In addition to this nonparametric appearance model, a prostate shape model is learned from the shape statistics using a novel approach, sparse shape composition, which can model nonGaussian distributions of shape variation and regularize the 3D mesh deformation by constraining it within the observed shape subspace. RESULTS: The proposed method has been evaluated on two datasets consisting of T2-weighted MR prostate images. For the first (internal) dataset, the classification effectiveness of the authors' improved dictionary learning has been validated by comparing it with three other variants of traditional dictionary learning methods. The experimental results show that the authors' method yields a Dice Ratio of 89.1% compared to the manual segmentation, which is more accurate than the three state-of-the-art MR prostate segmentation methods under comparison. For the second dataset, the MICCAI 2012 challenge dataset, the authors' proposed method yields a Dice Ratio of 87.4%, which also achieves better segmentation accuracy than other methods under comparison. CONCLUSIONS: A new magnetic resonance image prostate segmentation method is proposed based on the combination of deformable model and dictionary learning methods, which achieves more accurate segmentation performance on prostate T2 MR images.


Subject(s)
Artificial Intelligence , Image Interpretation, Computer-Assisted/methods , Imaging, Three-Dimensional/methods , Magnetic Resonance Imaging/methods , Prostate/pathology , Algorithms , Discriminant Analysis , Humans , Linear Models , Male , Models, Biological , Prostatic Neoplasms/pathology , Sensitivity and Specificity
16.
Mach Learn Med Imaging ; 8679: 93-100, 2014.
Article in English | MEDLINE | ID: mdl-30123893

ABSTRACT

Segmenting the prostate from CT images is a critical step in the radio-therapy planning for prostate cancer. The segmentation accuracy could largely affect the efficacy of radiation treatment. However, due to the touching boundaries with the bladder and the rectum, the prostate boundary is often ambiguous and hard to recognize, which leads to inconsistent manual delineations across different clinicians. In this paper, we propose a learning-based approach for boundary detection and deformable segmentation of the prostate. Our proposed method aims to learn a boundary distance transform, which maps an intensity image into a boundary distance map. To enforce the spatial consistency on the learned distance transform, we combine our approach with the auto-context model for iteratively refining the estimated distance map. After the refinement, the prostate boundaries can be readily detected by finding the valley in the distance map. In addition, the estimated distance map can also be used as a new external force for guiding the deformable segmentation. Specifically, to automatically segment the prostate, we integrate the estimated boundary distance map into a level set formulation. Experimental results on 73 CT planning images show that the proposed distance transform is more effective than the traditional classification-based method for driving the deformable segmentation. Also, our method can achieve more consistent segmentations than human raters, and more accurate results than the existing methods under comparison.

SELECTION OF CITATIONS
SEARCH DETAIL
...