Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 7 de 7
Filter
Add more filters










Database
Language
Publication year range
1.
Comput Biol Med ; 176: 108609, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38772056

ABSTRACT

Semi-supervised medical image segmentation presents a compelling approach to streamline large-scale image analysis, alleviating annotation burdens while maintaining comparable performance. Despite recent strides in cross-supervised training paradigms, challenges persist in addressing sub-network disagreement and training efficiency and reliability. In response, our paper introduces a novel cross-supervised learning framework, Quality-driven Deep Cross-supervised Learning Network (QDC-Net). QDC-Net incorporates both an evidential sub-network and an vanilla sub-network, leveraging their complementary strengths to effectively handle disagreement. To enable the reliability and efficiency of semi-supervised training, we introduce a real-time quality estimation of the model's segmentation performance and propose a directional cross-training approach through the design of directional weights. We further design a truncated form of sample-wise loss weighting to mitigate the impact of inaccurate predictions and collapsed samples in semi-supervised training. Extensive experiments on LA and Pancreas-CT datasets demonstrate that QDC-Net surpasses other state-of-the-art methods in semi-supervised medical image segmentation. Code release is available at https://github.com/Medsemiseg.


Subject(s)
Supervised Machine Learning , Humans , Deep Learning , Image Processing, Computer-Assisted/methods , Pancreas/diagnostic imaging , Tomography, X-Ray Computed
2.
IEEE J Biomed Health Inform ; 28(2): 929-940, 2024 Feb.
Article in English | MEDLINE | ID: mdl-37930923

ABSTRACT

Semi-supervised learning methods have been explored to mitigate the scarcity of pixel-level annotation in medical image segmentation tasks. Consistency learning, serving as a mainstream method in semi-supervised training, suffers from low efficiency and poor stability due to inaccurate supervision and insufficient feature representation. Prototypical learning is one potential and plausible way to handle this problem due to the nature of feature aggregation in prototype calculation. However, the previous works have not fully studied how to enhance the supervision quality and feature representation using prototypical learning under the semi-supervised condition. To address this issue, we propose an implicit-explicit alignment (IEPAlign) framework to foster semi-supervised consistency training. In specific, we develop an implicit prototype alignment method based on dynamic multiple prototypes on-the-fly. And then, we design a multiple prediction voting strategy for reliable unlabeled mask generation and prototype calculation to improve the supervision quality. Afterward, to boost the intra-class consistency and inter-class separability of pixel-wise features in semi-supervised segmentation, we construct a region-aware hierarchical prototype alignment, which transmits information from labeled to unlabeled and from certain regions to uncertain regions. We evaluate IEPAlign on three medical image segmentation tasks. The extensive experimental results demonstrate that the proposed method outperforms other popular semi-supervised segmentation methods and achieves comparable performance with fully-supervised training methods.


Subject(s)
Supervised Machine Learning , Voting , Image Processing, Computer-Assisted
3.
IEEE Trans Image Process ; 32: 2593-2607, 2023.
Article in English | MEDLINE | ID: mdl-37126632

ABSTRACT

Salient object detection (SOD) is an important task in computer vision that aims to identify visually conspicuous regions in images. RGB-Thermal SOD combines two spectra to achieve better segmentation results. However, most existing methods for RGB-T SOD use boundary maps to learn sharp boundaries, which lead to sub-optimal performance as they ignore the interactions between isolated boundary pixels and other confident pixels. To address this issue, we propose a novel position-aware relation learning network (PRLNet) for RGB-T SOD. PRLNet explores the distance and direction relationships between pixels by designing an auxiliary task and optimizing the feature structure to strengthen intra-class compactness and inter-class separation. Our method consists of two main components: A signed distance map auxiliary module (SDMAM), and a feature refinement approach with direction field (FRDF). SDMAM improves the encoder feature representation by considering the distance relationship between foreground-background pixels and boundaries, which increases the inter-class separation between foreground and background features. FRDF rectifies the features of boundary neighborhoods by exploiting the features inside salient objects. It utilizes the direction relationship of object pixels to enhance the intra-class compactness of salient features. In addition, we constitute a transformer-based decoder to decode multispectral feature representation. Experimental results on three public RGB-T SOD datasets demonstrate that our proposed method not only outperforms the state-of-the-art methods, but also can be integrated with different backbone networks in a plug-and-play manner. Ablation study and visualizations further prove the validity and interpretability of our method.

4.
Med Image Anal ; 79: 102458, 2022 07.
Article in English | MEDLINE | ID: mdl-35500497

ABSTRACT

Pixel-wise error correction of initial segmentation results provides an effective way for quality improvement. The additional error segmentation network learns to identify correct predictions and incorrect ones. The performance on error segmentation directly affects the accuracy on the test set and the subsequent self-training with the error-corrected pseudo labels. In this paper, we propose a novel label rectification method based on error correction, namely ECLR, which can be directly added after the fully-supervised segmentation framework. Moreover, it can be used to guide the semi-supervised learning (SSL) process, constituting an error correction guided SSL framework, called ECGSSL. Specifically, we analyze the types and causes of segmentation error, and divide it into intra-class error and inter-class error caused by intra-class inconsistency and inter-class similarity problems in segmentation, respectively. Further, we propose a collaborative multi-task discriminative error prediction network (DEP-Net) to highlight two error types. For better training of DEP-Net, we propose specific mask degradation methods representing typical segmentation errors. Under the fully-supervised regime, the pre-trained DEP-Net is used to directly rectify the initial segmentation results of the test set. While, under the semi-supervised regime, a dual error correction method is proposed for unlabeled data to obtain more reliable network re-training. Our method is easy to apply to different segmentation models. Extensive experiments on gland segmentation verify that ECLR yields substantial improvements based on initial segmentation predictions. ECGSSL shows consistent improvements over a supervised baseline learned only from labeled data and achieves competitive performance compared with other popular semi-supervised methods.


Subject(s)
Colon , Supervised Machine Learning , Humans
5.
J Microbiol Methods ; 153: 66-73, 2018 10.
Article in English | MEDLINE | ID: mdl-30195830

ABSTRACT

Counting colonies is usually used in microbiological analysis to assess if samples meet microbiological criteria. Although manual counting remains gold standard, the process is subjective, tedious, and time-consuming. Some developed automatic counting methods could save labors and time, but their results are easily affected by uneven illumination and reflection of visible light. To offer a method which counts colonies automatically and is robust to light, we constructed a convenient and cost-effective system to obtain images of colonies at near-infrared light, and proposed an automatic method to detect and count colonies by processing images. The colonies cultured by using raw cows' milk were used as identification objects. The developed system mainly consisted of a visible/near-infrared camera and a circular near-infrared illuminator. The automatic method proposed to count colonies includes four steps, i.e., eliminating noises outside agar plate, removing plate rim and wall, identifying and separating clustered or overlapped colonies, and counting colonies by using connected region labelling, distance transform, and watershed algorithms, etc. A user-friendly graphic user interface was also developed for the proposed method. The relative error and counting time of the automatic counting method were compared with those of manual counting. The results showed that the relative error of the automatic counting method was -7.4%~ + 8.3%, with average relative error of 0.2%, and the time used for counting colonies on each agar plate was 11-21 s, which was 15-75% of the time used in manual counting, depending on the numbers of colonies on agar plates. The proposed system and automatic counting method demonstrate promising performance in terms of precision, and they are robust and efficient in terms of labor- and time-savings.


Subject(s)
Automation, Laboratory/methods , Bacteria/growth & development , Colony Count, Microbial/instrumentation , Colony Count, Microbial/methods , Infrared Rays , Agar , Algorithms , Animals , Cattle , Image Processing, Computer-Assisted , Milk/microbiology , Raw Foods/microbiology
6.
IEEE Trans Image Process ; 24(12): 5236-48, 2015 Dec.
Article in English | MEDLINE | ID: mdl-26390460

ABSTRACT

This paper presents a new visual tracking framework based on an adaptive color attention tuned local sparse model. The histograms of sparse coefficients of all patches in an object are pooled together according to their spatial distribution. A particle filter methodology is used as the location model to predict candidates for object verification during tracking. Since color is an important visual clue to distinguish objects from background, we calculate the color similarity between objects in the previous frames and the candidates in current frame, which is adopted as color attention to tune the local sparse representation-based appearance similarity measurement between the object template and candidates. The color similarity can be calculated efficiently with hash coded color names, which helps the tracker find more reliable objects during tracking. We use a flexible local sparse coding of the object to evaluate the degeneration degree of the appearance model, based on which we build a model updating mechanism to alleviate drifting caused by temporal varying multi-factors. Experiments on 76 challenging benchmark color sequences and the evaluation under the object tracking benchmark protocol demonstrate the superiority of the proposed tracker over the state-of-the-art methods in accuracy.

7.
IEEE Trans Syst Man Cybern B Cybern ; 42(2): 320-33, 2012 Apr.
Article in English | MEDLINE | ID: mdl-22318490

ABSTRACT

Face images under uncontrolled environments suffer from the changes of multiple factors such as camera view, illumination, expression, etc. Tensor analysis provides a way of analyzing the influence of different factors on facial variation. However, the TensorFace model creates a difficulty in representing the nonlinearity of view subspace. In this paper, to break this limitation, we present a view-manifold-based TensorFace (V-TensorFace), in which the latent view manifold preserves the local distances in the multiview face space. Moreover, a kernelized TensorFace (K-TensorFace) for multiview face recognition is proposed to preserve the structure of the latent manifold in the image space. Both methods provide a generative model that involves a continuous view manifold for unseen view representation. Most importantly, we propose a unified framework to generalize TensorFace, V-TensorFace, and K-TensorFace. Finally, an expectation-maximization like algorithm is developed to estimate the identity and view parameters iteratively for a face image of an unknown/unseen view. The experiment on the PIE database shows the effectiveness of the manifold construction method. Extensive comparison experiments on Weizmann and Oriental Face databases for multiview face recognition demonstrate the superiority of the proposed V- and K-TensorFace methods over the view-based principal component analysis and other state-of-the-art approaches for such purpose.


Subject(s)
Biometric Identification/methods , Face/anatomy & histology , Algorithms , Artificial Intelligence , Databases, Factual , Humans , Linear Models , Nonlinear Dynamics
SELECTION OF CITATIONS
SEARCH DETAIL
...