Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 144.932
Filter
1.
J Biomed Opt ; 29(Suppl 2): S22702, 2025 Dec.
Article in English | MEDLINE | ID: mdl-38434231

ABSTRACT

Significance: Advancements in label-free microscopy could provide real-time, non-invasive imaging with unique sources of contrast and automated standardized analysis to characterize heterogeneous and dynamic biological processes. These tools would overcome challenges with widely used methods that are destructive (e.g., histology, flow cytometry) or lack cellular resolution (e.g., plate-based assays, whole animal bioluminescence imaging). Aim: This perspective aims to (1) justify the need for label-free microscopy to track heterogeneous cellular functions over time and space within unperturbed systems and (2) recommend improvements regarding instrumentation, image analysis, and image interpretation to address these needs. Approach: Three key research areas (cancer research, autoimmune disease, and tissue and cell engineering) are considered to support the need for label-free microscopy to characterize heterogeneity and dynamics within biological systems. Based on the strengths (e.g., multiple sources of molecular contrast, non-invasive monitoring) and weaknesses (e.g., imaging depth, image interpretation) of several label-free microscopy modalities, improvements for future imaging systems are recommended. Conclusion: Improvements in instrumentation including strategies that increase resolution and imaging speed, standardization and centralization of image analysis tools, and robust data validation and interpretation will expand the applications of label-free microscopy to study heterogeneous and dynamic biological systems.


Subject(s)
Histological Techniques , Microscopy , Animals , Flow Cytometry , Image Processing, Computer-Assisted
2.
Sci Rep ; 14(1): 12630, 2024 Jun 02.
Article in English | MEDLINE | ID: mdl-38824210

ABSTRACT

In this study, we present the development of a fine structural human phantom designed specifically for applications in dentistry. This research focused on assessing the viability of applying medical computer vision techniques to the task of segmenting individual teeth within a phantom. Using a virtual cone-beam computed tomography (CBCT) system, we generated over 170,000 training datasets. These datasets were produced by varying the elemental densities and tooth sizes within the human phantom, as well as varying the X-ray spectrum, noise intensity, and projection cutoff intensity in the virtual CBCT system. The deep-learning (DL) based tooth segmentation model was trained using the generated datasets. The results demonstrate an agreement with manual contouring when applied to clinical CBCT data. Specifically, the Dice similarity coefficient exceeded 0.87, indicating the robust performance of the developed segmentation model even when virtual imaging was used. The present results show the practical utility of virtual imaging techniques in dentistry and highlight the potential of medical computer vision for enhancing precision and efficiency in dental imaging processes.


Subject(s)
Cone-Beam Computed Tomography , Phantoms, Imaging , Tooth , Humans , Tooth/diagnostic imaging , Tooth/anatomy & histology , Cone-Beam Computed Tomography/methods , Dentistry/methods , Image Processing, Computer-Assisted/methods , Deep Learning
3.
Radiat Oncol ; 19(1): 69, 2024 May 31.
Article in English | MEDLINE | ID: mdl-38822385

ABSTRACT

BACKGROUND: Multiple artificial intelligence (AI)-based autocontouring solutions have become available, each promising high accuracy and time savings compared with manual contouring. Before implementing AI-driven autocontouring into clinical practice, three commercially available CT-based solutions were evaluated. MATERIALS AND METHODS: The following solutions were evaluated in this work: MIM-ProtégéAI+ (MIM), Radformation-AutoContour (RAD), and Siemens-DirectORGANS (SIE). Sixteen organs were identified that could be contoured by all solutions. For each organ, ten patients that had manually generated contours approved by the treating physician (AP) were identified, totaling forty-seven different patients. CT scans in the supine position were acquired using a Siemens-SOMATOMgo 64-slice helical scanner and used to generate autocontours. Physician scoring of contour accuracy was performed by at least three physicians using a five-point Likert scale. Dice similarity coefficient (DSC), Hausdorff distance (HD) and mean distance to agreement (MDA) were calculated comparing AI contours to "ground truth" AP contours. RESULTS: The average physician score ranged from 1.00, indicating that all physicians reviewed the contour as clinically acceptable with no modifications necessary, to 3.70, indicating changes are required and that the time taken to modify the structures would likely take as long or longer than manually generating the contour. When averaged across all sixteen structures, the AP contours had a physician score of 2.02, MIM 2.07, RAD 1.96 and SIE 1.99. DSC ranged from 0.37 to 0.98, with 41/48 (85.4%) contours having an average DSC ≥ 0.7. Average HD ranged from 2.9 to 43.3 mm. Average MDA ranged from 0.6 to 26.1 mm. CONCLUSIONS: The results of our comparison demonstrate that each vendor's AI contouring solution exhibited capabilities similar to those of manual contouring. There were a small number of cases where unusual anatomy led to poor scores with one or more of the solutions. The consistency and comparable performance of all three vendors' solutions suggest that radiation oncology centers can confidently choose any of the evaluated solutions based on individual preferences, resource availability, and compatibility with their existing clinical workflows. Although AI-based contouring may result in high-quality contours for the majority of patients, a minority of patients require manual contouring and more in-depth physician review.


Subject(s)
Artificial Intelligence , Radiotherapy Planning, Computer-Assisted , Tomography, X-Ray Computed , Humans , Radiotherapy Planning, Computer-Assisted/methods , Organs at Risk/radiation effects , Algorithms , Image Processing, Computer-Assisted/methods
4.
Neuroimage ; 294: 120631, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-38701993

ABSTRACT

INTRODUCTION: Spatial normalization is a prerequisite step for the quantitative analysis of SPECT or PET brain images using volume-of-interest (VOI) template or voxel-based analysis. MRI-guided spatial normalization is the gold standard, but the wide use of PET/CT or SPECT/CT in routine clinical practice makes CT-guided spatial normalization a necessary alternative. Ventricular enlargement is observed with aging, and it hampers the spatial normalization of the lateral ventricles and striatal regions, limiting their analysis. The aim of the present study was to propose a robust spatial normalization method based on CT scans that takes into account features of the aging brain to reduce bias in the CT-guided striatal analysis of SPECT images. METHODS: We propose an enhanced CT-guided spatial normalization pipeline based on SPM12. Performance of the proposed pipeline was assessed on visually normal [123I]-FP-CIT SPECT/CT images. SPM12 default CT-guided spatial normalization was used as reference method. The metrics assessed were the overlap between the spatially normalized lateral ventricles and caudate/putamen VOIs, and the computation of caudate and putamen specific binding ratios (SBR). RESULTS: In total 231 subjects (mean age ± SD = 61.9 ± 15.5 years) were included in the statistical analysis. The mean overlap between the spatially normalized lateral ventricles of subjects and the caudate VOI and the mean SBR of caudate were respectively 38.40 % (± SD = 19.48 %) of the VOI and 1.77 (± 0.79) when performing SPM12 default spatial normalization. The mean overlap decreased to 9.13 % (± SD = 1.41 %, P < 0.001) of the VOI and the SBR of caudate increased to 2.38 (± 0.51, P < 0.0001) when performing the proposed pipeline. Spatially normalized lateral ventricles did not overlap with putamen VOI using either method. The mean putamen SBR value derived from the proposed spatial normalization (2.75 ± 0.54) was not significantly different from that derived from the default SPM12 spatial normalization (2.83 ± 0.52, P > 0.05). CONCLUSION: The automatic CT-guided spatial normalization used herein led to a less biased spatial normalization of SPECT images, hence an improved semi-quantitative analysis. The proposed pipeline could be implemented in clinical routine to perform a more robust SBR computation using hybrid imaging.


Subject(s)
Corpus Striatum , Humans , Male , Female , Middle Aged , Aged , Adult , Corpus Striatum/diagnostic imaging , Corpus Striatum/metabolism , Tomography, X-Ray Computed/methods , Tomography, X-Ray Computed/standards , Tomography, Emission-Computed, Single-Photon/methods , Cerebral Ventricles/diagnostic imaging , Cerebral Ventricles/metabolism , Image Processing, Computer-Assisted/methods , Tropanes
5.
Comput Methods Programs Biomed ; 251: 108201, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38703719

ABSTRACT

BACKGROUND AND OBJECTIVE: Surgical robotics tends to develop cognitive control architectures to provide certain degree of autonomy to improve patient safety and surgery outcomes, while decreasing the required surgeons' cognitive load dedicated to low level decisions. Cognition needs workspace perception, which is an essential step towards automatic decision-making and task planning capabilities. Robust and accurate detection and tracking in minimally invasive surgery suffers from limited visibility, occlusions, anatomy deformations and camera movements. METHOD: This paper develops a robust methodology to detect and track anatomical structures in real time to be used in automatic control of robotic systems and augmented reality. The work focuses on the experimental validation in highly challenging surgery: fetoscopic repair of Open Spina Bifida. The proposed method is based on two sequential steps: first, selection of relevant points (contour) using a Convolutional Neural Network and, second, reconstruction of the anatomical shape by means of deformable geometric primitives. RESULTS: The methodology performance was validated with different scenarios. Synthetic scenario tests, designed for extreme validation conditions, demonstrate the safety margin offered by the methodology with respect to the nominal conditions during surgery. Real scenario experiments have demonstrated the validity of the method in terms of accuracy, robustness and computational efficiency. CONCLUSIONS: This paper presents a robust anatomical structure detection in present of abrupt camera movements, severe occlusions and deformations. Even though the paper focuses on a case study, Open Spina Bifida, the methodology is applicable in all anatomies which contours can be approximated by geometric primitives. The methodology is designed to provide effective inputs to cognitive robotic control and augmented reality systems that require accurate tracking of sensitive anatomies.


Subject(s)
Robotic Surgical Procedures , Humans , Robotic Surgical Procedures/methods , Neural Networks, Computer , Algorithms , Spinal Dysraphism/surgery , Spinal Dysraphism/diagnostic imaging , Image Processing, Computer-Assisted/methods , Robotics , Augmented Reality
6.
Nat Commun ; 15(1): 4598, 2024 May 30.
Article in English | MEDLINE | ID: mdl-38816394

ABSTRACT

Fluorescence microscopy has undergone rapid advancements, offering unprecedented visualization of biological events and shedding light on the intricate mechanisms governing living organisms. However, the exploration of rapid biological dynamics still poses a significant challenge due to the limitations of current digital camera architectures and the inherent compromise between imaging speed and other capabilities. Here, we introduce sHAPR, a high-speed acquisition technique that leverages the operating principles of sCMOS cameras to capture fast cellular and subcellular processes. sHAPR harnesses custom fiber optics to convert microscopy images into one-dimensional recordings, enabling acquisition at the maximum camera readout rate, typically between 25 and 250 kHz. We have demonstrated the utility of sHAPR with a variety of phantom and dynamic systems, including high-throughput flow cytometry, cardiomyocyte contraction, and neuronal calcium waves, using a standard epi-fluorescence microscope. sHAPR is highly adaptable and can be integrated into existing microscopy systems without requiring extensive platform modifications. This method pushes the boundaries of current fluorescence imaging capabilities, opening up new avenues for investigating high-speed biological phenomena.


Subject(s)
Microscopy, Fluorescence , Optical Imaging , Microscopy, Fluorescence/methods , Animals , Optical Imaging/methods , Optical Imaging/instrumentation , Humans , Myocytes, Cardiac/cytology , Phantoms, Imaging , Flow Cytometry/methods , Neurons , Image Processing, Computer-Assisted/methods
7.
PLoS Comput Biol ; 20(5): e1012075, 2024 May.
Article in English | MEDLINE | ID: mdl-38768230

ABSTRACT

Tracking body parts in behaving animals, extracting fluorescence signals from cells embedded in deforming tissue, and analyzing cell migration patterns during development all require tracking objects with partially correlated motion. As dataset sizes increase, manual tracking of objects becomes prohibitively inefficient and slow, necessitating automated and semi-automated computational tools. Unfortunately, existing methods for multiple object tracking (MOT) are either developed for specific datasets and hence do not generalize well to other datasets, or require large amounts of training data that are not readily available. This is further exacerbated when tracking fluorescent sources in moving and deforming tissues, where the lack of unique features and sparsely populated images create a challenging environment, especially for modern deep learning techniques. By leveraging technology recently developed for spatial transformer networks, we propose ZephIR, an image registration framework for semi-supervised MOT in 2D and 3D videos. ZephIR can generalize to a wide range of biological systems by incorporating adjustable parameters that encode spatial (sparsity, texture, rigidity) and temporal priors of a given data class. We demonstrate the accuracy and versatility of our approach in a variety of applications, including tracking the body parts of a behaving mouse and neurons in the brain of a freely moving C. elegans. We provide an open-source package along with a web-based graphical user interface that allows users to provide small numbers of annotations to interactively improve tracking results.


Subject(s)
Computational Biology , Animals , Mice , Computational Biology/methods , Caenorhabditis elegans/physiology , Imaging, Three-Dimensional/methods , Image Processing, Computer-Assisted/methods , Algorithms , Deep Learning
8.
Bioinformatics ; 40(5)2024 May 02.
Article in English | MEDLINE | ID: mdl-38775410

ABSTRACT

MOTIVATION: Accurate segmentation and recognition of C.elegans cells are critical for various biological studies, including gene expression, cell lineages, and cell fates analysis at single-cell level. However, the highly dense distribution, similar shapes, and inhomogeneous intensity profiles of whole-body cells in 3D fluorescence microscopy images make automatic cell segmentation and recognition a challenging task. Existing methods either rely on additional fiducial markers or only handle a subset of cells. Given the difficulty or expense associated with generating fiducial features in many experimental settings, a marker-free approach capable of reliably segmenting and recognizing C.elegans whole-body cells is highly desirable. RESULTS: We report a new pipeline, called automated segmentation and recognition (ASR) of cells, and applied it to 3D fluorescent microscopy images of L1-stage C.elegans with 558 whole-body cells. A novel displacement vector field based deep learning model is proposed to address the problem of reliable segmentation of highly crowded cells with blurred boundary. We then realize the cell recognition by encoding and exploiting statistical priors on cell positions and structural similarities of neighboring cells. To the best of our knowledge, this is the first method successfully applied to the segmentation and recognition of C.elegans whole-body cells. The ASR-segmentation module achieves an F1-score of 0.8956 on a dataset of 116 C.elegans image stacks with 64 728 cells (accuracy 0.9880, AJI 0.7813). Based on the segmentation results, the ASR recognition module achieved an average accuracy of 0.8879. We also show ASR's applicability to other cell types, e.g. platynereis and rat kidney cells. AVAILABILITY AND IMPLEMENTATION: The code is available at https://github.com/reaneyli/ASR.


Subject(s)
Caenorhabditis elegans , Caenorhabditis elegans/cytology , Animals , Microscopy, Fluorescence/methods , Imaging, Three-Dimensional/methods , Image Processing, Computer-Assisted/methods , Algorithms , Deep Learning
9.
Biomed Phys Eng Express ; 10(4)2024 May 31.
Article in English | MEDLINE | ID: mdl-38781934

ABSTRACT

Congenital heart defects (CHD) are one of the serious problems that arise during pregnancy. Early CHD detection reduces death rates and morbidity but is hampered by the relatively low detection rates (i.e., 60%) of current screening technology. The detection rate could be increased by supplementing ultrasound imaging with fetal ultrasound image evaluation (FUSI) using deep learning techniques. As a result, the non-invasive foetal ultrasound image has clear potential in the diagnosis of CHD and should be considered in addition to foetal echocardiography. This review paper highlights cutting-edge technologies for detecting CHD using ultrasound images, which involve pre-processing, localization, segmentation, and classification. Existing technique of preprocessing includes spatial domain filter, non-linear mean filter, transform domain filter, and denoising methods based on Convolutional Neural Network (CNN); segmentation includes thresholding-based techniques, region growing-based techniques, edge detection techniques, Artificial Neural Network (ANN) based segmentation methods, non-deep learning approaches and deep learning approaches. The paper also suggests future research directions for improving current methodologies.


Subject(s)
Deep Learning , Heart Defects, Congenital , Neural Networks, Computer , Ultrasonography, Prenatal , Humans , Heart Defects, Congenital/diagnostic imaging , Ultrasonography, Prenatal/methods , Pregnancy , Female , Image Processing, Computer-Assisted/methods , Echocardiography/methods , Algorithms , Fetal Heart/diagnostic imaging , Fetus/diagnostic imaging
10.
PLoS One ; 19(5): e0297244, 2024.
Article in English | MEDLINE | ID: mdl-38820354

ABSTRACT

Quantitative MRI (qMRI) has been shown to be clinically useful for numerous applications in the brain and body. The development of rapid, accurate, and reproducible qMRI techniques offers access to new multiparametric data, which can provide a comprehensive view of tissue pathology. This work introduces a multiparametric qMRI protocol along with full postprocessing pipelines, optimized for brain imaging at 3 Tesla and using state-of-the-art qMRI tools. The total scan time is under 50 minutes and includes eight pulse-sequences, which produce range of quantitative maps including T1, T2, and T2* relaxation times, magnetic susceptibility, water and macromolecular tissue fractions, mean diffusivity and fractional anisotropy, magnetization transfer ratio (MTR), and inhomogeneous MTR. Practical tips and limitations of using the protocol are also provided and discussed. Application of the protocol is presented on a cohort of 28 healthy volunteers and 12 brain regions-of-interest (ROIs). Quantitative values agreed with previously reported values. Statistical analysis revealed low variability of qMRI parameters across subjects, which, compared to intra-ROI variability, was x4.1 ± 0.9 times higher on average. Significant and positive linear relationship was found between right and left hemispheres' values for all parameters and ROIs with Pearson correlation coefficients of r>0.89 (P<0.001), and mean slope of 0.95 ± 0.04. Finally, scan-rescan stability demonstrated high reproducibility of the measured parameters across ROIs and volunteers, with close-to-zero mean difference and without correlation between the mean and difference values (across map types, mean P value was 0.48 ± 0.27). The entire quantitative data and postprocessing scripts described in the manuscript are publicly available under dedicated GitHub and Figshare repositories. The quantitative maps produced by the presented protocol can promote longitudinal and multi-center studies, and improve the biological interpretability of qMRI by integrating multiple metrics that can reveal information, which is not apparent when examined using only a single contrast mechanism.


Subject(s)
Brain , Magnetic Resonance Imaging , Humans , Brain/diagnostic imaging , Magnetic Resonance Imaging/methods , Adult , Male , Female , Image Processing, Computer-Assisted/methods , Young Adult
11.
PLoS One ; 19(5): e0304610, 2024.
Article in English | MEDLINE | ID: mdl-38820451

ABSTRACT

Face Morphing Attacks pose a threat to the security of identity documents, especially with respect to a subsequent access control process, because they allow both involved individuals to use the same document. Several algorithms are currently being developed to detect Morphing Attacks, often requiring large data sets of morphed face images for training. In the present study, face embeddings are used for two different purposes: first, to pre-select images for the subsequent large-scale generation of Morphing Attacks, and second, to detect potential Morphing Attacks. Previous studies have demonstrated the power of embeddings in both use cases. However, we aim to build on these studies by adding the more powerful MagFace model to both use cases, and by performing comprehensive analyses of the role of embeddings in pre-selection and attack detection in terms of the vulnerability of face recognition systems and attack detection algorithms. In particular, we use recent developments to assess the attack potential, but also investigate the influence of morphing algorithms. For the first objective, an algorithm is developed that pairs individuals based on the similarity of their face embeddings. Different state-of-the-art face recognition systems are used to extract embeddings in order to pre-select the face images and different morphing algorithms are used to fuse the face images. The attack potential of the differently generated morphed face images will be quantified to compare the usability of the embeddings for automatically generating a large number of successful Morphing Attacks. For the second objective, we compare the performance of the embeddings of two state-of-the-art face recognition systems with respect to their ability to detect morphed face images. Our results demonstrate that ArcFace and MagFace provide valuable face embeddings for image pre-selection. Various open-source and commercial-off-the-shelf face recognition systems are vulnerable to the generated Morphing Attacks, and their vulnerability increases when image pre-selection is based on embeddings compared to random pairing. In particular, landmark-based closed-source morphing algorithms generate attacks that pose a high risk to any tested face recognition system. Remarkably, more accurate face recognition systems show a higher vulnerability to Morphing Attacks. Among the systems tested, commercial-off-the-shelf systems were the most vulnerable to Morphing Attacks. In addition, MagFace embeddings stand out as a robust alternative for detecting morphed face images compared to the previously used ArcFace embeddings. The results endorse the benefits of face embeddings for more effective image pre-selection for face morphing and for more accurate detection of morphed face images, as demonstrated by extensive analysis of various designed attacks. The MagFace model is a powerful alternative to the often-used ArcFace model in detecting attacks and can increase performance depending on the use case. It also highlights the usability of embeddings to generate large-scale morphed face databases for various purposes, such as training Morphing Attack Detection algorithms as a countermeasure against attacks.


Subject(s)
Algorithms , Computer Security , Humans , Face , Image Processing, Computer-Assisted/methods , Automated Facial Recognition/methods , Facial Recognition
12.
PLoS One ; 19(5): e0303670, 2024.
Article in English | MEDLINE | ID: mdl-38820462

ABSTRACT

Breast cancer remains a critical global concern, underscoring the urgent need for early detection and accurate diagnosis to improve survival rates among women. Recent developments in deep learning have shown promising potential for computer-aided detection (CAD) systems to address this challenge. In this study, a novel segmentation method based on deep learning is designed to detect tumors in breast ultrasound images. Our proposed approach combines two powerful attention mechanisms: the novel Positional Convolutional Block Attention Module (PCBAM) and Shifted Window Attention (SWA), integrated into a Residual U-Net model. The PCBAM enhances the Convolutional Block Attention Module (CBAM) by incorporating the Positional Attention Module (PAM), thereby improving the contextual information captured by CBAM and enhancing the model's ability to capture spatial relationships within local features. Additionally, we employ SWA within the bottleneck layer of the Residual U-Net to further enhance the model's performance. To evaluate our approach, we perform experiments using two widely used datasets of breast ultrasound images and the obtained results demonstrate its capability in accurately detecting tumors. Our approach achieves state-of-the-art performance with dice score of 74.23% and 78.58% on BUSI and UDIAT datasets, respectively in segmenting the breast tumor region, showcasing its potential to help with precise tumor detection. By leveraging the power of deep learning and integrating innovative attention mechanisms, our study contributes to the ongoing efforts to improve breast cancer detection and ultimately enhance women's survival rates. The source code of our work can be found here: https://github.com/AyushRoy2001/DAUNet.


Subject(s)
Breast Neoplasms , Deep Learning , Ultrasonography, Mammary , Humans , Breast Neoplasms/diagnostic imaging , Female , Ultrasonography, Mammary/methods , Neural Networks, Computer , Algorithms , Image Interpretation, Computer-Assisted/methods , Breast/diagnostic imaging , Breast/pathology , Image Processing, Computer-Assisted/methods
13.
PLoS One ; 19(5): e0303744, 2024.
Article in English | MEDLINE | ID: mdl-38820479

ABSTRACT

During the machine vision inspection of the inner section of bottle caps within pharmaceutical packaging, the unique conca bottom and convex side walls often create obstructions to the illumination. Consequently, this results in challenges such as irregular background and diminished feature contrast in the image, ultimately leading to the misidentification of defects. As a solution, a vision system characterized by a Low-Angle and Large Divergence Angle (LALDA) is presented in this paper. Using the large divergence angle of LED, combined with low-angle illumination, a uniform image of the side wall region with bright-field characteristics and a uniform image of inner circle region at the bottom with dark-field characteristics are obtained, thus solving the problems of light being obscured and brightness overexposure of the background. Based on the imaging characteristics of LALDA, a multi-channel segmentation (MCS) algorithm is designed. The HSV color space has been transformed, and the image is automatically segmented into multiple sub-regions by mutual calculation of different channels. Further, image homogenization and enhancement are used to eliminate fluctuations in the background and to enhance the contrast of defects. In addition, a variety of defect extraction methods are designed based on the imaging characteristics of different sub-regions, which can avoid the problem of over-segmentation in detection. In this paper, the LALDA is applied to the defect detection inside the cap of capsule medicine bottle, the detection speed is better than 400 pcs/min and the detection accuracy is better than 95%, which can meet the actual production line capacity and detection requirements.


Subject(s)
Algorithms , Drug Packaging/methods , Image Processing, Computer-Assisted/methods , Lighting
14.
Zhonghua Wei Chang Wai Ke Za Zhi ; 27(5): 464-470, 2024 May 25.
Article in Chinese | MEDLINE | ID: mdl-38778686

ABSTRACT

Objective: To investigate the feasibility and accuracy of computer vision-based artificial intelligence technology in detecting and recognizing instruments and organs in the scenario of radical laparoscopic gastrectomy for gastric cancer. Methods: Eight complete laparoscopic distal radical gastrectomy surgery videos were collected from four large tertiary hospitals in China (First Medical Center of Chinese PLA General Hospital [three cases], Liaoning Cancer Hospital [two cases], Liyang Branch of Jiangsu Province People's Hospital [two cases], and Fudan University Shanghai Cancer Center [one case]). PR software was used to extract frames every 5-10 seconds and convert them into image frames. To ensure quality, deduplication was performed manually to remove obvious duplication and blurred image frames. After conversion and deduplication, there were 3369 frame images with a resolution of 1,920×1,080 PPI. LabelMe was used for instance segmentation of the images into the following 23 categories: veins, arteries, sutures, needle holders, ultrasonic knives, suction devices, bleeding, colon, forceps, gallbladder, small gauze, Hem-o-lok, Hem-o-lok appliers, electrocautery hooks, small intestine, hepatogastric ligaments, liver, omentum, pancreas, spleen, surgical staplers, stomach, and trocars. The frame images were randomly allocated to training and validation sets in a 9:1 ratio. The YOLOv8 deep learning framework was used for model training and validation. Precision, recall, average precision (AP), and mean average precision (mAP) were used to evaluate detection and recognition accuracy. Results: The training set contained 3032 frame images comprising 30 895 instance segmentation counts across 23 categories. The validation set contained 337 frame images comprising 3407 instance segmentation counts. The YOLOv8m model was used for training. The loss curve of the training set showed a smooth gradual decrease in loss value as the number of iteration calculations increased. In the training set, the AP values of all 23 categories were above 0.90, with a mAP of 0.99, whereas in the validation set, the mAP of the 23 categories was 0.82. As to individual categories, the AP values for ultrasonic knives, needle holders, forceps, gallbladders, small pieces of gauze, and surgical staplers were 0.96, 0.94, 0.91, 0.91, 0.91, and 0.91, respectively. The model successfully inferred and applied to a 5-minutes video segment of laparoscopic gastroenterostomy suturing. Conclusion: The primary finding of this multicenter study is that computer vision can efficiently, accurately, and in real-time detect organs and instruments in various scenarios of radical laparoscopic gastrectomy for gastric cancer.


Subject(s)
Artificial Intelligence , Gastrectomy , Laparoscopy , Stomach Neoplasms , Humans , Stomach Neoplasms/surgery , Laparoscopy/methods , Gastrectomy/methods , Image Processing, Computer-Assisted/methods
15.
PLoS One ; 19(5): e0302880, 2024.
Article in English | MEDLINE | ID: mdl-38718092

ABSTRACT

Gastrointestinal (GI) cancer is leading general tumour in the Gastrointestinal tract, which is fourth significant reason of tumour death in men and women. The common cure for GI cancer is radiation treatment, which contains directing a high-energy X-ray beam onto the tumor while avoiding healthy organs. To provide high dosages of X-rays, a system needs for accurately segmenting the GI tract organs. The study presents a UMobileNetV2 model for semantic segmentation of small and large intestine and stomach in MRI images of the GI tract. The model uses MobileNetV2 as an encoder in the contraction path and UNet layers as a decoder in the expansion path. The UW-Madison database, which contains MRI scans from 85 patients and 38,496 images, is used for evaluation. This automated technology has the capability to enhance the pace of cancer therapy by aiding the radio oncologist in the process of segmenting the organs of the GI tract. The UMobileNetV2 model is compared to three transfer learning models: Xception, ResNet 101, and NASNet mobile, which are used as encoders in UNet architecture. The model is analyzed using three distinct optimizers, i.e., Adam, RMS, and SGD. The UMobileNetV2 model with the combination of Adam optimizer outperforms all other transfer learning models. It obtains a dice coefficient of 0.8984, an IoU of 0.8697, and a validation loss of 0.1310, proving its ability to reliably segment the stomach and intestines in MRI images of gastrointestinal cancer patients.


Subject(s)
Gastrointestinal Neoplasms , Gastrointestinal Tract , Magnetic Resonance Imaging , Humans , Magnetic Resonance Imaging/methods , Gastrointestinal Neoplasms/diagnostic imaging , Gastrointestinal Neoplasms/pathology , Gastrointestinal Tract/diagnostic imaging , Semantics , Image Processing, Computer-Assisted/methods , Female , Male , Stomach/diagnostic imaging , Stomach/pathology
16.
Sci Rep ; 14(1): 10560, 2024 05 08.
Article in English | MEDLINE | ID: mdl-38720020

ABSTRACT

The research on video analytics especially in the area of human behavior recognition has become increasingly popular recently. It is widely applied in virtual reality, video surveillance, and video retrieval. With the advancement of deep learning algorithms and computer hardware, the conventional two-dimensional convolution technique for training video models has been replaced by three-dimensional convolution, which enables the extraction of spatio-temporal features. Specifically, the use of 3D convolution in human behavior recognition has been the subject of growing interest. However, the increased dimensionality has led to challenges such as the dramatic increase in the number of parameters, increased time complexity, and a strong dependence on GPUs for effective spatio-temporal feature extraction. The training speed can be considerably slow without the support of powerful GPU hardware. To address these issues, this study proposes an Adaptive Time Compression (ATC) module. Functioning as an independent component, ATC can be seamlessly integrated into existing architectures and achieves data compression by eliminating redundant frames within video data. The ATC module effectively reduces GPU computing load and time complexity with negligible loss of accuracy, thereby facilitating real-time human behavior recognition.


Subject(s)
Algorithms , Data Compression , Video Recording , Humans , Data Compression/methods , Human Activities , Deep Learning , Image Processing, Computer-Assisted/methods , Pattern Recognition, Automated/methods
17.
Platelets ; 35(1): 2344512, 2024 Dec.
Article in English | MEDLINE | ID: mdl-38722090

ABSTRACT

The last decade has seen increasing use of advanced imaging techniques in platelet research. However, there has been a lag in the development of image analysis methods, leaving much of the information trapped in images. Herein, we present a robust analytical pipeline for finding and following individual platelets over time in growing thrombi. Our pipeline covers four steps: detection, tracking, estimation of tracking accuracy, and quantification of platelet metrics. We detect platelets using a deep learning network for image segmentation, which we validated with proofreading by multiple experts. We then track platelets using a standard particle tracking algorithm and validate the tracks with custom image sampling - essential when following platelets within a dense thrombus. We show that our pipeline is more accurate than previously described methods. To demonstrate the utility of our analytical platform, we use it to show that in vivo thrombus formation is much faster than that ex vivo. Furthermore, platelets in vivo exhibit less passive movement in the direction of blood flow. Our tools are free and open source and written in the popular and user-friendly Python programming language. They empower researchers to accurately find and follow platelets in fluorescence microscopy experiments.


In this paper we describe computational tools to find and follow individual platelets in blood clots recorded with fluorescence microscopy. Our tools work in a diverse range of conditions, both in living animals and in artificial flow chamber models of thrombosis. Our work uses deep learning methods to achieve excellent accuracy. We also provide tools for visualizing data and estimating error rates, so you don't have to just trust the output. Our workflow measures platelet density, shape, and speed, which we use to demonstrate differences in the kinetics of clotting in living vessels versus a synthetic environment. The tools we wrote are open source, written in the popular Python programming language, and freely available to all. We hope they will be of use to other platelet researchers.


Subject(s)
Blood Platelets , Deep Learning , Thrombosis , Blood Platelets/metabolism , Thrombosis/blood , Humans , Image Processing, Computer-Assisted/methods , Animals , Mice , Algorithms
18.
PLoS One ; 19(5): e0300924, 2024.
Article in English | MEDLINE | ID: mdl-38768105

ABSTRACT

The identification research of hydrogenation catalyst information has always been one of the most important businesses in the chemical industry. In order to aid researchers in efficiently screening high-performance catalyst carriers and tackle the pressing challenge at hand, it is imperative to find a solution for the intelligent recognition of hydrogenation catalyst images. To address the issue of low recognition accuracy caused by adhesion and stacking of hydrogenation catalysts, An image recognition algorithm of hydrogenation catalyst based on FPNC Net was proposed in this paper. In the present study, Resnet50 backbone network was used to extract the features, and spatially-separable convolution kernel was used to extract the multi-scale features of catalyst fringe. In addition, to effectively segment the adhesive regions of stripes, FPN (Feature Pyramid Network) is added to the backbone network for deep and shallow feature fusion. Introducing an attention module to adaptively adjust weights can effectively highlight the target features of the catalyst. The experimental results showed that the FPNC Net model achieved an accuracy of 94.2% and an AP value improvement of 19.37% compared to the original CenterNet model. The improved model demonstrates a significant enhancement in detection accuracy, indicating a high capability for detecting hydrogenation catalyst targets.


Subject(s)
Algorithms , Deep Learning , Catalysis , Hydrogenation , Image Processing, Computer-Assisted/methods , Neural Networks, Computer
19.
PLoS One ; 19(5): e0300017, 2024.
Article in English | MEDLINE | ID: mdl-38768119

ABSTRACT

With the increasing applications of traffic scene image classification in intelligent transportation systems, there is a growing demand for improved accuracy and robustness in this classification task. However, due to weather conditions, time, lighting variations, and annotation costs, traditional deep learning methods still have limitations in extracting complex traffic scene features and achieving higher recognition accuracy. The previous classification methods for traffic scene images had gaps in multi-scale feature extraction and the combination of frequency domain, spatial, and channel attention. To address these issues, this paper proposes a multi-scale and multi-attention model based on Res2Net. Our proposed framework introduces an Adaptive Feature Refinement Pyramid Module (AFRPM) to enhance multi-scale feature extraction, thus improving the accuracy of traffic scene image classification. Additionally, we integrate frequency domain and spatial-channel attention mechanisms to develop recognition capabilities for complex backgrounds, objects of different scales, and local details in traffic scene images. The paper conducts the task of classifying traffic scene images using the Traffic-Net dataset. The experimental results demonstrate that our model achieves an accuracy of 96.88% on this dataset, which is an improvement of approximately 2% compared to the baseline Res2Net network. Furthermore, we validate the effectiveness of the proposed modules through ablation experiments.


Subject(s)
Image Processing, Computer-Assisted , Image Processing, Computer-Assisted/methods , Algorithms , Deep Learning , Neural Networks, Computer , Humans
20.
Sci Data ; 11(1): 483, 2024 May 10.
Article in English | MEDLINE | ID: mdl-38729970

ABSTRACT

The Sparsely Annotated Region and Organ Segmentation (SAROS) dataset was created using data from The Cancer Imaging Archive (TCIA) to provide a large open-access CT dataset with high-quality annotations of body landmarks. In-house segmentation models were employed to generate annotation proposals on randomly selected cases from TCIA. The dataset includes 13 semantic body region labels (abdominal/thoracic cavity, bones, brain, breast implant, mediastinum, muscle, parotid/submandibular/thyroid glands, pericardium, spinal cord, subcutaneous tissue) and six body part labels (left/right arm/leg, head, torso). Case selection was based on the DICOM series description, gender, and imaging protocol, resulting in 882 patients (438 female) for a total of 900 CTs. Manual review and correction of proposals were conducted in a continuous quality control cycle. Only every fifth axial slice was annotated, yielding 20150 annotated slices from 28 data collections. For the reproducibility on downstream tasks, five cross-validation folds and a test set were pre-defined. The SAROS dataset serves as an open-access resource for training and evaluating novel segmentation models, covering various scanner vendors and diseases.


Subject(s)
Tomography, X-Ray Computed , Whole Body Imaging , Female , Humans , Male , Image Processing, Computer-Assisted
SELECTION OF CITATIONS
SEARCH DETAIL
...