Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 173
Filter
1.
ISA Trans ; 2024 Jul 06.
Article in English | MEDLINE | ID: mdl-38987042

ABSTRACT

To guarantee the safety and reliability of equipment operation, such as liquid rocket engine (LRE), carrying out system-level anomaly detection (AD) is crucial. However, current methods ignore the prior knowledge of mechanical system itself, and seldom unite the observations with the inherent relation in data tightly. Meanwhile, they neglect the weakness and nonindependence of system-level anomaly which is different from component fault. To overcome above limitations, we propose a separate reconstruction framework using worsened tendency for system-level AD. To prevent anomalous feature being attenuated, we first propose to divide single sample into two equal-length parts along the temporal dimension. And we maximize the mean maximum discrepancy (MMD) between feature segments to force encoders to learn normal features with different distributions. Then, to fully explore the multivariate time series, we model temporal-spatial dependence by temporal convolution and graph attention. Besides, a joint graph learning strategy is proposed to handle prior knowledge and data characteristics simultaneously. Finally, the proposed method is evaluated on two real multi-sensor datasets from LRE and the results demonstrate the effectiveness and potential of the proposed method on system-level AD.

2.
Heliyon ; 10(11): e31836, 2024 Jun 15.
Article in English | MEDLINE | ID: mdl-38947471

ABSTRACT

Electric truck platooning offers a promising solution to extend the range of electric vehicles during long-haul operations. However, optimizing the platoon speed to ensure efficient energy utilization remains a critical challenge. The existing research on implementing data-driven solutions for truck platooning remains limited and implementing first principles solution is still a challenge. However, recognizing the resemblance of truck platoon data to a time series serves as a compelling motivation to explore suitable analytical techniques to address the problem. This paper presents a novel deep learning approach using a sequence-to-sequence encoder-decoder model to obtain the speed profile to be followed by an autonomous electric truck platoon considering various constraints such as the available state of charge (SOC) in the batteries along with other vehicles and road conditions while ensuring that the platoon is string stable. To ensure that the framework is suitable for long-haul highway operation, the model has been trained using various known highway drive cycles. Encoder-decoder models were trained and hyperparameter tuning was performed for the same. Finally, the most suitable model has been chosen for the application. For testing the entire framework, drive cycle/speed prediction corresponding to different desired SOC profiles has been presented. A case study showing the relevance of the proposed framework in predicting the drive cycle on various routes and its impact on taking critical policy decisions during the planning of electric truck platoons has also been presented. This study would help to efficiently plan the feasible routes for electric trucks considering multiple constraints such as battery capacity, expected discharge rate, charging infrastructure availability, route length/travel time, and other on-road operating conditions while also maintaining stability.

3.
Sci Rep ; 14(1): 14375, 2024 Jun 22.
Article in English | MEDLINE | ID: mdl-38909068

ABSTRACT

During nighttime road scenes, images are often affected by contrast distortion, loss of detailed information, and a significant amount of noise. These factors can negatively impact the accuracy of segmentation and object detection in nighttime road scenes. A cycle-consistent generative adversarial network has been proposed to address this issue to improve the quality of nighttime road scene images. The network includes two generative networks with identical structures and two adversarial networks with identical structures. The generative network comprises an encoder network and a corresponding decoder network. A context feature extraction module is designed as the foundational element of the encoder-decoder network to capture more contextual semantic information with different receptive fields. A receptive field residual module is also designed to increase the receptive field in the encoder network.The illumination attention module is inserted between the encoder and decoder to transfer critical features extracted by the encoder to the decoder. The network also includes a multiscale discriminative network to discriminate better whether the image is a real high-quality or generated image. Additionally, an improved loss function is proposed to enhance the efficacy of image enhancement. Compared to state-of-the-art methods, the proposed approach achieves the highest performance in enhancing nighttime images, making them clearer and more natural.

4.
Comput Biol Med ; 178: 108773, 2024 Jun 25.
Article in English | MEDLINE | ID: mdl-38925090

ABSTRACT

Extracting global and local feature information is still challenging due to the problems of retinal blood vessel medical images like fuzzy edge features, noise, difficulty in distinguishing between lesion regions and background information, and loss of low-level feature information, which leads to insufficient extraction of feature information. To better solve these problems and fully extract the global and local feature information of the image, we propose a novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation, which consists of an encoder and a decoder and is connected between the encoder and decoder by a transscale transformer cascade module. Among them, the encoder consists of a local-global transscale transformer module, a multi-head layered transscale adaptive embedding module, and a local context(LCNet) module. The transscale transformer cascade module learns local and global feature information from the first three layers of the encoder, and multi-scale dependent features, fuses the hierarchical feature information from the skip connection block and the channel-token interaction fusion block, respectively, and inputs it to the decoder. The decoder includes a decoding module for the local context network and a transscale position transformer module to input the local and global feature information extracted from the encoder with retained key position information into the decoding module and the position embedding transformer module for recovery and output of the prediction results that are consistent with the input feature information. In addition, we propose an improved cross-entropy loss function based on the difference between the deterministic observation samples and the prediction results with the deviation distance, which is validated on the DRIVE and STARE datasets combined with the proposed network model based on the dual transformer structure in this paper, and the segmentation accuracies are 97.26% and 97.87%, respectively. Compared with other state-of-the-art networks, the results show that the proposed network model has a significant competitive advantage in improving the segmentation performance of retinal blood vessel images.

5.
J Struct Biol ; 216(3): 108107, 2024 Jun 19.
Article in English | MEDLINE | ID: mdl-38906499

ABSTRACT

Atomic force microscope enables ultra-precision imaging of living cells. However, atomic force microscope imaging is a complex and time-consuming process. The obtained images of living cells usually have low resolution and are easily influenced by noise leading to unsatisfactory imaging quality, obstructing the research and analysis based on cell images. Herein, an adaptive attention image reconstruction network based on residual encoder-decoder was proposed, through the combination of deep learning technology and atomic force microscope imaging supporting high-quality cell image acquisition. Compared with other learning-based methods, the proposed network showed higher peak signal-to-noise ratio, higher structural similarity and better image reconstruction performances. In addition, the cell images reconstructed by each method were used for cell recognition, and the cell images reconstructed by the proposed network had the highest cell recognition rate. The proposed network has brought insights into the atomic force microscope-based imaging of living cells and cell image reconstruction, which is of great significance in biological and medical research.

6.
Int J Med Inform ; 188: 105462, 2024 Aug.
Article in English | MEDLINE | ID: mdl-38733641

ABSTRACT

OBJECTIVE: For ICD-10 coding causes of death in France in 2018 and 2019, predictions by deep neural networks (DNNs) are employed in addition to fully automatic batch coding by a rule-based expert system and to interactive coding by the coding team focused on certificates with a special public health interest and those for which DNNs have a low confidence index. METHODS: Supervised seq-to-seq DNNs are trained on previously coded data to ICD-10 code multiple causes and underlying causes of death. The DNNs are then used to target death certificates to be sent to the coding team and to predict multiple causes and underlying causes of death for part of the certificates. Hence, the coding campaign for 2018 and 2019 combines three modes of coding and a loop of interaction between the three. FINDINGS: In this campaign, 62% of the certificates are automatically batch coded by the expert system, 3% by the coding team, and the remainder by DNNs. Compared to a traditional campaign that would have relied on automatic batch coding and manual coding, the present campaign reaches an accuracy of 93.4% for ICD-10 coding of the underlying cause (95.6% at the European shortlist level). Some limitations (risks of under- or overestimation) appear for certain ICD categories, with the advantage of being quantifiable. CONCLUSION: The combination of the three coding methods illustrates how artificial intelligence, automated and human codings are mutually enriching. Quantified limitations on some chapters of ICD codes encourage an increase in the volume of certificates sent for manual coding from 2021 onward.


Subject(s)
Cause of Death , Clinical Coding , Death Certificates , International Classification of Diseases , Neural Networks, Computer , France , Humans , Clinical Coding/standards , Clinical Coding/methods , Expert Systems , Male , Infant , Female , Child , Aged , Child, Preschool
7.
Environ Monit Assess ; 196(6): 527, 2024 May 09.
Article in English | MEDLINE | ID: mdl-38722419

ABSTRACT

Understanding the connections between human activities and the natural environment depends heavily on information about land use and land cover (LULC) in the form of accurate LULC maps. Environmental monitoring using deep learning (DL) is rapidly growing to preserve a sustainable environment in the long term. For establishing effective policies, regulations, and implementation, DL can be a valuable tool for assessing environmental conditions and natural resources that will positively impact the ecosystem. This paper presents the assessment of land use and land cover change detection (LULCCD) and prediction using DL techniques for the southwestern coastal region, Goa, also known as the tourist destination of India. It consists of three components: (i) change detection (CD), (ii) quantification of LULC changes, and (iii) prediction. A new CD assessment framework, Spatio-Temporal Encoder-Decoder Self Attention Network (STEDSAN), is proposed for the LULCCD process. A dual branch encoder-decoder network is constructed using strided convolution with downsampling for the encoder and transpose convolution with upsampling for the decoder to assess the bitemporal images spatially. The self-attention (SA) mechanism captures the complex global spatial-temporal (ST) interactions between individual pixels over space-time to produce more distinct features. Each branch accepts the LULC map of 2 years as one of its inputs to determine binary and multiclass changes among the bitemporal images. The STEDSAN model determines the patterns, trends, and conversion from one LULC type to another for the assessment period from 2005 to 2018. The binary change maps were also compared with the existing state of the art (SOTA) CD methods, with STEDSAN having an overall accuracy of 94.93%. The prediction was made using an recurrent neural network (RNN) known as long short term memory network (LSTM) for the year 2025. Experiments were conducted to determine area-wise changes in several LULC classes, such as built-up (BU), crops (kharif crop (KC), rabi crop (RC), zaid crop (ZC), double/triple (D/T C)), current fallow (CF), plantation (PL), forests (evergreen forest (EF), deciduous forest (DF), degraded/scurb forest (D/SF) ), littoral swamp (LS), grassland (GL), wasteland (WL), waterbodies max (Wmx), and waterbodies min (Wmn). As per the analysis, over the period of 13 years, there has been a net increase in the amount of BU (1.25%), RC (1.17%), and D/TC( 2.42%) and a net decrease in DF (3.29%) and WL(1.44%) being the most dominant classes being changed. These findings will offer a thorough description of identifying trends in coastal areas that may incorporate methodological hints for future studies. This study will also promote handling the spatial and temporal complexity of remotely sensed data employed in categorizing the coastal LULC of a heterogeneous landscape.


Subject(s)
Conservation of Natural Resources , Deep Learning , Environmental Monitoring , India , Environmental Monitoring/methods , Conservation of Natural Resources/methods , Ecosystem , Agriculture/methods
8.
Comput Biol Med ; 173: 108293, 2024 May.
Article in English | MEDLINE | ID: mdl-38574528

ABSTRACT

Accurately identifying the Kirsten rat sarcoma virus (KRAS) gene mutation status in colorectal cancer (CRC) patients can assist doctors in deciding whether to use specific targeted drugs for treatment. Although deep learning methods are popular, they are often affected by redundant features from non-lesion areas. Moreover, existing methods commonly extract spatial features from imaging data, which neglect important frequency domain features and may degrade the performance of KRAS gene mutation status identification. To address this deficiency, we propose a segmentation-guided Transformer U-Net (SG-Transunet) model for KRAS gene mutation status identification in CRC. Integrating the strength of convolutional neural networks (CNNs) and Transformers, SG-Transunet offers a unique approach for both lesion segmentation and KRAS mutation status identification. Specifically, for precise lesion localization, we employ an encoder-decoder to obtain segmentation results and guide the KRAS gene mutation status identification task. Subsequently, a frequency domain supplement block is designed to capture frequency domain features, integrating it with high-level spatial features extracted in the encoding path to derive advanced spatial-frequency domain features. Furthermore, we introduce a pre-trained Xception block to mitigate the risk of overfitting associated with small-scale datasets. Following this, an aggregate attention module is devised to consolidate spatial-frequency domain features with global information extracted by the Transformer at shallow and deep levels, thereby enhancing feature discriminability. Finally, we propose a mutual-constrained loss function that simultaneously constrains the segmentation mask acquisition and gene status identification process. Experimental results demonstrate the superior performance of SG-Transunet over state-of-the-art methods in discriminating KRAS gene mutation status.


Subject(s)
Colorectal Neoplasms , Proto-Oncogene Proteins p21(ras) , Humans , Proto-Oncogene Proteins p21(ras)/genetics , Drug Delivery Systems , Mutation/genetics , Neural Networks, Computer , Colorectal Neoplasms/diagnostic imaging , Colorectal Neoplasms/genetics , Image Processing, Computer-Assisted
9.
Front Neurosci ; 18: 1363930, 2024.
Article in English | MEDLINE | ID: mdl-38680446

ABSTRACT

Introduction: In neurological diagnostics, accurate detection and segmentation of brain lesions is crucial. Identifying these lesions is challenging due to its complex morphology, especially when using traditional methods. Conventional methods are either computationally demanding with a marginal impact/enhancement or sacrifice fine details for computational efficiency. Therefore, balancing performance and precision in compute-intensive medical imaging remains a hot research topic. Methods: We introduce a novel encoder-decoder network architecture named the Adaptive Feature Medical Segmentation Network (AFMS-Net) with two encoder variants: the Single Adaptive Encoder Block (SAEB) and the Dual Adaptive Encoder Block (DAEB). A squeeze-and-excite mechanism is employed in SAEB to identify significant data while disregarding peripheral details. This approach is best suited for scenarios requiring quick and efficient segmentation, with an emphasis on identifying key lesion areas. In contrast, the DAEB utilizes an advanced channel spatial attention strategy for fine-grained delineation and multiple-class classifications. Additionally, both architectures incorporate a Segmentation Path (SegPath) module between the encoder and decoder, refining segmentation, enhancing feature extraction, and improving model performance and stability. Results: AFMS-Net demonstrates exceptional performance across several notable datasets, including BRATs 2021, ATLAS 2021, and ISLES 2022. Its design aims to construct a lightweight architecture capable of handling complex segmentation challenges with high precision. Discussion: The proposed AFMS-Net addresses the critical balance issue between performance and computational efficiency in the segmentation of brain lesions. By introducing two tailored encoder variants, the network adapts to varying requirements of speed and feature. This approach not only advances the state-of-the-art in lesion segmentation but also provides a scalable framework for future research in medical image processing.

10.
Sci Rep ; 14(1): 9336, 2024 04 23.
Article in English | MEDLINE | ID: mdl-38653997

ABSTRACT

Skin cancer is the most prevalent kind of cancer in people. It is estimated that more than 1 million people get skin cancer every year in the world. The effectiveness of the disease's therapy is significantly impacted by early identification of this illness. Preprocessing is the initial detecting stage in enhancing the quality of skin images by removing undesired background noise and objects. This study aims is to compile preprocessing techniques for skin cancer imaging that are currently accessible. Researchers looking into automated skin cancer diagnosis might use this article as an excellent place to start. The fully convolutional encoder-decoder network and Sparrow search algorithm (FCEDN-SpaSA) are proposed in this study for the segmentation of dermoscopic images. The individual wolf method and the ensemble ghosting technique are integrated to generate a neighbour-based search strategy in SpaSA for stressing the correct balance between navigation and exploitation. The classification procedure is accomplished by using an adaptive CNN technique to discriminate between normal skin and malignant skin lesions suggestive of disease. Our method provides classification accuracies comparable to commonly used incremental learning techniques while using less energy, storage space, memory access, and training time (only network updates with new training samples, no network sharing). In a simulation, the segmentation performance of the proposed technique on the ISBI 2017, ISIC 2018, and PH2 datasets reached accuracies of 95.28%, 95.89%, 92.70%, and 98.78%, respectively, on the same dataset and assessed the classification performance. It is accurate 91.67% of the time. The efficiency of the suggested strategy is demonstrated through comparisons with cutting-edge methodologies.


Subject(s)
Algorithms , Dermoscopy , Neural Networks, Computer , Skin Neoplasms , Humans , Skin Neoplasms/diagnosis , Skin Neoplasms/diagnostic imaging , Skin Neoplasms/classification , Skin Neoplasms/pathology , Dermoscopy/methods , Image Processing, Computer-Assisted/methods , Image Interpretation, Computer-Assisted/methods , Skin/pathology , Skin/diagnostic imaging
11.
Micron ; 180: 103615, 2024 05.
Article in English | MEDLINE | ID: mdl-38471391

ABSTRACT

Medical imaging plays a critical role in diagnosing and treating various medical conditions. However, interpreting medical images can be challenging even for expert clinicians, as they are often degraded by noise and artifacts that can hinder the accurate identification and analysis of diseases, leading to severe consequences such as patient misdiagnosis or mortality. Various types of noise, including Gaussian, Rician, and Salt-pepper noise, can corrupt the area of interest, limiting the precision and accuracy of algorithms. Denoising algorithms have shown the potential in improving the quality of medical images by removing noise and other artifacts that obscure essential information. Deep learning has emerged as a powerful tool for image analysis and has demonstrated promising results in denoising different medical images such as MRIs, CT scans, PET scans, etc. This review paper provides a comprehensive overview of state-of-the-art deep learning algorithms used for denoising medical images. A total of 120 relevant papers were reviewed, and after screening with specific inclusion and exclusion criteria, 104 papers were selected for analysis. This study aims to provide a thorough understanding for researchers in the field of intelligent denoising by presenting an extensive survey of current techniques and highlighting significant challenges that remain to be addressed. The findings of this review are expected to contribute to the development of intelligent models that enable timely and accurate diagnoses of medical disorders. It was found that 40% of the researchers used models based on Deep convolutional neural networks to denoise the images, followed by encoder-decoder (18%) and other artificial intelligence-based techniques (15%) (Like DIP, etc.). Generative adversarial network was used by 12%, transformer-based approaches (13%) and multilayer perceptron was used by 2% of the researchers. Moreover, Gaussian noise was present in 35% of the images, followed by speckle noise (16%), poisson noise (14%), artifacts (10%), rician noise (7%), Salt-pepper noise (6%), Impulse noise (3%) and other types of noise (9%). While the progress in developing novel models for the denoising of medical images is evident, significant work remains to be done in creating standardized denoising models that perform well across a wide spectrum of medical images. Overall, this review highlights the importance of denoising medical images and provides a comprehensive understanding of the current state-of-the-art deep learning algorithms in this field.


Subject(s)
Deep Learning , Humans , Artificial Intelligence , Signal-To-Noise Ratio , Neural Networks, Computer , Image Processing, Computer-Assisted/methods , Algorithms
12.
Comput Biol Med ; 173: 108354, 2024 May.
Article in English | MEDLINE | ID: mdl-38522251

ABSTRACT

Colorectal cancer (CRC) is a leading cause of cancer-related deaths, with colonic crypts (CC) being crucial in its development. Accurate segmentation of CC is essential for decisions CRC and developing diagnostic strategies. However, colonic crypts' blurred boundaries and morphological diversity bring substantial challenges for automatic segmentation. To mitigate this problem, we proposed the Dual-Branch Asymmetric Encoder-Decoder Segmentation Network (DAUNet), a novel and efficient model tailored for confocal laser endomicroscopy (CLE) CC images. In DAUNet, we crafted a dual-branch feature extraction module (DFEM), employing Focus operations and dense depth-wise separable convolution (DDSC) to extract multiscale features, boosting semantic understanding and coping with the morphological diversity of CC. We also introduced the feature fusion guided module (FFGM) to adaptively combine features from both branches using cross-group spatial and channel attention to improve the model representation in focusing on specific lesion features. These modules are seamlessly integrated into the encoder for effective multiscale information extraction and fusion, and DDSC is further introduced in the decoder to provide rich representations for precise segmentation. Moreover, the local multi-layer perceptron (LMLP) module is designed to decouple and recalibrate features through a local linear transformation that filters out the noise and refines features to provide edge-enriched representation. Experimental evaluations on two datasets demonstrate that the proposed method achieves Intersection over Union (IoU) scores of 81.54% and 84.83%, respectively, which are on par with state-of-the-art methods, exhibiting its effectiveness for CC segmentation. The proposed method holds great potential in assisting physicians with precise lesion localization and region analysis, thereby improving the diagnostic accuracy of CRC.


Subject(s)
Colon , Coping Skills , Colon/diagnostic imaging , Information Storage and Retrieval , Neural Networks, Computer , Semantics , Image Processing, Computer-Assisted
13.
Int J Mol Sci ; 25(4)2024 Feb 17.
Article in English | MEDLINE | ID: mdl-38397034

ABSTRACT

The receptor tyrosine kinase RET (rearranged during transfection) plays a vital role in various cell signaling pathways and is a critical factor in the development of the nervous system. Abnormal activation of the RET kinase can lead to several cancers, including thyroid cancer and non-small-cell lung cancer. However, most RET kinase inhibitors are multi-kinase inhibitors. Therefore, the development of an effective RET-specific inhibitor continues to present a significant challenge. To address this issue, we built a molecular generation model based on fragment-based drug design (FBDD) and a long short-term memory (LSTM) encoder-decoder structure to generate receptor-specific molecules with novel scaffolds. Remarkably, our model was trained with a molecular assembly accuracy of 98.4%. Leveraging the pre-trained model, we rapidly generated a RET-specific-candidate active-molecule library by transfer learning. Virtual screening based on our molecular generation model was performed, combined with molecular dynamics simulation and binding energy calculation, to discover specific RET inhibitors, and five novel molecules were selected. Further analyses indicated that two of these molecules have good binding affinities and synthesizability, exhibiting high selectivity. Overall, this investigation demonstrates the capacity of our model to generate novel receptor-specific molecules and provides a rapid method to discover potential drugs.


Subject(s)
Carcinoma, Non-Small-Cell Lung , Lung Neoplasms , Protein Kinase Inhibitors , Proto-Oncogene Proteins c-ret , Humans , Molecular Dynamics Simulation , Protein Kinase Inhibitors/pharmacology , Protein Kinase Inhibitors/chemistry , Proto-Oncogene Proteins c-ret/antagonists & inhibitors
14.
Artif Intell Med ; 148: 102781, 2024 02.
Article in English | MEDLINE | ID: mdl-38325926

ABSTRACT

The Concordance Index (C-index) is a commonly used metric in Survival Analysis for evaluating the performance of a prediction model. In this paper, we propose a decomposition of the C-index into a weighted harmonic mean of two quantities: one for ranking observed events versus other observed events, and the other for ranking observed events versus censored cases. This decomposition enables a finer-grained analysis of the relative strengths and weaknesses between different survival prediction methods. The usefulness of this decomposition is demonstrated through benchmark comparisons against classical models and state-of-the-art methods, together with the new variational generative neural-network-based method (SurVED) proposed in this paper. The performance of the models is assessed using four publicly available datasets with varying levels of censoring. Using the C-index decomposition and synthetic censoring, the analysis shows that deep learning models utilize the observed events more effectively than other models. This allows them to keep a stable C-index in different censoring levels. In contrast to such deep learning methods, classical machine learning models deteriorate when the censoring level decreases due to their inability to improve on ranking the events versus other events.


Subject(s)
Machine Learning , Neural Networks, Computer , Survival Analysis
15.
Med Biol Eng Comput ; 62(6): 1751-1762, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38372910

ABSTRACT

In a treatment or diagnosis related to oral health conditions such as oral cancer and oropharyngeal cancer, an investigation of tongue's movements is a major part. In an automatic measurement of such movement, it must first start with a task of tongue segmentation. This paper proposes a solution of tongue segmentation based on a decoder-encoder CNN-based structure i.e., U-Net. However, it could suffer from a problem of feature loss in deep layers. This paper proposes a Deep Upscale U-Net (DU-UNET). An additional up-sampling of the feature map from a contracting path is concatenated to an upper layer of an expansive path, based on an original U-Net structure. The segmentation model is constructed by training DU-UNET on the two publicly available datasets, and transferred to the self-collected dataset of tongue images with five tongue postures which were recorded at a far distance from a camera under a real-world scenario. The proposed DU-UNET outperforms the other existing methods in our literature reviews, with accuracy of 99.2%, mean IoU of 97.8%, Dice score of 96.8%, and Jaccard score of 96.8%.


Subject(s)
Image Processing, Computer-Assisted , Neural Networks, Computer , Tongue , Tongue/diagnostic imaging , Humans , Image Processing, Computer-Assisted/methods , Algorithms , Deep Learning
16.
Front Plant Sci ; 15: 1328075, 2024.
Article in English | MEDLINE | ID: mdl-38362454

ABSTRACT

In order to effectively support wheat breeding, farmland ridge segmentation can be used to visualize the size and spacing of a wheat field. At the same time, accurate ridge information collecting can deliver useful data support for farmland management. However, in the farming ridge segmentation scenarios based on remote sensing photos, the commonly used semantic segmentation methods tend to overlook the ridge edges and ridge strip features, which impair the segmentation effect. In order to efficiently collect ridge information, this paper proposes a segmentation method based on encoder-decoder of network with strip pooling module and ASPP module. First, in order to extract context information for multi-scale features, ASPP module are integrated in the deepest feature map. Second, the remote dependence of the ridge features is improved in both horizontal and vertical directions by using the strip pooling module. The final segmentation map is generated by fusing the boundary features and semantic features using an encoder and decoder architecture. As a result, the accuracy of the proposed method in the validation set is 98.0% and mIoU is 94.6%. The results of the experiments demonstrate that the method suggested in this paper can precisely segment the ridge information, as well as its value in obtaining data on the distribution of farmland and its potential for practical application.

17.
J Appl Clin Med Phys ; 25(3): e14297, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38373289

ABSTRACT

PURPOSE: Deep learning-based auto-segmentation algorithms can improve clinical workflow by defining accurate regions of interest while reducing manual labor. Over the past decade, convolutional neural networks (CNNs) have become prominent in medical image segmentation applications. However, CNNs have limitations in learning long-range spatial dependencies due to the locality of the convolutional layers. Transformers were introduced to address this challenge. In transformers with self-attention mechanism, even the first layer of information processing makes connections between distant image locations. Our paper presents a novel framework that bridges these two unique techniques, CNNs and transformers, to segment the gross tumor volume (GTV) accurately and efficiently in computed tomography (CT) images of non-small cell-lung cancer (NSCLC) patients. METHODS: Under this framework, input of multiple resolution images was used with multi-depth backbones to retain the benefits of high-resolution and low-resolution images in the deep learning architecture. Furthermore, a deformable transformer was utilized to learn the long-range dependency on the extracted features. To reduce computational complexity and to efficiently process multi-scale, multi-depth, high-resolution 3D images, this transformer pays attention to small key positions, which were identified by a self-attention mechanism. We evaluated the performance of the proposed framework on a NSCLC dataset which contains 563 training images and 113 test images. Our novel deep learning algorithm was benchmarked against five other similar deep learning models. RESULTS: The experimental results indicate that our proposed framework outperforms other CNN-based, transformer-based, and hybrid methods in terms of Dice score (0.92) and Hausdorff Distance (1.33). Therefore, our proposed model could potentially improve the efficiency of auto-segmentation of early-stage NSCLC during the clinical workflow. This type of framework may potentially facilitate online adaptive radiotherapy, where an efficient auto-segmentation workflow is required. CONCLUSIONS: Our deep learning framework, based on CNN and transformer, performs auto-segmentation efficiently and could potentially assist clinical radiotherapy workflow.


Subject(s)
Carcinoma, Non-Small-Cell Lung , Deep Learning , Lung Neoplasms , Humans , Lung Neoplasms/diagnostic imaging , Lung Neoplasms/radiotherapy , Tomography, X-Ray Computed , Neural Networks, Computer , Algorithms , Carcinoma, Non-Small-Cell Lung/diagnostic imaging , Carcinoma, Non-Small-Cell Lung/radiotherapy , Image Processing, Computer-Assisted/methods
18.
Article in English | MEDLINE | ID: mdl-38321907

ABSTRACT

Traditional molecular de novo generation methods, such as evolutionary algorithms, generate new molecules mainly by linking existing atomic building blocks. The challenging issues in these methods include difficulty in synthesis, failure to achieve desired properties, and structural optimization requirements. Advances in deep learning offer new ideas for rational and robust de novo drug design. Deep learning, a branch of machine learning, is more efficient than traditional methods for processing problems, such as speech, image, and translation. This study provides a comprehensive overview of the current state of research in de novo drug design based on deep learning and identifies key areas for further development. Deep learning-based de novo drug design is pivotal in four key dimensions. Molecular databases form the basis for model training, while effective molecular representations impact model performance. Common DL models (GANs, RNNs, VAEs, CNNs, DMs) generate drug molecules with desired properties. The evaluation metrics guide research directions by determining the quality and applicability of generated molecules. This abstract highlights the foundational aspects of DL-based de novo drug design, offering a concise overview of its multifaceted contributions. Consequently, deep learning in de novo molecule generation has attracted more attention from academics and industry. As a result, many deep learning-based de novo molecule generation types have been actively proposed.

19.
Pharmaceuticals (Basel) ; 17(2)2024 Jan 27.
Article in English | MEDLINE | ID: mdl-38399376

ABSTRACT

The discovery of novel therapeutic compounds through de novo drug design represents a critical challenge in the field of pharmaceutical research. Traditional drug discovery approaches are often resource intensive and time consuming, leading researchers to explore innovative methods that harness the power of deep learning and reinforcement learning techniques. Here, we introduce a novel drug design approach called drugAI that leverages the Encoder-Decoder Transformer architecture in tandem with Reinforcement Learning via a Monte Carlo Tree Search (RL-MCTS) to expedite the process of drug discovery while ensuring the production of valid small molecules with drug-like characteristics and strong binding affinities towards their targets. We successfully integrated the Encoder-Decoder Transformer architecture, which generates molecular structures (drugs) from scratch with the RL-MCTS, serving as a reinforcement learning framework. The RL-MCTS combines the exploitation and exploration capabilities of a Monte Carlo Tree Search with the machine translation of a transformer-based Encoder-Decoder model. This dynamic approach allows the model to iteratively refine its drug candidate generation process, ensuring that the generated molecules adhere to essential physicochemical and biological constraints and effectively bind to their targets. The results from drugAI showcase the effectiveness of the proposed approach across various benchmark datasets, demonstrating a significant improvement in both the validity and drug-likeness of the generated compounds, compared to two existing benchmark methods. Moreover, drugAI ensures that the generated molecules exhibit strong binding affinities to their respective targets. In summary, this research highlights the real-world applications of drugAI in drug discovery pipelines, potentially accelerating the identification of promising drug candidates for a wide range of diseases.

20.
Med Image Anal ; 93: 103072, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38176356

ABSTRACT

Accurate quantification of cerebral blood flow (CBF) is essential for the diagnosis and assessment of a wide range of neurological diseases. Positron emission tomography (PET) with radiolabeled water (15O-water) is the gold-standard for the measurement of CBF in humans, however, it is not widely available due to its prohibitive costs and the use of short-lived radiopharmaceutical tracers that require onsite cyclotron production. Magnetic resonance imaging (MRI), in contrast, is more accessible and does not involve ionizing radiation. This study presents a convolutional encoder-decoder network with attention mechanisms to predict the gold-standard 15O-water PET CBF from multi-contrast MRI scans, thus eliminating the need for radioactive tracers. The model was trained and validated using 5-fold cross-validation in a group of 126 subjects consisting of healthy controls and cerebrovascular disease patients, all of whom underwent simultaneous 15O-water PET/MRI. The results demonstrate that the model can successfully synthesize high-quality PET CBF measurements (with an average SSIM of 0.924 and PSNR of 38.8 dB) and is more accurate compared to concurrent and previous PET synthesis methods. We also demonstrate the clinical significance of the proposed algorithm by evaluating the agreement for identifying the vascular territories with impaired CBF. Such methods may enable more widespread and accurate CBF evaluation in larger cohorts who cannot undergo PET imaging due to radiation concerns, lack of access, or logistic challenges.


Subject(s)
Brain , Positron-Emission Tomography , Humans , Brain/diagnostic imaging , Magnetic Resonance Imaging , Cerebrovascular Circulation , Algorithms
SELECTION OF CITATIONS
SEARCH DETAIL
...