Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 3 de 3
Filter
Add more filters










Database
Language
Publication year range
1.
Int J Med Robot ; 19(3): e2487, 2023 Jun.
Article in English | MEDLINE | ID: mdl-36478373

ABSTRACT

BACKGROUND: Segmentation of brain tumours is a complex problem in medical image processing and analysis. It is a time-consuming and error-prone task. Therefore, computer-aided detection systems need to be developed to decrease physicians' workload and improve the accuracy of segmentation. METHODS: This paper proposes a level set method constrained by an intuitive artificial intelligence-based approach to perform brain tumour segmentation. By studying 3D brain tumour images, a new segmentation technique based on the Modified Particle Swarm Optimisation (MPSO), Darwin Particle Swarm Optimisation (DPSO), and Fractional Order Darwinian Particle Swarm Optimisation (FODPSO) algorithms were developed. RESULTS: The introduced technique was verified according to the MICCAI RASTS 2013 database for high-grade glioma patients. The three algorithms were evaluated using different performance measures: accuracy, sensitivity, specificity, and Dice similarity coefficient to prove the performance and robustness of our 3D segmentation technique. CONCLUSION: The result is that the MPSO algorithm consistently outperforms the DPSO and FO DPSO.


Subject(s)
Artificial Intelligence , Brain Neoplasms , Humans , Brain Neoplasms/diagnostic imaging , Algorithms , Magnetic Resonance Imaging/methods , Image Processing, Computer-Assisted/methods
2.
Vis Comput ; 38(8): 2939-2970, 2022.
Article in English | MEDLINE | ID: mdl-34131356

ABSTRACT

The research progress in multimodal learning has grown rapidly over the last decade in several areas, especially in computer vision. The growing potential of multimodal data streams and deep learning algorithms has contributed to the increasing universality of deep multimodal learning. This involves the development of models capable of processing and analyzing the multimodal information uniformly. Unstructured real-world data can inherently take many forms, also known as modalities, often including visual and textual content. Extracting relevant patterns from this kind of data is still a motivating goal for researchers in deep learning. In this paper, we seek to improve the understanding of key concepts and algorithms of deep multimodal learning for the computer vision community by exploring how to generate deep models that consider the integration and combination of heterogeneous visual cues across sensory modalities. In particular, we summarize six perspectives from the current literature on deep multimodal learning, namely: multimodal data representation, multimodal fusion (i.e., both traditional and deep learning-based schemes), multitask learning, multimodal alignment, multimodal transfer learning, and zero-shot learning. We also survey current multimodal applications and present a collection of benchmark datasets for solving problems in various vision domains. Finally, we highlight the limitations and challenges of deep multimodal learning and provide insights and directions for future research.

3.
Phys Eng Sci Med ; 43(4): 1415-1431, 2020 Dec.
Article in English | MEDLINE | ID: mdl-33301073

ABSTRACT

The novel Coronavirus disease (COVID-19), which first appeared at the end of December 2019, continues to spread rapidly in most countries of the world. Respiratory infections occur primarily in the majority of patients treated with COVID-19. In light of the growing number of COVID-19 cases, the need for diagnostic tools to identify COVID-19 infection at early stages is of vital importance. For decades, chest X-ray (CXR) technologies have proven their ability to accurately detect respiratory diseases. More recently, with the availability of COVID-19 CXR scans, deep learning algorithms have played a critical role in the healthcare arena by allowing radiologists to recognize COVID-19 patients from their CXR images. However, the majority of screening methods for COVID-19 reported in recent studies are based on 2D convolutional neural networks (CNNs). Although 3D CNNs are capable of capturing contextual information compared to their 2D counterparts, their use is limited due to their increased computational cost (i.e. requires much extra memory and much more computing power). In this study, a transfer learning-based hybrid 2D/3D CNN architecture for COVID-19 screening using CXRs has been developed. The proposed architecture consists of the incorporation of a pre-trained deep model (VGG16) and a shallow 3D CNN, combined with a depth-wise separable convolution layer and a spatial pyramid pooling module (SPP). Specifically, the depth-wise separable convolution helps to preserve the useful features while reducing the computational burden of the model. The SPP module is designed to extract multi-level representations from intermediate ones. Experimental results show that the proposed framework can achieve reasonable performances when evaluated on a collected dataset (3 classes to be predicted: COVID-19, Pneumonia, and Normal). Notably, it achieved a sensitivity of 98.33%, a specificity of 98.68% and an overall accuracy of 96.91.


Subject(s)
Algorithms , COVID-19 Testing , COVID-19/diagnostic imaging , COVID-19/diagnosis , Mass Screening , Neural Networks, Computer , Thorax/diagnostic imaging , Aged , COVID-19/virology , Databases as Topic , Humans , Lung/diagnostic imaging , Lung/pathology , Male , SARS-CoV-2/physiology , Time Factors , X-Rays
SELECTION OF CITATIONS
SEARCH DETAIL
...