Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 17 de 17
Filter
Add more filters










Publication year range
1.
Article in English | MEDLINE | ID: mdl-38922721

ABSTRACT

OBJECTIVE: Segmentation, the partitioning of patient imaging into multiple, labeled segments, has several potential clinical benefits but when performed manually is tedious and resource intensive. Automated deep learning (DL)-based segmentation methods can streamline the process. The objective of this study was to evaluate a label-efficient DL pipeline that requires only a small number of annotated scans for semantic segmentation of sinonasal structures in CT scans. STUDY DESIGN: Retrospective cohort study. SETTING: Academic institution. METHODS: Forty CT scans were used in this study including 16 scans in which the nasal septum (NS), inferior turbinate (IT), maxillary sinus (MS), and optic nerve (ON) were manually annotated using an open-source software. A label-efficient DL framework was used to train jointly on a few manually labeled scans and the remaining unlabeled scans. Quantitative analysis was then performed to obtain the number of annotated scans needed to achieve submillimeter average surface distances (ASDs). RESULTS: Our findings reveal that merely four labeled scans are necessary to achieve median submillimeter ASDs for large sinonasal structures-NS (0.96 mm), IT (0.74 mm), and MS (0.43 mm), whereas eight scans are required for smaller structures-ON (0.80 mm). CONCLUSION: We have evaluated a label-efficient pipeline for segmentation of sinonasal structures. Empirical results demonstrate that automated DL methods can achieve submillimeter accuracy using a small number of labeled CT scans. Our pipeline has the potential to improve pre-operative planning workflows, robotic- and image-guidance navigation systems, computer-assisted diagnosis, and the construction of statistical shape models to quantify population variations. LEVEL OF EVIDENCE: N/A.

2.
Int J Comput Assist Radiol Surg ; 19(7): 1359-1366, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38753135

ABSTRACT

PURPOSE: Preoperative imaging plays a pivotal role in sinus surgery where CTs offer patient-specific insights of complex anatomy, enabling real-time intraoperative navigation to complement endoscopy imaging. However, surgery elicits anatomical changes not represented in the preoperative model, generating an inaccurate basis for navigation during surgery progression. METHODS: We propose a first vision-based approach to update the preoperative 3D anatomical model leveraging intraoperative endoscopic video for navigated sinus surgery where relative camera poses are known. We rely on comparisons of intraoperative monocular depth estimates and preoperative depth renders to identify modified regions. The new depths are integrated in these regions through volumetric fusion in a truncated signed distance function representation to generate an intraoperative 3D model that reflects tissue manipulation RESULTS: We quantitatively evaluate our approach by sequentially updating models for a five-step surgical progression in an ex vivo specimen. We compute the error between correspondences from the updated model and ground-truth intraoperative CT in the region of anatomical modification. The resulting models show a decrease in error during surgical progression as opposed to increasing when no update is employed. CONCLUSION: Our findings suggest that preoperative 3D anatomical models can be updated using intraoperative endoscopy video in navigated sinus surgery. Future work will investigate improvements to monocular depth estimation as well as removing the need for external navigation systems. The resulting ability to continuously update the patient model may provide surgeons with a more precise understanding of the current anatomical state and paves the way toward a digital twin paradigm for sinus surgery.


Subject(s)
Endoscopy , Imaging, Three-Dimensional , Models, Anatomic , Surgery, Computer-Assisted , Tomography, X-Ray Computed , Imaging, Three-Dimensional/methods , Humans , Endoscopy/methods , Tomography, X-Ray Computed/methods , Surgery, Computer-Assisted/methods , Paranasal Sinuses/surgery , Paranasal Sinuses/diagnostic imaging
3.
Int J Comput Assist Radiol Surg ; 19(7): 1259-1266, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38775904

ABSTRACT

PURPOSE: Monocular SLAM algorithms are the key enabling technology for image-based surgical navigation systems for endoscopic procedures. Due to the visual feature scarcity and unique lighting conditions encountered in endoscopy, classical SLAM approaches perform inconsistently. Many of the recent approaches to endoscopic SLAM rely on deep learning models. They show promising results when optimized on singular domains such as arthroscopy, sinus endoscopy, colonoscopy or laparoscopy, but are limited by an inability to generalize to different domains without retraining. METHODS: To address this generality issue, we propose OneSLAM a monocular SLAM algorithm for surgical endoscopy that works out of the box for several endoscopic domains, including sinus endoscopy, colonoscopy, arthroscopy and laparoscopy. Our pipeline builds upon robust tracking any point (TAP) foundation models to reliably track sparse correspondences across multiple frames and runs local bundle adjustment to jointly optimize camera poses and a sparse 3D reconstruction of the anatomy. RESULTS: We compare the performance of our method against three strong baselines previously proposed for monocular SLAM in endoscopy and general scenes. OneSLAM presents better or comparable performance over existing approaches targeted to that specific data in all four tested domains, generalizing across domains without the need for retraining. CONCLUSION: OneSLAM benefits from the convincing performance of TAP foundation models but generalizes to endoscopic sequences of different anatomies all while demonstrating better or comparable performance over domain-specific SLAM approaches. Future research on global loop closure will investigate how to reliably detect loops in endoscopic scenes to reduce accumulated drift and enhance long-term navigation capabilities.


Subject(s)
Algorithms , Endoscopy , Humans , Endoscopy/methods , Imaging, Three-Dimensional/methods , Surgery, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods
4.
Int J Comput Assist Radiol Surg ; 19(7): 1273-1280, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38816649

ABSTRACT

PURPOSE: Skullbase surgery demands exceptional precision when removing bone in the lateral skull base. Robotic assistance can alleviate the effect of human sensory-motor limitations. However, the stiffness and inertia of the robot can significantly impact the surgeon's perception and control of the tool-to-tissue interaction forces. METHODS: We present a situational-aware, force control technique aimed at regulating interaction forces during robot-assisted skullbase drilling. The contextual interaction information derived from the digital twin environment is used to enhance sensory perception and suppress undesired high forces. RESULTS: To validate our approach, we conducted initial feasibility experiments involving a medical and two engineering students. The experiment focused on further drilling around critical structures following cortical mastoidectomy. The experiment results demonstrate that robotic assistance coupled with our proposed control scheme effectively limited undesired interaction forces when compared to robotic assistance without the proposed force control. CONCLUSIONS: The proposed force control techniques show promise in significantly reducing undesired interaction forces during robot-assisted skullbase surgery. These findings contribute to the ongoing efforts to enhance surgical precision and safety in complex procedures involving the lateral skull base.


Subject(s)
Robotic Surgical Procedures , Skull Base , Humans , Skull Base/surgery , Robotic Surgical Procedures/methods , Feasibility Studies , Mastoidectomy/methods
5.
Article in English | MEDLINE | ID: mdl-38686594

ABSTRACT

OBJECTIVE: Obtaining automated, objective 3-dimensional (3D) models of the Eustachian tube (ET) and the internal carotid artery (ICA) from computed tomography (CT) scans could provide useful navigational and diagnostic information for ET pathologies and interventions. We aim to develop a deep learning (DL) pipeline to automatically segment the ET and ICA and use these segmentations to compute distances between these structures. STUDY DESIGN: Retrospective cohort. SETTING: Tertiary referral center. METHODS: From a database of 30 CT scans, 60 ET and ICA pairs were manually segmented and used to train an nnU-Net model, a DL segmentation framework. These segmentations were also used to develop a quantitative tool to capture the magnitude and location of the minimum distance point (MDP) between ET and ICA. Performance metrics for the nnU-Net automated segmentations were calculated via the average Hausdorff distance (AHD) and dice similarity coefficient (DSC). RESULTS: The AHD for the ET and ICA were 0.922 and 0.246 mm, respectively. Similarly, the DSC values for the ET and ICA were 0.578 and 0.884. The mean MDP from ET to ICA in the cartilaginous region was 2.6 mm (0.7-5.3 mm) and was located on average 1.9 mm caudal from the bony cartilaginous junction. CONCLUSION: This study describes the first end-to-end DL pipeline for automated ET and ICA segmentation and analyzes distances between these structures. In addition to helping to ensure the safe selection of patients for ET dilation, this method can facilitate large-scale studies exploring the relationship between ET pathologies and the 3D shape of the ET.

6.
Otolaryngol Head Neck Surg ; 171(1): 188-196, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38488231

ABSTRACT

OBJECTIVE: Use microscopic video-based tracking of laryngeal surgical instruments to investigate the effect of robot assistance on instrument tremor. STUDY DESIGN: Experimental trial. SETTING: Tertiary Academic Medical Center. METHODS: In this randomized cross-over trial, 36 videos were recorded from 6 surgeons performing left and right cordectomies on cadaveric pig larynges. These recordings captured 3 distinct conditions: without robotic assistance, with robot-assisted scissors, and with robot-assisted graspers. To assess tool tremor, we employed computer vision-based algorithms for tracking surgical tools. Absolute tremor bandpower and normalized path length were utilized as quantitative measures. Wilcoxon rank sum exact tests were employed for statistical analyses and comparisons between trials. Additionally, surveys were administered to assess the perceived ease of use of the robotic system. RESULTS: Absolute tremor bandpower showed a significant decrease when using robot-assisted instruments compared to freehand instruments (P = .012). Normalized path length significantly decreased with robot-assisted compared to freehand trials (P = .001). For the scissors, robot-assisted trials resulted in a significant decrease in absolute tremor bandpower (P = .002) and normalized path length (P < .001). For the graspers, there was no significant difference in absolute tremor bandpower (P = .4), but there was a significantly lower normalized path length in the robot-assisted trials (P = .03). CONCLUSION: This study demonstrated that computer-vision-based approaches can be used to assess tool motion in simulated microlaryngeal procedures. The results suggest that robot assistance is capable of reducing instrument tremor.


Subject(s)
Microsurgery , Robotic Surgical Procedures , Swine , Animals , Robotic Surgical Procedures/methods , Microsurgery/methods , Tremor/surgery , Cross-Over Studies , Video Recording , Cadaver , Humans
7.
Int J Comput Assist Radiol Surg ; 18(7): 1303-1310, 2023 Jul.
Article in English | MEDLINE | ID: mdl-37266885

ABSTRACT

PURPOSE: Tracking the 3D motion of the surgical tool and the patient anatomy is a fundamental requirement for computer-assisted skull-base surgery. The estimated motion can be used both for intra-operative guidance and for downstream skill analysis. Recovering such motion solely from surgical videos is desirable, as it is compliant with current clinical workflows and instrumentation. METHODS: We present Tracker of Anatomy and Tool (TAToo). TAToo jointly tracks the rigid 3D motion of the patient skull and surgical drill from stereo microscopic videos. TAToo estimates motion via an iterative optimization process in an end-to-end differentiable form. For robust tracking performance, TAToo adopts a probabilistic formulation and enforces geometric constraints on the object level. RESULTS: We validate TAToo on both simulation data, where ground truth motion is available, as well as on anthropomorphic phantom data, where optical tracking provides a strong baseline. We report sub-millimeter and millimeter inter-frame tracking accuracy for skull and drill, respectively, with rotation errors below [Formula: see text]. We further illustrate how TAToo may be used in a surgical navigation setting. CONCLUSIONS: We present TAToo, which simultaneously tracks the surgical tool and the patient anatomy in skull-base surgery. TAToo directly predicts the motion from surgical videos, without the need of any markers. Our results show that the performance of TAToo compares favorably to competing approaches. Future work will include fine-tuning of our depth network to reach a 1 mm clinical accuracy goal desired for surgical applications in the skull base.


Subject(s)
Neurosurgical Procedures , Surgery, Computer-Assisted , Humans , Neurosurgical Procedures/methods , Surgery, Computer-Assisted/methods , Computer Simulation , Skull Base/diagnostic imaging , Skull Base/surgery
8.
Int J Comput Assist Radiol Surg ; 18(7): 1135-1142, 2023 Jul.
Article in English | MEDLINE | ID: mdl-37160580

ABSTRACT

PURPOSE: Recent advances in computer vision and machine learning have resulted in endoscopic video-based solutions for dense reconstruction of the anatomy. To effectively use these systems in surgical navigation, a reliable image-based technique is required to constantly track the endoscopic camera's position within the anatomy, despite frequent removal and re-insertion. In this work, we investigate the use of recent learning-based keypoint descriptors for six degree-of-freedom camera pose estimation in intraoperative endoscopic sequences and under changes in anatomy due to surgical resection. METHODS: Our method employs a dense structure from motion (SfM) reconstruction of the preoperative anatomy, obtained with a state-of-the-art patient-specific learning-based descriptor. During the reconstruction step, each estimated 3D point is associated with a descriptor. This information is employed in the intraoperative sequences to establish 2D-3D correspondences for Perspective-n-Point (PnP) camera pose estimation. We evaluate this method in six intraoperative sequences that include anatomical modifications obtained from two cadaveric subjects. RESULTS: Show that this approach led to translation and rotation errors of 3.9 mm and 0.2 radians, respectively, with 21.86% of localized cameras averaged over the six sequences. In comparison to an additional learning-based descriptor (HardNet++), the selected descriptor can achieve a better percentage of localized cameras with similar pose estimation performance. We further discussed potential error causes and limitations of the proposed approach. CONCLUSION: Patient-specific learning-based descriptors can relocalize images that are well distributed across the inspected anatomy, even where the anatomy is modified. However, camera relocalization in endoscopic sequences remains a persistently challenging problem, and future research is necessary to increase the robustness and accuracy of this technique.


Subject(s)
Endoscopy , Surgery, Computer-Assisted , Humans , Endoscopy/methods , Rotation
9.
Int J Comput Assist Radiol Surg ; 18(6): 1077-1084, 2023 Jun.
Article in English | MEDLINE | ID: mdl-37160583

ABSTRACT

PURPOSE: Digital twins are virtual replicas of real-world objects and processes, and they have potential applications in the field of surgical procedures, such as enhancing situational awareness. We introduce Twin-S, a digital twin framework designed specifically for skull base surgeries. METHODS: Twin-S is a novel framework that combines high-precision optical tracking and real-time simulation, making it possible to integrate it into image-guided interventions. To guarantee accurate representation, Twin-S employs calibration routines to ensure that the virtual model precisely reflects all real-world processes. Twin-S models and tracks key elements of skull base surgery, including surgical tools, patient anatomy, and surgical cameras. Importantly, Twin-S mirrors real-world drilling and updates the virtual model at frame rate of 28. RESULTS: Our evaluation of Twin-S demonstrates its accuracy, with an average error of 1.39 mm during the drilling process. Our study also highlights the benefits of Twin-S, such as its ability to provide augmented surgical views derived from the continuously updated virtual model, thus offering additional situational awareness to the surgeon. CONCLUSION: We present Twin-S, a digital twin environment for skull base surgery. Twin-S captures the real-world surgical progresses and updates the virtual model in real time through the use of modern tracking technologies. Future research that integrates vision-based techniques could further increase the accuracy of Twin-S.


Subject(s)
Surgery, Computer-Assisted , Humans , Surgery, Computer-Assisted/methods , Imaging, Three-Dimensional/methods , Neurosurgical Procedures , Computer Simulation , Skull Base/surgery
10.
Int J Comput Assist Radiol Surg ; 18(7): 1167-1174, 2023 Jul.
Article in English | MEDLINE | ID: mdl-37171660

ABSTRACT

PURPOSE: Robotic assistance in otologic surgery can reduce the task load of operating surgeons during the removal of bone around the critical structures in the lateral skull base. However, safe deployment into the anatomical passageways necessitates the development of advanced sensing capabilities to actively limit the interaction forces between the surgical tools and critical anatomy. METHODS: We introduce a surgical drill equipped with a force sensor that is capable of measuring accurate tool-tissue interaction forces to enable force control and feedback to surgeons. The design, calibration and validation of the force-sensing surgical drill mounted on a cooperatively controlled surgical robot are described in this work. RESULTS: The force measurements on the tip of the surgical drill are validated with raw-egg drilling experiments, where a force sensor mounted below the egg serves as ground truth. The average root mean square error for points and path drilling experiments is 41.7 (± 12.2) mN and 48.3 (± 13.7) mN, respectively. CONCLUSION: The force-sensing prototype measures forces with sub-millinewton resolution and the results demonstrate that the calibrated force-sensing drill generates accurate force measurements with minimal error compared to the measured drill forces. The development of such sensing capabilities is crucial for the safe use of robotic systems in a clinical context.


Subject(s)
Robotic Surgical Procedures , Robotics , Surgery, Computer-Assisted , Humans , Mastoidectomy , Surgery, Computer-Assisted/methods , Feedback
11.
Otolaryngol Head Neck Surg ; 169(4): 988-998, 2023 10.
Article in English | MEDLINE | ID: mdl-36883992

ABSTRACT

OBJECTIVE: Preoperative planning for otologic or neurotologic procedures often requires manual segmentation of relevant structures, which can be tedious and time-consuming. Automated methods for segmenting multiple geometrically complex structures can not only streamline preoperative planning but also augment minimally invasive and/or robot-assisted procedures in this space. This study evaluates a state-of-the-art deep learning pipeline for semantic segmentation of temporal bone anatomy. STUDY DESIGN: A descriptive study of a segmentation network. SETTING: Academic institution. METHODS: A total of 15 high-resolution cone-beam temporal bone computed tomography (CT) data sets were included in this study. All images were co-registered, with relevant anatomical structures (eg, ossicles, inner ear, facial nerve, chorda tympani, bony labyrinth) manually segmented. Predicted segmentations from no new U-Net (nnU-Net), an open-source 3-dimensional semantic segmentation neural network, were compared against ground-truth segmentations using modified Hausdorff distances (mHD) and Dice scores. RESULTS: Fivefold cross-validation with nnU-Net between predicted and ground-truth labels were as follows: malleus (mHD: 0.044 ± 0.024 mm, dice: 0.914 ± 0.035), incus (mHD: 0.051 ± 0.027 mm, dice: 0.916 ± 0.034), stapes (mHD: 0.147 ± 0.113 mm, dice: 0.560 ± 0.106), bony labyrinth (mHD: 0.038 ± 0.031 mm, dice: 0.952 ± 0.017), and facial nerve (mHD: 0.139 ± 0.072 mm, dice: 0.862 ± 0.039). Comparison against atlas-based segmentation propagation showed significantly higher Dice scores for all structures (p < .05). CONCLUSION: Using an open-source deep learning pipeline, we demonstrate consistently submillimeter accuracy for semantic CT segmentation of temporal bone anatomy compared to hand-segmented labels. This pipeline has the potential to greatly improve preoperative planning workflows for a variety of otologic and neurotologic procedures and augment existing image guidance and robot-assisted systems for the temporal bone.


Subject(s)
Deep Learning , Ear, Inner , Humans , Temporal Bone/diagnostic imaging , Cone-Beam Computed Tomography , Tomography, X-Ray Computed/methods , Image Processing, Computer-Assisted/methods
12.
Comput Intell Neurosci ; 2022: 9653513, 2022.
Article in English | MEDLINE | ID: mdl-36105634

ABSTRACT

The capacity to carry out one's regular tasks is affected to varying degrees by hearing difficulties. Poorer understanding, slower learning, and an overall reduction in efficiency in academic endeavours are just a few of the negative impacts of hearing impairments on children's performance, which may range from mild to severe. A significant factor in determining whether or not there will be a decrease in performance is the kind and source of impairment. Research has shown that the Artificial Neural Network technique is capable of modelling both linear and nonlinear solution surfaces in a trustworthy way, as demonstrated in previous studies. To improve the precision with which hearing impairment challenges are diagnosed, a neural network backpropagation approach has been developed with the purpose of fine-tuning the diagnostic process. In particular, it highlights the vital role performed by medical informatics in supporting doctors in the identification of diseases as well as the formulation of suitable choices via the use of data management and knowledge discovery. As part of the intelligent control method, it is proposed in this research to construct a Histogram Equalization (HE)-based Adaptive Center-Weighted Median (ACWM) filter, which is then used to segment/detect the OM in tympanic membrane images using different segmentation methods in order to minimise noise and improve the image quality. A tympanic membrane dataset, which is freely accessible, was used in all experiments.


Subject(s)
Algorithms , Otitis , Child , Humans , Neural Networks, Computer
13.
Med Image Anal ; 73: 102166, 2021 10.
Article in English | MEDLINE | ID: mdl-34340104

ABSTRACT

Vertebral labelling and segmentation are two fundamental tasks in an automated spine processing pipeline. Reliable and accurate processing of spine images is expected to benefit clinical decision support systems for diagnosis, surgery planning, and population-based analysis of spine and bone health. However, designing automated algorithms for spine processing is challenging predominantly due to considerable variations in anatomy and acquisition protocols and due to a severe shortage of publicly available data. Addressing these limitations, the Large Scale Vertebrae Segmentation Challenge (VerSe) was organised in conjunction with the International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) in 2019 and 2020, with a call for algorithms tackling the labelling and segmentation of vertebrae. Two datasets containing a total of 374 multi-detector CT scans from 355 patients were prepared and 4505 vertebrae have individually been annotated at voxel level by a human-machine hybrid algorithm (https://osf.io/nqjyw/, https://osf.io/t98fz/). A total of 25 algorithms were benchmarked on these datasets. In this work, we present the results of this evaluation and further investigate the performance variation at the vertebra level, scan level, and different fields of view. We also evaluate the generalisability of the approaches to an implicit domain shift in data by evaluating the top-performing algorithms of one challenge iteration on data from the other iteration. The principal takeaway from VerSe: the performance of an algorithm in labelling and segmenting a spine scan hinges on its ability to correctly identify vertebrae in cases of rare anatomical variations. The VerSe content and code can be accessed at: https://github.com/anjany/verse.


Subject(s)
Benchmarking , Tomography, X-Ray Computed , Algorithms , Humans , Image Processing, Computer-Assisted , Spine/diagnostic imaging
14.
Int J Comput Assist Radiol Surg ; 16(5): 849-859, 2021 May.
Article in English | MEDLINE | ID: mdl-33982232

ABSTRACT

PURPOSE: Segmentation of surgical instruments in endoscopic video streams is essential for automated surgical scene understanding and process modeling. However, relying on fully supervised deep learning for this task is challenging because manual annotation occupies valuable time of the clinical experts. METHODS: We introduce a teacher-student learning approach that learns jointly from annotated simulation data and unlabeled real data to tackle the challenges in simulation-to-real unsupervised domain adaptation for endoscopic image segmentation. RESULTS: Empirical results on three datasets highlight the effectiveness of the proposed framework over current approaches for the endoscopic instrument segmentation task. Additionally, we provide analysis of major factors affecting the performance on all datasets to highlight the strengths and failure modes of our approach. CONCLUSIONS: We show that our proposed approach can successfully exploit the unlabeled real endoscopic video frames and improve generalization performance over pure simulation-based training and the previous state-of-the-art. This takes us one step closer to effective segmentation of surgical instrument in the annotation scarce setting.


Subject(s)
Computer Simulation , Data Curation , Endoscopy/methods , Image Processing, Computer-Assisted/methods , Algorithms , Artifacts , Humans , Learning , Software , Students , Video Recording
15.
Med Image Anal ; 52: 24-41, 2019 02.
Article in English | MEDLINE | ID: mdl-30468970

ABSTRACT

Surgical tool detection is attracting increasing attention from the medical image analysis community. The goal generally is not to precisely locate tools in images, but rather to indicate which tools are being used by the surgeon at each instant. The main motivation for annotating tool usage is to design efficient solutions for surgical workflow analysis, with potential applications in report generation, surgical training and even real-time decision support. Most existing tool annotation algorithms focus on laparoscopic surgeries. However, with 19 million interventions per year, the most common surgical procedure in the world is cataract surgery. The CATARACTS challenge was organized in 2017 to evaluate tool annotation algorithms in the specific context of cataract surgery. It relies on more than nine hours of videos, from 50 cataract surgeries, in which the presence of 21 surgical tools was manually annotated by two experts. With 14 participating teams, this challenge can be considered a success. As might be expected, the submitted solutions are based on deep learning. This paper thoroughly evaluates these solutions: in particular, the quality of their annotations are compared to that of human interpretations. Next, lessons learnt from the differential analysis of these solutions are discussed. We expect that they will guide the design of efficient surgery monitoring tools in the near future.


Subject(s)
Cataract Extraction/instrumentation , Deep Learning , Surgical Instruments , Algorithms , Humans , Video Recording
16.
Int J Comput Assist Radiol Surg ; 12(6): 1013-1020, 2017 Jun.
Article in English | MEDLINE | ID: mdl-28357628

ABSTRACT

PURPOSE: A fully automated surgical tool detection framework is proposed for endoscopic video streams. State-of-the-art surgical tool detection methods rely on supervised one-vs-all or multi-class classification techniques, completely ignoring the co-occurrence relationship of the tools and the associated class imbalance. METHODS: In this paper, we formulate tool detection as a multi-label classification task where tool co-occurrences are treated as separate classes. In addition, imbalance on tool co-occurrences is analyzed and stratification techniques are employed to address the imbalance during convolutional neural network (CNN) training. Moreover, temporal smoothing is introduced as an online post-processing step to enhance runtime prediction. RESULTS: Quantitative analysis is performed on the M2CAI16 tool detection dataset to highlight the importance of stratification, temporal smoothing and the overall framework for tool detection. CONCLUSION: The analysis on tool imbalance, backed by the empirical results, indicates the need and superiority of the proposed framework over state-of-the-art techniques.


Subject(s)
Endoscopy/methods , Foreign Bodies/diagnostic imaging , Image Processing, Computer-Assisted/methods , Neural Networks, Computer , Pattern Recognition, Automated/methods , Algorithms , Humans
17.
J Clin Diagn Res ; 9(10): ZC49-52, 2015 Oct.
Article in English | MEDLINE | ID: mdl-26557616

ABSTRACT

BACKGROUND: Oral submucous fibrosis (OSMF) is a chronic progressive debilitating disease affecting the oral, oropharyngeal and sometimes the oesophageal mucosa resulting in inability to eat due to burning, ulcers and stiffness. AIM: The study was undertaken, to evaluate the correlation of clinical staging, histological grading and nutritional status using body mass index (BMI) with gutkha (habit) index in OSMF patients. MATERIALS AND METHODS: The study group comprised of 50 patients clinically diagnosed and histopathologically confirmed cases of OSMF. Habit (gutkha) index was calculated by multiplying duration and frequency. Body mass index was calculated by dividing weight in kilograms and height in centimetres of the patient. RESULTS: Male to female ratio was 2.8:1. Clinical grading increased with increase in gutkha index, patients with gutkha index 1-50, maximum were in mild stage; with gutkha index 51-100, maximum in moderate and patients with gutkha index 101-150, all were in severe stage. Histological staging showed direct correlation with gutkha index, it increased with increase in gutkha index with p <0.05. Site analysis showed that buccal mucosa and retromolar area were involved in all the patient and floor of mouth in 46% of patients Body mass index analysis revealed that out of 27 patients with moderate clinical staging 3 was underweight; out of 3 with severe clinical staging, 2 was underweight. CONCLUSION: The duration and frequency of areca nut product use effects on the incidence and severity of OSMF and the patient becomes unable to eat due to burning, ulcers and inability to open mouth which affect the health of the individual. Thus it is important to access the nutritional status to improve the survival rate of patients.

SELECTION OF CITATIONS
SEARCH DETAIL
...