Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 32
Filter
1.
IEEE Rev Biomed Eng ; 10: 213-234, 2017.
Article in English | MEDLINE | ID: mdl-28092576

ABSTRACT

Multiple-instance learning (MIL) is a recent machine-learning paradigm that is particularly well suited to medical image and video analysis (MIVA) tasks. Based solely on class labels assigned globally to images or videos, MIL algorithms learn to detect relevant patterns locally in images or videos. These patterns are then used for classification at a global level. Because supervision relies on global labels, manual segmentations are not needed to train MIL algorithms, unlike traditional single-instance learning (SIL) algorithms. Consequently, these solutions are attracting increasing interest from the MIVA community: since the term was coined by Dietterich et al. in 1997, 73 research papers about MIL have been published in the MIVA literature. This paper reviews the existing strategies for modeling MIVA tasks as MIL problems, recommends general-purpose MIL algorithms for each type of MIVA tasks, and discusses MIVA-specific MIL algorithms. Various experiments performed in medical image and video datasets are compiled in order to back up these discussions. This meta-analysis shows that, besides being more convenient than SIL solutions, MIL algorithms are also more accurate in many cases. In other words, MIL is the ideal solution for many MIVA tasks. Recent trends are discussed, and future directions are proposed for this emerging paradigm.


Subject(s)
Algorithms , Image Interpretation, Computer-Assisted , Video Recording , Databases as Topic , Humans , Models, Theoretical
2.
Transl Vis Sci Technol ; 5(2): 16, 2016 Apr.
Article in English | MEDLINE | ID: mdl-27134775

ABSTRACT

PURPOSE: We assessed the suitability of a low-cost, handheld, nonmydriatic retinograph, namely the Horus DEC 200, for diabetic retinopathy (DR) diagnosis. Two factors were considered: ease of image acquisition and image quality. METHODS: One operator acquired fundus photographs from 54 patients using the Horus and AFC-330, a more expensive, nonportable retinograph. Satisfaction surveys were filled out by patients. Then, two retinologists subjectively assessed image quality and graded DR severity in one eye of each patient. Objective image quality indices also were computed. RESULTS: During image acquisitions, patients had difficulty locating the fixation target inside the Horus: by default, 53.7% of them had to fixate external points with the contralateral eye, as opposed to none of them using the AFC-330 (P < 0.0001). This issue impacted the duration of image acquisitions. Images obtained by the Horus were of significantly lower quality according to the experts (P = 0.0002 and P = 0.0004) and to the objective criterion (P < 0.0001). As a result, up to 20.4% of eyes were inadequate for interpretation, as opposed to 9.3% using the AFC-330. However, no significant difference was found in terms of DR severity according to both experts (P = 0.557 and P = 0.156). CONCLUSIONS: The Horus can be used to screen DR, but at the cost of longer examination times and higher proportions of patients referred to an ophthalmologist due to inadequate image quality. TRANSLATIONAL RELEVANCE: The Horus is adequate to screen DR, for instance in primary care centers or in mobile imaging units.

3.
IEEE Trans Med Imaging ; 35(7): 1604-14, 2016 07.
Article in English | MEDLINE | ID: mdl-26829783

ABSTRACT

This paper describes a computer-aided detection and diagnosis system for breast cancer, the most common form of cancer among women, using mammography. The system relies on the Multiple-Instance Learning (MIL) paradigm, which has proven useful for medical decision support in previous works from our team. In the proposed framework, breasts are first partitioned adaptively into regions. Then, features derived from the detection of lesions (masses and microcalcifications) as well as textural features, are extracted from each region and combined in order to classify mammography examinations as "normal" or "abnormal". Whenever an abnormal examination record is detected, the regions that induced that automated diagnosis can be highlighted. Two strategies are evaluated to define this anomaly detector. In a first scenario, manual segmentations of lesions are used to train an SVM that assigns an anomaly index to each region; local anomaly indices are then combined into a global anomaly index. In a second scenario, the local and global anomaly detectors are trained simultaneously, without manual segmentations, using various MIL algorithms (DD, APR, mi-SVM, MI-SVM and MILBoost). Experiments on the DDSM dataset show that the second approach, which is only weakly-supervised, surprisingly outperforms the first approach, even though it is strongly-supervised. This suggests that anomaly detectors can be advantageously trained on large medical image archives, without the need for manual segmentation.


Subject(s)
Mammography , Algorithms , Breast , Breast Neoplasms , Calcinosis , Female , Humans
4.
Med Image Anal ; 29: 47-64, 2016 Apr.
Article in English | MEDLINE | ID: mdl-26774796

ABSTRACT

With the increased prevalence of retinal pathologies, automating the detection of these pathologies is becoming more and more relevant. In the past few years, many algorithms have been developed for the automated detection of a specific pathology, typically diabetic retinopathy, using eye fundus photography. No matter how good these algorithms are, we believe many clinicians would not use automatic detection tools focusing on a single pathology and ignoring any other pathology present in the patient's retinas. To solve this issue, an algorithm for characterizing the appearance of abnormal retinas, as well as the appearance of the normal ones, is presented. This algorithm does not focus on individual images: it considers examination records consisting of multiple photographs of each retina, together with contextual information about the patient. Specifically, it relies on data mining in order to learn diagnosis rules from characterizations of fundus examination records. The main novelty is that the content of examination records (images and context) is characterized at multiple levels of spatial and lexical granularity: 1) spatial flexibility is ensured by an adaptive decomposition of composite retinal images into a cascade of regions, 2) lexical granularity is ensured by an adaptive decomposition of the feature space into a cascade of visual words. This multigranular representation allows for great flexibility in automatically characterizing normality and abnormality: it is possible to generate diagnosis rules whose precision and generalization ability can be traded off depending on data availability. A variation on usual data mining algorithms, originally designed to mine static data, is proposed so that contextual and visual data at adaptive granularity levels can be mined. This framework was evaluated in e-ophtha, a dataset of 25,702 examination records from the OPHDIAT screening network, as well as in the publicly-available Messidor dataset. It was successfully applied to the detection of patients that should be referred to an ophthalmologist and also to the specific detection of several pathologies.


Subject(s)
Data Mining/methods , Decision Support Systems, Clinical , Electronic Health Records , Pattern Recognition, Automated/methods , Retinal Diseases/diagnosis , Retinoscopy/methods , Humans , Image Interpretation, Computer-Assisted/methods , Machine Learning , Natural Language Processing , Photography/methods , Referral and Consultation , Reproducibility of Results , Sensitivity and Specificity
5.
Annu Int Conf IEEE Eng Med Biol Soc ; 2016: 3239-3242, 2016 Aug.
Article in English | MEDLINE | ID: mdl-28268998

ABSTRACT

With the increased prevalence of retinal pathologies, automating the detection and progression measurement of these pathologies is becoming more and more relevant. Color fundus photography is the leading modality for assessing retinal pathologies. Because eye fundus cameras have a limited field of view, multiple photographs are taken from each retina during an eye fundus examination. However, operators usually don't indicate which photographs are from the left retina and which ones are from the right retina. This paper presents a novel algorithm that automatically assigns each photograph to one retina and builds a composite image (or "mosaic") per retina, which is expected to push the performance of automated diagnosis forward. The algorithm starts by jointly forming two mosaics, one per retina, using a novel graph theoretic approach. Then, in order to determine which mosaic corresponds to the left retina and which one corresponds to the right retina, two retinal landmarks are detected robustly in each mosaic: the main vessel arch surrounding the macula and the optic disc. The laterality of each mosaic derives from their relative location. Experiments on 2790 manually annotated images validate the very good performance of the proposed framework even for highly pathological images.


Subject(s)
Fundus Oculi , Photography/methods , Retina/physiology , Algorithms , Databases as Topic , Diabetic Retinopathy/diagnosis , Diabetic Retinopathy/physiopathology , Humans , Image Processing, Computer-Assisted
6.
Article in English | MEDLINE | ID: mdl-26737917

ABSTRACT

This paper describes an experimental computer-aided detection and diagnosis system for breast cancer, the most common form of cancer among women, using mammography. The system relies on the Multiple-Instance Learning (MIL) paradigm, which has proven useful for medical decision support in previous works from our team. In the proposed framework, the breasts are first partitioned adaptively into regions. Then, either textural features, or features derived from the detection of masses and microcalcifications, are extracted from each region. Finally, feature vectors extracted from each region are combined using an MIL algorithm (Citation k-NN or mi-Graph), in order to recognize "normal" mammography examinations or to categorize examinations as "normal", "benign" or "cancer". An accuracy of 91.1% (respectively 62.1%) was achieved for normality recognition (respectively three-class categorization) in a subset of 720 mammograms from the DDSM dataset. The paper also discusses future improvements, that will make the most of the MIL paradigm, in order to improve "benign" versus "cancer" discrimination in particular.


Subject(s)
Algorithms , Breast Neoplasms/diagnostic imaging , Mammography/methods , Breast/pathology , Calcinosis/diagnostic imaging , Databases as Topic , Female , Humans
7.
IEEE Trans Med Imaging ; 34(4): 877-87, 2015 Apr.
Article in English | MEDLINE | ID: mdl-25373078

ABSTRACT

This paper introduces a new algorithm for recognizing surgical tasks in real-time in a video stream. The goal is to communicate information to the surgeon in due time during a video-monitored surgery. The proposed algorithm is applied to cataract surgery, which is the most common eye surgery. To compensate for eye motion and zoom level variations, cataract surgery videos are first normalized. Then, the motion content of short video subsequences is characterized with spatiotemporal polynomials: a multiscale motion characterization based on adaptive spatiotemporal polynomials is presented. The proposed solution is particularly suited to characterize deformable moving objects with fuzzy borders, which are typically found in surgical videos. Given a target surgical task, the system is trained to identify which spatiotemporal polynomials are usually extracted from videos when and only when this task is being performed. These key spatiotemporal polynomials are then searched in new videos to recognize the target surgical task. For improved performances, the system jointly adapts the spatiotemporal polynomial basis and identifies the key spatiotemporal polynomials using the multiple-instance learning paradigm. The proposed system runs in real-time and outperforms the previous solution from our group, both for surgical task recognition ( Az = 0.851 on average, as opposed to Az = 0.794 previously) and for the joint segmentation and recognition of surgical tasks ( Az = 0.856 on average, as opposed to Az = 0.832 previously).


Subject(s)
Cataract Extraction/methods , Pattern Recognition, Automated/methods , Surgery, Computer-Assisted/methods , Video Recording/methods , Humans , Image Processing, Computer-Assisted , Machine Learning
8.
IEEE Trans Med Imaging ; 33(12): 2352-60, 2014 Dec.
Article in English | MEDLINE | ID: mdl-25055383

ABSTRACT

In ophthalmology, it is now common practice to record every surgical procedure and to archive the resulting videos for documentation purposes. In this paper, we present a solution to automatically segment and categorize surgical tasks in real-time during the surgery, using the video recording. The goal would be to communicate information to the surgeon in due time, such as recommendations to the less experienced surgeons. The proposed solution relies on the content-based video retrieval paradigm: it reuses previously archived videos to automatically analyze the current surgery, by analogy reasoning. Each video is segmented, in real-time, into an alternating sequence of idle phases, during which no clinically-relevant motions are visible, and action phases. As soon as an idle phase is detected, the previous action phase is categorized and the next action phase is predicted. A conditional random field is used for categorization and prediction. The proposed system was applied to the automatic segmentation and categorization of cataract surgery tasks. A dataset of 186 surgeries, performed by ten different surgeons, was manually annotated: ten possibly overlapping surgical tasks were delimited in each surgery. Using the content of action phases and the duration of idle phases as sources of evidence, an average recognition performance of Az = 0.832 ± 0.070 was achieved.


Subject(s)
Cataract Extraction/methods , Image Processing, Computer-Assisted/methods , Surgery, Computer-Assisted/methods , Video-Assisted Surgery/methods , Algorithms , Databases, Factual , Female , Humans , Male , ROC Curve
9.
Med Image Anal ; 18(7): 1026-43, 2014 Oct.
Article in English | MEDLINE | ID: mdl-24972380

ABSTRACT

The automatic detection of exudates in color eye fundus images is an important task in applications such as diabetic retinopathy screening. The presented work has been undertaken in the framework of the TeleOphta project, whose main objective is to automatically detect normal exams in a tele-ophthalmology network, thus reducing the burden on the readers. A new clinical database, e-ophtha EX, containing precisely manually contoured exudates, is introduced. As opposed to previously available databases, e-ophtha EX is very heterogeneous. It contains images gathered within the OPHDIAT telemedicine network for diabetic retinopathy screening. Image definition, quality, as well as patients condition or the retinograph used for the acquisition, for example, are subject to important changes between different examinations. The proposed exudate detection method has been designed for this complex situation. We propose new preprocessing methods, which perform not only normalization and denoising tasks, but also detect reflections and artifacts in the image. A new candidates segmentation method, based on mathematical morphology, is proposed. These candidates are characterized using classical features, but also novel contextual features. Finally, a random forest algorithm is used to detect the exudates among the candidates. The method has been validated on the e-ophtha EX database, obtaining an AUC of 0.95. It has been also validated on other databases, obtaining an AUC between 0.93 and 0.95, outperforming state-of-the-art methods.


Subject(s)
Diabetic Retinopathy/diagnosis , Exudates and Transudates , Image Interpretation, Computer-Assisted/methods , Mass Screening/methods , Algorithms , Artifacts , Calibration , Color , Humans , Reproducibility of Results , Sensitivity and Specificity
10.
Med Image Anal ; 18(3): 579-90, 2014 Apr.
Article in English | MEDLINE | ID: mdl-24637155

ABSTRACT

Nowadays, many surgeries, including eye surgeries, are video-monitored. We present in this paper an automatic video analysis system able to recognize surgical tasks in real-time. The proposed system relies on the Content-Based Video Retrieval (CBVR) paradigm. It characterizes short subsequences in the video stream and searches for video subsequences with similar structures in a video archive. Fixed-length feature vectors are built for each subsequence: the feature vectors are unchanged by variations in duration and temporal structure among the target surgical tasks. Therefore, it is possible to perform fast nearest neighbor searches in the video archive. The retrieved video subsequences are used to recognize the current surgical task by analogy reasoning. The system can be trained to recognize any surgical task using weak annotations only. It was applied to a dataset of 23 epiretinal membrane surgeries and a dataset of 100 cataract surgeries. Three surgical tasks were annotated in the first dataset. Nine surgical tasks were annotated in the second dataset. To assess its generality, the system was also applied to a dataset of 1,707 movie clips in which 12 human actions were annotated. High task recognition scores were measured in all three datasets. Real-time task recognition will be used in future works to communicate with surgeons (trainees in particular) or with surgical devices.


Subject(s)
Image Interpretation, Computer-Assisted/methods , Imaging, Three-Dimensional/methods , Ophthalmologic Surgical Procedures/methods , Pattern Recognition, Automated/methods , Photography/methods , Surgery, Computer-Assisted/methods , Video Recording/methods , Algorithms , Artificial Intelligence , Computer Systems , Eye Diseases/surgery , Humans , Image Enhancement/methods , Reproducibility of Results , Sensitivity and Specificity
11.
Article in English | MEDLINE | ID: mdl-25569912

ABSTRACT

Anterior eye segment surgeries are usually video-recorded. If we are able to efficiently analyze surgical videos in real-time, new decision support tools will emerge. The main anatomical landmarks in these videos are the pupil boundaries and the limbus, but segmenting them is challenging due to the variety of colors and textures in the pupil, the iris, the sclera and the lids. In this paper, we present a solution to reliably normalize the center and the scale in videos, without explicitly segmenting these landmarks. First, a robust solution to track the pupil center is presented: it uses the fact that the pupil boundaries, the limbus and the sclera / lid interface are concentric. Second, a solution to estimate the zoom level is presented: it relies on the illumination pattern reflected on the cornea. The proposed solution was assessed in a dataset of 186 real-live cataract surgery videos. The distance between the true and estimated pupil centers was equal to 8.0 ± 6.9% of the limbus radius. The correlation between the estimated zoom level and the true limbus size in images was high: R = 0.834.


Subject(s)
Cataract Extraction/methods , Cataract/diagnosis , Cornea/surgery , Cornea/pathology , Decision Support Techniques , Humans , Image Interpretation, Computer-Assisted , Sclera/pathology , Video Recording/methods
12.
Article in English | MEDLINE | ID: mdl-25571028

ABSTRACT

Huge amounts of surgical data are recorded during video-monitored surgery. Content-based video retrieval systems intent to reuse those data for computer-aided surgery. In this paper, we focus on real-time recognition of cataract surgery steps: the goal is to retrieve from a database surgery videos that were recorded during the same surgery step. The proposed system relies on motion features for video characterization. Motion features are usually impacted by eye motion or zoom level variations, which are not necessarily relevant for surgery step recognition. Those problems certainly limit the performance of the retrieval system. We therefore propose to refine motion feature extraction by applying pre-processing steps based on a novel pupil center and scale tracking method. Those pre-processing steps are evaluated for two different motion features. In this paper, a similarity measure adapted from Piciarelli's video surveillance system is evaluated for the first time in a surgery dataset. This similarity measure provides good results and for both motion features, the proposed preprocessing steps improved the retrieval performance of the system significantly.


Subject(s)
Cataract Extraction , Pattern Recognition, Automated/methods , Video Recording , Algorithms , Automation , Databases, Factual , Eye Movements , Humans
13.
Article in English | MEDLINE | ID: mdl-24110967

ABSTRACT

Breast mass segmentation in mammography plays a crucial role in Computer-Aided Diagnosis (CAD) systems. In this paper a Bidimensional Emperical Mode Decomposition (BEMD) method is introduced for the mass segmentation in mammography images. This method is used to decompose images into a set of functions named Bidimensional Intrinsic Mode Functions (BIMF) and a residue. Our approach consists of three steps: 1) the regions of interest (ROIs) were identified by using iterative thresholding; 2) the contour of the regions of interest (ROI) was extracted from the first BIMF by using the (BEMD) method; 3) the region of interest was finally refined by the extracted contour. The proposed approach is tested on (MIAS) database and the obtained results demonstrate the efficacy of the proposed approach.


Subject(s)
Breast Neoplasms/diagnostic imaging , Databases, Factual , Image Processing, Computer-Assisted/methods , Mammography/methods , Algorithms , Diagnosis, Computer-Assisted/methods , Female , Humans , Mammary Glands, Human/pathology
14.
Article in English | MEDLINE | ID: mdl-24111392

ABSTRACT

This paper presents TeleOphta, an automatic system for screening diabetic retinopathy in teleophthalmology networks. Its goal is to reduce the burden on ophthalmologists by automatically detecting non referable examination records, i.e. examination records presenting no image quality problems and no pathological signs related to diabetic retinopathy or any other retinal pathology. TeleOphta is an attempt to put into practice years of algorithmic developments from our groups. It combines image quality metrics, specific lesion detectors and a generic pathological pattern miner to process the visual content of eye fundus photographs. This visual information is further combined with contextual data in order to compute an abnormality risk for each examination record. The TeleOphta system was trained and tested on a large dataset of 25,702 examination records from the OPHDIAT screening network in Paris. It was able to automatically detect 68% of the non referable examination records while achieving the same sensitivity as a second ophthalmologist. This suggests that it could safely reduce the burden on ophthalmologists by 56%.


Subject(s)
Data Mining , Diabetic Retinopathy/pathology , Algorithms , Aneurysm/pathology , Databases, Factual , Diabetic Retinopathy/diagnosis , Diabetic Retinopathy/epidemiology , Exudates and Transudates/metabolism , Humans , Multimedia , Photography , ROC Curve , Retina/pathology , Sensitivity and Specificity , Telemedicine
15.
Med Image Anal ; 16(6): 1228-40, 2012 Aug.
Article in English | MEDLINE | ID: mdl-22850462

ABSTRACT

A novel multiple-instance learning framework, for automated image classification, is presented in this paper. Given reference images marked by clinicians as relevant or irrelevant, the image classifier is trained to detect patterns, of arbitrary size, that only appear in relevant images. After training, similar patterns are sought in new images in order to classify them as either relevant or irrelevant images. Therefore, no manual segmentations are required. As a consequence, large image datasets are available for training. The proposed framework was applied to diabetic retinopathy screening in 2-D retinal image datasets: Messidor (1200 images) and e-ophtha, a dataset of 25,702 examination records from the Ophdiat screening network (107,799 images). In this application, an image (or an examination record) is relevant if the patient should be referred to an ophthalmologist. Trained on one half of Messidor, the classifier achieved high performance on the other half of Messidor (A(z)=0.881) and on e-ophtha (A(z)=0.761). We observed, in a subset of 273 manually segmented images from e-ophtha, that all eight types of diabetic retinopathy lesions are detected.


Subject(s)
Algorithms , Artificial Intelligence , Diabetic Retinopathy/pathology , Image Interpretation, Computer-Assisted/methods , Mass Screening/methods , Pattern Recognition, Automated/methods , Retinoscopy/methods , Humans , Image Enhancement/methods , Reproducibility of Results , Sensitivity and Specificity
16.
Article in English | MEDLINE | ID: mdl-23367041

ABSTRACT

In this paper, we address the problem of computer-aided ophthalmic surgery. In particular, a novel Content-Based Video Retrieval (CBVR) system is presented : given a video stream captured by a digital camera monitoring the current surgery, the system retrieves, within digital archives, videos that resemble the current surgery monitoring video. The search results may be used to guide surgeons' decisions, for example, let the surgeon know what a more experienced fellow worker would do in a similar situation. With this goal, we propose to use motion information contained in MPEG- 4 AVC/H.264 video standard to extract features from videos. We propose two approaches, one of which is based on motion histogram created for every frame of a compressed video sequence to extract motion direction and intensity statistics. The other combine segmentation and tracking to extract region displacements between consecutive frames and therefore characterize region trajectories. To compare videos, an extension of the fast dynamic time warping to multidimensional time series was adopted. The system is applied to a dataset of 69 video-recorded retinal surgery steps. Results are promising: the retrieval efficiency is higher than 69%.


Subject(s)
Epiretinal Membrane/pathology , Epiretinal Membrane/surgery , Image Interpretation, Computer-Assisted/methods , Information Storage and Retrieval/methods , Ophthalmologic Surgical Procedures/methods , Surgery, Computer-Assisted/methods , Video Recording/methods , Algorithms , Databases, Factual , Humans , Motion , Subtraction Technique
17.
Article in English | MEDLINE | ID: mdl-23367286

ABSTRACT

In recent years, many image analysis algorithms have been presented to assist Diabetic Retinopathy (DR) screening. The goal was usually to detect healthy examination records automatically, in order to reduce the number of records that should be analyzed by retinal experts. In this paper, a novel application is presented: these algorithms are used to 1) discover image characteristics that sometimes cause an expert to disagree with his/her peers and 2) warn the expert whenever these characteristics are detected in an examination record. In a DR screening program, each examination record is only analyzed by one expert, therefore analyzing disagreements among experts is challenging. A statistical framework, based on Parzen-windowing and the Patrick-Fischer distance, is presented to solve this problem. Disagreements among eleven experts from the Ophdiat screening program were analyzed, using an archive of 25,702 examination records.


Subject(s)
Diabetic Retinopathy/physiopathology , Image Processing, Computer-Assisted , Retina/physiology , Algorithms , Humans
18.
IEEE Trans Image Process ; 21(4): 1613-23, 2012 Apr.
Article in English | MEDLINE | ID: mdl-22194244

ABSTRACT

Adaptive wavelet-based image characterizations have been proposed in previous works for content-based image retrieval (CBIR) applications. In these applications, the same wavelet basis was used to characterize each query image: This wavelet basis was tuned to maximize the retrieval performance in a training data set. We take it one step further in this paper: A different wavelet basis is used to characterize each query image. A regression function, which is tuned to maximize the retrieval performance in the training data set, is used to estimate the best wavelet filter, i.e., in terms of expected retrieval performance, for each query image. A simple image characterization, which is based on the standardized moments of the wavelet coefficient distributions, is presented. An algorithm is proposed to compute this image characterization almost instantly for every possible separable or nonseparable wavelet filter. Therefore, using a different wavelet basis for each query image does not considerably increase computation times. On the other hand, significant retrieval performance increases were obtained in a medical image data set, a texture data set, a face recognition data set, and an object picture data set. This additional flexibility in wavelet adaptation paves the way to relevance feedback on image characterization itself and not simply on the way image characterizations are combined.


Subject(s)
Algorithms , Artificial Intelligence , Image Interpretation, Computer-Assisted/methods , Information Storage and Retrieval/methods , Pattern Recognition, Automated/methods , Subtraction Technique , Wavelet Analysis , Image Enhancement/methods , Radiology Information Systems
19.
Invest Ophthalmol Vis Sci ; 52(11): 8342-8, 2011 Oct 21.
Article in English | MEDLINE | ID: mdl-21896872

ABSTRACT

PURPOSE: Recent studies on diabetic retinopathy (DR) screening in fundus photographs suggest that disagreements between algorithms and clinicians are now comparable to disagreements among clinicians. The purpose of this study is to (1) determine whether this observation also holds for automated DR severity assessment algorithms, and (2) show the interest of such algorithms in clinical practice. METHODS: A dataset of 85 consecutive DR examinations (168 eyes, 1176 multimodal eye fundus photographs) was collected at Brest University Hospital (Brest, France). Two clinicians with different experience levels determined DR severity in each eye, according to the International Clinical Diabetic Retinopathy Disease Severity (ICDRS) scale. Based on Cohen's kappa (κ) measurements, the performance of clinicians at assessing DR severity was compared to the performance of state-of-the-art content-based image retrieval (CBIR) algorithms from our group. RESULTS: At assessing DR severity in each patient, intraobserver agreement was κ = 0.769 for the most experienced clinician. Interobserver agreement between clinicians was κ = 0.526. Interobserver agreement between the most experienced clinicians and the most advanced algorithm was κ = 0.592. Besides, the most advanced algorithm was often able to predict agreements and disagreements between clinicians. CONCLUSIONS: Automated DR severity assessment algorithms, trained to imitate experienced clinicians, can be used to predict when young clinicians would agree or disagree with their more experienced fellow members. Such algorithms may thus be used in clinical practice to help validate or invalidate their diagnoses. CBIR algorithms, in particular, may also be used for pooling diagnostic knowledge among peers, with applications in training and coordination of clinicians' prescriptions.


Subject(s)
Algorithms , Diabetes Mellitus, Type 1/complications , Diabetes Mellitus, Type 2/complications , Diabetic Retinopathy/diagnosis , Diagnosis, Computer-Assisted/methods , Photography/methods , Aged , Artificial Intelligence , Databases, Factual , Diagnosis, Computer-Assisted/statistics & numerical data , Diagnostic Techniques, Ophthalmological/statistics & numerical data , Female , Humans , Male , Middle Aged , Observer Variation , Photography/statistics & numerical data , Severity of Illness Index
20.
IEEE Trans Med Imaging ; 30(1): 108-18, 2011 Jan.
Article in English | MEDLINE | ID: mdl-20693107

ABSTRACT

A novel content-based heterogeneous information retrieval framework, particularly well suited to browse medical databases and support new generation computer aided diagnosis (CADx) systems, is presented in this paper. It was designed to retrieve possibly incomplete documents, consisting of several images and semantic information, from a database; more complex data types such as videos can also be included in the framework. The proposed retrieval method relies on image processing, in order to characterize each individual image in a document by their digital content, and information fusion. Once the available images in a query document are characterized, a degree of match, between the query document and each reference document stored in the database, is defined for each attribute (an image feature or a metadata). A Bayesian network is used to recover missing information if need be. Finally, two novel information fusion methods are proposed to combine these degrees of match, in order to rank the reference documents by decreasing relevance for the query. In the first method, the degrees of match are fused by the Bayesian network itself. In the second method, they are fused by the Dezert-Smarandache theory: the second approach lets us model our confidence in each source of information (i.e., each attribute) and take it into account in the fusion process for a better retrieval performance. The proposed methods were applied to two heterogeneous medical databases, a diabetic retinopathy database and a mammography screening database, for computer aided diagnosis. Precisions at five of 0.809 ± 0.158 and 0.821 ± 0.177, respectively, were obtained for these two databases, which is very promising.


Subject(s)
Databases, Factual , Diagnosis, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods , Information Storage and Retrieval/methods , Algorithms , Bayes Theorem , Database Management Systems , Diabetic Retinopathy/diagnosis , Diagnostic Imaging/methods , Humans , Mammography/methods , Pattern Recognition, Automated/methods , Radiology Information Systems
SELECTION OF CITATIONS
SEARCH DETAIL
...