ABSTRACT
INTRODUCTION: The rise of new technologies in the field of health is yielding promising results. In certain chronic conditions such as type 2 diabetes mellitus, which ranks among the top five causes of global mortality, it could be useful in supporting patient management. MATERIALS AND METHODS: A systematic review will be conducted on scientific publications from the last 5 years (January 2019 to October 2023) to describe the effect of mobile app usage on glycated hemoglobin for the management of adult patients with type 2 diabetes mellitus who participated in randomized controlled clinical trials. The search will be carried out in the databases of MEDLINE (Ovid), Embase (Ovid), CINAHL (EBSCOhost), CENTRAL, WoS, Scopus, Epistemonikos, and LILACS. The search strategy will be constructed using both controlled and natural language. Additionally, the Cochrane filter will be applied to identify randomized controlled trials. The review will include scientific articles reporting studies that present results from randomized controlled trials, with texts in Spanish, English, or French, utilizing mobile applications for the management of adult individuals (over 18 years) with type 2 diabetes mellitus, and whose outcomes report the effects on glycated hemoglobin. The Cochrane Risk of Bias Tool will be used to assess the quality of the studies, and the Grading of Recommendations, Assessment, Development, and Evaluation (GRADE) methodology will be implemented to evaluate the certainty of the evidence. RESULTS: The analysis will be conducted by observing the value of the glycated hemoglobin levels of the participants. Given that this data is a quantitative and continuous value, it facilitates the identification of the effects of the mobile applications used for the management of type 2 diabetes mellitus (T2DM) in adults. Furthermore, if sufficient data are available, a meta-analysis will be conducted using IBM-SPSS. The effect of the intervention will be estimated by the mean difference. All point estimates will be accompanied by 95% confidence intervals. A random effects model will be used. The heterogeneity of the results will be assessed using Cochrane's Q and I2 statistics. DISCUSSION: Considering that the quality of content and functionality of certain applications in the healthcare field is highly variable, it is necessary to evaluate the scientific evidence reported on the effect of the use of this type of technology in people with T2DM.
Subject(s)
Diabetes Mellitus, Type 2 , Mobile Applications , Systematic Reviews as Topic , Diabetes Mellitus, Type 2/therapy , Humans , Glycated Hemoglobin/analysis , Glycated Hemoglobin/metabolism , Randomized Controlled Trials as TopicABSTRACT
Obesity is a chronic disease with an increasing impact on the world's population. In this work, we present a method of identifying obesity automatically using text mining techniques and information related to body weight measures and obesity comorbidities. We used a dataset of 3015 de-identified medical records that contain labels for two classification problems. The first classification problem distinguishes between obesity, overweight, normal weight, and underweight. The second classification problem differentiates between obesity types: super obesity, morbid obesity, severe obesity and moderate obesity. We used a Bag of Words approach to represent the records together with unigram and bigram representations of the features. We implemented two approaches: a hierarchical method and a nonhierarchical one. We used Support Vector Machine and Naïve Bayes together with ten-fold cross validation to evaluate and compare performances. Our results indicate that the hierarchical approach does not work as well as the nonhierarchical one. In general, our results show that Support Vector Machine obtains better performances than Naïve Bayes for both classification problems. We also observed that bigram representation improves performance compared with unigram representation.
Subject(s)
Artificial Intelligence , Data Mining/methods , Electronic Health Records/organization & administration , Obesity/diagnosis , Bayes Theorem , Comorbidity , Humans , Natural Language Processing , Overweight/diagnosis , Support Vector MachineABSTRACT
In this work we present a system to identify and extract patient's smoking status from clinical narrative text in Spanish. The clinical narrative text was processed using natural language processing techniques, and annotated by four people with a biomedical background. The dataset used for classification had 2,465 documents, each one annotated with one of the four smoking status categories. We used two feature representations: single word token and bigrams. The classification problem was divided in two levels. First recognizing between smoker (S) and non-smoker (NS); second recognizing between current smoker (CS) and past smoker (PS). For each feature representation and classification level, we used two classifiers: Support Vector Machines (SVM) and Bayesian Networks (BN). We split our dataset as follows: a training set containing 66% of the available documents that was used to build classifiers and a test set containing the remaining 34% of the documents that was used to test and evaluate the model. Our results show that SVM together with the bigram representation performed better in both classification levels. For S vs NS classification level performance measures were: ACC=85%, Precision=85%, and Recall=90%. For CS vs PS classification level performance measures were: ACC=87%, Precision=91%, and Recall=94%.
Subject(s)
Databases, Factual , Electronic Health Records/classification , Natural Language Processing , Smoking , Bayes Theorem , Chile , Humans , Narration , Support Vector MachineABSTRACT
The use of text mining and supervised machine learning algorithms on biomedical databases has become increasingly common. However, a question remains: How much data must be annotated to create a suitable training set for a machine learning classifier? In prior research with active learning in medical text classification, we found evidence that not only sample size but also some of the intrinsic characteristics of the texts being analyzed-such as the size of the vocabulary and the length of a document-may also influence the resulting classifier's performance. This study is an attempt to create a regression model to predict performance based on sample size and other text features. While the model needs to be trained on existing datasets, we believe it is feasible to predict performance without obtaining annotations from new datasets once the model is built.
Subject(s)
Artificial Intelligence , Documentation/classification , Documentation/statistics & numerical data , Meaningful Use/statistics & numerical data , Natural Language Processing , Terminology as Topic , Vocabulary, Controlled , Data Curation/methods , Data Mining/statistics & numerical data , Pattern Recognition, Automated/methods , Pattern Recognition, Automated/statistics & numerical data , Sample SizeABSTRACT
OBJECTIVE: This study explores active learning algorithms as a way to reduce the requirements for large training sets in medical text classification tasks. DESIGN: Three existing active learning algorithms (distance-based (DIST), diversity-based (DIV), and a combination of both (CMB)) were used to classify text from five datasets. The performance of these algorithms was compared to that of passive learning on the five datasets. We then conducted a novel investigation of the interaction between dataset characteristics and the performance results. MEASUREMENTS: Classification accuracy and area under receiver operating characteristics (ROC) curves for each algorithm at different sample sizes were generated. The performance of active learning algorithms was compared with that of passive learning using a weighted mean of paired differences. To determine why the performance varies on different datasets, we measured the diversity and uncertainty of each dataset using relative entropy and correlated the results with the performance differences. RESULTS: The DIST and CMB algorithms performed better than passive learning. With a statistical significance level set at 0.05, DIST outperformed passive learning in all five datasets, while CMB was found to be better than passive learning in four datasets. We found strong correlations between the dataset diversity and the DIV performance, as well as the dataset uncertainty and the performance of the DIST algorithm. CONCLUSION: For medical text classification, appropriate active learning algorithms can yield performance comparable to that of passive learning with considerably smaller training sets. In particular, our results suggest that DIV performs better on data with higher diversity and DIST on data with lower uncertainty.
Subject(s)
Data Mining/methods , Natural Language Processing , Algorithms , Artificial Intelligence , Humans , ROC CurveABSTRACT
BACKGROUND: Supervised learning methods need annotated data in order to generate efficient models. Annotated data, however, is a relatively scarce resource and can be expensive to obtain. For both passive and active learning methods, there is a need to estimate the size of the annotated sample required to reach a performance target. METHODS: We designed and implemented a method that fits an inverse power law model to points of a given learning curve created using a small annotated training set. Fitting is carried out using nonlinear weighted least squares optimization. The fitted model is then used to predict the classifier's performance and confidence interval for larger sample sizes. For evaluation, the nonlinear weighted curve fitting method was applied to a set of learning curves generated using clinical text and waveform classification tasks with active and passive sampling methods, and predictions were validated using standard goodness of fit measures. As control we used an un-weighted fitting method. RESULTS: A total of 568 models were fitted and the model predictions were compared with the observed performances. Depending on the data set and sampling method, it took between 80 to 560 annotated samples to achieve mean average and root mean squared error below 0.01. Results also show that our weighted fitting method outperformed the baseline un-weighted method (p < 0.05). CONCLUSIONS: This paper describes a simple and effective sample size prediction algorithm that conducts weighted fitting of learning curves. The algorithm outperformed an un-weighted algorithm described in previous literature. It can help researchers determine annotation sample size for supervised machine learning.
Subject(s)
Algorithms , Learning Curve , Problem-Based Learning/methods , Sample Size , Data Interpretation, Statistical , Diagnosis, Computer-Assisted , Humans , Models, Statistical , Nonlinear Dynamics , Pattern Recognition, Automated , Predictive Value of Tests , Probability Learning , Reproducibility of Results , Stochastic ProcessesABSTRACT
We developed a method to help tailor a comprehensive vocabulary system (e.g. the UMLS) for a sub-domain (e.g. clinical reports) in support of natural language processing (NLP). The method detects unused sense in a sub-domain by comparing the relational neighborhood of a word/term in the vocabulary with the semantic neighborhood of the word/term in the sub-domain. The semantic neighborhood of the word/term in the sub-domain is determined using latent semantic analysis (LSA). We trained and tested the unused sense detection on two clinical text corpora: one contains discharge summaries and the other outpatient visit notes. We were able to detect unused senses with precision from 79% to 87%, recall from 48% to 74%, and an area under receiver operation curve (AUC) of 72% to 87%.