Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 3 de 3
Filter
Add more filters










Database
Language
Publication year range
1.
J Voice ; 2022 Sep 05.
Article in English | MEDLINE | ID: mdl-36075802

ABSTRACT

OBJECTIVES: The purpose of study is to improve the classification accuracy by comparing the results obtained by applying decision tree ensemble learning, which is one of the methods to increase the classification accuracy for a relatively small dataset, with the results obtained by the convolutional neural network (CNN) algorithm for the diagnosis of glottal cancer. METHODS: Pusan National University Hospital (PNUH) dataset were used to establish classifiers and Pusan National University Yangsan Hospital (PNUYH) dataset were used to verify the classifier's performance in the generated model. For the diagnosis of glottic cancer, deep learning-based CNN models were established and classified using laryngeal image and voice data. Classification accuracy was obtained by performing decision tree ensemble learning using probability through CNN classification algorithm. In this process, the classification and regression tree (CART) method was used. Then, we compared the classification accuracy of decision tree ensemble learning with CNN individual classifiers by fusing the laryngeal image with the voice decision tree classifier. RESULTS: We obtained classification accuracy of 81.03 % and 99.18 % in the established laryngeal image and voice classification models using PNUH training dataset, respectively. However, the classification accuracy of CNN classifiers decreased to 73.88 % in voice and 68.92 % in laryngeal image when using an external dataset of PNUYH. To solve this problem, decision tree ensemble learning of laryngeal image and voice was used, and the classification accuracy was improved by integrating data of laryngeal image and voice of the same person. The classification accuracy was 87.88 % and 89.06 % for the individualized laryngeal image and voice decision tree model respectively, and the fusion of the laryngeal image and voice decision tree results represented a classification accuracy of 95.31 %. CONCLUSION: The results of our study suggest that decision tree ensemble learning aimed at training multiple classifiers is useful to obtain an increased classification accuracy despite a small dataset. Although a large data amount is essential for AI analysis, when an integrated approach is taken by combining various input data high diagnostic classification accuracy can be expected.

2.
Article in English | MEDLINE | ID: mdl-18002689

ABSTRACT

Diagnosis of pathological voice is one of the most important issues in biomedical applications of speech technology. This study focuses on the classification of pathological voice using the HMM(Hidden Markov Model), the GMM(Gaussian Mixture Model) and a SVM (Support Vector Machine), and then compares the results to work done previously using an ANN (Artificial Neural Network). Speech data were collected from those without and those with vocal disorders. Normal and pathological speech data were mixed in out experiment. Six characteristic parameters (Jitter, Shimmer, NHR, SPI, APQ and RAP) were chosen. Then the pattern recognition methods (HMM, GMM and SVM) were used to distinguish the mixed data into categories of normal and pathological speech. We found that the GMM-based method can give us superior classification rates compared to the other classification methods.


Subject(s)
Artificial Intelligence , Diagnosis, Computer-Assisted/methods , Pattern Recognition, Automated/methods , Sound Spectrography/methods , Speech Disorders/diagnosis , Speech Production Measurement/methods , Vocal Cord Paralysis/diagnosis , Algorithms , Humans , Reproducibility of Results , Sensitivity and Specificity , Speech Disorders/etiology , Vocal Cord Paralysis/complications
3.
Conf Proc IEEE Eng Med Biol Soc ; 2005: 4678-81, 2005.
Article in English | MEDLINE | ID: mdl-17281284

ABSTRACT

The aim of this paper is to analyze and discriminate the pathological voice by separating signal into periodic and aperiodic parts. Separation was performed recursively from the residual signal of voice signal. Based on initial estimation of aperiodic part of spectrum, aperiodic part is decided from the extrapolation method. Periodic part is decided by subtracting aperiodic part from the original spectrum. A parameter HNR is derived based on the separation. Parameter value statistics are compared with those of Jitter and Shimmer for normal, benign and malignant cases.

SELECTION OF CITATIONS
SEARCH DETAIL
...