Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 13 de 13
Filter
Add more filters










Publication year range
1.
JMIR Biomed Eng ; 8: e50924, 2023.
Article in English | MEDLINE | ID: mdl-37982072

ABSTRACT

Background: In Japan, individuals with mild COVID-19 illness previously required to be monitored in designated areas and were hospitalized only if their condition worsened to moderate illness or worse. Daily monitoring using a pulse oximeter was a crucial indicator for hospitalization. However, a drastic increase in the number of patients resulted in a shortage of pulse oximeters for monitoring. Therefore, an alternative and cost-effective method for monitoring patients with mild illness was required. Previous studies have shown that voice biomarkers for Parkinson disease or Alzheimer disease are useful for classifying or monitoring symptoms; thus, we tried to adapt voice biomarkers for classifying the severity of COVID-19 using a dynamic time warping (DTW) algorithm where voice wavelets can be treated as 2D features; the differences between wavelet features are calculated as scores. Objective: This feasibility study aimed to test whether DTW-based indices can generate voice biomarkers for a binary classification model using COVID-19 patients' voices to distinguish moderate illness from mild illness at a significant level. Methods: We conducted a cross-sectional study using voice samples of COVID-19 patients. Three kinds of long vowels were processed into 10-cycle waveforms with standardized power and time axes. The DTW-based indices were generated by all pairs of waveforms and tested with the Mann-Whitney U test (α<.01) and verified with a linear discrimination analysis and confusion matrix to determine which indices were better for binary classification of disease severity. A binary classification model was generated based on a generalized linear model (GLM) using the most promising indices as predictors. The receiver operating characteristic curve/area under the curve (ROC/AUC) validated the model performance, and the confusion matrix calculated the model accuracy. Results: Participants in this study (n=295) were infected with COVID-19 between June 2021 and March 2022, were aged 20 years or older, and recuperated in Kanagawa prefecture. Voice samples (n=110) were selected from the participants' attribution matrix based on age group, sex, time of infection, and whether they had mild illness (n=61) or moderate illness (n=49). The DTW-based variance indices were found to be significant (P<.001, except for 1 of 6 indices), with a balanced accuracy in the range between 79% and 88.6% for the /a/, /e/, and /u/ vowel sounds. The GLM achieved a high balance accuracy of 86.3% (for /a/), 80.2% (for /e/), and 88% (for /u/) and ROC/AUC of 94.8% (95% CI 90.6%-94.8%) for /a/, 86.5% (95% CI 79.8%-86.5%) for /e/, and 95.6% (95% CI 92.1%-95.6%) for /u/. Conclusions: The proposed model can be a voice biomarker for an alternative and cost-effective method of monitoring the progress of COVID-19 patients in care.

2.
Front Neurol ; 14: 1197840, 2023.
Article in English | MEDLINE | ID: mdl-37416305

ABSTRACT

In this study, the technique associated with the capturing involuntary changes in voice elements caused by diseases is applied to diagnose them and a voice index is proposed to discriminate mild cognitive impairments. The participants in this study included 399 elderly people aged 65 years or older living in Matsumoto City, Nagano Prefecture, Japan. The participants were categorized into healthy and mild cognitive impairment groups based on clinical evaluation. It was hypothesized that as dementia progressed, task performance would become more challenging, and the effects on vocal cords and prosody would change significantly. In the study, voice samples of the participants were recorded while they were engaged in mental calculational tasks and during the reading of the results of the calculations written on paper. The change in prosody during the calculation from that during reading was expressed based on the difference in the acoustics. Principal component analysis was used to aggregate groups of voice features with similar characteristics of feature differences into several principal components. These principal components were combined with logistic regression analysis to propose a voice index to discriminate different mild cognitive impairment types. Discrimination accuracies of 90% and 65% were obtained for discriminations using the proposed index on the training and verification data (obtained from a population different from the training data), respectively. Therefore, it is suggested that the proposed index may be utilized as a means for discriminating mild cognitive impairments.

3.
Article in English | MEDLINE | ID: mdl-36900976

ABSTRACT

Voice-based depression detection methods have been studied worldwide as an objective and easy method to detect depression. Conventional studies estimate the presence or severity of depression. However, an estimation of symptoms is a necessary technique not only to treat depression, but also to relieve patients' distress. Hence, we studied a method for clustering symptoms from HAM-D scores of depressed patients and by estimating patients in different symptom groups based on acoustic features of their speech. We could separate different symptom groups with an accuracy of 79%. The results suggest that voice from speech can estimate the symptoms associated with depression.


Subject(s)
Depressive Disorder, Major , Voice , Humans , Depression , Depressive Disorder, Major/diagnosis , Speech , Acoustics
4.
Article in English | MEDLINE | ID: mdl-36834110

ABSTRACT

The authors are currently conducting research on methods to estimate psychiatric and neurological disorders from a voice by focusing on the features of speech. It is empirically known that numerous psychosomatic symptoms appear in voice biomarkers; in this study, we examined the effectiveness of distinguishing changes in the symptoms associated with novel coronavirus infection using speech features. Multiple speech features were extracted from the voice recordings, and, as a countermeasure against overfitting, we selected features using statistical analysis and feature selection methods utilizing pseudo data and built and verified machine learning algorithm models using LightGBM. Applying 5-fold cross-validation, and using three types of sustained vowel sounds of /Ah/, /Eh/, and /Uh/, we achieved a high performance (accuracy and AUC) of over 88% in distinguishing "asymptomatic or mild illness (symptoms)" and "moderate illness 1 (symptoms)". Accordingly, the results suggest that the proposed index using voice (speech features) can likely be used in distinguishing the symptoms associated with novel coronavirus infection.


Subject(s)
COVID-19 , Coronavirus , Humans , Speech , Voice Quality , Speech Acoustics , Patient Acuity , Severity of Illness Index
5.
Article in English | MEDLINE | ID: mdl-36141675

ABSTRACT

In general, it is common knowledge that people's feelings are reflected in their voice and facial expressions. This research work focuses on developing techniques for diagnosing depression based on acoustic properties of the voice. In this study, we developed a composite index of vocal acoustic properties that can be used for depression detection. Voice recordings were collected from patients undergoing outpatient treatment for major depressive disorder at a hospital or clinic following a physician's diagnosis. Numerous features were extracted from the collected audio data using openSMILE software. Furthermore, qualitatively similar features were combined using principal component analysis. The resulting components were incorporated as parameters in a logistic regression based classifier, which achieved a diagnostic accuracy of ~90% on the training set and ~80% on the test set. Lastly, the proposed metric could serve as a new measure for evaluation of major depressive disorder.


Subject(s)
Depressive Disorder, Major , Voice Disorders , Voice , Acoustics , Depressive Disorder, Major/diagnosis , Humans , Logistic Models
6.
Sci Rep ; 11(1): 13615, 2021 06 30.
Article in English | MEDLINE | ID: mdl-34193915

ABSTRACT

In this research, we propose a new index of emotional arousal level using sound pressure change acceleration, called the emotional arousal level voice index (EALVI), and investigate the relationship between this index and depression severity. First, EALVI values were calculated from various speech recordings in the interactive emotional dyadic motion capture database, and the correlation with the emotional arousal level of each voice was examined. The resulting correlation coefficient was 0.52 (n = 10,039, p < 2.2 × 10-16). We collected a total of 178 datasets comprising 10 speech phrases and the Hamilton Rating Scale for Depression (HAM-D) score of outpatients with major depression at the Ginza Taimei Clinic (GTC) and the National Defense Medical College (NDMC) Hospital. The correlation coefficients between the EALVI and HAM-D scores were - 0.33 (n = 88, p = 1.8 × 10-3) and - 0.43 (n = 90, p = 2.2 × 10-5) at the GTC and NDMC, respectively. Next, the dataset was divided into "no depression" (HAM-D < 8) and "depression" groups (HAM-D ≥ 8) according to the HAM-D score. The number of patients in the "no depression" and "depression" groups were 10 and 78 in the GTC data, and 65 and 25 in the NDMC data, respectively. There was a significant difference in the mean EALVI values between the two groups in both the GTC and NDMC data (p = 8.9 × 10-3, Cliff's delta = 0.51 and p = 1.6 × 10-3; Cliff's delta = 0.43, respectively). The area under the curve of the receiver operating characteristic curve when discriminating both groups by EALVI was 0.76 in GTC data and 0.72 in NDMC data. Indirectly, the data suggest that there is some relationship between emotional arousal level and depression severity.


Subject(s)
Arousal , Databases, Factual , Depression/physiopathology , Depressive Disorder, Major/physiopathology , Emotions , Voice , Adult , Female , Humans , Male , Middle Aged , Severity of Illness Index
7.
Article in English | MEDLINE | ID: mdl-34069609

ABSTRACT

BACKGROUND: In many developed countries, mood disorders have become problematic, and the economic loss due to treatment costs and interference with work is immeasurable. Therefore, a simple technique to determine individuals' depressive state and stress level is desired. METHODS: We developed a method to assess specific the psychological issues of individuals with major depressive disorders using emotional components contained in their voice. We propose two indices: vitality, a short-term index, and mental activity, a long-term index capturing trends in vitality. To evaluate our method, we used the voices of healthy individuals (n = 14) and patients with major depression (n = 30). The patients were also assessed by specialists using the Hamilton Rating Scale for Depression (HAM-D). RESULTS: A significant negative correlation existed between the vitality extracted from the voices and HAM-D scores (r = -0.33, p < 0.05). Furthermore, we could discriminate the voice data of healthy individuals and patients with depression with a high accuracy using the vitality indicator (p = 0.0085, area under the curve of the receiver operating characteristic curve = 0.76).


Subject(s)
Depressive Disorder, Major , Affect , Depression , Depressive Disorder, Major/diagnosis , Humans , Mood Disorders , Psychiatric Status Rating Scales
8.
Sensors (Basel) ; 22(1)2021 Dec 23.
Article in English | MEDLINE | ID: mdl-35009610

ABSTRACT

It is empirically known that mood changes affect facial expressions and voices. In this study, the authors have focused on the voice to develop a method for estimating depression in individuals from their voices. A short input voice is ideal for applying the proposed method to a wide range of applications. Therefore, we evaluated this method using multiple input utterances while assuming a unit utterance input. The experimental results revealed that depressive states could be estimated with sufficient accuracy using the smallest number of utterances when positive utterances were included in three to four input utterances.


Subject(s)
Depression , Voice , Humans
10.
Sensors (Basel) ; 20(18)2020 Sep 04.
Article in English | MEDLINE | ID: mdl-32899881

ABSTRACT

Recently, the relationship between emotional arousal and depression has been studied. Focusing on this relationship, we first developed an arousal level voice index (ALVI) to measure arousal levels using the Interactive Emotional Dyadic Motion Capture database. Then, we calculated ALVI from the voices of depressed patients from two hospitals (Ginza Taimei Clinic (H1) and National Defense Medical College hospital (H2)) and compared them with the severity of depression as measured by the Hamilton Rating Scale for Depression (HAM-D). Depending on the HAM-D score, the datasets were classified into a no depression (HAM-D < 8) and a depression group (HAM-D ≥ 8) for each hospital. A comparison of the mean ALVI between the groups was performed using the Wilcoxon rank-sum test and a significant difference at the level of 10% (p = 0.094) at H1 and 1% (p = 0.0038) at H2 was determined. The area under the curve (AUC) of the receiver operating characteristic was 0.66 when categorizing between the two groups for H1, and the AUC for H2 was 0.70. The relationship between arousal level and depression severity was indirectly suggested via the ALVI.


Subject(s)
Arousal , Depressive Disorder, Major , Voice Recognition , Adult , Aged , Depression/diagnosis , Depressive Disorder, Major/diagnosis , Female , Humans , Male , Middle Aged , Psychiatric Status Rating Scales , Severity of Illness Index , Young Adult
11.
JMIR Form Res ; 4(7): e16455, 2020 Jul 20.
Article in English | MEDLINE | ID: mdl-32554367

ABSTRACT

BACKGROUND: We developed a system for monitoring mental health using voice data from daily phone calls, termed Mind Monitoring System (MIMOSYS), by implementing a method for estimating mental health status from voice data. OBJECTIVE: The objective of this study was to evaluate the potential of this system for detecting depressive states and monitoring stress-induced mental changes. METHODS: We opened our system to the public in the form of a prospective study in which data were collected over 2 years from a large, unspecified sample of users. We used these data to analyze the relationships between the rate of continued use, the men-to-women ratio, and existing psychological tests for this system over the study duration. Moreover, we analyzed changes in mental data over time under stress from particular life events. RESULTS: The system had a high rate of continued use. Voice indicators showed that women have more depressive tendencies than men, matching the rate of depression in Japan. The system's voice indicators and the scores on classical psychological tests were correlated. We confirmed deteriorating mental health for users in areas affected by major earthquakes in Japan around the time of the earthquakes. CONCLUSIONS: The results suggest that although this system is insufficient for detecting depression, it may be effective for monitoring changes in mental health due to stress. The greatest feature of our system is mental health monitoring, which is most effectively accomplished by performing long-term time-series analysis of the acquired data considering the user's life events. Such a system can improve the implementation of patient interventions by evaluating objective data along with life events.

12.
Am J Disaster Med ; 15(4): 251-259, 2020.
Article in English | MEDLINE | ID: mdl-33428196

ABSTRACT

OBJECTIVE: The mental health issues of personnel dealing with the deceased at times of disasters is a problem and techniques are needed that allow for real-time, easy-to-use stress checks. We have studied techniques for measuring mental state using voice analysis which has the benefit of being non-invasive, easy-to-use, and can be performed in real-time. For this study, we used voice measurement to determine the stress experienced during body identification training workshops for dentists. We studied whether or not stress levels were affected by having previous experience with body identification either in actual disaster settings or during training. DESIGN: Since participants training using actual dead bodies in particular are expected to suffer higher stress exposure, we also assessed their mental state pre- and post-training using actual dead bodies. RESULTS: The results confirmed marked differences in the mental state between before and after training in participants without any actual experience, between participants who engaged in training using manikins before actual dead bodies and participants who did not. CONCLUSIONS: These results suggest that, in body identification training, the level of stress when coming into contact with dead bodies varies depending on participants' experience and the training sequence. Moreover, it is believed that voice-based stress assessment can be conducted in the limited time during training sessions and that it can be usefully implemented in actual disaster response settings.


Subject(s)
Disasters , Humans , Time Factors
13.
Disaster Mil Med ; 3: 4, 2017.
Article in English | MEDLINE | ID: mdl-28405348

ABSTRACT

BACKGROUND: Disaster relief personnel tend to be exposed to excessive stress, which can be a cause of mental disorders. To prevent from mental disorders, frequent assessment of mental status is important. This pilot study aimed to examine feasibility of stress assessment using vocal affect display (VAD) indices as calculated by our proposed algorithms in a situation of comparison between different durations of stay in stricken area as disaster relief operation, which is an environment highly likely to induce stress. METHODS: We used Sensibility Technology (ST) software to analyze VAD from voices of participants exposed to extreme stress for either long or short durations, and we proposed algorithms for indices of low VAD (VAD-L), high VAD (VAD-H), and VAD ratio (VAD-R), calculated from the intensity of emotions as measured by voice emotion analysis. As a preliminary validation, 12 members of Japan Self-Defense Forces dispatched overseas for long (3 months or more) or short (about a week) durations were asked to record their voices saying 11 phrases repeatedly across 6 days during their dispatch. RESULTS: In the validation, the two groups showed an inverse relationship in VAD-L and VAD-H, in that long durations in disaster zones resulted in higher values of both VAD-L and VAD-R, and lower values of VAD-H, compared with short durations. Interestingly, phrases produced varied results in terms of group differences and VAD indices, demonstrating the sensitivity of the ST. CONCLUSIONS: A comparison of the values obtained for the different groups of subjects clarified that there were tendencies of the VAD-L, VAD-H, and VAD-R indices observed for each group of participants. The results suggest the possibility of using ST software in the measurement of affective aspects related to mental health from vocal behavior.

SELECTION OF CITATIONS
SEARCH DETAIL
...