Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 32
Filtrar
1.
Stud Health Technol Inform ; 310: 124-128, 2024 Jan 25.
Artigo em Inglês | MEDLINE | ID: mdl-38269778

RESUMO

Creating notes in the EHR is one of the most problematic aspects for health professionals. The main challenges are the time spent on this task and the quality of the records. Automatic speech recognition technologies aim to facilitate clinical documentation for users, optimizing their workflow. In our hospital, we internally developed an automatic speech recognition system (ASR) to record progress notes in a mobile EHR. The objective of this article is to describe the pilot study carried out to evaluate the implementation of ASR to record progress notes in a mobile EHR application. As a result, the specialty that used ASR the most was Home Medicine. The lack of access to a computer at the time of care and the need to perform short and fast evolutions were the main reasons for users to use the system.


Assuntos
Documentação , Interface para o Reconhecimento da Fala , Humanos , Projetos Piloto , Pessoal de Saúde , Hospitais
2.
Res Dev Disabil ; 135: 104466, 2023 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-36863156

RESUMO

This article reports the first group-based intervention study in the UK of using speech to-text (STT) technology to improve the writing of children with special educational needs and disabilities (SEND). Over a period of five years, thirty children took part in total from three settings; a mainstream school, a special school and a special unit of a different mainstream school. All children had Education, Health and Care Plans because of their difficulties in spoken and written communication. Children were trained to use the Dragon STT system, and used it on set tasks for 16-18 weeks. Handwritten text and self-esteem were assessed before and after the intervention, and screen-written text at the end. The results showed that this approach had boosted the quantity and quality of handwritten text, with post-test screen-written text significantly better than handwritten at post-test. The self-esteem instrument also showed positive and statistically significant results. The findings support the feasibility of using STT to support children with writing difficulties. All the data were gathered before the Covid-19 pandemic; the implications of this, and of the innovative research design, are discussed.


Assuntos
COVID-19 , Pessoas com Deficiência , Criança , Humanos , Fala , Pandemias , Redação
3.
J Am Med Inform Assoc ; 30(4): 703-711, 2023 03 16.
Artigo em Inglês | MEDLINE | ID: mdl-36688526

RESUMO

OBJECTIVES: Ambient clinical documentation technology uses automatic speech recognition (ASR) and natural language processing (NLP) to turn patient-clinician conversations into clinical documentation. It is a promising approach to reducing clinician burden and improving documentation quality. However, the performance of current-generation ASR remains inadequately validated. In this study, we investigated the impact of non-lexical conversational sounds (NLCS) on ASR performance. NLCS, such as Mm-hm and Uh-uh, are commonly used to convey important information in clinical conversations, for example, Mm-hm as a "yes" response from the patient to the clinician question "are you allergic to antibiotics?" MATERIALS AND METHODS: In this study, we evaluated 2 contemporary ASR engines, Google Speech-to-Text Clinical Conversation ("Google ASR"), and Amazon Transcribe Medical ("Amazon ASR"), both of which have their language models specifically tailored to clinical conversations. The empirical data used were from 36 primary care encounters. We conducted a series of quantitative and qualitative analyses to examine the word error rate (WER) and the potential impact of misrecognized NLCS on the quality of clinical documentation. RESULTS: Out of a total of 135 647 spoken words contained in the evaluation data, 3284 (2.4%) were NLCS. Among these NLCS, 76 (0.06% of total words, 2.3% of all NLCS) were used to convey clinically relevant information. The overall WER, of all spoken words, was 11.8% for Google ASR and 12.8% for Amazon ASR. However, both ASR engines demonstrated poor performance in recognizing NLCS: the WERs across frequently used NLCS were 40.8% (Google) and 57.2% (Amazon), respectively; and among the NLCS that conveyed clinically relevant information, 94.7% and 98.7%, respectively. DISCUSSION AND CONCLUSION: Current ASR solutions are not capable of properly recognizing NLCS, particularly those that convey clinically relevant information. Although the volume of NLCS in our evaluation data was very small (2.4% of the total corpus; and for NLCS that conveyed clinically relevant information: 0.06%), incorrect recognition of them could result in inaccuracies in clinical documentation and introduce new patient safety risks.


Assuntos
Idioma , Interface para o Reconhecimento da Fala , Humanos , Fala/fisiologia , Tecnologia , Documentação
4.
JMIR Aging ; 5(3): e33460, 2022 Sep 21.
Artigo em Inglês | MEDLINE | ID: mdl-36129754

RESUMO

BACKGROUND: Speech data for medical research can be collected noninvasively and in large volumes. Speech analysis has shown promise in diagnosing neurodegenerative disease. To effectively leverage speech data, transcription is important, as there is valuable information contained in lexical content. Manual transcription, while highly accurate, limits the potential scalability and cost savings associated with language-based screening. OBJECTIVE: To better understand the use of automatic transcription for classification of neurodegenerative disease, namely, Alzheimer disease (AD), mild cognitive impairment (MCI), or subjective memory complaints (SMC) versus healthy controls, we compared automatically generated transcripts against transcripts that went through manual correction. METHODS: We recruited individuals from a memory clinic ("patients") with a diagnosis of mild-to-moderate AD, (n=44, 30%), MCI (n=20, 13%), SMC (n=8, 5%), as well as healthy controls (n=77, 52%) living in the community. Participants were asked to describe a standardized picture, read a paragraph, and recall a pleasant life experience. We compared transcripts generated using Google speech-to-text software to manually verified transcripts by examining transcription confidence scores, transcription error rates, and machine learning classification accuracy. For the classification tasks, logistic regression, Gaussian naive Bayes, and random forests were used. RESULTS: The transcription software showed higher confidence scores (P<.001) and lower error rates (P>.05) for speech from healthy controls compared with patients. Classification models using human-verified transcripts significantly (P<.001) outperformed automatically generated transcript models for both spontaneous speech tasks. This comparison showed no difference in the reading task. Manually adding pauses to transcripts had no impact on classification performance. However, manually correcting both spontaneous speech tasks led to significantly higher performances in the machine learning models. CONCLUSIONS: We found that automatically transcribed speech data could be used to distinguish patients with a diagnosis of AD, MCI, or SMC from controls. We recommend a human verification step to improve the performance of automatic transcripts, especially for spontaneous tasks. Moreover, human verification can focus on correcting errors and adding punctuation to transcripts. However, manual addition of pauses is not needed, which can simplify the human verification step to more efficiently process large volumes of speech data.

5.
BMC Med Inform Decis Mak ; 22(1): 96, 2022 04 08.
Artigo em Inglês | MEDLINE | ID: mdl-35395798

RESUMO

BACKGROUND: Despite the rapid expansion of electronic health records, the use of computer mouse and keyboard, challenges the data entry into these systems. Speech recognition software is one of the substitutes for the mouse and keyboard. The objective of this study was to evaluate the use of online and offline speech recognition software on spelling errors in nursing reports and to compare them with errors in handwritten reports. METHODS: For this study, online and offline speech recognition software were selected and customized based on unrecognized terms by these softwares. Two groups of 35 nurses provided the admission notes of hospitalized patients upon their arrival using three data entry methods (using the handwritten method or two types of speech recognition software). After at least a month, they created the same reports using the other methods. The number of spelling errors in each method was determined. These errors were compared between the paper method and the two electronic methods before and after the correction of errors. RESULTS: The lowest accuracy was related to online software with 96.4% and accuracy. On the average per report, the online method 6.76, and the offline method 4.56 generated more errors than the paper method. After correcting the errors by the participants, the number of errors in the online reports decreased by 94.75% and the number of errors in the offline reports decreased by 97.20%. The highest number of reports with errors was related to reports created by online software. CONCLUSION: Although two software had relatively high accuracy, they created more errors than the paper method that can be lowered by optimizing and upgrading these softwares. The results showed that error correction by users significantly reduced the documentation errors caused by the software.


Assuntos
Percepção da Fala , Documentação , Humanos , Fala , Interface para o Reconhecimento da Fala , Tecnologia
6.
JMIR Mhealth Uhealth ; 9(10): e32301, 2021 10 12.
Artigo em Inglês | MEDLINE | ID: mdl-34636729

RESUMO

BACKGROUND: Prehospitalization documentation is a challenging task and prone to loss of information, as paramedics operate under disruptive environments requiring their constant attention to the patients. OBJECTIVE: The aim of this study is to develop a mobile platform for hands-free prehospitalization documentation to assist first responders in operational medical environments by aggregating all existing solutions for noise resiliency and domain adaptation. METHODS: The platform was built to extract meaningful medical information from the real-time audio streaming at the point of injury and transmit complete documentation to a field hospital prior to patient arrival. To this end, the state-of-the-art automatic speech recognition (ASR) solutions with the following modular improvements were thoroughly explored: noise-resilient ASR, multi-style training, customized lexicon, and speech enhancement. The development of the platform was strictly guided by qualitative research and simulation-based evaluation to address the relevant challenges through progressive improvements at every process step of the end-to-end solution. The primary performance metrics included medical word error rate (WER) in machine-transcribed text output and an F1 score calculated by comparing the autogenerated documentation to manual documentation by physicians. RESULTS: The total number of 15,139 individual words necessary for completing the documentation were identified from all conversations that occurred during the physician-supervised simulation drills. The baseline model presented a suboptimal performance with a WER of 69.85% and an F1 score of 0.611. The noise-resilient ASR, multi-style training, and customized lexicon improved the overall performance; the finalized platform achieved a medical WER of 33.3% and an F1 score of 0.81 when compared to manual documentation. The speech enhancement degraded performance with medical WER increased from 33.3% to 46.33% and the corresponding F1 score decreased from 0.81 to 0.78. All changes in performance were statistically significant (P<.001). CONCLUSIONS: This study presented a fully functional mobile platform for hands-free prehospitalization documentation in operational medical environments and lessons learned from its implementation.


Assuntos
Interface para o Reconhecimento da Fala , Fala , Documentação , Humanos , Tecnologia
7.
JMIR Res Protoc ; 10(7): e27227, 2021 Jul 28.
Artigo em Inglês | MEDLINE | ID: mdl-34319248

RESUMO

BACKGROUND: Due to an aging population, the demand for many services is exceeding the capacity of the clinical workforce. As a result, staff are facing a crisis of burnout from being pressured to deliver high-volume workloads, driving increasing costs for providers. Artificial intelligence (AI), in the form of conversational agents, presents a possible opportunity to enable efficiency in the delivery of care. OBJECTIVE: This study aims to evaluate the effectiveness, usability, and acceptability of Dora agent: Ufonia's autonomous voice conversational agent, an AI-enabled autonomous telemedicine call for the detection of postoperative cataract surgery patients who require further assessment. The objectives of this study are to establish Dora's efficacy in comparison with an expert clinician, determine baseline sensitivity and specificity for the detection of true complications, evaluate patient acceptability, collect evidence for cost-effectiveness, and capture data to support further development and evaluation. METHODS: Using an implementation science construct, the interdisciplinary study will be a mixed methods phase 1 pilot establishing interobserver reliability of the system, usability, and acceptability. This will be done using the following scales and frameworks: the system usability scale; assessment of Health Information Technology Interventions in Evidence-Based Medicine Evaluation Framework; the telehealth usability questionnaire; and the Non-Adoption, Abandonment, and Challenges to the Scale-up, Spread and Suitability framework. RESULTS: The evaluation is expected to show that conversational technology can be used to conduct an accurate assessment and that it is acceptable to different populations with different backgrounds. In addition, the results will demonstrate how successfully the system can be delivered in organizations with different clinical pathways and how it can be integrated with their existing platforms. CONCLUSIONS: The project's key contributions will be evidence of the effectiveness of AI voice conversational agents and their associated usability and acceptability. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID): PRR1-10.2196/27227.

8.
Health Technol (Berl) ; 11(4): 803-809, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34094806

RESUMO

Electronic health records (EHRs) allow for meaningful usage of healthcare data. Their adoption provides clinicians with a central location to access and share data, write notes, order labs and prescriptions, and bill for patient visits. However, as non-clinical requirements have increased, time spent using EHRs eclipsed time spent on direct patient care. Several solutions have been proposed to minimize the time spent using EHRs, though each have limitations. Digital scribe technology uses voice-to-text software to convert ambient listening to meaningful medical notes and may eliminate the physical task of documentation, allowing physicians to spend less time on EHR engagement and more time with patients. However, adoption of digital scribe technology poses many barriers for physicians. In this study, we perform a scoping review of the literature to identify barriers to digital scribe implementation and provide solutions to address these barriers. We performed a literature review of digital scribe technology and voice-to-text conversion and information extraction as a scope for future research. Fifteen articles met inclusion criteria. Of the articles included, four were comparative studies, three were reviews, three were original investigations, two were perspective pieces, one was a cost-effectiveness study, one was a keynote address, and one was an observational study. The published articles on digital scribe technology and voice-to-text conversion highlight digital scribe technology as a solution to the inefficient interaction with EHRs. Benefits of digital scribe technologies included enhancing clinician ability to navigate charts, write notes, use decision support tools, and improve the quality of time spent with patients. Digital scribe technologies can improve clinic efficiency and increase patient access to care while simultaneously reducing physician burnout. Implementation barriers include upfront costs, integration with existing technology, and time-intensive training. Technological barriers include adaptability to linguistic differences, compatibility across different clinical encounters, and integration of medical jargon into the note. Broader risks include automation bias and risks to data privacy. Overcoming significant barriers to implementation will facilitate more widespread adoption. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s12553-021-00568-0.

10.
J Med Internet Res ; 22(10): e20346, 2020 10 22.
Artigo em Inglês | MEDLINE | ID: mdl-33090118

RESUMO

BACKGROUND: The high demand for health care services and the growing capability of artificial intelligence have led to the development of conversational agents designed to support a variety of health-related activities, including behavior change, treatment support, health monitoring, training, triage, and screening support. Automation of these tasks could free clinicians to focus on more complex work and increase the accessibility to health care services for the public. An overarching assessment of the acceptability, usability, and effectiveness of these agents in health care is needed to collate the evidence so that future development can target areas for improvement and potential for sustainable adoption. OBJECTIVE: This systematic review aims to assess the effectiveness and usability of conversational agents in health care and identify the elements that users like and dislike to inform future research and development of these agents. METHODS: PubMed, Medline (Ovid), EMBASE (Excerpta Medica dataBASE), CINAHL (Cumulative Index to Nursing and Allied Health Literature), Web of Science, and the Association for Computing Machinery Digital Library were systematically searched for articles published since 2008 that evaluated unconstrained natural language processing conversational agents used in health care. EndNote (version X9, Clarivate Analytics) reference management software was used for initial screening, and full-text screening was conducted by 1 reviewer. Data were extracted, and the risk of bias was assessed by one reviewer and validated by another. RESULTS: A total of 31 studies were selected and included a variety of conversational agents, including 14 chatbots (2 of which were voice chatbots), 6 embodied conversational agents (3 of which were interactive voice response calls, virtual patients, and speech recognition screening systems), 1 contextual question-answering agent, and 1 voice recognition triage system. Overall, the evidence reported was mostly positive or mixed. Usability and satisfaction performed well (27/30 and 26/31), and positive or mixed effectiveness was found in three-quarters of the studies (23/30). However, there were several limitations of the agents highlighted in specific qualitative feedback. CONCLUSIONS: The studies generally reported positive or mixed evidence for the effectiveness, usability, and satisfactoriness of the conversational agents investigated, but qualitative user perceptions were more mixed. The quality of many of the studies was limited, and improved study design and reporting are necessary to more accurately evaluate the usefulness of the agents in health care and identify key areas for improvement. Further research should also analyze the cost-effectiveness, privacy, and security of the agents. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID): RR2-10.2196/16934.


Assuntos
Inteligência Artificial/normas , Comunicação , Atenção à Saúde , Feminino , Humanos , Masculino
11.
JMIR Form Res ; 4(8): e18751, 2020 Aug 13.
Artigo em Inglês | MEDLINE | ID: mdl-32788153

RESUMO

BACKGROUND: Objective and continuous severity measures of anxiety and depression are highly valuable and would have many applications in psychiatry and psychology. A collective source of data for objective measures are the sensors in a person's smartphone, and a particularly rich source is the microphone that can be used to sample the audio environment. This may give broad insight into activity, sleep, and social interaction, which may be associated with quality of life and severity of anxiety and depression. OBJECTIVE: This study aimed to explore the properties of passively recorded environmental audio from a subject's smartphone to find potential correlates of symptom severity of social anxiety disorder, generalized anxiety disorder, depression, and general impairment. METHODS: An Android app was designed, together with a centralized server system, to collect periodic measurements of the volume of sounds in the environment and to detect the presence or absence of English-speaking voices. Subjects were recruited into a 2-week observational study during which the app was run on their personal smartphone to collect audio data. Subjects also completed self-report severity measures of social anxiety disorder, generalized anxiety disorder, depression, and functional impairment. Participants were 112 Canadian adults from a nonclinical population. High-level features were extracted from the environmental audio of 84 participants with sufficient data, and correlations were measured between the 4 audio features and the 4 self-report measures. RESULTS: The regularity in daily patterns of activity and inactivity inferred from the environmental audio volume was correlated with the severity of depression (r=-0.37; P<.001). A measure of sleep disturbance inferred from the environmental audio volume was also correlated with the severity of depression (r=0.23; P=.03). A proxy measure of social interaction based on the detection of speaking voices in the environmental audio was correlated with depression (r=-0.37; P<.001) and functional impairment (r=-0.29; P=.01). None of the 4 environmental audio-based features tested showed significant correlations with the measures of generalized anxiety or social anxiety. CONCLUSIONS: In this study group, the environmental audio was shown to contain signals that were associated with the severity of depression and functional impairment. Associations with the severity of social anxiety disorder and generalized anxiety disorder were much weaker in comparison and not statistically significant at the 5% significance level. This work also confirmed previous work showing that the presence of voices is associated with depression. Furthermore, this study suggests that sparsely sampled audio volume could provide potentially relevant insight into subjects' mental health.

12.
Int J Med Inform ; 141: 104178, 2020 09.
Artigo em Inglês | MEDLINE | ID: mdl-32521449

RESUMO

IMPORTANCE: Speech recognition (SR) is increasingly used directly by clinicians for electronic health record (EHR) documentation. Its usability and effect on quality and efficiency versus other documentation methods remain unclear. OBJECTIVE: To study usability and quality of documentation with SR versus typing. DESIGN: In this controlled observational study, each subject participated in two of five simulated outpatient scenarios. Sessions were recorded with Morae® usability software. Two notes were documented into the EHR per encounter (one dictated, one typed) in randomized order. Participants were interviewed about each method's perceived advantages and disadvantages. Demographics and documentation habits were collected via survey. Data collection occurred between January 8 and February 8, 2019, and data analysis was conducted from February through September of 2019. SETTING: Brigham and Women's Hospital, Boston, Massachusetts, USA. PARTICIPANTS: Ten physicians who had used SR for at least six months. MAIN OUTCOMES AND MEASURES: Documentation time, word count, vocabulary size, number of errors, number of corrections and quality (clarity, completeness, concision, information sufficiency and prioritization). RESULTS: Dictated notes were longer than typed notes (320.6 vs. 180.8 words; p = 0.004) with more unique words (170.9 vs. 120.4; p = 0.01). Documentation time was similar between methods, with dictated notes taking slightly less time to complete than typed notes. Typed notes had more uncorrected errors per note than dictated notes (2.9 vs. 1.5), although most were minor misspellings. Dictated notes had a higher mean quality score (7.7 vs. 6.6; p = 0.04), were more complete and included more sufficient information. CONCLUSIONS AND RELEVANCE: Participants felt that SR saves them time, increases their efficiency and allows them to quickly document more relevant details. Quality analysis supports the perception that SR allows for more detailed notes, but whether dictation is objectively faster than typing remains unclear, and participants described some scenarios where typing is still preferred. Dictation can be effective for creating comprehensive documentation, especially when physicians like and feel comfortable using SR. Research is needed to further improve integration of SR with EHR systems and assess its impact on clinical practice, workflows, provider and patient experience, and costs.


Assuntos
Médicos , Percepção da Fala , Boston , Documentação , Registros Eletrônicos de Saúde , Feminino , Humanos , Massachusetts , Interface para o Reconhecimento da Fala
13.
J Med Internet Res ; 22(6): e14827, 2020 06 01.
Artigo em Inglês | MEDLINE | ID: mdl-32442129

RESUMO

BACKGROUND: Recent advances in natural language processing and artificial intelligence have led to widespread adoption of speech recognition technologies. In consumer health applications, speech recognition is usually applied to support interactions with conversational agents for data collection, decision support, and patient monitoring. However, little is known about the use of speech recognition in consumer health applications and few studies have evaluated the efficacy of conversational agents in the hands of consumers. In other consumer-facing tools, cognitive load has been observed to be an important factor affecting the use of speech recognition technologies in tasks involving problem solving and recall. Users find it more difficult to think and speak at the same time when compared to typing, pointing, and clicking. However, the effects of speech recognition on cognitive load when performing health tasks has not yet been explored. OBJECTIVE: The aim of this study was to evaluate the use of speech recognition for documentation in consumer digital health tasks involving problem solving and recall. METHODS: Fifty university staff and students were recruited to undertake four documentation tasks with a simulated conversational agent in a computer laboratory. The tasks varied in complexity determined by the amount of problem solving and recall required (simple and complex) and the input modality (speech recognition vs keyboard and mouse). Cognitive load, task completion time, error rate, and usability were measured. RESULTS: Compared to using a keyboard and mouse, speech recognition significantly increased the cognitive load for complex tasks (Z=-4.08, P<.001) and simple tasks (Z=-2.24, P=.03). Complex tasks took significantly longer to complete (Z=-2.52, P=.01) and speech recognition was found to be overall less usable than a keyboard and mouse (Z=-3.30, P=.001). However, there was no effect on errors. CONCLUSIONS: Use of a keyboard and mouse was preferable to speech recognition for complex tasks involving problem solving and recall. Further studies using a broader variety of consumer digital health tasks of varying complexity are needed to investigate the contexts in which use of speech recognition is most appropriate. The effects of cognitive load on task performance and its significance also need to be investigated.


Assuntos
Informática Aplicada à Saúde dos Consumidores/métodos , Laboratórios/normas , Resolução de Problemas/fisiologia , Interface para o Reconhecimento da Fala/normas , Adolescente , Adulto , Feminino , Humanos , Masculino , Pessoa de Meia-Idade , Adulto Jovem
14.
J Am Med Inform Assoc ; 27(5): 808-817, 2020 05 01.
Artigo em Inglês | MEDLINE | ID: mdl-32181812

RESUMO

OBJECTIVE: Use of medical scribes reduces clinician burnout by sharing the burden of clinical documentation. However, medical scribes are cost-prohibitive for most settings, prompting a growing interest in developing ambient, speech-based technologies capable of automatically generating clinical documentation based on patient-provider conversation. Through a systematic review, we aimed to develop a thorough understanding of the work performed by medical scribes in order to inform the design of such technologies. MATERIALS AND METHODS: Relevant articles retrieved by searching in multiple literature databases. We conducted the screening process following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) in guidelines, and then analyzed the data using qualitative methods to identify recurring themes. RESULTS: The literature search returned 854 results, 65 of which met the inclusion criteria. We found that there is significant variation in scribe expectations and responsibilities across healthcare organizations; scribes also frequently adapt their work based on the provider's style and preferences. Further, scribes' job extends far beyond capturing conversation in the exam room; they also actively interact with patients and the care team and integrate data from other sources such as prior charts and lab test results. DISCUSSION: The results of this study provide several implications for designing technologies that can generate clinical documentation based on naturalistic conversations taking place in the exam room. First, a one-size-fits-all solution will be unlikely to work because of the significant variation in scribe work. Second, technology designers need to be aware of the limited role that their solution can fulfill. Third, to produce comprehensive clinical documentation, such technologies will likely have to incorporate information beyond the exam room conversation. Finally, issues of patient consent and privacy have yet to be adequately addressed, which could become paramount barriers to implementing such technologies in realistic clinical settings. CONCLUSIONS: Medical scribes perform complex and delicate work. Further research is needed to better understand their roles in a clinical setting in order to inform the development of speech-based clinical documentation technologies.


Assuntos
Pessoal Técnico de Saúde , Documentação/métodos , Registros Eletrônicos de Saúde , Reconhecimento de Voz , Pessoal Técnico de Saúde/economia , Pessoal Técnico de Saúde/educação , Humanos , Interface para o Reconhecimento da Fala
15.
JMIR Res Protoc ; 9(3): e16934, 2020 Mar 09.
Artigo em Inglês | MEDLINE | ID: mdl-32149717

RESUMO

BACKGROUND: Conversational agents (also known as chatbots) have evolved in recent decades to become multimodal, multifunctional platforms with potential to automate a diverse range of health-related activities supporting the general public, patients, and physicians. Multiple studies have reported the development of these agents, and recent systematic reviews have described the scope of use of conversational agents in health care. However, there is scarce research on the effectiveness of these systems; thus, their viability and applicability are unclear. OBJECTIVE: The objective of this systematic review is to assess the effectiveness of conversational agents in health care and to identify limitations, adverse events, and areas for future investigation of these agents. METHODS: The Preferred Reporting Items for Systematic Reviews and Meta-Analyses Protocols will be used to structure this protocol. The focus of the systematic review is guided by a population, intervention, comparator, and outcome framework. A systematic search of the PubMed (Medline), EMBASE, CINAHL, and Web of Science databases will be conducted. Two authors will independently screen the titles and abstracts of the identified references and select studies according to the eligibility criteria. Any discrepancies will then be discussed and resolved. Two reviewers will independently extract and validate data from the included studies into a standardized form and conduct quality appraisal. RESULTS: As of January 2020, we have begun a preliminary literature search and piloting of the study selection process. CONCLUSIONS: This systematic review aims to clarify the effectiveness, limitations, and future applications of conversational agents in health care. Our findings may be useful to inform the future development of conversational agents and promote the personalization of patient care. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID): PRR1-10.2196/16934.

16.
Stud Health Technol Inform ; 264: 1761-1762, 2019 Aug 21.
Artigo em Inglês | MEDLINE | ID: mdl-31438331

RESUMO

Clinical documentation in healthcare institutions is one of the daily tasks that consumes most of the time for those involved. The adoption of mobile devices in medical practice increases efficiency among healthcare professionals. We describe the design and evaluation of an automatic speech recognition system that enables the transcription of audio to text of clinical notes in a mobile environment. Our system achieved 94.1% word accuracy when evaluated on pediatrics, internal medicine and surgery services.


Assuntos
Percepção da Fala , Documentação , Eficiência , Pessoal de Saúde , Humanos
17.
Stud Health Technol Inform ; 264: 1787-1788, 2019 Aug 21.
Artigo em Inglês | MEDLINE | ID: mdl-31438344

RESUMO

Voice technology offers a range of novel and promising strategies for clinical diabetes management. Incorporation of voice-powered virtual assistants (such as Apple Siri and Microsoft Cortana) into diabetes care programs has the potential to improve patient awareness and adherence; facilitate comprehensive provider-patient integration and data collection; and expedite consultations, procedures, and meal preparations. This study will present a qualitative literature review on existing and speculative applications of voice technology in diabetes care.


Assuntos
Diabetes Mellitus , Conscientização , Diabetes Mellitus/terapia , Humanos , Tecnologia
18.
JMIR Form Res ; 3(3): e13898, 2019 Jul 26.
Artigo em Inglês | MEDLINE | ID: mdl-31350840

RESUMO

BACKGROUND: Persistent cognitive impairment is prevalent in unipolar and bipolar disorders and is associated with decreased quality of life and psychosocial dysfunction. The screen for cognitive impairment in psychiatry (SCIP) test is a validated paper-and-pencil instrument for the assessment of cognition in affective disorders. However, there is no digital cognitive screening tool for the brief and accurate assessment of cognitive impairments in this patient group. OBJECTIVE: In this paper, we present the design process and feasibility study of the internet-based cognitive assessment tool (ICAT) that is designed based on the cognitive tasks of the SCIP. The aims of this feasibility study were to perform the following tasks among healthy individuals: (1) evaluate the usability of the ICAT, (2) investigate the feasibility of the ICAT as a patient-administered cognitive assessment tool, and (3) examine the performance of automatic speech recognition (ASR) for the assessment of verbal recall. METHODS: The ICAT was developed in a user-centered design process. The cognitive measures of the ICAT were immediate and delayed recall, working memory, and psychomotor speed. Usability and feasibility studies were conducted separately with 2 groups of healthy individuals (N=21 and N=19, respectively). ICAT tests were available in the English and Danish languages. The participants were asked to fill in the post study system usability questionnaire (PSSUQ) upon completing the ICAT test. Verbal recall in the ICAT was assessed using ASR, and the performance evaluation criterion was word error rate (WER). A Pearson 2-tailed correlation analysis significant at the .05 level was applied to investigate the association between the SCIP and ICAT scores. RESULTS: The overall psychometric factors of PSSUQ for both studies gave scores above 4 (out of 5). The analysis of the feasibility study revealed a moderate to strong correlation between the total scores of the SCIP and ICAT (r=0.63; P=.009). There were also moderate to strong correlations between the SCIP and ICAT subtests for immediate verbal recall (r=0.67; P=.002) and psychomotor speed (r=0.71; P=.001). The associations between the respective subtests for working memory, executive function, and delayed recall, however, were not statistically significant. The corresponding WER for English and Danish responses were 17.8% and 6.3%, respectively. CONCLUSIONS: The ICAT is the first digital screening instrument modified from the SCIP using Web-based technology and ASR. There was good accuracy of the ASR for verbal memory assessment. The moderate correlation between the ICAT and SCIP scores suggests that the ICAT is a valid tool for assessing cognition, although this should be confirmed in a larger study with greater statistical power. Taken together, the ICAT seems to be a valid Web-based cognitive assessment tool that, after some minor modifications and further validation, may be used to screen for cognitive impairment in clinical settings.

19.
J Am Med Inform Assoc ; 26(4): 324-338, 2019 04 01.
Artigo em Inglês | MEDLINE | ID: mdl-30753666

RESUMO

OBJECTIVE: The study sought to review recent literature regarding use of speech recognition (SR) technology for clinical documentation and to understand the impact of SR on document accuracy, provider efficiency, institutional cost, and more. MATERIALS AND METHODS: We searched 10 scientific and medical literature databases to find articles about clinician use of SR for documentation published between January 1, 1990, and October 15, 2018. We annotated included articles with their research topic(s), medical domain(s), and SR system(s) evaluated and analyzed the results. RESULTS: One hundred twenty-two articles were included. Forty-eight (39.3%) involved the radiology department exclusively and 10 (8.2%) involved emergency medicine; 10 (8.2%) mentioned multiple departments. Forty-eight (39.3%) articles studied productivity; 20 (16.4%) studied the effect of SR on documentation time, with mixed findings. Decreased turnaround time was reported in all 19 (15.6%) studies in which it was evaluated. Twenty-nine (23.8%) studies conducted error analyses, though various evaluation metrics were used. Reported percentage of documents with errors ranged from 4.8% to 71%; reported word error rates ranged from 7.4% to 38.7%. Seven (5.7%) studies assessed documentation-associated costs; 5 reported decreases and 2 reported increases. Many studies (44.3%) used products by Nuance Communications. Other vendors included IBM (9.0%) and Philips (6.6%); 7 (5.7%) used self-developed systems. CONCLUSION: Despite widespread use of SR for clinical documentation, research on this topic remains largely heterogeneous, often using different evaluation metrics with mixed findings. Further, that SR-assisted documentation has become increasingly common in clinical settings beyond radiology warrants further investigation of its use and effectiveness in these settings.


Assuntos
Documentação/métodos , Eficiência , Interface para o Reconhecimento da Fala , Pesquisa Biomédica , Documentação/economia , Registros Eletrônicos de Saúde , Humanos , Sistemas de Informação em Radiologia , Interface para o Reconhecimento da Fala/economia , Fatores de Tempo , Estudos de Tempo e Movimento
20.
J Craniomaxillofac Surg ; 46(12): 2022-2026, 2018 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-30420149

RESUMO

An automated cleft speech evaluator, available globally, has the potential to dramatically improve quality of life for children born with a cleft palate, as well as eliminating bias for outcome collaboration between cleft centers in the developed world. Our automated cleft speech evaluator interprets resonance and articulatory cleft speech errors to distinguish between normal speech, velopharyngeal dysfunction and articulatory speech errors. This article describes a significant update in the efficiency of our evaluator. Speech samples from our Craniofacial Team clinic were recorded and rated independently by two experienced speech pathologists: 60 patients were used to train the evaluator, and the evaluator was tested on the 13 subsequent patients. All sounds from 6 of the CAPS-A-AM sentences were used to train the system. The inter-speech pathologist agreement rate was 79%. Our cleft speech evaluator achieved 85% agreement with the combined speech pathologist rating, compared with 65% agreement using the previous training model. This automated cleft speech evaluator demonstrates good accuracy despite low training numbers. We anticipate that as the training samples increase, the accuracy will match human listeners.


Assuntos
Fissura Palatina/fisiopatologia , Inteligibilidade da Fala , Insuficiência Velofaríngea/fisiopatologia , Adolescente , Criança , Pré-Escolar , Feminino , Humanos , Masculino , Cadeias de Markov
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...