Your browser doesn't support javascript.
loading
ChatGPT vs. sleep disorder specialist responses to common sleep queries: Ratings by experts and laypeople.
Kim, Jiyoung; Lee, Seo-Young; Kim, Jee Hyun; Shin, Dong-Hyeon; Oh, Eun Hye; Kim, Jin A; Cho, Jae Wook.
Affiliation
  • Kim J; Department of Neurology and Sleep Disorder Center, Bio Medical Research Institute, Pusan National University School of Medicine, Pusan National University Hospital, Busan, South Korea.
  • Lee SY; Department of Neurology, School of Medicine, Kangwon National University, Chuncheon, South Korea.
  • Kim JH; Department of Neurology, Ewha Womans University Seoul Hospital, Ewha Womans University College of Medicine, Seoul, South Korea.
  • Shin DH; Department of Neurology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan, South Korea.
  • Oh EH; Department of Neurology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan, South Korea.
  • Kim JA; Department of Neurology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan, South Korea.
  • Cho JW; Department of Neurology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan, South Korea; Sleep Tech Research Center, bitsensing Inc., Seongnam, South Korea. Electronic address: sleepcho@pusan.ac.kr.
Sleep Health ; 2024 Sep 21.
Article in En | MEDLINE | ID: mdl-39307579
ABSTRACT

BACKGROUND:

Many individuals use the Internet, including generative artificial intelligence like ChatGPT, for sleep-related information before consulting medical professionals. This study compared responses from sleep disorder specialists and ChatGPT to common sleep queries, with experts and laypersons evaluating the responses' accuracy and clarity.

METHODS:

We assessed responses from sleep medicine specialists and ChatGPT-4 to 140 sleep-related questions from the Korean Sleep Research Society's website. In a blinded study design, sleep disorder experts and laypersons rated the medical helpfulness, emotional supportiveness, and sentence comprehensibility of the responses on a 1-5 scale.

RESULTS:

Laypersons rated ChatGPT higher for medical helpfulness (3.79 ± 0.90 vs. 3.44 ± 0.99, p < .001), emotional supportiveness (3.48 ± 0.79 vs. 3.12 ± 0.98, p < .001), and sentence comprehensibility (4.24 ± 0.79 vs. 4.14 ± 0.96, p = .028). Experts also rated ChatGPT higher for emotional supportiveness (3.33 ± 0.62 vs. 3.01 ± 0.67, p < .001) but preferred specialists' responses for sentence comprehensibility (4.15 ± 0.74 vs. 3.94 ± 0.90, p < .001). When it comes to medical helpfulness, the experts rated the specialists' answers slightly higher than the laypersons did (3.70 ± 0.84 vs. 3.63 ± 0.87, p = .109). Experts slightly preferred specialist responses overall (56.0%), while laypersons favored ChatGPT (54.3%; p < .001). ChatGPT's responses were significantly longer (186.76 ± 39.04 vs. 113.16 ± 95.77 words, p < .001).

DISCUSSION:

Generative artificial intelligence like ChatGPT may help disseminate sleep-related medical information online. Laypersons appear to prefer ChatGPT's detailed, emotionally supportive responses over those from sleep disorder specialists.
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Language: En Journal: Sleep Health Year: 2024 Document type: Article Affiliation country: South Korea Country of publication: United States

Full text: 1 Collection: 01-internacional Database: MEDLINE Language: En Journal: Sleep Health Year: 2024 Document type: Article Affiliation country: South Korea Country of publication: United States