Appropriateness of Frequently Asked Patient Questions Following Total Hip Arthroplasty From ChatGPT Compared to Arthroplasty-Trained Nurses

Jeremy A. Dubin,Sandeep S. Bains, Michael J. DeRegoatis, Mallory C. Moore,Daniel Hameed,Michael A. Mont,James Nace,Ronald E. Delanois

The Journal of Arthroplasty(2024)

引用 0|浏览1
暂无评分
摘要
Introduction The use of ChatGPT (Generative Pretrained Transformer), which is a natural language artificial intelligence model, has gained unparalleled attention with the accumulation of over 100 million users within months of launching. As such, we aimed to compare: 1) orthopaedic surgeons’ evaluation of the appropriateness of the answers to the most frequently asked patient questions after total hip arthroplasty (THA); and 2) patients’ evaluation of ChatGPT and arthroplasty-trained nurses responses to answer their postoperative questions. Methods We prospectively created 60 questions to address the most commonly asked patient questions following THA. We obtained answers from arthroplasty-trained nurses and from the ChatGPT-3.5 version for each of the questions. Surgeons graded each set of responses based on clinical judgment as 1) “appropriate,” 2) “inappropriate” if the response contained inappropriate information, or 3) “unreliable” if the responses provided inconsistent content. Each patient was given a randomly selected question from the 60 aforementioned questions, with responses provided by ChatGPT and arthroplasty-trained nurses, using a Research Electronic Data Capture (REDCap) survey hosted at our local hospital. Results The three fellowship-trained surgeons graded 56 out of 60 (93.3%) responses for the arthroplasty-trained nurses and 57 out of 60 (95.0%) for ChatGPT to be “appropriate.” There were 175 out of 252 (69.4%) patients who were more comfortable following the ChatGPT responses and 77 out of 252 (30.6%) who preferred arthroplasty-trained nurses’ responses. However, 199 out of 252 patients (79.0%) responded that they were “uncertain” in regards to trusting AI to answer their postoperative questions. Conclusion ChatGPT provided appropriate answers from a physician perspective. Patients were also more comfortable with the ChatGPT responses than those from arthroplasty-trained nurses. Inevitably, its successful implementation is dependent on its ability to provide credible information that is consistent with the goals of the physician and patient alike.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要