Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

AI-Driven Health Advice: Evaluating the Potential of Large Language Models as Health Assistants

View through CrossRef
This study aims to evaluate whether the GPT model can be a health assistant by addressing health concerns from three aspects: providing preliminary guidance, clarifying information, and offering accessible recommendations. 31 questions in total were collected from multiple online health platforms, which included diverse health concerns across different age ranges and genders. A tailored system prompt was built to guide GPT model GPT-3.5-turbo generating responses. The evaluation metrics are designed based on 3 metrics: “Preliminary Guidance”, “Clarifying Information”, and “Accessibility and Convenience”, which is used to evaluate responses with score method from 0 to 5. Lastly, the generated responses were evaluated using established metrics by an experienced medical doctor with over 20 years of experience in the fields of general and preventive care. The results indicate that LLMs demonstrated moderate performance in both the ‘preliminary guidance’ and ‘clarifying information’ aspects. Specifically, the mean score for ‘preliminary guidance’ was 3.65, implying that LLMs are capable of offering valuable insights when symptoms indicate the need for urgent or emergency care, as well as providing reassurance to patients for minor symptoms. In a similar manner, the mean score for ‘clarifying information’ was 3.87, demonstrating that LLMs effectively provide supplementary information to aid patients in making informed decisions. However, the mean score for ‘accessibility and convenience’ was notably lower at 2.65, highlighting a deficiency in LLMs’ ability to offer advice customized to the specific needs of individual patients.
Title: AI-Driven Health Advice: Evaluating the Potential of Large Language Models as Health Assistants
Description:
This study aims to evaluate whether the GPT model can be a health assistant by addressing health concerns from three aspects: providing preliminary guidance, clarifying information, and offering accessible recommendations.
31 questions in total were collected from multiple online health platforms, which included diverse health concerns across different age ranges and genders.
A tailored system prompt was built to guide GPT model GPT-3.
5-turbo generating responses.
The evaluation metrics are designed based on 3 metrics: “Preliminary Guidance”, “Clarifying Information”, and “Accessibility and Convenience”, which is used to evaluate responses with score method from 0 to 5.
Lastly, the generated responses were evaluated using established metrics by an experienced medical doctor with over 20 years of experience in the fields of general and preventive care.
The results indicate that LLMs demonstrated moderate performance in both the ‘preliminary guidance’ and ‘clarifying information’ aspects.
Specifically, the mean score for ‘preliminary guidance’ was 3.
65, implying that LLMs are capable of offering valuable insights when symptoms indicate the need for urgent or emergency care, as well as providing reassurance to patients for minor symptoms.
In a similar manner, the mean score for ‘clarifying information’ was 3.
87, demonstrating that LLMs effectively provide supplementary information to aid patients in making informed decisions.
However, the mean score for ‘accessibility and convenience’ was notably lower at 2.
65, highlighting a deficiency in LLMs’ ability to offer advice customized to the specific needs of individual patients.

Related Results

Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
<p><em><span style="font-size: 11.0pt; font-family: 'Times New Roman',serif; mso-fareast-font-family: 'Times New Roman'; mso-ansi-language: EN-US; mso-fareast-langua...
Investigating user perceptions of commercial virtual assistants: A qualitative study
Investigating user perceptions of commercial virtual assistants: A qualitative study
As commercial virtual assistants become an integrated part of almost every smart device that we use on a daily basis, including but not limited to smartphones, speakers, personal c...
A Wideband mm-Wave Printed Dipole Antenna for 5G Applications
A Wideband mm-Wave Printed Dipole Antenna for 5G Applications
<span lang="EN-MY">In this paper, a wideband millimeter-wave (mm-Wave) printed dipole antenna is proposed to be used for fifth generation (5G) communications. The single elem...
Barriers and facilitators for providing self-care advice in community pharmacies: a qualitative study
Barriers and facilitators for providing self-care advice in community pharmacies: a qualitative study
Abstract Background Community pharmacies are easily accessible for self-care advice. Guidelines for providing self-care advice were introduced in se...
Comparative Methods for Building Chatbots: Open Source, Hybrid, and Fully Integrated Large Language Models
Comparative Methods for Building Chatbots: Open Source, Hybrid, and Fully Integrated Large Language Models
In the complex and dynamic realm of biodiversity informatics, the accessibility and comprehension of standards and vocabularies are pivotal for, but not limited to, effective data ...
Linguistic Influences on the Outcomes of Imposed Advice
Linguistic Influences on the Outcomes of Imposed Advice
This study examines how language used by advice recipients and whether the peer apologized before offering advice affected advice outcomes. Participants wrote about a problem and t...
Barriers and Facilitators for Evidence-Based Self-Care Counselling in Community Pharmacy, a qualitative study
Barriers and Facilitators for Evidence-Based Self-Care Counselling in Community Pharmacy, a qualitative study
Abstract Background Community pharmacy is easily accessible for self-care advice, supported by self-care guidelines. Such guidelines were introduced in several countries, ...

Back to Top