Research Article |
Corresponding author: Ali H. Salama ( asalama@meu.edu ) Academic editor: Valentina Petkova
© 2024 Ali H. Salama.
This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Citation:
Salama AH (2024) The promise and challenges of ChatGPT in community pharmacy: A comparative analysis of response accuracy. Pharmacia 71: 1-5. https://doi.org/10.3897/pharmacia.71.e116927
|
This study evaluates ChatGPT, an AI-based language model, in addressing common pharmacist inquiries in community pharmacies. The assessment encompasses Drug-Drug Interactions, Adverse Drug Effects, Drug Dosage, and Alternative Therapies, each comprising 20 questions, totaling 80 questions. Responses from ChatGPT were compared against standard answers, generating textual and chart scores. Textual score was computed by relating correct answers to the total questions within each category, while chart score involved the total correct answers multiplied by the chart-type questions. ChatGPT exhibited distinct performance rates: 30% for Drug-Drug Interactions, 65% for Adverse Drug Effects, 35% for Drug Dosage, and an impressive 85% for Alternative Therapies. While Alternative Therapies displayed high accuracy, challenges arose in accurately addressing Drug Dosage and Drug-Drug Interactions. Conclusion: The study underscores the complexity of pharmacy-related inquiries and the necessity for AI model enhancement. Despite promising accuracy in certain categories, like Alternative Therapies, improvements are crucial for Drug Dosage and Drug-Drug Interactions. The findings emphasize the need for ongoing AI model development to optimize integration into community pharmacy settings.
ChatGPT, community pharmacy, healthcare, effectiveness and technology
In recent years, the integration of artificial intelligence (AI) and natural language processing (NLP) technologies has revolutionized various sectors, including healthcare (
In this study, we aim to evaluate the efficacy of ChatGPT in providing accurate responses to a diverse range of inquiries commonly encountered by pharmacists in community pharmacy settings.
The inquiry into four distinct categories–Drug-Drug Interactions, Adverse Drug Effects, Drug Dosage, and Alternative Therapies–involved the manual input of questions into ChatGPT, as depicted in Fig.
Step 2 involved consolidating each question along with its corresponding response into a singular entry. Step 3 expanded the dataset by obtaining 20 questions from each category through collaboration with pharmacists in community pharmacies. This process helped establish a standardized question format.
To evaluate performance, the answers generated by ChatGPT were compared against standard responses, and scores were tallied. Textual scores were determined by dividing the total number of questions in each category by the score (ranging from 0 to 100 points) and then multiplying the result by the number of accurate responses. Chart scores were computed by multiplying the total count of correct answers by the quantity of chart-type questions within each category.
The mean score for each stage surpassed 60 points, meeting the stipulated passing threshold (a score of 60 points or higher) for both stages. The model test utilizing ChatGPT 3.5 was conducted from September 12 to 15, 2023.
The relevant data for this study were collected and analyzed as percentages by using Microsoft Excel (Microsoft, Redmond, WA, USA).
The correlation between perceived benefits and concerns was evaluated using Spearman’s rho correlation due to the data’s non-normal distribution.
In this study, we have expanded the scope by introducing four categories, encompassing a grand total of 80 questions, with each category comprising 20 multiple-choice questions, offering respondents a wider array of options to choose from (A, B, C, or D). Fig.
The results were shown in Table
Category | Number of questions | Number of correct questions | % of correct questions |
---|---|---|---|
Drug –Drug interactions | 20 | 6 | 30% |
Adverse drug effects | 20 | 13 | 65% |
Drug dosing | 20 | 7 | 35% |
Alternative therapy | 20 | 17 | 85% |
Table
The research paper presents a comprehensive assessment of ChatGPT’s effectiveness in providing accurate responses to diverse inquiries commonly encountered by pharmacists in community pharmacy settings. The study employed a well-structured methodology, categorizing questions into five distinct groups, each consisting of twenty questions, to evaluate ChatGPT’s performance. These categories encompassed fundamental aspects of pharmacy practice, including drug-drug interactions, adverse drug effects, drug dosage, and alternative therapies. The study’s results indicate varying performance levels of ChatGPT across these categories. Let us delve deeper into these findings and discuss their implications. Drug-Drug Interactions (Category A): Within this category, ChatGPT demonstrated a correct response rate of 30%. While this performance appears relatively low, it is essential to recognize the intricacy of drug-drug interactions and the challenges associated with providing precise responses (
In conclusion, this research paper provides valuable insights into the use of ChatGPT in a community pharmacy context. The study’s findings shed light on the AI model’s strengths and weaknesses in addressing different categories of inquiries. By understanding these results and their implications, researchers and healthcare professionals can work together to further refine AI models, ensuring they become reliable and valuable assets in delivering patient-centered care.
The author is grateful to the Middle East University (MEU), Amman, Jordan, for the financial support granted to cover the publication fee of this research article.