2024 | Reinhard Chun Wang Chau, Khaing Myat Thu, Ollie Yiru Yu, Richard Tai-Chiu Hsung, Edward Chin Man Lo, Walter Yu Hang Lam
This study investigates the performance of generative artificial intelligence (GenAI), specifically ChatGPT versions 3.5 and 4.0, in answering questions from dental licensing examinations. The research aims to assess the accuracy and proficiency of these models in providing dental knowledge, which is crucial for patient management and dental education.
**Methods:**
- **Data Selection:** 1461 multiple-choice questions from US and UK dental licensing examinations were selected.
- **Input into GenAI:** Questions were input into both versions of ChatGPT (3.5 and 4.0) by an independent assessor.
- **Performance Analysis:** The performance of the AI systems was compared to the passing rates of the US (75.0%) and UK (50.0%) dental examinations.
**Results:**
- **ChatGPT 3.5:** Correctly answered 68.3% of US questions and 43.3% of UK questions.
- **ChatGPT 4.0:** Correctly answered 80.7% of US questions and 62.7% of UK questions.
- **Passing Rates:** ChatGPT 4.0 passed both examinations, while ChatGPT 3.5 failed.
**Discussion:**
- **Implications:** The study highlights the potential of GenAI in supporting dental professionals and patients with accurate oral health information.
- **Challenges:** The results should be interpreted with caution due to the single model used and the limitations of multiple-choice questions.
- **Future Research:** Further research is needed to explore the long-term impact of GenAI on dentistry and to address implementation challenges.
**Conclusion:**
- **Proficiency:** GPT-4 demonstrated proficiency in passing dental licensing examinations and performed well in various dental subjects.
- **Implications:** The integration of GenAI in healthcare and dental education has significant implications, but further research is required to address challenges and ensure reliable and accurate responses.This study investigates the performance of generative artificial intelligence (GenAI), specifically ChatGPT versions 3.5 and 4.0, in answering questions from dental licensing examinations. The research aims to assess the accuracy and proficiency of these models in providing dental knowledge, which is crucial for patient management and dental education.
**Methods:**
- **Data Selection:** 1461 multiple-choice questions from US and UK dental licensing examinations were selected.
- **Input into GenAI:** Questions were input into both versions of ChatGPT (3.5 and 4.0) by an independent assessor.
- **Performance Analysis:** The performance of the AI systems was compared to the passing rates of the US (75.0%) and UK (50.0%) dental examinations.
**Results:**
- **ChatGPT 3.5:** Correctly answered 68.3% of US questions and 43.3% of UK questions.
- **ChatGPT 4.0:** Correctly answered 80.7% of US questions and 62.7% of UK questions.
- **Passing Rates:** ChatGPT 4.0 passed both examinations, while ChatGPT 3.5 failed.
**Discussion:**
- **Implications:** The study highlights the potential of GenAI in supporting dental professionals and patients with accurate oral health information.
- **Challenges:** The results should be interpreted with caution due to the single model used and the limitations of multiple-choice questions.
- **Future Research:** Further research is needed to explore the long-term impact of GenAI on dentistry and to address implementation challenges.
**Conclusion:**
- **Proficiency:** GPT-4 demonstrated proficiency in passing dental licensing examinations and performed well in various dental subjects.
- **Implications:** The integration of GenAI in healthcare and dental education has significant implications, but further research is required to address challenges and ensure reliable and accurate responses.