RAGCare-QA: A Benchmark Dataset for Evaluating Retrieval-Augmented Generation Pipelines in Theoretical Medical Knowledge
Journal
medRxiv
Date Issued
2025
Author(s)
Dobreva, Jovana
Karasmanakis, Ivana
Ivanisevic, Filip
Horvat, Tadej
Gams, Matjaz
Simjanoska Misheva, Monika
Abstract
The paper introduces RAGCare-QA, an extensive dataset of 420 theoretical medical knowledge questions for assessing Retrieval-Augmented Generation (RAG) pipelines in medical education and evaluation settings. The dataset includes one-choice-only questions from six medical specialties (Cardiology, Endocrinology, Gastroenterology, Family Medicine, Oncology, and Neurology) with three levels of complexity (Basic, Intermediate, and Advanced). Each question is accompanied by the best fit of RAG implementation complexity level, such as Basic RAG (315 questions, 75.0%), Multi-vector RAG (82 questions, 19.5%), and Graph-enhanced RAG (23 questions, 5.5%). The questions emphasize theoretical medical knowledge on fundamental concepts, pathophysiology, diagnostic criteria, and treatment principles important in medical education. The dataset is a useful tool for the assessment of RAG- based medical education systems, allowing researchers to fine-tune retrieval methods for various categories of theoretical medical knowledge questions.
Subjects
File(s)![Thumbnail Image]()
Loading...
Name
2025.08.15.25333718.full.pdf
Size
3.01 MB
Format
Adobe PDF
Checksum
(MD5):40a8e75200720dac8ab4a28545c677ec
