RxSafeBench.zip is a benchmark dataset designed to evaluate the medication safety capabilities of large language models (LLMs) in simulated clinical consultation scenarios. The final benchmark consists of 2,443 high-quality consultation scenarios, evenly split across different types of medication risks. This dataset is essential for advancing research on LLM-based systems in healthcare, particularly in improving medication safety.