Evaluation metrics in medical imaging AI: fundamentals, pitfalls, misapplications, and recommendations

Publication date

2026-03-04T12:03:38Z

2026-03-04T12:03:38Z

2025-09

2026-03-04T12:03:38Z



Abstract

Robust assessment of artificial intelligence (AI) models in medical imaging is paramount for reliable clinical integration. This international collaborative review paper provides an overview of key evaluation metrics across diverse tasks, including classification, regression, survival analysis, detection, and segmentation, as well as specialized metrics for calibration, foundation models, large language models, and synthetic images. Challenges of comparing models statistically and translating metric scores to clinical practice are also discussed. For each section, the paper outlines fundamental metrics, identifies common pitfalls and misapplications, and offers recommendations for more robust evaluations. Key recommendations often involve utilizing multiple, complementary metrics tailored to the specific task and dataset properties, transparent reporting of methodology, and critically, considering the clinical utility and real-world implications of model performance. Ultimately, effective evaluation requires a comprehensive, context-aware approach that goes beyond statistical metrics to ensure.

Document Type

Article


Published version

Language

English

Publisher

Elsevier B.V.

Related items

Reproducció del document publicat a: https://doi.org/10.1016/j.ejrai.2025.100030

European Journal of Radiology Artificial Intelligence, 2025, vol. 3, p. 100030

https://doi.org/10.1016/j.ejrai.2025.100030

Recommended citation

This citation was generated automatically.

Rights

cc-by (c) Burak Kocak et al., 2025

http://creativecommons.org/licenses/by/4.0/

This item appears in the following Collection(s)