Metrics reloaded: recommendations for image analysis validation.

Abstract

Increasing evidence shows that flaws in machine learning (ML) algorithm validation are an underestimated global problem. In biomedical image analysis, chosen performance metrics often do not reflect the domain interest, and thus fail to adequately measure scientific progress and hinder translation of ML techniques into practice. To overcome this, we created Metrics Reloaded, a comprehensive framework guiding researchers in the problem-aware selection of metrics. Developed by a large international consortium in a multistage Delphi process, it is based on the novel concept of a problem fingerprint-a structured representation of the given problem that captures all aspects that are relevant for metric selection, from the domain interest to the properties of the target structure(s), dataset and algorithm output. On the basis of the problem fingerprint, users are guided through the process of choosing and applying appropriate validation metrics while being made aware of potential pitfalls. Metrics Reloaded targets image analysis problems that can be interpreted as classification tasks at image, object or pixel level, namely image-level classification, object detection, semantic segmentation and instance segmentation tasks. To improve the user experience, we implemented the framework in the Metrics Reloaded online tool. Following the convergence of ML methodology across application domains, Metrics Reloaded fosters the convergence of validation methodology. Its applicability is demonstrated for various biomedical use cases.

Department

Description

Provenance

Subjects

Algorithms, Semantics, Image Processing, Computer-Assisted, Machine Learning

Citation

Published Version (Please cite this version)

10.1038/s41592-023-02151-z

Publication Info

Maier-Hein, Lena, Annika Reinke, Patrick Godau, Minu D Tizabi, Florian Buettner, Evangelia Christodoulou, Ben Glocker, Fabian Isensee, et al. (2024). Metrics reloaded: recommendations for image analysis validation. Nature methods, 21(2). pp. 195–212. 10.1038/s41592-023-02151-z Retrieved from https://hdl.handle.net/10161/33085.

This is constructed from limited available data and may be imprecise. To cite this article, please review & use the official citation provided by the journal.

Scholars@Duke

Benis

Arriel Benis

Adjunct Associate Professor in the Department of Biomedical Engineering

Dr. Arriel Benis is a researcher and educator working at the intersection of medical informatics, digital health, and artificial intelligence, advancing health systems and biomedical engineering innovation. His work leverages AI, data science, and knowledge management to improve health-related decision-making at the individual, population, and public health levels.

His research focuses on developing data-driven healthcare solutions that enhance patient care, optimize clinical processes, and promote sustainable systems. Dr. Benis has engineered (a) clinical decision support systems with direct patient and healthcare partitioners impact such as ADHD, PTSD, and diabetes patient management and health communication, (b) MIMO -the Medical Informatics and Digital Health Multilingual Ontology- integrating more than 3500 terms and concepts across 30+ languages, actively deployed in healthcare organizations for AI-powered training and international projects support, (c) smart home and smart city health monitoring approach from a One Health viewpoint. Dr. Benis is a pioneer of the One Digital Health framework, which strategically links digital health innovation with environmental monitoring.

His past academic positions include serving as a department head and track director in biomedical and health informatics. He holds various leadership roles in the international medical informatics community, is a fellow of the International Academy for Health Sciences Informatics, and is the Editor-in-Chief of JMIR Medical Informatics. Dr. Benis is committed to training the next generation of innovators in digital health and medical informatics.


Unless otherwise indicated, scholarly articles published by Duke faculty members are made available here with a CC-BY-NC (Creative Commons Attribution Non-Commercial) license, as enabled by the Duke Open Access Policy. If you wish to use the materials in ways not already permitted under CC-BY-NC, please consult the copyright owner. Other materials are made available here through the author’s grant of a non-exclusive license to make their work openly accessible.