Interpretable Similarity of Synthetic Image Utility

AI in healthcare
Published: arXiv: 2512.17080v1
Authors

Panagiota Gatoula George Dimas Dimitris K. Iakovidis

Abstract

Synthetic medical image data can unlock the potential of deep learning (DL)-based clinical decision support (CDS) systems through the creation of large scale, privacy-preserving, training sets. Despite the significant progress in this field, there is still a largely unanswered research question: "How can we quantitatively assess the similarity of a synthetically generated set of images with a set of real images in a given application domain?". Today, answers to this question are mainly provided via user evaluation studies, inception-based measures, and the classification performance achieved on synthetic images. This paper proposes a novel measure to assess the similarity between synthetically generated and real sets of images, in terms of their utility for the development of DL-based CDS systems. Inspired by generalized neural additive models, and unlike inception-based measures, the proposed measure is interpretable (Interpretable Utility Similarity, IUS), explaining why a synthetic dataset could be more useful than another one in the context of a CDS system based on clinically relevant image features. The experimental results on publicly available datasets from various color medical imaging modalities including endoscopic, dermoscopic and fundus imaging, indicate that selecting synthetic images of high utility similarity using IUS can result in relative improvements of up to 54.6% in terms of classification performance. The generality of IUS for synthetic data assessment is demonstrated also for greyscale X-ray and ultrasound imaging modalities. IUS implementation is available at https://github.com/innoisys/ius

Paper Summary

Problem
The main problem this paper addresses is the difficulty in assessing the similarity between synthetic medical images and real images in terms of their utility for developing Deep Learning (DL)-based Clinical Decision Support (CDS) systems. Currently, answers to this question are mainly provided through user evaluation studies, inception-based measures, and classification performance achieved on synthetic images, which are time-consuming, costly, and subjective.
Key Innovation
This paper proposes a novel measure called Interpretable Utility Similarity (IUS) to assess the similarity between synthetically generated and real sets of images in terms of their utility for DL-based CDS systems. IUS is inspired by generalized neural additive models and is interpretable, explaining why a synthetic dataset could be more useful than another one in the context of a CDS system based on clinically relevant image features.
Practical Impact
The proposed IUS measure has the potential to significantly improve the development of DL-based CDS systems by allowing researchers to select synthetic images that are more useful and relevant for their specific application. This can lead to better classification performance, improved accuracy, and more reliable clinical decision-making. The IUS implementation is available at https://github.com/innoisys/ius1, making it accessible to researchers and developers.
Analogy / Intuitive Explanation
Think of IUS like a quality control system in a factory. Just as a quality control system checks the quality of products before they are released to the market, IUS checks the quality of synthetic medical images before they are used to train DL models. But instead of just checking for defects, IUS assesses the utility of the synthetic images in terms of their relevance to the specific clinical application. This ensures that only the most useful and relevant synthetic images are used to develop DL-based CDS systems, leading to better clinical outcomes.
Paper Information
Categories:
cs.CV
Published Date:

arXiv ID:

2512.17080v1

Quick Actions