Surrogate Supervision for Robust and Generalizable Deformable Image Registration

Computer Vision & MultiModal AI
Published: arXiv: 2509.09869v1
Authors

Yihao Liu Junyu Chen Lianrui Zuo Shuwen Wei Brian D. Boyd Carmen Andreescu Olusola Ajilore Warren D. Taylor Aaron Carass Bennett A. Landman

Abstract

Objective: Deep learning-based deformable image registration has achieved strong accuracy, but remains sensitive to variations in input image characteristics such as artifacts, field-of-view mismatch, or modality difference. We aim to develop a general training paradigm that improves the robustness and generalizability of registration networks. Methods: We introduce surrogate supervision, which decouples the input domain from the supervision domain by applying estimated spatial transformations to surrogate images. This allows training on heterogeneous inputs while ensuring supervision is computed in domains where similarity is well defined. We evaluate the framework through three representative applications: artifact-robust brain MR registration, mask-agnostic lung CT registration, and multi-modal MR registration. Results: Across tasks, surrogate supervision demonstrated strong resilience to input variations including inhomogeneity field, inconsistent field-of-view, and modality differences, while maintaining high performance on well-curated data. Conclusions: Surrogate supervision provides a principled framework for training robust and generalizable deep learning-based registration models without increasing complexity. Significance: Surrogate supervision offers a practical pathway to more robust and generalizable medical image registration, enabling broader applicability in diverse biomedical imaging scenarios.

Paper Summary

Problem
Deformable image registration is a crucial task in medical image analysis that enables the alignment of anatomical structures across images and subjects. However, deep learning-based approaches to this task remain sensitive to variations in input image characteristics, such as artifacts, field-of-view mismatch, or modality difference. This limits their ability to generalize across datasets, scanners, and institutions.
Key Innovation
The researchers introduce a new training paradigm called surrogate supervision, which decouples the input domain from the supervision domain by applying estimated spatial transformations to surrogate images. This allows training on heterogeneous inputs while ensuring supervision is computed in domains where similarity is well defined. Surrogate supervision can be applied in a generalized framework that unifies and extends prior works.
Practical Impact
Surrogate supervision has the potential to improve the robustness and generalizability of deep learning-based deformable image registration models. This can lead to more accurate and reliable image registration results, even in the presence of artifacts, field-of-view mismatch, or modality difference. The approach can also be extended to other applications, such as hybrid imaging (e.g., nuclear medicine/CT) or image harmonization.
Analogy / Intuitive Explanation
Think of surrogate supervision as a way to "translate" the input images into a more familiar and consistent language, allowing the registration model to learn from them more effectively. Just as a translator can help communicate between two people who speak different languages, surrogate supervision can help the registration model understand the input images and align them accurately, even if they have different characteristics or modalities.
Paper Information
Categories:
cs.CV cs.AI
Published Date:

arXiv ID:

2509.09869v1

Quick Actions