Recent research in medical image analysis with deep learning almost exclusively focuses on grid- or voxel-based data representations. We challenge this common choice by introducing MedFuncta, a modality-agnostic continuous data representation based on neural fields. We demonstrate how to scale neural fields from single instances to large datasets by exploiting redundancy in medical signals and by applying an efficient meta-learning approach with a context reduction scheme. We further address the spectral bias in commonly used SIREN activations, by introducing an \(\omega_0\)-schedule, improving reconstruction quality and convergence speed. We validate our proposed approach on a large variety of medical signals of different dimensions and modalities (1D: ECG; 2D: Chest X-ray, Retinal OCT, Fundus Camera, Dermatoscope, Colon Histopathology, Cell Microscopy; 3D: Brain MRI, Lung CT) and successfully demonstrate that we can solve relevant downstream tasks on these representations. We additionally release a large-scale dataset of > 550 k annotated neural fields to promote research in this direction.
It is a common choice to represent data on discretized grids, e.g., to represent an image as a grid of pixels. While this data representation is widely explored, it poorly scales with grid resolution and ignores the often continuous nature of the underlying signal. Recent research has shown that neural fields (NFs) provide an interesting, continuous alternative to represent different kinds of data modalities like sound, images, shapes, or 3D scenes, by treating data as a neural function that takes a spatial position (e.g., a pixel coordinate) as input and outputs the appropriate measurement (e.g., an image intensity value). This work investigates how to find meaningful functional representations of medical data, allowing relevant downstream tasks to be solved on this compressed, modality-agnostic representation rather than the original signal, mitigating the need to design modality-specific networks.
We argue that most signals, especially in medicine, contain large amounts of redundant information or structure that we can learn over an entire set of signals. We therefore define a neural network \(f_{\theta,\phi^{(i)}}:\mathbb{R}^C\rightarrow\mathbb{R}^D\) with shared network parameters \(\theta\) that represents this redundant information and additional signal specific parameters \(\phi^{(i)}\in\mathbb{R}^{P}\) that condition the base network to represent a specific signal \(s_i\). The proposed network can handle data of different dimensionalities (from 1D ECG data to 3D MRI), by simply changing input and/or output dimensions for the specific data type. The representation of a single datapoint, namely the signal-specific parameters \(\phi^{(i)}\), always remains a single 1D vector.
@article{friedrich2025medfuncta,
title={MedFuncta: Modality-Agnostic Representations Based on Efficient Neural Fields},
author={Friedrich, Paul and Bieder, Florentin and Cattin, Philippe C},
journal={arXiv preprint arXiv:2502.14401},
year={2025}
}
This work was financially supported by the Werner Siemens Foundation through the MIRACLE II project.