Beyond Uniformity: Regularizing Implicit Neural Representations through a Lipschitz Lens

ICLR 2026

*Equal contribution Equal senior contribution
1Technical University of Munich 2MCML 3University of Zurich
4ETH AI Centre 5Hasso-Plattner-Institute 6University of Basel 7Imperial College London

Bridging the gap between theory and practice in Lipschitz regularization. While Lipschitz continuity offers a principled form of implicit regularization, selecting and distributing the budget K remains non-trivial. We propose a novel, data-driven approach to derive K from interpretable, signal- or domain-specific properties, such as tissue compressibility in deformable registration, and strategically distribute this budget across layers. This methodology allows to better balance smoothness and expressiveness compared to uniform allocation strategies.

Abstract

Implicit Neural Representations (INRs) have shown great promise in solving inverse problems, but their lack of inherent regularization often leads to a trade-off between expressiveness and smoothness. While Lipschitz continuity presents a principled form of implicit regularization, it is often applied as a rigid, uniform 1-Lipschitz constraint, limiting its potential in inverse problems.

In this work, we reframe Lipschitz regularization as a flexible Lipschitz budget framework. We propose a method to first derive a principled, task-specific total budget K, then proceed to distribute this budget non-uniformly across all network components, including linear weights, activations, and embeddings.

Across extensive experiments on deformable registration and image inpainting, we show that non-uniform allocation strategies provide a measure to balance regularization and expressiveness within the specified global budget. Our Lipschitz lens introduces an alternative, interpretable perspective to Neural Tangent Kernel (NTK) and Fourier analysis frameworks in INRs, offering practitioners actionable principles for improving network architecture and performance.

Method Overview

Layer-wise Lipschitz Composition

For a network \(f_\theta\) with L layers, the overall Lipschitz constant K is bounded by the product of individual layer and activation Lipschitz constants:

$$ K = \text{Lip}(f_\theta) \leq \prod_{i=1}^{L} \text{Lip}(\phi_i) \cdot \text{Lip}(\mathbf{W}_i) $$

This compositional property reveals that a global budget K can be achieved through various combinations of layer-wise constants, motivating our exploration of different budget allocation strategies.

Budget Allocation Strategies

Given the total Lipschitz budget \(K_B\), we investigate multiple allocation strategies:

  • (A) Uniform: Each component receives equal contribution: \(K_i = \sqrt[M]{K_B}\)
  • (B) All-first: Allocate \(K_1 = K_B\) and \(K_i = 1\) for remaining layers
  • (C) Linear: Monotonically increasing allocation from first to last layer
  • (D) Exponential: Front-heavy ramp in log space
  • (E) Cosine-annealed: Smooth transition using cosine schedule

Results: 1-Lipschitz Signed Distance Fields

We investigate how different spectral normalization techniques and gradient-preserving activation functions affect SDF learning quality. Our key finding: approaching the upper Lipschitz bound correlates with better perceptual quality.

Learning the Stanford bunny with different spectral normalization techniques (first row) and different gradient-preserving activation functions with Bjorck normalized layers (second row). We report Chamfer Distance (CD↓) and empirically estimated Lipschitz constant \(K_m\). Networks that approach the 1-Lipschitz budget more closely achieve better reconstruction quality.

Budget allocation experiments for 1-Lipschitz SDFs. Training with Householder (HH) and GroupSort (MaxMin) activations using Bjorck and SLL normalization shows that non-uniform allocation strategies perform on par with uniform approaches—suggesting the benefits may be stifled by the overly restrictive unit budget.

Results: Deformable Image Registration

For lung CT registration, we derive the Lipschitz budget from clinical evidence: strain approaching 2.0 marks a threshold for tissue failure. This motivates our choice of \(K = 2\) as a principled, domain-driven budget.

Comparison of smoothness (Folding Ratio↓) and expressiveness (TRE↓) across Lipschitz-regularized INR architectures. Non-uniform allocations (e.g., exponential) can improve target registration error while maintaining comparable folding ratio to uniform allocation.

Results: Image Inpainting

For image inpainting, we derive the Lipschitz budget using a data-driven oracle based on the maximum spectral norm of the image Jacobian. This provides an interpretable upper bound on expected signal variation.

Quantitative results for different allocation strategies in inpainting with FFNs. Performance peaks near the oracle estimate and degrades when the Lipschitz budget deviates, demonstrating the oracle provides a meaningful approximation.

Qualitative examples from CelebA inpainting. Non-uniform budget allocation strategies (First, Linear, Exponential) yield statistically significant improvements over uniform allocation, particularly near the oracle estimate (Distance=0).

A Novel Perspective: Weight Scaling through a Lipschitz Lens

Recent work shows that scaling SIREN weights by a factor \(\alpha\) improves accuracy and convergence. While existing explanations rely on NTK theory, we show that Lipschitz theory offers a complementary perspective: for a SIREN layer \(f(x) = \sin(\omega(\alpha \mathbf{W}x + b))\), the induced Lipschitz constant is \(\text{Lip}(f) = |\omega\alpha\|\mathbf{W}\||\), scaling linearly with \(\alpha\).

Visualization of the upper induced Lipschitz bound for weight scaling in SIREN. Scaling the initialization leads to direct scaling of layer Lipschitz bounds, allowing the network to increase its capacity for high-frequency content. Layer-wise analysis shows self-regulating capacity when weights are sufficiently scaled.

Practical Guidelines

Estimating Budget K

  • Domain-driven: Use physical constraints (e.g., max tissue strain ≈2.0 for lung registration, max cardiac contraction) or intensity bounds as interpretable upper bounds.
  • Data-driven: Compute an oracle from representative high-resolution samples based on the maximum spectral norm of the image Jacobian. See the paper for the full derivation.
  • Signal-theoretic: Use known bandlimits or sampling rates (e.g., 44.1 kHz for audio, ≈150 Hz for ECG) as conservative, noise-suppressing estimates.
Tip: For noisy images, use percentile-based estimates (99.9%) rather than the absolute maximum to improve robustness.

Allocating Budget K

  • Treat allocation as a hyperparameter search centered on network expressivity
  • Analyze performance with respect to \(K_{min}\), the minimum imposed Lipschitz bound
  • Non-uniform strategies enable balancing regularization with expressiveness under the same global budget

Key Insights

  • For 1-Lipschitz SDFs, non-uniform allocation performs on par with uniform—benefits emerge in general K-Lipschitz settings where the budget itself is task-dependent.
  • Non-uniform strategies (e.g., exponential) can improve TRE in registration while maintaining comparable folding ratio.
  • Performance peaks near the oracle estimate and degrades when the budget deviates significantly.
  • Different architectures exhibit varying degrees of self-regulation: SIREN shows superior self-regulation compared to FFNs.

BibTeX

      
@inproceedings{mcginnis2026beyond,
    title={Beyond Uniformity: Regularizing Implicit Neural Representations through a Lipschitz Lens},
    author={Julian McGinnis and Suprosanna Shit and Florian A. H{\"o}lzl and Paul Friedrich and Paul B{\"u}schl and Vasiliki Sideri-Lampretsa and Mark M{\"u}hlau and Philippe C. Cattin and Bj{\"o}rn Menze and Daniel Rueckert and Benedikt Wiestler},
    booktitle={International Conference on Learning Representations},
    year={2026},
    url={https://openreview.net/forum?id=XXXXX}
}
      

Acknowledgements

We thank Patricia Pauli for helpful discussions. This work is funded by the Munich Center for Machine Learning. Julian McGinnis and Mark Mühlau are supported by Bavarian State Ministry for Science and Art (Collaborative Bilateral Research Program Bavaria – Quebec: AI in medicine, grant F.4-V0134.K5.1/86/34). Suprosanna Shit is supported by the UZH Postdoc Grant (K-74851-03-01). Suprosanna Shit and Björn Menze acknowledge support by the Helmut Horten Foundation.