Papers
Topics
Authors
Recent
Search
2000 character limit reached

I-INR: Iterative Implicit Neural Representations

Published 24 Apr 2025 in cs.CV | (2504.17364v2)

Abstract: Implicit Neural Representations (INRs) have revolutionized signal processing and computer vision by modeling signals as continuous, differentiable functions parameterized by neural networks. However, their inherent formulation as a regression problem makes them prone to regression to the mean, limiting their ability to capture fine details, retain high-frequency information, and handle noise effectively. To address these challenges, we propose Iterative Implicit Neural Representations (I-INRs) a novel plug-and-play framework that enhances signal reconstruction through an iterative refinement process. I-INRs effectively recover high-frequency details, improve robustness to noise, and achieve superior reconstruction quality. Our framework seamlessly integrates with existing INR architectures, delivering substantial performance gains across various tasks. Extensive experiments show that I-INRs outperform baseline methods, including WIRE, SIREN, and Gauss, in diverse computer vision applications such as image restoration, image denoising, and object occupancy prediction.

Summary

  • The paper proposes an iterative framework that incrementally refines implicit neural representations to capture high-frequency details and mitigate noise.
  • The method leverages a base INR and an auxiliary network in a stepwise reconstruction process, improving performance on image fitting, super-resolution, and denoising tasks.
  • Experiments demonstrate significant PSNR gains and enhanced fidelity, setting a new benchmark for robustness and detail preservation in neural signal reconstruction.

Iterative Implicit Neural Representations

Introduction

Implicit Neural Representations (INRs) have become integral to signal processing and computer vision by offering a framework where signals are modeled as continuous functions, realized through neural networks like MLPs. However, traditional INRs face challenges in capturing fine details and managing noise due to their predisposition towards low-frequency components. The "I-INR: Iterative Implicit Neural Representations" paper proposes the Iterative Implicit Neural Representations (I-INRs) framework, which aims to address these issues by refining reconstruction iteratively. Figure 1

Figure 1: Effectiveness of the proposed methods across multiple tasks, showcasing improvements in detail preservation and high-frequency reconstruction.

Proposed Method

I-INRs are designed to be an enhancement to existing INRs without requiring substantial modifications. The framework introduces an iterative mechanism where signals are reconstructed over multiple steps, effectively capturing high-frequency details and improving noise robustness. The proposed method involves a formulation that transitions the target signal from an initial state Z\mathcal{Z} towards the desired output progressively. Figure 2

Figure 2

Figure 2: (a) Proposed architecture of the I-INR model featuring a Base INR and an Auxiliary Network. (b) Iterative reconstruction process.

This iterative approach can be mathematically represented using a linear interpolation between the target signal and a known initial state, evolving as follows:

g(x)t=I(x)(1−t)+Zt,g(x)_t = \mathcal{I}(x)(1 - t) + \mathcal{Z}t,

where Z\mathcal{Z} can be sampled from a distribution, such as Gaussian noise.

Experiments and Results

Extensive experimentation demonstrates that I-INRs significantly outperform baseline models like SIREN, WIRE, and Gauss across tasks such as image fitting, super-resolution, and denoising. For image fitting, the iterative models enhanced fidelity and reduced reconstruction errors across test datasets. Figure 3

Figure 3: Improved image fitting results using iterative models with reported PSNR gains.

The super-resolution tasks show consistent improvements in perceptual quality metrics alongside fidelity, illustrating the capability of I-INRs to generalize beyond trained scales. Figure 4

Figure 4: Visual comparison of super-resolution results, highlighting fewer artifacts and better detail preservation with iterative methods.

Similarly, in denoising, I-INRs exhibit a marked reduction in artifacts compared to traditional methods. Figure 5

Figure 5: Denoising results where iterative approaches achieve superior detail preservation and artifact reduction.

The reconstruction quality for 3D occupancy further supports I-INR's superiority in capturing intricate spatial details. Figure 6

Figure 6: Iterative methods offer enhanced geometric detail in 3D occupancy reconstruction.

Ablation Studies

The ablation studies confirm the robustness of the I-INRs across different settings, including the choice of initial states and its effect on reconstruction quality. These experiments underscore the importance of iterations in maximizing PSNR and perceptual quality while balancing computational efficiency. Figure 7

Figure 7: Impact of reconstruction steps on super-resolution performance, showing improved perceptual quality with increased iteration steps.

Conclusion

In conclusion, I-INRs introduce a significant advancement in implicit neural signal representation by refining the reconstruction process iteratively. This framework surmounts traditional limitations in capturing high-frequency details and robustness to noise, setting a new standard for quality in computer vision applications. The iterative nature of I-INRs holds promise for further advancements in AI research, particularly in tasks requiring precision and resilience against noisy data environments.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 3 likes about this paper.