The Surprising Effectiveness of Noise Pretraining for Implicit Neural Representations

CVPR, 2026

Coming soon!

Abstract: The approximation and convergence properties of implicit neural representations (INRs) are known to be highly sensitive to parameter initialization strategies. Several data-driven INR parameter initialization methods demonstrate significant improvement over standard random sampling, but the reason for their successes – whether they encode classical statistical signal priors or something more sophisticated – is not well understood. In this study, we explore this topic with a series of experimental analyses leveraging noise pretraining. In particular, we pretrain INRs on noise signals of different classes (e.g., Gaussian, Dead Leaves, Spectral), and measure their abilities at both fitting unseen signals and encoding priors for an inverse imaging task (denoising). Our analyses on image and video data reveal the highly surprising finding that simply pretraining on unstructured noise (Uniform, Gaussian) results in a dramatic improvement in signal fitting capacity compared to all other baselines. However, unstructured noise also yields poor deep image priors for denoising. In contrast, noise with the classic 1/f spectral structure of natural images yields an excellent balance of both signal fitting and inverse imaging capabilities on par with the best data-driven initialization methods. This finding can enable more efficient training of INRs in applications without sufficient prior domain-specific data.



noisepretraining cover fig
Overview of Noise Pretraining for INRs.

In this study, we explore the effectiveness of initializing INR parameters with signals from noise classes with different properties.: (a) Samples of a few of the signal types we consider in this study: natural images, noise with the classic ``\(1/|f^\alpha|\)’’ spectral signature of natural images, Dead Leaves noise, and Uniform noise. (b) We propose SNP, a simple but powerful data-driven INR parameter initialization method using noise samples leveraging the STRAINER initialization framework. We pretrain an INR encoder with signal-specific decoder heads on noise samples from one class. (c) At test time, the pretrained encoder may be fine-tuned with a randomized decoder on a new (real) signal. (d) Surprising test-time signal fitting results on Celeb-A-HQ: SNP initialization using Uniform noise performs dramatically better on signal fitting compared to all other methods.

On high-frequency reconstruction abilities of noise pretrained INRs

highfreq snp
STRAINER, SNP (Spectrum and Uniform) learn high quality reconstructions in $200$ steps (PSNR$\approx40db+$), while the SIREN output remains blurry. Power spectrum insets show that: SNP (Uniform) matches closely with the ground truth, STRAINER has a noisy high-frequency coverage, and SNP (Spectrum) recovers most high frequencies while SIREN lacks critical frequencies.

We find that pretraining INRs on uniform noise yields to quickest convergence. Power spectrum insets show reconstructed images have near perfect frequency spectrum. INRs pretraied on Spectrum noise also depict similar results. Siren , on the other hand, suffers from spectral bias and converges to lower frequencies while taking longer to fit to higher frequencies.

Implicit Prior v/s signal fitting ability

prior trade off
Inverse trade off between an implicit prior v/s fitting capacity as a function of noise-based initialization

Tradeoff in signal representation and denoising ability of SNP INRs on Celeb-A-HQ images. At early stages of signal fitting (\(T=500\)) and denoising (\(T=50\)), SNP (Spectrum) and STRAINER perform well. In contrast, we find that SNP (Uniform,Gaussian) excel at signal fitting, but are poor denoisers. SIREN is a good denoiser but a poor signal fitter. Hence, there is a tradeoff to consider when choosing an appropriate initialization method for a task.

Noise Pretraining, INR’s internal geometry, and connection to Hash-encoding type approaches!

splines and geometry
Internal geometry of INR for pretraining on uniform noise reveals complex random encoding of input space whereas data-driven or spectrum-based pretraining showcases more priors

In our study, we show that INRs pretrained on unstructured noise such as Uniform and Gaussian noise causes initial layers of the INR to behave as complex random encoding functions. We observe similar parallels in Instant-NGP or hash-encoding based models - both, demonstrate significantly increased fitting performance. We also show that by pretraining on structured spectrum noise / in-domain face images/ ImageNet natural images, the INRs exhibit strong spatial and frequency guided internal geometry useful for inverse problems.

Noise pretraining for Video INRs

splines and geometry
Noise pretraining for video INRs

We also demonstrate that noise pretraining is effective for video INRs. We use ResFields[29], a popular video INR method. We show frames from representation of two example videos with randomly initialized (vanilla) ResFields~\cite{resfield}, and SNP-Spectrum-initialized Resfields. In both examples, SNP-initialization results in a higher accuracy with clearer visible details.

For more details, please refer full paper!


Citation

            @inproceedings{vyas2026cvpr_snp,
                author={Vyas, Kushal
                    and Kayabasi, Alper
                    and Kim, Daniel
                    and Saragadam, Vishwanath,
                    and Veeraraghavan, Ashok
                    and Balakrishnan, Guha},
                title={The Surprising Effectiveness of Noise Pretraining for Implicit Neural Representations},
                booktitle={Computer Vision and Pattern Recognition (CVPR)},
                year={2026},
            }