TeCoNeRV Model Checkpoints

TeCoNeRV uses hypernetworks to predict implicit neural representation (INR) weights for video compression. A patch-tubelet decomposition enables hypernetworks to scale to high-resolution video prediction. The temporal coherence objective reduces redundancy across consecutive clips, enabling compact residual encoding of per-clip parameters.

This repository contains hypernetwork training checkpoints for the three model families described in the paper.

Model families

nervenc — Baseline NeRVEnc hypernetwork. Predicts full-resolution clip reconstructions directly.

patch_tubelet — Proposed patch-tubelet hypernetwork. Predicts parameters for spatial tubelets; full frames are reconstructed by tiling. Supports resolution-independent inference.

teconerv — Proposed method. Initialized from a patch_tubelet checkpoint and finetuned with a temporal coherence objective.

Getting started

See the GitHub repository for full documentation on setup, training, and evaluation. Checkpoint download instructions are in docs/models.md.

git lfs install
git clone https://huggingface.co/namithap/teconerv-models

Citation

@article{padmanabhan2026teconerv,
  title={TeCoNeRV: Leveraging Temporal Coherence for Compressible Neural Representations for Videos},
  author={Padmanabhan, Namitha and Gwilliam, Matthew and Shrivastava, Abhinav},
  journal={arXiv preprint arXiv:2602.16711},
  year={2026}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Paper for namithap/teconerv-models