tttLRM pushes long-context 3D reconstruction
tttLRM is a CVPR 2026 paper and project from researchers at UPenn, Adobe Research, and UCI that uses a test-time training layer to handle long-context autoregressive 3D reconstruction with linear complexity. The system compresses many posed images into fast weights, decodes them into explicit 3D formats like Gaussian splats, and supports progressive reconstruction from streaming observations.
This is a meaningful systems paper, not just another prettier 3D demo — the real contribution is making longer visual sequences computationally tractable for reconstruction. If the approach holds up broadly, it points toward 3D models that can keep absorbing views instead of choking on context length.
- –The paper explicitly targets the long-context bottleneck in LRM-style 3D reconstruction, which is one of the main reasons these models struggle to scale beyond short image sets
- –Its online-learning variant is notable because it supports progressive refinement from streaming observations, making it more relevant to robotics, embodied AI, and capture pipelines than static one-shot reconstruction papers
- –The project page and GitHub repo show code, checkpoints, and inference scripts, which gives the work a better shot at adoption than paper-only CVPR releases
- –The repo says the implementation builds on LongLRM and LaCT, so the project looks like a substantive extension of existing long-sequence reconstruction ideas rather than a totally isolated architecture
- –For AI developers, the practical caveat is that this is still research infrastructure: heavy GPU requirements, custom data formatting, and no released training code yet limit immediate production use
DISCOVERED
36d ago
2026-03-06
PUBLISHED
36d ago
2026-03-06
RELEVANCE
AUTHOR
AI Search