🐢 Learning Truncated Causal History Model for Video Restoration

1ECE Department, University of Alberta, Canada
2Huawei Technologies, Canada
Neural Information Processing Systems (NeurIPS), 2024

*indicates equal contribution

TL;DR - Hear the Paper Summary 🎧

Turtle in Action 🔥

Abstract

One key challenge to video restoration is to model the transition dynamics of video frames governed by motion. In this work, we propose Turtle to learn the truncated causal history model for efficient and high-performing video restoration. Unlike traditional methods that process a range of contextual frames in parallel, Turtle enhances efficiency by storing and summarizing a truncated history of the input frame latent representation into an evolving historical state. This is achieved through a sophisticated similarity-based retrieval mechanism that implicitly accounts for inter-frame motion and alignment. The causal design in Turtle enables recurrence in inference through state-memorized historical features while allowing parallel training by sampling truncated video clips. We report new state-of-the-art results on a multitude of video restoration benchmark tasks, including video desnowing, nighttime video deraining, video raindrops and rain streak removal, video super-resolution, real-world and synthetic video deblurring, and blind video denoising while reducing the computational cost compared to existing best contextual methods on all these tasks.

Turtle (🐢) is a U-Net style architecture, wherein the encoder blocks are historyless feedforward blocks, while the decoder couples the causal history model (CHM) to condition the restoration procedure on truncated history of the input.

Causal History Model (CHM) 💕

The powerhouse of Turtle (🐢)

🔥 Results on Video Restoration Tasks 🔥

BibTeX

BibTex Code Here