A slimmable framework for practical neural video compression

Zhaocheng Liu, Fei Yang, Defa Wang, Marc Górriz Blanch, Luka Murn, Shuai Wan, Saiping Zhang, Marta Mrak, Luis Herranz

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Deep learning is being increasingly applied to image and video compression in a new paradigm known as neural video compression. While achieving impressive rate–distortion (RD) performance, neural video codecs (NVC) require heavy neural networks, which in turn have large memory and computational costs and often lack important functionalities such as variable rate. These are significant limitations to their practical application. Addressing these problems, recent slimmable image codecs can dynamically adjust their model capacity to elegantly reduce the memory and computation requirements, without harming RD performance. However, the extension to video is not straightforward due to the non-trivial interplay with complex motion estimation and compensation modules in most NVC architectures. In this paper we propose the slimmable video codec framework (SlimVC) that integrates an slimmable autoencoder and a motion-free conditional entropy model. We show that the slimming mechanism is also applicable to the more complex case of video architectures, providing SlimVC with simultaneous control of the computational cost, memory and rate, which are all important requirements in practice. We further provide detailed experimental analysis, and describe application scenarios that can benefit from slimmable video codecs.

Original languageEnglish
Article number128525
JournalNeurocomputing
Volume610
DOIs
StatePublished - 28 Dec 2024

Keywords

  • Deep learning
  • Feature modulation
  • Neural video compression
  • Slimmable codec
  • Slimmable network
  • Variable rate

Fingerprint

Dive into the research topics of 'A slimmable framework for practical neural video compression'. Together they form a unique fingerprint.

Cite this