Improving Learned Video Compression by Exploring Spatial Redundancy

ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2024)

引用 0|浏览0
暂无评分
摘要
Learned video compression has developed rapidly and shown promising rate-distortion performance recently. Existing works have made great progress on removing temporal redundancy between inter-frames, while neglecting spatial redundancy within a frame. In this paper, we propose to explore spatial redundancy to further improve compression efficiency and reduce computation complexity by extending inter-frame prediction from temporal prediction to spatial-temporal prediction. Specifically, the current frame is first down-sampled to reduce spatial redundancy, which is then temporally predicted to compensate motions and spatially predicted to restore resolutions. The spatial prediction module is designed in a reference-based manner, which improves prediction performance without additional bit rate cost by mining high-frequency details from a neighboring high-resolution decoded frame. Experimental results demonstrate that our proposed module can improve compression efficiency and reduce encoding/decoding time. Our framework achieves competitive compression efficiency compared with recent learned methods and traditional codecs.
更多
查看译文
关键词
Learned video compression,spatial prediction
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要