CSVideoNet: A Real-time End-to-end Learning Framework for High-frame-rate Video Compressive Sensing

2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018)(2016)

引用 35|浏览19
暂无评分
摘要
This paper addresses the real-time encoding-decoding problem for high-frame-rate video compressive sensing (CS). Unlike prior works that perform reconstruction using iterative optimization-based approaches, we propose a non-iterative model, named "CSVideoNet". CSVideoNet directly learns the inverse mapping of CS and reconstructs the original input in a single forward propagation. To overcome the limitations of existing CS cameras, we propose a multi-rate CNN and a synthesizing RNN to improve the trade-off between compression ratio (CR) and spatial-temporal resolution of the reconstructed videos. The experiment results demonstrate that CSVideoNet significantly outperforms the state-of-the-art approaches. With no pre/post-processing, we achieve 25dB PSNR recovery quality at 100x CR, with a frame rate of 125 fps on a Titan X GPU. Due to the feedforward and high-data-concurrency natures of CSVideoNet, it can take advantage of GPU acceleration to achieve three orders of magnitude speed-up over conventional iterative-based approaches. We share the source code at https://github.com/PSCLab-ASU/CSVideoNet.
更多
查看译文
关键词
high-frame-rate video compressive sensing,real-time encoding-decoding problem,multirate CNN,compression ratio,frame rate,feedforward,video reconstruction,CSVideoNet,CS cameras,real-time end-to-end learning framework,forward propagation,synthesizing RNN,video spatial-temporal resolution,peak signal-to-noise ratio,PSNR recovery quality,Titan X GPU
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要