论文标题
学习端到端有损图像压缩:基准测试
Learning End-to-End Lossy Image Compression: A Benchmark
论文作者
论文摘要
图像压缩是图像和视频处理字段中最基本的技术和常用应用之一。较早的方法建立了精心设计的管道,并努力通过手工调整来改善管道的所有模块。后来,做出了巨大的贡献,尤其是当数据驱动的方法以出色的建模能力和灵活性在结合新设计的模块和约束时振兴域时。尽管取得了长足的进步,但缺乏对端到端学习图像压缩方法的系统基准和全面分析。在本文中,我们首先对学习的图像压缩方法进行了全面的文献调查。文献是基于几个方面组织的,以共同优化神经网络,即网络体系结构,熵模型和速率控制的速率延伸性能。我们描述了尖端学习的图像压缩方法的里程碑,回顾了广泛的现有作品,并提供了对其历史发展路线的见解。通过这项调查,揭示了图像压缩方法的主要挑战,以及最近高级学习方法解决相关问题的机会。该分析为迈向更高效率的图像压缩提供了进一步的一步。通过引入熵估计和信号重建的粗到五个高位模型,我们实现了改善的速率 - 延伸性能,尤其是在高分辨率图像上。广泛的基准实验证明了我们的模型在多核CPU和GPU上的速率延伸性能和时间复杂性方面的优势。我们的项目网站可在https://huzi96.github.io/compression-bench.html上找到。
Image compression is one of the most fundamental techniques and commonly used applications in the image and video processing field. Earlier methods built a well-designed pipeline, and efforts were made to improve all modules of the pipeline by handcrafted tuning. Later, tremendous contributions were made, especially when data-driven methods revitalized the domain with their excellent modeling capacities and flexibility in incorporating newly designed modules and constraints. Despite great progress, a systematic benchmark and comprehensive analysis of end-to-end learned image compression methods are lacking. In this paper, we first conduct a comprehensive literature survey of learned image compression methods. The literature is organized based on several aspects to jointly optimize the rate-distortion performance with a neural network, i.e., network architecture, entropy model and rate control. We describe milestones in cutting-edge learned image-compression methods, review a broad range of existing works, and provide insights into their historical development routes. With this survey, the main challenges of image compression methods are revealed, along with opportunities to address the related issues with recent advanced learning methods. This analysis provides an opportunity to take a further step towards higher-efficiency image compression. By introducing a coarse-to-fine hyperprior model for entropy estimation and signal reconstruction, we achieve improved rate-distortion performance, especially on high-resolution images. Extensive benchmark experiments demonstrate the superiority of our model in rate-distortion performance and time complexity on multi-core CPUs and GPUs. Our project website is available at https://huzi96.github.io/compression-bench.html.