Abstract
This work proposes to augment the lifting steps of the conventional wavelet transform with additional neural network assisted lifting steps. These additional steps reduce residual redundancy (notably aliasing information) amongst the wavelet subbands, and also improve the visual quality of reconstructed images at reduced resolutions. The proposed approach involves two steps, a high-to-low step followed by a low-to-high step. The high-to-low step suppresses aliasing in the low-pass band by using the detail bands at the same resolution, while the low-to-high step aims to further remove redundancy from detail bands, so as to achieve higher energy compaction. The proposed two lifting steps are trained in an end-to-end fashion; we employ a backward annealing approach to overcome the non-differentiability of the quantization and cost functions during back-propagation. Importantly, the networks employed in this paper are compact and with limited non-linearities, allowing a fully scalable system; one pair of trained network parameters are applied for all levels of decomposition and for all bit-rates of interest. By employing the proposed approach within the JPEG 2000 image coding standard, our method can achieve up to 17.4% average BD bit-rate saving over a wide range of bit-rates, while retaining quality and resolution scalability features of JPEG 2000.
Abstract (translated)
本文提出了一种通过添加神经网络辅助提升传统波浪变换的 lifting 步骤来增强其提升步数的方法。这些额外的步骤减少了波浪子带之间的残余冗余(显著是 aliasing 信息),并且还改善了在低分辨率下重构图像的视觉效果。所提出的方法包括两个步骤:从高到低的步骤和从低到高的步骤。从高到低的步骤通过在同一分辨率下使用详细波浪带来抑制低通带中的 aliasing,而从低到高的步骤旨在进一步消除详细波浪带中的冗余,以实现更高的能量压缩。与传统的提升步骤相比,本文提出的两个提升步骤在端到端的方式下进行训练;我们采用反向退化方法来克服在反向传播过程中量化和非线性函数的不可导性。重要的是,本文使用的网络具有紧凑的模型和有限的非线性,允许实现完全可扩展的系统;对于所有分解级别和感兴趣的比特率,我们采用一对训练好的网络参数。通过将所提出的提升方法应用于 JPEG 2000 图像编码标准,我们的方法在广泛的比特率范围内可以实现最高 17.4% 的平均 BD 位率节省,同时保留 JPEG 2000 的质量和分辨率可扩展性特征。
URL
https://arxiv.org/abs/2403.01647