Zhang Yichi, Duan Zhihao, Huang Yuning, Zhu Fengqing
Purdue University.
Transact Mach Learn Res. 2025 Apr;2025.
As learned image compression (LIC) methods become increasingly computationally demanding, enhancing their training efficiency is crucial. This paper takes a step forward in accelerating the training of LIC methods by modeling the neural training dynamics. We first propose a Sensitivity-aware True and Dummy Embedding Training mechanism (STDET) that clusters LIC model parameters into few separate modes where parameters are expressed as affine transformations of reference parameters within the same mode. By further utilizing the stable intra-mode correlations throughout training and parameter sensitivities, we gradually embed non-reference parameters, reducing the number of trainable parameters. Additionally, we incorporate a Sampling-then-Moving Average (SMA) technique, interpolating sampled weights from stochastic gradient descent (SGD) training to obtain the moving average weights, ensuring smooth temporal behavior and minimizing training state variances. Overall, our method significantly reduces training space dimensions and the number of trainable parameters without sacrificing model performance, thus accelerating model convergence. We also provide a theoretical analysis on the Noisy quadratic model, showing that the proposed method achieves a lower training variance than standard SGD. Our approach offers valuable insights for further developing efficient training methods for LICs.
随着有监督图像压缩(LIC)方法的计算需求日益增加,提高其训练效率至关重要。本文通过对神经训练动力学进行建模,在加速LIC方法的训练方面迈出了一步。我们首先提出了一种灵敏度感知的真嵌入和虚拟嵌入训练机制(STDET),该机制将LIC模型参数聚类为几个单独的模式,其中参数表示为同一模式内参考参数的仿射变换。通过在整个训练过程中进一步利用稳定的模式内相关性和参数灵敏度,我们逐步嵌入非参考参数,减少了可训练参数的数量。此外,我们引入了一种先采样后移动平均(SMA)技术,对随机梯度下降(SGD)训练中的采样权重进行插值以获得移动平均权重,确保时间行为平滑并最小化训练状态方差。总体而言,我们的方法在不牺牲模型性能的情况下显著降低了训练空间维度和可训练参数的数量,从而加速了模型收敛。我们还对噪声二次模型进行了理论分析,表明所提出的方法比标准SGD具有更低的训练方差。我们的方法为进一步开发LIC的高效训练方法提供了有价值的见解。