Qiu Senhui, Bhattacharyya Saugat, Coyle Damien, Dora Shirin
Intelligent Systems Research Centre, School of Computing, Engineering and Intelligent Systems, Ulster University, Londonderry, BT48 7JL, UK.
Intelligent Systems Research Centre, School of Computing, Engineering and Intelligent Systems, Ulster University, Londonderry, BT48 7JL, UK; Bath Institute for the Augmented Human, University of Bath, Bath, BA2 7AY, UK.
Neural Netw. 2025 Nov;191:107785. doi: 10.1016/j.neunet.2025.107785. Epub 2025 Jul 3.
Predictive Coding (PC) has emerged as a prominent theory underlying information processing in the brain. The general concept for learning in PC is that each layer learns to predict the activities of neurons in the previous layer, which enables local computation of error as well as in-parallel learning across layers. Deep Bi-directional Predictive Coding (DBPC) is proposed here as a new learning algorithm that enables neural networks to simultaneously perform classification and reconstruction tasks using the same learned weights. Building on existing PC approaches, DBPC supports both feedforward and feedback propagation of information. Each layer in the network trained using DBPC learns to predict the activities of neurons in the previous and next layers, enabling the network to simultaneously perform classification and reconstruction tasks using feedforward and feedback propagation, respectively. DBPC also relies on locally available information for learning, thus enabling in-parallel learning across all layers in the network. DBPC enables the training of both fully connected networks and convolutional neural networks. The classification accuracies of DBPC on the MNIST, Fashion-MNIST, and CIFAR-10 datasets (99.58%, 92.42%, and 74.29%, respectively) exceed those of well-established PC-based benchmark approaches (including FIPC and iPC) and are competitive with state-of-the-art Error-Backpropagation-based methods (including ResNet and DenseNet) on MNIST, Fashion-MNIST, and EuroSAT datasets. Importantly, DBPC achieves these results using significantly smaller networks for MNIST, Fashion-MNIST, and CIFAR-10 datasets (0.425, 1.004, and 1.109 million parameters), and every representation estimated in DBPC can be used for the reconstruction of inputs. The significant benefit of DBPC is its ability to achieve this performance using locally available information and in-parallel learning mechanisms, which results in an efficient training protocol. Overall, we demonstrate that DBPC is a much more efficient approach for training networks that can perform both classification and reconstruction simultaneously.
预测编码(PC)已成为大脑信息处理背后的一个重要理论。PC中学习的一般概念是,每一层都学习预测前一层神经元的活动,这使得能够局部计算误差以及跨层并行学习。本文提出了深度双向预测编码(DBPC)作为一种新的学习算法,它使神经网络能够使用相同的学习权重同时执行分类和重建任务。基于现有的PC方法,DBPC支持信息的前向和反馈传播。使用DBPC训练的网络中的每一层都学习预测前一层和下一层神经元的活动,从而使网络能够分别使用前向和反馈传播同时执行分类和重建任务。DBPC还依赖于局部可用信息进行学习,从而能够在网络的所有层中并行学习。DBPC能够训练全连接网络和卷积神经网络。DBPC在MNIST、Fashion-MNIST和CIFAR-10数据集上的分类准确率(分别为99.58%、92.42%和74.29%)超过了成熟的基于PC的基准方法(包括FIPC和iPC),并且在MNIST、Fashion-MNIST和EuroSAT数据集上与基于误差反向传播的最新方法(包括ResNet和DenseNet)具有竞争力。重要的是,DBPC在MNIST、Fashion-MNIST和CIFAR-10数据集上使用显著更小的网络(0.四百二十五万、1.004万和1.109万个参数)取得了这些结果,并且DBPC中估计的每个表示都可用于输入的重建。DBPC的显著优势在于它能够使用局部可用信息和并行学习机制实现这种性能,这导致了一种高效的训练协议。总体而言,我们证明DBPC是一种训练能够同时执行分类和重建的网络的更高效方法。