Dube Swaraj, Wong Yee Wan, Nugroho Hermawan
Department of Electrical and Electronic Engineering, University of Nottingham - Malaysia Campus, Semenyih, Selangor, Malaysia.
PeerJ Comput Sci. 2021 Jul 15;7:e633. doi: 10.7717/peerj-cs.633. eCollection 2021.
Incremental learning evolves deep neural network knowledge over time by learning continuously from new data instead of training a model just once with all data present before the training starts. However, in incremental learning, new samples are always streaming in whereby the model to be trained needs to continuously adapt to new samples. Images are considered to be high dimensional data and thus training deep neural networks on such data is very time-consuming. Fog computing is a paradigm that uses fog devices to carry out computation near data sources to reduce the computational load on the server. Fog computing allows democracy in deep learning by enabling intelligence at the fog devices, however, one of the main challenges is the high communication costs between fog devices and the centralized servers especially in incremental learning where data samples are continuously arriving and need to be transmitted to the server for training. While working with Convolutional Neural Networks (CNN), we demonstrate a novel data sampling algorithm that discards certain training images per class before training even starts which reduces the transmission cost from the fog device to the server and the model training time while maintaining model learning performance both for static and incremental learning. Results show that our proposed method can effectively perform data sampling regardless of the model architecture, dataset, and learning settings.
增量学习通过不断从新数据中学习来随时间演进深度神经网络知识,而不是在训练开始前一次性使用所有现有数据训练模型。然而,在增量学习中,新样本总是不断涌入,因此待训练的模型需要持续适应新样本。图像被视为高维数据,因此在这类数据上训练深度神经网络非常耗时。雾计算是一种利用雾设备在数据源附近进行计算以减轻服务器计算负载的范式。雾计算通过在雾设备上实现智能,使深度学习更加灵活,然而,主要挑战之一是雾设备与集中式服务器之间的通信成本高昂,特别是在增量学习中,数据样本不断到达且需要传输到服务器进行训练。在使用卷积神经网络(CNN)时,我们展示了一种新颖的数据采样算法,该算法在训练开始前就丢弃每个类别的某些训练图像,这降低了从雾设备到服务器的传输成本以及模型训练时间,同时在静态学习和增量学习中都保持了模型的学习性能。结果表明,无论模型架构、数据集和学习设置如何,我们提出的方法都能有效地进行数据采样。