Suppr超能文献

线性递归网络中记忆与预测的区别。

Difference between memory and prediction in linear recurrent networks.

机构信息

Department of Physics, Physics of Living Systems, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA.

出版信息

Phys Rev E. 2017 Sep;96(3-1):032308. doi: 10.1103/PhysRevE.96.032308. Epub 2017 Sep 11.

Abstract

Recurrent networks are trained to memorize their input better, often in the hopes that such training will increase the ability of the network to predict. We show that networks designed to memorize input can be arbitrarily bad at prediction. We also find, for several types of inputs, that one-node networks optimized for prediction are nearly at upper bounds on predictive capacity given by Wiener filters and are roughly equivalent in performance to randomly generated five-node networks. Our results suggest that maximizing memory capacity leads to very different networks than maximizing predictive capacity and that optimizing recurrent weights can decrease reservoir size by half an order of magnitude.

摘要

递归网络经过训练可以更好地记忆它们的输入,通常是希望这种训练能够提高网络的预测能力。我们表明,设计用于记忆输入的网络在预测方面可能非常差。我们还发现,对于几种类型的输入,针对预测进行优化的单节点网络几乎达到了 Wiener 滤波器给出的预测能力上限,并且在性能上大致相当于随机生成的五节点网络。我们的结果表明,最大化记忆容量会导致与最大化预测能力非常不同的网络,并且优化递归权重可以将储层大小减小一个数量级。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验