Huang Haiping
PMI Lab, School of Physics, Sun Yat-sen University, Guangzhou, China.
Front Comput Neurosci. 2024 Jul 24;18:1388166. doi: 10.3389/fncom.2024.1388166. eCollection 2024.
A good theory of mathematical beauty is more practical than any current observation, as new predictions about physical reality can be self-consistently verified. This belief applies to the current status of understanding deep neural networks including large language models and even the biological intelligence. Toy models provide a metaphor of physical reality, allowing mathematically formulating the reality (i.e., the so-called theory), which can be updated as more conjectures are justified or refuted. One does not need to present all details in a model, but rather, more abstract models are constructed, as complex systems such as the brains or deep networks have many sloppy dimensions but much less stiff dimensions that strongly impact macroscopic observables. This type of bottom-up mechanistic modeling is still promising in the modern era of understanding the natural or artificial intelligence. Here, we shed light on eight challenges in developing theory of intelligence following this theoretical paradigm. Theses challenges are representation learning, generalization, adversarial robustness, continual learning, causal learning, internal model of the brain, next-token prediction, and the mechanics of subjective experience.
一个好的数学美理论比任何当前的观察都更具实用性,因为关于物理现实的新预测可以自洽地得到验证。这种信念适用于当前对深度神经网络(包括大语言模型)乃至生物智能的理解现状。玩具模型提供了物理现实的一种隐喻,使人们能够以数学方式表述现实(即所谓的理论),随着更多猜想得到证实或被反驳,该理论可以不断更新。人们无需在模型中呈现所有细节,而是构建更抽象的模型,因为诸如大脑或深度网络这样的复杂系统有许多模糊维度,但对宏观可观测量有强烈影响的刚性维度要少得多。在理解自然或人工智能的现代时代,这种自下而上的机制建模仍然很有前景。在此,我们根据这一理论范式阐明了发展智能理论过程中的八个挑战。这些挑战包括表征学习、泛化、对抗鲁棒性、持续学习、因果学习、大脑的内部模型、下一个token预测以及主观体验的机制。