Neural Networks2022,Vol.14822.DOI:10.1016/j.neunet.2022.01.017

The emergence of a concept in shallow neural networks

Agliari, Elena Alemanno, Francesco Barra, Adriano De Marzo, Giordano
Neural Networks2022,Vol.14822.DOI:10.1016/j.neunet.2022.01.017

The emergence of a concept in shallow neural networks

Agliari, Elena 1Alemanno, Francesco 2Barra, Adriano 2De Marzo, Giordano3
扫码查看

作者信息

  • 1. Dipartimento Matemat,Sapienza Univ Roma
  • 2. Dipartimento Matemat & Fis,Univ Salento
  • 3. Dipartimento Fis,Sapienza Univ Roma
  • 折叠

Abstract

We consider restricted Boltzmann machine (RBMs) trained over an unstructured dataset made of blurred copies of definite but unavailable "archetypes " and we show that there exists a critical sample size beyond which the RBM can learn archetypes, namely the machine can successfully play as a generative model or as a classifier, according to the operational routine. In general, assessing a critical sample size (possibly in relation to the quality of the dataset) is still an open problem in machine learning. Here, restricting to the random theory, where shallow networks suffice and the "grandmother-cell " scenario is correct, we leverage the formal equivalence between RBMs and Hopfield networks, to obtain a phase diagram for both the neural architectures which highlights regions, in the space of the control parameters (i.e., number of archetypes, number of neurons, size and quality of the training set), where learning can be accomplished. Our investigations are led by analytical methods based on the statistical-mechanics of disordered systems and results are further corroborated by extensive Monte Carlo simulations. (C) 2022 Elsevier Ltd. All rights reserved.

Key words

Neural networks/Machine learning/Glassy statistical mechanics/PATTERNS

引用本文复制引用

出版年

2022
Neural Networks

Neural Networks

EISCI
ISSN:0893-6080
被引量10
参考文献量39
段落导航相关论文