中国物理快报(英文版)2024,Vol.41Issue(12) :35-45.DOI:10.1088/0256-307X/41/12/120701

Universal Scaling Laws in Quantum-Probabilistic Machine Learning by Tensor Network:Toward Interpreting Representation and Generalization Powers

Sheng-Chen Bai Shi-Ju Ran
中国物理快报(英文版)2024,Vol.41Issue(12) :35-45.DOI:10.1088/0256-307X/41/12/120701

Universal Scaling Laws in Quantum-Probabilistic Machine Learning by Tensor Network:Toward Interpreting Representation and Generalization Powers

Sheng-Chen Bai 1Shi-Ju Ran1
扫码查看

作者信息

  • 1. Center for Quantum Physics and Intelligent Sciences,Department of Physics,Capital Normal University,Beijing 100048,China
  • 折叠

Abstract

The interpretation of representations and generalization powers has been a long-standing challenge in the fields of machine learning(ML)and artificial intelligence.This study contributes to understanding the emergence of universal scaling laws in quantum-probabilistic ML.We consider the generative tensor network(GTN)in the form of a matrix-product state as an example and show that with an untrained GTN(such as a random TN state),the negative logarithmic likelihood(NLL)L generally increases linearly with the number of features M,that is,L~kM+const.This is a consequence of the so-called"catastrophe of orthogonality,"which states that quantum many-body states tend to become exponentially orthogonal to each other as M increases.This study reveals that,while gaining information through training,the linear-scaling law is suppressed by a negative quadratic correction,leading to L(~)βM-αM2+const.The scaling coefficients exhibit logarithmic relationships with the number of training samples and quantum channels The emergence of a quadratic correction term in the NLL for the testing(training)set can be regarded as evidence of the generalization(representation)power of the GTN.Over-parameterization can be identified by the deviation in the values of α between the training and testing sets while increasing x.We further investigate how orthogonality in the quantum-feature map relates to the satisfaction of quantum-probabilistic interpretation and the representation and generalization powers of the GTN.Unveiling universal scaling laws in quantum-probabilistic ML would be a valuable step toward establishing a white-box ML scheme interpreted within the quantum-probabilistic framework.

引用本文复制引用

出版年

2024
中国物理快报(英文版)
中国科学院物理研究所,中国物理学会

中国物理快报(英文版)

CSTPCDCSCDEI
影响因子:0.515
ISSN:0256-307X
段落导航相关论文