Quick Navigation
Topics
Quantum Error Correction Fault Tolerance
Unlearnable phases of matter
arXiv
Authors: Tarun Advaith Kumar, Yijian Zou, Amir-Reza Negari, Roger G. Melko, Timothy H. Hsieh
Year
2026
Paper ID
112
Status
Preprint
Abstract Read
~2 min
Abstract Words
165
Citations
N/A
Abstract
We identify fundamental limitations in machine learning by demonstrating that non-trivial mixed-state phases of matter are computationally hard to learn. Focusing on unsupervised learning of distributions, we show that autoregressive neural networks fail to learn global properties of distributions characterized by locally indistinguishable (LI) states. We demonstrate that conditional mutual information (CMI) is a useful diagnostic for LI: we show that for classical distributions, long-range CMI of a state implies a spatially LI partner. By introducing a restricted statistical query model, we prove that nontrivial phases with long-range CMI, such as strong-to-weak spontaneous symmetry breaking phases, are hard to learn. We validate our claims by using recurrent, convolutional, and Transformer neural networks to learn the syndrome and physical distributions of toric/surface code under bit flip noise. Our findings suggest hardness of learning as a diagnostic tool for detecting mixed-state phases and transitions and error-correction thresholds, and they suggest CMI and more generally ``non-local Gibbsness'' as metrics for how hard a distribution is to learn.
Paper Tools
Category Correction Request
Help us improve classification quality by proposing a better category. Every request is reviewed by an admin.
Sign in to submit a category correction request for this paper.
Log In to SubmitReferences & Citation Signals
Community Reactions
Quick sentiment from readers on this paper.
Score:
0
Likes: 0
Dislikes: 0
Sign in to react to this paper.
Discussion & Reviews (Moderated)
Average Rating: 0.0 / 5 (0 ratings)
No written reviews yet.