Quick Navigation
Topics
Quantum Error Correction Fault Tolerance
A Mixture of Experts Vision Transformer for High-Fidelity Surface Code Decoding
arXiv
Authors: Hoang Viet Nguyen, Manh Hung Nguyen, Hoang Ta, Van Khu Vu, Yeow Meng Chee
Year
2026
Paper ID
3659
Status
Preprint
Abstract Read
~2 min
Abstract Words
193
Citations
N/A
Abstract
Quantum error correction is a key ingredient for large scale quantum computation, protecting logical information from physical noise by encoding it into many physical qubits. Topological stabilizer codes are particularly appealing due to their geometric locality and practical relevance. In these codes, stabilizer measurements yield a syndrome that must be decoded into a recovery operation, making decoding a central bottleneck for scalable real time operation. Existing decoders are commonly classified into two categories. Classical algorithmic decoders provide strong and well established baselines, but may incur substantial computational overhead at large code distances or under stringent latency constraints. Machine learning based decoders offer fast GPU inference and flexible function approximation, yet many approaches do not explicitly exploit the lattice geometry and local structure of topological codes, which can limit performance. In this work, we propose QuantumSMoE, a quantum vision transformer based decoder that incorporates code structure through plus shaped embeddings and adaptive masking to capture local interactions and lattice connectivity, and improves scalability via a mixture of experts layer with a novel auxiliary loss. Experiments on the toric code demonstrate that QuantumSMoE outperforms state-of-the-art machine learning decoders as well as widely used classical baselines.
Paper Tools
Category Correction Request
Help us improve classification quality by proposing a better category. Every request is reviewed by an admin.
Sign in to submit a category correction request for this paper.
Log In to SubmitReferences & Citation Signals
Community Reactions
Quick sentiment from readers on this paper.
Score:
0
Likes: 0
Dislikes: 0
Sign in to react to this paper.
Discussion & Reviews (Moderated)
Average Rating: 0.0 / 5 (0 ratings)
No written reviews yet.