167
v1v2 (latest)

Efficient Turing Machine Simulation with Transformers

Main:11 Pages
2 Figures
Bibliography:4 Pages
1 Tables
Appendix:4 Pages
Abstract

Constant bit-size Transformers are known to be Turing complete, but existing constructions require Ω(s(n))\Omega(s(n)) chain-of-thought (CoT) steps per simulated Turing machine (TM) step, leading to impractical reasoning lengths. In this paper, we significantly reduce this efficiency gap by proving that any (t(n),s(n))(t(n),s(n))-bounded multi-tape TM can be simulated by a constant bit-size Transformer with an optimal O(s(n))O(s(n))-long context window and only O(s(n)c)O(s(n)^c) CoT steps per TM step, where c>0c>0 can be made arbitrarily small by letting the Transformers' head-layer product sufficiently large. In addition, our construction shows that sparse attention with fixed geometric offsets suffices for efficient universal computation. Our proof leverages multi-queue TMs as a bridge. The main technical novelty is a more efficient simulation of multi-tape TMs by synchronous multi-queue TMs, improving both time and space complexity under stricter model assumptions.

View on arXiv
Comments on this paper