Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1901.03429
Cited By
On the Turing Completeness of Modern Neural Network Architectures
10 January 2019
Jorge A. Pérez
Javier Marinkovic
Pablo Barceló
BDL
Re-assign community
ArXiv
PDF
HTML
Papers citing
"On the Turing Completeness of Modern Neural Network Architectures"
35 / 35 papers shown
Title
Can Large Language Models Learn Formal Logic? A Data-Driven Training and Evaluation Framework
Yuan Xia
Akanksha Atrey
Fadoua Khmaissia
Kedar S. Namjoshi
LRM
ELM
47
0
0
28 Apr 2025
Looped ReLU MLPs May Be All You Need as Practical Programmable Computers
Yingyu Liang
Zhizhou Sha
Zhenmei Shi
Zhao Song
Yufa Zhou
101
18
0
21 Feb 2025
Lower Bounds for Chain-of-Thought Reasoning in Hard-Attention Transformers
Alireza Amiri
Xinting Huang
Mark Rofin
Michael Hahn
LRM
260
0
0
04 Feb 2025
Are Transformers Able to Reason by Connecting Separated Knowledge in Training Data?
Yutong Yin
Zhaoran Wang
LRM
ReLM
238
0
0
27 Jan 2025
Circuit Complexity Bounds for Visual Autoregressive Model
Yekun Ke
Xiaoyu Li
Yingyu Liang
Zhenmei Shi
Zhao Song
45
5
0
08 Jan 2025
Can Transformers Reason Logically? A Study in SAT Solving
Leyan Pan
Vijay Ganesh
Jacob Abernethy
Chris Esposo
Wenke Lee
ReLM
LRM
33
1
0
09 Oct 2024
On the Complexity of Neural Computation in Superposition
Micah Adler
Nir Shavit
123
3
0
05 Sep 2024
LifeGPT: Topology-Agnostic Generative Pretrained Transformer Model for Cellular Automata
Jaime Berkovich
Markus J. Buehler
AI4CE
34
2
0
03 Sep 2024
Representing Rule-based Chatbots with Transformers
Dan Friedman
Abhishek Panigrahi
Danqi Chen
71
1
0
15 Jul 2024
Separations in the Representational Capabilities of Transformers and Recurrent Architectures
S. Bhattamishra
Michael Hahn
Phil Blunsom
Varun Kanade
GNN
44
9
0
13 Jun 2024
Limits of Deep Learning: Sequence Modeling through the Lens of Complexity Theory
Nikola Zubić
Federico Soldá
Aurelio Sulser
Davide Scaramuzza
LRM
BDL
57
6
0
26 May 2024
Chain of Thought Empowers Transformers to Solve Inherently Serial Problems
Zhiyuan Li
Hong Liu
Denny Zhou
Tengyu Ma
LRM
AI4CE
30
101
0
20 Feb 2024
Sample, estimate, aggregate: A recipe for causal discovery foundation models
Menghua Wu
Yujia Bao
Regina Barzilay
Tommi Jaakkola
CML
56
7
0
02 Feb 2024
Transformers as Decision Makers: Provable In-Context Reinforcement Learning via Supervised Pretraining
Licong Lin
Yu Bai
Song Mei
OffRL
37
45
0
12 Oct 2023
Self-attention Dual Embedding for Graphs with Heterophily
Yurui Lai
Taiyan Zhang
Rui Fan
GNN
48
0
0
28 May 2023
Looped Transformers as Programmable Computers
Angeliki Giannou
Shashank Rajput
Jy-yong Sohn
Kangwook Lee
Jason D. Lee
Dimitris Papailiopoulos
15
96
0
30 Jan 2023
E(n)-equivariant Graph Neural Cellular Automata
G. Gala
Daniele Grattarola
Erik Quaeghebeur
GNN
48
3
0
25 Jan 2023
Memory Augmented Large Language Models are Computationally Universal
Dale Schuurmans
35
45
0
10 Jan 2023
Attention-based Neural Cellular Automata
Mattie Tesfaldet
Derek Nowrouzezahrai
C. Pal
ViT
37
17
0
02 Nov 2022
HYPRO: A Hybridly Normalized Probabilistic Model for Long-Horizon Prediction of Event Sequences
Siqiao Xue
Xiaoming Shi
James Y. Zhang
Hongyuan Mei
AI4TS
19
34
0
04 Oct 2022
Provably expressive temporal graph networks
Amauri Souza
Diego Mesquita
Samuel Kaski
Vikas K. Garg
89
55
0
29 Sep 2022
What Can Transformers Learn In-Context? A Case Study of Simple Function Classes
Shivam Garg
Dimitris Tsipras
Percy Liang
Gregory Valiant
39
459
0
01 Aug 2022
Neural Networks and the Chomsky Hierarchy
Grégoire Delétang
Anian Ruoss
Jordi Grau-Moya
Tim Genewein
L. Wenliang
...
Chris Cundy
Marcus Hutter
Shane Legg
Joel Veness
Pedro A. Ortega
UQCV
109
133
0
05 Jul 2022
Formal Language Recognition by Hard Attention Transformers: Perspectives from Circuit Complexity
Sophie Hao
Dana Angluin
Robert Frank
16
71
0
13 Apr 2022
Can Vision Transformers Perform Convolution?
Shanda Li
Xiangning Chen
Di He
Cho-Jui Hsieh
ViT
49
19
0
02 Nov 2021
Pairing Conceptual Modeling with Machine Learning
W. Maass
V. Storey
HAI
27
33
0
27 Jun 2021
Vector Symbolic Architectures as a Computing Framework for Emerging Hardware
Denis Kleyko
Mike Davies
E. P. Frady
P. Kanerva
Spencer J. Kent
...
Evgeny Osipov
J. Rabaey
D. Rachkovskij
Abbas Rahimi
Friedrich T. Sommer
40
57
0
09 Jun 2021
On the Expressive Power of Self-Attention Matrices
Valerii Likhosherstov
K. Choromanski
Adrian Weller
37
34
0
07 Jun 2021
Attention is Not All You Need: Pure Attention Loses Rank Doubly Exponentially with Depth
Yihe Dong
Jean-Baptiste Cordonnier
Andreas Loukas
52
373
0
05 Mar 2021
Transformers in Vision: A Survey
Salman Khan
Muzammal Naseer
Munawar Hayat
Syed Waqas Zamir
Fahad Shahbaz Khan
M. Shah
ViT
233
2,434
0
04 Jan 2021
On the Computational Power of Transformers and its Implications in Sequence Modeling
S. Bhattamishra
Arkil Patel
Navin Goyal
33
66
0
16 Jun 2020
How hard is to distinguish graphs with graph neural networks?
Andreas Loukas
GNN
25
6
0
13 May 2020
It's Not What Machines Can Learn, It's What We Cannot Teach
Gal Yehuda
Moshe Gabel
Assaf Schuster
FaML
19
37
0
21 Feb 2020
Are Transformers universal approximators of sequence-to-sequence functions?
Chulhee Yun
Srinadh Bhojanapalli
A. S. Rawat
Sashank J. Reddi
Sanjiv Kumar
26
336
0
20 Dec 2019
On the Relationship between Self-Attention and Convolutional Layers
Jean-Baptiste Cordonnier
Andreas Loukas
Martin Jaggi
45
524
0
08 Nov 2019
1