Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2103.05247
Cited By
v1
v2 (latest)
Pretrained Transformers as Universal Computation Engines
9 March 2021
Kevin Lu
Aditya Grover
Pieter Abbeel
Igor Mordatch
Re-assign community
ArXiv (abs)
PDF
HTML
HuggingFace (1 upvotes)
Papers citing
"Pretrained Transformers as Universal Computation Engines"
50 / 151 papers shown
Selection-Inference: Exploiting Large Language Models for Interpretable Logical Reasoning
International Conference on Learning Representations (ICLR), 2022
Antonia Creswell
Murray Shanahan
I. Higgins
ReLM
LRM
313
433
0
19 May 2022
Clinical Prompt Learning with Frozen Language Models
Niall Taylor
Yi Zhang
Dan W Joyce
A. Nevado-Holgado
Andrey Kormilitzin
VLM
LM&MA
144
37
0
11 May 2022
StorSeismic: A new paradigm in deep learning for seismic processing
IEEE Transactions on Geoscience and Remote Sensing (IEEE TGRS), 2022
R. Harsuko
T. Alkhalifah
131
44
0
30 Apr 2022
Efficient Architecture Search for Diverse Tasks
Neural Information Processing Systems (NeurIPS), 2022
Jun Shen
M. Khodak
Ameet Talwalkar
293
38
0
15 Apr 2022
Causal Transformer for Estimating Counterfactual Outcomes
International Conference on Machine Learning (ICML), 2022
Valentyn Melnychuk
Dennis Frauen
Stefan Feuerriegel
CML
275
131
0
14 Apr 2022
Single-Stream Multi-Level Alignment for Vision-Language Pretraining
European Conference on Computer Vision (ECCV), 2022
Zaid Khan
B. Vijaykumar
Xiang Yu
S. Schulter
Manmohan Chandraker
Y. Fu
CLIP
VLM
356
22
0
27 Mar 2022
Linking Emergent and Natural Languages via Corpus Transfer
International Conference on Learning Representations (ICLR), 2022
Shunyu Yao
Mo Yu
Yang Zhang
Karthik Narasimhan
J. Tenenbaum
Chuang Gan
235
19
0
24 Mar 2022
Hyperbolic Vision Transformers: Combining Improvements in Metric Learning
Computer Vision and Pattern Recognition (CVPR), 2022
Aleksandr Ermolov
L. Mirvakhabova
Valentin Khrulkov
Andrii Zadaianchuk
Ivan Oseledets
319
145
0
21 Mar 2022
Pretraining with Artificial Language: Studying Transferable Knowledge in Language Models
Annual Meeting of the Association for Computational Linguistics (ACL), 2022
Ryokan Ri
Yoshimasa Tsuruoka
231
36
0
19 Mar 2022
Summarizing a virtual robot's past actions in natural language
Chad DeChant
Daniel Bauer
LM&Ro
172
4
0
13 Mar 2022
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Interspeech (Interspeech), 2022
Guan-Ting Lin
Yung-Sung Chuang
Ho-Lam Chung
Shu-Wen Yang
Hsuan-Jui Chen
Shuyan Dong
Shang-Wen Li
Abdel-rahman Mohamed
Hung-yi Lee
Lin-Shan Lee
329
25
0
09 Mar 2022
High-Modality Multimodal Transformer: Quantifying Modality & Interaction Heterogeneity for High-Modality Representation Learning
Paul Pu Liang
Yiwei Lyu
Xiang Fan
Jeffrey Tsaw
Yudong Liu
Shentong Mo
Dani Yogatama
Louis-Philippe Morency
Ruslan Salakhutdinov
230
43
0
02 Mar 2022
Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning
AAAI Conference on Artificial Intelligence (AAAI), 2022
Pin-Yu Chen
VLM
425
77
0
22 Feb 2022
MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Neural Information Processing Systems (NeurIPS), 2022
Anssi Kanervisto
Stephanie Milani
Karolis Ramanauskas
Nicholay Topin
Zichuan Lin
...
Franccois Fleuret
Alexander Nikulin
Yury Belousov
Oleg Svidchenko
A. Shpilman
OffRL
245
35
0
17 Feb 2022
Online Decision Transformer
International Conference on Machine Learning (ICML), 2022
Qinqing Zheng
Amy Zhang
Aditya Grover
OffRL
284
242
0
11 Feb 2022
TTS-GAN: A Transformer-based Time-Series Generative Adversarial Network
Conference on Artificial Intelligence in Medicine in Europe (AIME), 2022
Xiaomin Li
V. Metsis
Huan Wang
Anne H. H. Ngu
139
140
0
06 Feb 2022
Pre-Trained Language Models for Interactive Decision-Making
Neural Information Processing Systems (NeurIPS), 2022
Shuang Li
Xavier Puig
Chris Paxton
Yilun Du
Clinton Jia Wang
...
Anima Anandkumar
Jacob Andreas
Igor Mordatch
Antonio Torralba
Yuke Zhu
LM&Ro
423
307
0
03 Feb 2022
Regression Transformer: Concurrent sequence regression and generation for molecular language modeling
Nature Machine Intelligence (Nat. Mach. Intell.), 2022
Jannis Born
Matteo Manica
299
127
0
01 Feb 2022
Can Wikipedia Help Offline Reinforcement Learning?
Machel Reid
Yutaro Yamada
S. Gu
3DV
RALM
OffRL
656
101
0
28 Jan 2022
Pre-Trained Language Transformers are Universal Image Classifiers
Rahul Goel
Modar Sulaiman
Kimia Noorbakhsh
Mahdieh Sharifi
Ragunath R. Sharma
Pooyan Jamshidi
Kallol Roy
ViT
134
3
0
25 Jan 2022
Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents
International Conference on Machine Learning (ICML), 2022
Wenlong Huang
Pieter Abbeel
Deepak Pathak
Igor Mordatch
LM&Ro
322
1,419
0
18 Jan 2022
Video Transformers: A Survey
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
Javier Selva
A. S. Johansen
Sergio Escalera
Kamal Nasrollahi
T. Moeslund
Albert Clapés
ViT
458
139
0
16 Jan 2022
CLIP-TD: CLIP Targeted Distillation for Vision-Language Tasks
Zhecan Wang
Noel Codella
Yen-Chun Chen
Luowei Zhou
Jianwei Yang
Xiyang Dai
Bin Xiao
Haoxuan You
Shih-Fu Chang
Lu Yuan
CLIP
VLM
213
44
0
15 Jan 2022
Improved Input Reprogramming for GAN Conditioning
Tuan Dinh
Daewon Seo
Zhixu Du
Liang Shang
Kangwook Lee
AI4CE
261
8
0
07 Jan 2022
Discrete and continuous representations and processing in deep learning: Looking forward
AI Open (AO), 2022
Ruben Cartuyvels
Graham Spinks
Marie-Francine Moens
OCL
301
30
0
04 Jan 2022
FLAVA: A Foundational Language And Vision Alignment Model
Amanpreet Singh
Ronghang Hu
Vedanuj Goswami
Guillaume Couairon
Wojciech Galuba
Marcus Rohrbach
Douwe Kiela
CLIP
VLM
383
873
0
08 Dec 2021
Emojich -- zero-shot emoji generation using Russian language: a technical report
Alex Shonenkov
Daria Bakshandaeva
Denis Dimitrov
Aleks D. Nikolich
VLM
213
5
0
04 Dec 2021
Show Your Work: Scratchpads for Intermediate Computation with Language Models
Maxwell Nye
Anders Andreassen
Guy Gur-Ari
Henryk Michalewski
Jacob Austin
...
Aitor Lewkowycz
Maarten Bosma
D. Luan
Charles Sutton
Augustus Odena
ReLM
LRM
547
927
0
30 Nov 2021
Many Heads but One Brain: Fusion Brain -- a Competition and a Single Multimodal Multitask Architecture
Daria Bakshandaeva
Denis Dimitrov
V.Ya. Arkhipkin
Alex Shonenkov
M. Potanin
...
Mikhail Martynov
Anton Voronov
Vera Davydova
E. Tutubalina
Aleksandr Petiushko
379
0
0
22 Nov 2021
Attention Approximates Sparse Distributed Memory
Neural Information Processing Systems (NeurIPS), 2021
Trenton Bricken
Cengiz Pehlevan
316
37
0
10 Nov 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
355
150
0
19 Oct 2021
The Power of Prompt Tuning for Low-Resource Semantic Parsing
Nathan Schucher
Siva Reddy
H. D. Vries
VLM
243
36
0
16 Oct 2021
Scaling Laws for the Few-Shot Adaptation of Pre-trained Image Classifiers
Gabriele Prato
Simon Guiroy
Ethan Caballero
Irina Rish
Sarath Chandar
VLM
239
12
0
13 Oct 2021
Pretrained Language Models are Symbolic Mathematics Solvers too!
Kimia Noorbakhsh
Modar Sulaiman
M. Sharifi
Kallol Roy
Pooyan Jamshidi
LRM
292
22
0
07 Oct 2021
Compositional generalization in semantic parsing with pretrained transformers
A. Orhan
248
8
0
30 Sep 2021
Cross-lingual Transfer of Monolingual Models
Evangelia Gogoulou
Ariel Ekgren
T. Isbister
Magnus Sahlgren
256
20
0
15 Sep 2021
TEASEL: A Transformer-Based Speech-Prefixed Language Model
Mehdi Arjmand
M. Dousti
H. Moradi
140
23
0
12 Sep 2021
Exploiting BERT For Multimodal Target Sentiment Classification Through Input Space Translation
Zaid Khan
Y. Fu
174
178
0
03 Aug 2021
Transformer-based Map Matching Model with Limited Ground-Truth Data using Transfer-Learning Approach
Zhixiong Jin
Jiwon Kim
H. Yeo
Seongjin Choi
235
33
0
01 Aug 2021
Perceiver IO: A General Architecture for Structured Inputs & Outputs
International Conference on Learning Representations (ICLR), 2021
Andrew Jaegle
Sebastian Borgeaud
Jean-Baptiste Alayrac
Carl Doersch
Catalin Ionescu
...
Olivier J. Hénaff
M. Botvinick
Andrew Zisserman
Oriol Vinyals
João Carreira
MLLM
VLM
GNN
659
730
0
30 Jul 2021
Don't Sweep your Learning Rate under the Rug: A Closer Look at Cross-modal Transfer of Pretrained Transformers
Dan Rothermel
Margaret Li
Tim Rocktaschel
Jakob N. Foerster
251
6
0
26 Jul 2021
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised Speech Representation Learning
IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2021
Samuel Kessler
Bethan Thomas
S. Karout
SSL
167
31
0
26 Jul 2021
Multimodal Few-Shot Learning with Frozen Language Models
Neural Information Processing Systems (NeurIPS), 2021
Maria Tsimpoukelli
Jacob Menick
Serkan Cabi
S. M. Ali Eslami
Oriol Vinyals
Felix Hill
MLLM
548
907
0
25 Jun 2021
Going Beyond Linear Transformers with Recurrent Fast Weight Programmers
Neural Information Processing Systems (NeurIPS), 2021
Kazuki Irie
Imanol Schlag
Róbert Csordás
Jürgen Schmidhuber
254
77
0
11 Jun 2021
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning
Computer Vision and Pattern Recognition (CVPR), 2021
Liangqiong Qu
Yuyin Zhou
Paul Pu Liang
Yingda Xia
Feifei Wang
Ehsan Adeli
L. Fei-Fei
D. Rubin
FedML
AI4CE
414
215
0
10 Jun 2021
Image2Point: 3D Point-Cloud Understanding with 2D Image Pretrained Models
European Conference on Computer Vision (ECCV), 2021
Chenfeng Xu
Shijia Yang
Tomer Galanti
Bichen Wu
Xiangyu Yue
Bohan Zhai
Wei Zhan
Peter Vajda
Kurt Keutzer
Masayoshi Tomizuka
3DPC
206
62
0
08 Jun 2021
Decision Transformer: Reinforcement Learning via Sequence Modeling
Neural Information Processing Systems (NeurIPS), 2021
Lili Chen
Kevin Lu
Aravind Rajeswaran
Kimin Lee
Aditya Grover
Michael Laskin
Pieter Abbeel
A. Srinivas
Igor Mordatch
OffRL
563
2,023
0
02 Jun 2021
Choose a Transformer: Fourier or Galerkin
Neural Information Processing Systems (NeurIPS), 2021
Shuhao Cao
444
349
0
31 May 2021
Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Mozhdeh Gheini
Xiang Ren
Jonathan May
LRM
319
165
0
18 Apr 2021
Generalization over different cellular automata rules learned by a deep feed-forward neural network
Marcel Aach
J. Goebbert
J. Jitsev
AI4CE
145
6
0
27 Mar 2021
Previous
1
2
3
4
Next
Page 3 of 4