Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2103.05247
Cited By
v1
v2 (latest)
Pretrained Transformers as Universal Computation Engines
9 March 2021
Kevin Lu
Aditya Grover
Pieter Abbeel
Igor Mordatch
Re-assign community
ArXiv (abs)
PDF
HTML
HuggingFace (1 upvotes)
Papers citing
"Pretrained Transformers as Universal Computation Engines"
50 / 151 papers shown
Selection-Inference: Exploiting Large Language Models for Interpretable Logical Reasoning
International Conference on Learning Representations (ICLR), 2022
Antonia Creswell
Murray Shanahan
I. Higgins
ReLM
LRM
418
451
0
19 May 2022
Clinical Prompt Learning with Frozen Language Models
Niall Taylor
Yi Zhang
Dan W Joyce
A. Nevado-Holgado
Andrey Kormilitzin
VLM
LM&MA
189
39
0
11 May 2022
StorSeismic: A new paradigm in deep learning for seismic processing
IEEE Transactions on Geoscience and Remote Sensing (IEEE TGRS), 2022
R. Harsuko
T. Alkhalifah
173
48
0
30 Apr 2022
Efficient Architecture Search for Diverse Tasks
Neural Information Processing Systems (NeurIPS), 2022
Jun Shen
M. Khodak
Ameet Talwalkar
372
38
0
15 Apr 2022
Causal Transformer for Estimating Counterfactual Outcomes
International Conference on Machine Learning (ICML), 2022
Valentyn Melnychuk
Dennis Frauen
Stefan Feuerriegel
CML
296
137
0
14 Apr 2022
Single-Stream Multi-Level Alignment for Vision-Language Pretraining
European Conference on Computer Vision (ECCV), 2022
Zaid Khan
B. Vijaykumar
Xiang Yu
S. Schulter
Manmohan Chandraker
Y. Fu
CLIP
VLM
399
23
0
27 Mar 2022
Linking Emergent and Natural Languages via Corpus Transfer
International Conference on Learning Representations (ICLR), 2022
Shunyu Yao
Mo Yu
Yang Zhang
Karthik Narasimhan
J. Tenenbaum
Chuang Gan
343
20
0
24 Mar 2022
Hyperbolic Vision Transformers: Combining Improvements in Metric Learning
Computer Vision and Pattern Recognition (CVPR), 2022
Aleksandr Ermolov
L. Mirvakhabova
Valentin Khrulkov
Andrii Zadaianchuk
Ivan Oseledets
347
157
0
21 Mar 2022
Pretraining with Artificial Language: Studying Transferable Knowledge in Language Models
Annual Meeting of the Association for Computational Linguistics (ACL), 2022
Ryokan Ri
Yoshimasa Tsuruoka
285
37
0
19 Mar 2022
Summarizing a virtual robot's past actions in natural language
Chad DeChant
Daniel Bauer
LM&Ro
210
4
0
13 Mar 2022
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Interspeech (Interspeech), 2022
Guan-Ting Lin
Yung-Sung Chuang
Ho-Lam Chung
Shu-Wen Yang
Hsuan-Jui Chen
Shuyan Dong
Shang-Wen Li
Abdel-rahman Mohamed
Hung-yi Lee
Lin-Shan Lee
392
25
0
09 Mar 2022
High-Modality Multimodal Transformer: Quantifying Modality & Interaction Heterogeneity for High-Modality Representation Learning
Paul Pu Liang
Yiwei Lyu
Xiang Fan
Jeffrey Tsaw
Yudong Liu
Shentong Mo
Dani Yogatama
Louis-Philippe Morency
Ruslan Salakhutdinov
291
47
0
02 Mar 2022
Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning
AAAI Conference on Artificial Intelligence (AAAI), 2022
Pin-Yu Chen
VLM
565
80
0
22 Feb 2022
MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Neural Information Processing Systems (NeurIPS), 2022
Anssi Kanervisto
Stephanie Milani
Karolis Ramanauskas
Nicholay Topin
Zichuan Lin
...
Franccois Fleuret
Alexander Nikulin
Yury Belousov
Oleg Svidchenko
A. Shpilman
OffRL
315
35
0
17 Feb 2022
Online Decision Transformer
International Conference on Machine Learning (ICML), 2022
Qinqing Zheng
Amy Zhang
Aditya Grover
OffRL
403
250
0
11 Feb 2022
TTS-GAN: A Transformer-based Time-Series Generative Adversarial Network
Conference on Artificial Intelligence in Medicine in Europe (AIME), 2022
Xiaomin Li
V. Metsis
Huan Wang
Anne H. H. Ngu
236
150
0
06 Feb 2022
Pre-Trained Language Models for Interactive Decision-Making
Neural Information Processing Systems (NeurIPS), 2022
Shuang Li
Xavier Puig
Chris Paxton
Yilun Du
Clinton Jia Wang
...
Anima Anandkumar
Jacob Andreas
Igor Mordatch
Antonio Torralba
Yuke Zhu
LM&Ro
482
317
0
03 Feb 2022
Regression Transformer: Concurrent sequence regression and generation for molecular language modeling
Nature Machine Intelligence (Nat. Mach. Intell.), 2022
Jannis Born
Matteo Manica
315
135
0
01 Feb 2022
Can Wikipedia Help Offline Reinforcement Learning?
Machel Reid
Yutaro Yamada
S. Gu
3DV
RALM
OffRL
726
103
0
28 Jan 2022
Pre-Trained Language Transformers are Universal Image Classifiers
Rahul Goel
Modar Sulaiman
Kimia Noorbakhsh
Mahdieh Sharifi
Ragunath R. Sharma
Pooyan Jamshidi
Kallol Roy
ViT
174
3
0
25 Jan 2022
Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents
International Conference on Machine Learning (ICML), 2022
Wenlong Huang
Pieter Abbeel
Deepak Pathak
Igor Mordatch
LM&Ro
402
1,480
0
18 Jan 2022
Video Transformers: A Survey
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
Javier Selva
A. S. Johansen
Sergio Escalera
Kamal Nasrollahi
T. Moeslund
Albert Clapés
ViT
550
152
0
16 Jan 2022
CLIP-TD: CLIP Targeted Distillation for Vision-Language Tasks
Zhecan Wang
Noel Codella
Yen-Chun Chen
Luowei Zhou
Jianwei Yang
Xiyang Dai
Bin Xiao
Haoxuan You
Shih-Fu Chang
Lu Yuan
CLIP
VLM
277
44
0
15 Jan 2022
Improved Input Reprogramming for GAN Conditioning
Tuan Dinh
Daewon Seo
Zhixu Du
Liang Shang
Kangwook Lee
AI4CE
313
8
0
07 Jan 2022
Discrete and continuous representations and processing in deep learning: Looking forward
AI Open (AO), 2022
Ruben Cartuyvels
Graham Spinks
Marie-Francine Moens
OCL
328
30
0
04 Jan 2022
FLAVA: A Foundational Language And Vision Alignment Model
Amanpreet Singh
Ronghang Hu
Vedanuj Goswami
Guillaume Couairon
Wojciech Galuba
Marcus Rohrbach
Douwe Kiela
CLIP
VLM
489
912
0
08 Dec 2021
Emojich -- zero-shot emoji generation using Russian language: a technical report
Alex Shonenkov
Daria Bakshandaeva
Denis Dimitrov
Aleks D. Nikolich
VLM
245
5
0
04 Dec 2021
Show Your Work: Scratchpads for Intermediate Computation with Language Models
Maxwell Nye
Anders Andreassen
Guy Gur-Ari
Henryk Michalewski
Jacob Austin
...
Aitor Lewkowycz
Maarten Bosma
D. Luan
Charles Sutton
Augustus Odena
ReLM
LRM
681
976
0
30 Nov 2021
Many Heads but One Brain: Fusion Brain -- a Competition and a Single Multimodal Multitask Architecture
Daria Bakshandaeva
Denis Dimitrov
V.Ya. Arkhipkin
Alex Shonenkov
M. Potanin
...
Mikhail Martynov
Anton Voronov
Vera Davydova
E. Tutubalina
Aleksandr Petiushko
405
0
0
22 Nov 2021
Attention Approximates Sparse Distributed Memory
Neural Information Processing Systems (NeurIPS), 2021
Trenton Bricken
Cengiz Pehlevan
340
37
0
10 Nov 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
455
152
0
19 Oct 2021
The Power of Prompt Tuning for Low-Resource Semantic Parsing
Nathan Schucher
Siva Reddy
H. D. Vries
VLM
263
37
0
16 Oct 2021
Scaling Laws for the Few-Shot Adaptation of Pre-trained Image Classifiers
Gabriele Prato
Simon Guiroy
Ethan Caballero
Irina Rish
Sarath Chandar
VLM
286
12
0
13 Oct 2021
Pretrained Language Models are Symbolic Mathematics Solvers too!
Kimia Noorbakhsh
Modar Sulaiman
M. Sharifi
Kallol Roy
Pooyan Jamshidi
LRM
375
22
0
07 Oct 2021
Compositional generalization in semantic parsing with pretrained transformers
A. Orhan
303
8
0
30 Sep 2021
Cross-lingual Transfer of Monolingual Models
Evangelia Gogoulou
Ariel Ekgren
T. Isbister
Magnus Sahlgren
289
21
0
15 Sep 2021
TEASEL: A Transformer-Based Speech-Prefixed Language Model
Mehdi Arjmand
M. Dousti
H. Moradi
167
24
0
12 Sep 2021
Exploiting BERT For Multimodal Target Sentiment Classification Through Input Space Translation
Zaid Khan
Y. Fu
195
188
0
03 Aug 2021
Transformer-based Map Matching Model with Limited Ground-Truth Data using Transfer-Learning Approach
Zhixiong Jin
Jiwon Kim
H. Yeo
Seongjin Choi
286
34
0
01 Aug 2021
Perceiver IO: A General Architecture for Structured Inputs & Outputs
International Conference on Learning Representations (ICLR), 2021
Andrew Jaegle
Sebastian Borgeaud
Jean-Baptiste Alayrac
Carl Doersch
Catalin Ionescu
...
Olivier J. Hénaff
M. Botvinick
Andrew Zisserman
Oriol Vinyals
João Carreira
MLLM
VLM
GNN
825
773
0
30 Jul 2021
Don't Sweep your Learning Rate under the Rug: A Closer Look at Cross-modal Transfer of Pretrained Transformers
Dan Rothermel
Margaret Li
Tim Rocktaschel
Jakob N. Foerster
303
6
0
26 Jul 2021
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised Speech Representation Learning
IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2021
Samuel Kessler
Bethan Thomas
S. Karout
SSL
194
31
0
26 Jul 2021
Multimodal Few-Shot Learning with Frozen Language Models
Neural Information Processing Systems (NeurIPS), 2021
Maria Tsimpoukelli
Jacob Menick
Serkan Cabi
S. M. Ali Eslami
Oriol Vinyals
Felix Hill
MLLM
661
938
0
25 Jun 2021
Going Beyond Linear Transformers with Recurrent Fast Weight Programmers
Neural Information Processing Systems (NeurIPS), 2021
Kazuki Irie
Imanol Schlag
Róbert Csordás
Jürgen Schmidhuber
278
79
0
11 Jun 2021
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning
Computer Vision and Pattern Recognition (CVPR), 2021
Liangqiong Qu
Yuyin Zhou
Paul Pu Liang
Yingda Xia
Feifei Wang
Ehsan Adeli
L. Fei-Fei
D. Rubin
FedML
AI4CE
476
219
0
10 Jun 2021
Image2Point: 3D Point-Cloud Understanding with 2D Image Pretrained Models
European Conference on Computer Vision (ECCV), 2021
Chenfeng Xu
Shijia Yang
Tomer Galanti
Bichen Wu
Xiangyu Yue
Bohan Zhai
Wei Zhan
Peter Vajda
Kurt Keutzer
Masayoshi Tomizuka
3DPC
241
62
0
08 Jun 2021
Decision Transformer: Reinforcement Learning via Sequence Modeling
Neural Information Processing Systems (NeurIPS), 2021
Lili Chen
Kevin Lu
Aravind Rajeswaran
Kimin Lee
Aditya Grover
Michael Laskin
Pieter Abbeel
A. Srinivas
Igor Mordatch
OffRL
698
2,126
0
02 Jun 2021
Choose a Transformer: Fourier or Galerkin
Neural Information Processing Systems (NeurIPS), 2021
Shuhao Cao
555
373
0
31 May 2021
Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Mozhdeh Gheini
Xiang Ren
Jonathan May
LRM
388
174
0
18 Apr 2021
Generalization over different cellular automata rules learned by a deep feed-forward neural network
Marcel Aach
J. Goebbert
J. Jitsev
AI4CE
185
6
0
27 Mar 2021
Previous
1
2
3
4
Next
Page 3 of 4