ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.05247
  4. Cited By
Pretrained Transformers as Universal Computation Engines
v1v2 (latest)

Pretrained Transformers as Universal Computation Engines

9 March 2021
Kevin Lu
Aditya Grover
Pieter Abbeel
Igor Mordatch
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)

Papers citing "Pretrained Transformers as Universal Computation Engines"

50 / 151 papers shown
Selection-Inference: Exploiting Large Language Models for Interpretable
  Logical Reasoning
Selection-Inference: Exploiting Large Language Models for Interpretable Logical ReasoningInternational Conference on Learning Representations (ICLR), 2022
Antonia Creswell
Murray Shanahan
I. Higgins
ReLMLRM
313
433
0
19 May 2022
Clinical Prompt Learning with Frozen Language Models
Clinical Prompt Learning with Frozen Language Models
Niall Taylor
Yi Zhang
Dan W Joyce
A. Nevado-Holgado
Andrey Kormilitzin
VLMLM&MA
144
37
0
11 May 2022
StorSeismic: A new paradigm in deep learning for seismic processing
StorSeismic: A new paradigm in deep learning for seismic processingIEEE Transactions on Geoscience and Remote Sensing (IEEE TGRS), 2022
R. Harsuko
T. Alkhalifah
131
44
0
30 Apr 2022
Efficient Architecture Search for Diverse Tasks
Efficient Architecture Search for Diverse TasksNeural Information Processing Systems (NeurIPS), 2022
Jun Shen
M. Khodak
Ameet Talwalkar
293
38
0
15 Apr 2022
Causal Transformer for Estimating Counterfactual Outcomes
Causal Transformer for Estimating Counterfactual OutcomesInternational Conference on Machine Learning (ICML), 2022
Valentyn Melnychuk
Dennis Frauen
Stefan Feuerriegel
CML
275
131
0
14 Apr 2022
Single-Stream Multi-Level Alignment for Vision-Language Pretraining
Single-Stream Multi-Level Alignment for Vision-Language PretrainingEuropean Conference on Computer Vision (ECCV), 2022
Zaid Khan
B. Vijaykumar
Xiang Yu
S. Schulter
Manmohan Chandraker
Y. Fu
CLIPVLM
356
22
0
27 Mar 2022
Linking Emergent and Natural Languages via Corpus Transfer
Linking Emergent and Natural Languages via Corpus TransferInternational Conference on Learning Representations (ICLR), 2022
Shunyu Yao
Mo Yu
Yang Zhang
Karthik Narasimhan
J. Tenenbaum
Chuang Gan
235
19
0
24 Mar 2022
Hyperbolic Vision Transformers: Combining Improvements in Metric
  Learning
Hyperbolic Vision Transformers: Combining Improvements in Metric LearningComputer Vision and Pattern Recognition (CVPR), 2022
Aleksandr Ermolov
L. Mirvakhabova
Valentin Khrulkov
Andrii Zadaianchuk
Ivan Oseledets
319
145
0
21 Mar 2022
Pretraining with Artificial Language: Studying Transferable Knowledge in
  Language Models
Pretraining with Artificial Language: Studying Transferable Knowledge in Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Ryokan Ri
Yoshimasa Tsuruoka
231
36
0
19 Mar 2022
Summarizing a virtual robot's past actions in natural language
Summarizing a virtual robot's past actions in natural language
Chad DeChant
Daniel Bauer
LM&Ro
172
4
0
13 Mar 2022
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken
  Question Answering
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question AnsweringInterspeech (Interspeech), 2022
Guan-Ting Lin
Yung-Sung Chuang
Ho-Lam Chung
Shu-Wen Yang
Hsuan-Jui Chen
Shuyan Dong
Shang-Wen Li
Abdel-rahman Mohamed
Hung-yi Lee
Lin-Shan Lee
329
25
0
09 Mar 2022
High-Modality Multimodal Transformer: Quantifying Modality & Interaction
  Heterogeneity for High-Modality Representation Learning
High-Modality Multimodal Transformer: Quantifying Modality & Interaction Heterogeneity for High-Modality Representation Learning
Paul Pu Liang
Yiwei Lyu
Xiang Fan
Jeffrey Tsaw
Yudong Liu
Shentong Mo
Dani Yogatama
Louis-Philippe Morency
Ruslan Salakhutdinov
230
43
0
02 Mar 2022
Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning
Model Reprogramming: Resource-Efficient Cross-Domain Machine LearningAAAI Conference on Artificial Intelligence (AAAI), 2022
Pin-Yu Chen
VLM
425
77
0
22 Feb 2022
MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
MineRL Diamond 2021 Competition: Overview, Results, and Lessons LearnedNeural Information Processing Systems (NeurIPS), 2022
Anssi Kanervisto
Stephanie Milani
Karolis Ramanauskas
Nicholay Topin
Zichuan Lin
...
Franccois Fleuret
Alexander Nikulin
Yury Belousov
Oleg Svidchenko
A. Shpilman
OffRL
245
35
0
17 Feb 2022
Online Decision Transformer
Online Decision TransformerInternational Conference on Machine Learning (ICML), 2022
Qinqing Zheng
Amy Zhang
Aditya Grover
OffRL
284
242
0
11 Feb 2022
TTS-GAN: A Transformer-based Time-Series Generative Adversarial Network
TTS-GAN: A Transformer-based Time-Series Generative Adversarial NetworkConference on Artificial Intelligence in Medicine in Europe (AIME), 2022
Xiaomin Li
V. Metsis
Huan Wang
Anne H. H. Ngu
139
140
0
06 Feb 2022
Pre-Trained Language Models for Interactive Decision-Making
Pre-Trained Language Models for Interactive Decision-MakingNeural Information Processing Systems (NeurIPS), 2022
Shuang Li
Xavier Puig
Chris Paxton
Yilun Du
Clinton Jia Wang
...
Anima Anandkumar
Jacob Andreas
Igor Mordatch
Antonio Torralba
Yuke Zhu
LM&Ro
423
307
0
03 Feb 2022
Regression Transformer: Concurrent sequence regression and generation
  for molecular language modeling
Regression Transformer: Concurrent sequence regression and generation for molecular language modelingNature Machine Intelligence (Nat. Mach. Intell.), 2022
Jannis Born
Matteo Manica
299
127
0
01 Feb 2022
Can Wikipedia Help Offline Reinforcement Learning?
Can Wikipedia Help Offline Reinforcement Learning?
Machel Reid
Yutaro Yamada
S. Gu
3DVRALMOffRL
656
101
0
28 Jan 2022
Pre-Trained Language Transformers are Universal Image Classifiers
Pre-Trained Language Transformers are Universal Image Classifiers
Rahul Goel
Modar Sulaiman
Kimia Noorbakhsh
Mahdieh Sharifi
Ragunath R. Sharma
Pooyan Jamshidi
Kallol Roy
ViT
134
3
0
25 Jan 2022
Language Models as Zero-Shot Planners: Extracting Actionable Knowledge
  for Embodied Agents
Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied AgentsInternational Conference on Machine Learning (ICML), 2022
Wenlong Huang
Pieter Abbeel
Deepak Pathak
Igor Mordatch
LM&Ro
322
1,419
0
18 Jan 2022
Video Transformers: A Survey
Video Transformers: A SurveyIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
Javier Selva
A. S. Johansen
Sergio Escalera
Kamal Nasrollahi
T. Moeslund
Albert Clapés
ViT
458
139
0
16 Jan 2022
CLIP-TD: CLIP Targeted Distillation for Vision-Language Tasks
CLIP-TD: CLIP Targeted Distillation for Vision-Language Tasks
Zhecan Wang
Noel Codella
Yen-Chun Chen
Luowei Zhou
Jianwei Yang
Xiyang Dai
Bin Xiao
Haoxuan You
Shih-Fu Chang
Lu Yuan
CLIPVLM
213
44
0
15 Jan 2022
Improved Input Reprogramming for GAN Conditioning
Improved Input Reprogramming for GAN Conditioning
Tuan Dinh
Daewon Seo
Zhixu Du
Liang Shang
Kangwook Lee
AI4CE
261
8
0
07 Jan 2022
Discrete and continuous representations and processing in deep learning:
  Looking forward
Discrete and continuous representations and processing in deep learning: Looking forwardAI Open (AO), 2022
Ruben Cartuyvels
Graham Spinks
Marie-Francine Moens
OCL
301
30
0
04 Jan 2022
FLAVA: A Foundational Language And Vision Alignment Model
FLAVA: A Foundational Language And Vision Alignment Model
Amanpreet Singh
Ronghang Hu
Vedanuj Goswami
Guillaume Couairon
Wojciech Galuba
Marcus Rohrbach
Douwe Kiela
CLIPVLM
383
873
0
08 Dec 2021
Emojich -- zero-shot emoji generation using Russian language: a
  technical report
Emojich -- zero-shot emoji generation using Russian language: a technical report
Alex Shonenkov
Daria Bakshandaeva
Denis Dimitrov
Aleks D. Nikolich
VLM
213
5
0
04 Dec 2021
Show Your Work: Scratchpads for Intermediate Computation with Language
  Models
Show Your Work: Scratchpads for Intermediate Computation with Language Models
Maxwell Nye
Anders Andreassen
Guy Gur-Ari
Henryk Michalewski
Jacob Austin
...
Aitor Lewkowycz
Maarten Bosma
D. Luan
Charles Sutton
Augustus Odena
ReLMLRM
547
927
0
30 Nov 2021
Many Heads but One Brain: Fusion Brain -- a Competition and a Single
  Multimodal Multitask Architecture
Many Heads but One Brain: Fusion Brain -- a Competition and a Single Multimodal Multitask Architecture
Daria Bakshandaeva
Denis Dimitrov
V.Ya. Arkhipkin
Alex Shonenkov
M. Potanin
...
Mikhail Martynov
Anton Voronov
Vera Davydova
E. Tutubalina
Aleksandr Petiushko
379
0
0
22 Nov 2021
Attention Approximates Sparse Distributed Memory
Attention Approximates Sparse Distributed MemoryNeural Information Processing Systems (NeurIPS), 2021
Trenton Bricken
Cengiz Pehlevan
316
37
0
10 Nov 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
355
150
0
19 Oct 2021
The Power of Prompt Tuning for Low-Resource Semantic Parsing
The Power of Prompt Tuning for Low-Resource Semantic Parsing
Nathan Schucher
Siva Reddy
H. D. Vries
VLM
243
36
0
16 Oct 2021
Scaling Laws for the Few-Shot Adaptation of Pre-trained Image
  Classifiers
Scaling Laws for the Few-Shot Adaptation of Pre-trained Image Classifiers
Gabriele Prato
Simon Guiroy
Ethan Caballero
Irina Rish
Sarath Chandar
VLM
239
12
0
13 Oct 2021
Pretrained Language Models are Symbolic Mathematics Solvers too!
Pretrained Language Models are Symbolic Mathematics Solvers too!
Kimia Noorbakhsh
Modar Sulaiman
M. Sharifi
Kallol Roy
Pooyan Jamshidi
LRM
292
22
0
07 Oct 2021
Compositional generalization in semantic parsing with pretrained
  transformers
Compositional generalization in semantic parsing with pretrained transformers
A. Orhan
248
8
0
30 Sep 2021
Cross-lingual Transfer of Monolingual Models
Cross-lingual Transfer of Monolingual Models
Evangelia Gogoulou
Ariel Ekgren
T. Isbister
Magnus Sahlgren
256
20
0
15 Sep 2021
TEASEL: A Transformer-Based Speech-Prefixed Language Model
TEASEL: A Transformer-Based Speech-Prefixed Language Model
Mehdi Arjmand
M. Dousti
H. Moradi
140
23
0
12 Sep 2021
Exploiting BERT For Multimodal Target Sentiment Classification Through
  Input Space Translation
Exploiting BERT For Multimodal Target Sentiment Classification Through Input Space Translation
Zaid Khan
Y. Fu
174
178
0
03 Aug 2021
Transformer-based Map Matching Model with Limited Ground-Truth Data
  using Transfer-Learning Approach
Transformer-based Map Matching Model with Limited Ground-Truth Data using Transfer-Learning Approach
Zhixiong Jin
Jiwon Kim
H. Yeo
Seongjin Choi
235
33
0
01 Aug 2021
Perceiver IO: A General Architecture for Structured Inputs & Outputs
Perceiver IO: A General Architecture for Structured Inputs & OutputsInternational Conference on Learning Representations (ICLR), 2021
Andrew Jaegle
Sebastian Borgeaud
Jean-Baptiste Alayrac
Carl Doersch
Catalin Ionescu
...
Olivier J. Hénaff
M. Botvinick
Andrew Zisserman
Oriol Vinyals
João Carreira
MLLMVLMGNN
659
730
0
30 Jul 2021
Don't Sweep your Learning Rate under the Rug: A Closer Look at
  Cross-modal Transfer of Pretrained Transformers
Don't Sweep your Learning Rate under the Rug: A Closer Look at Cross-modal Transfer of Pretrained Transformers
Dan Rothermel
Margaret Li
Tim Rocktaschel
Jakob N. Foerster
251
6
0
26 Jul 2021
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised
  Speech Representation Learning
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised Speech Representation LearningIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2021
Samuel Kessler
Bethan Thomas
S. Karout
SSL
167
31
0
26 Jul 2021
Multimodal Few-Shot Learning with Frozen Language Models
Multimodal Few-Shot Learning with Frozen Language ModelsNeural Information Processing Systems (NeurIPS), 2021
Maria Tsimpoukelli
Jacob Menick
Serkan Cabi
S. M. Ali Eslami
Oriol Vinyals
Felix Hill
MLLM
548
907
0
25 Jun 2021
Going Beyond Linear Transformers with Recurrent Fast Weight Programmers
Going Beyond Linear Transformers with Recurrent Fast Weight ProgrammersNeural Information Processing Systems (NeurIPS), 2021
Kazuki Irie
Imanol Schlag
Róbert Csordás
Jürgen Schmidhuber
254
77
0
11 Jun 2021
Rethinking Architecture Design for Tackling Data Heterogeneity in
  Federated Learning
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated LearningComputer Vision and Pattern Recognition (CVPR), 2021
Liangqiong Qu
Yuyin Zhou
Paul Pu Liang
Yingda Xia
Feifei Wang
Ehsan Adeli
L. Fei-Fei
D. Rubin
FedMLAI4CE
414
215
0
10 Jun 2021
Image2Point: 3D Point-Cloud Understanding with 2D Image Pretrained
  Models
Image2Point: 3D Point-Cloud Understanding with 2D Image Pretrained ModelsEuropean Conference on Computer Vision (ECCV), 2021
Chenfeng Xu
Shijia Yang
Tomer Galanti
Bichen Wu
Xiangyu Yue
Bohan Zhai
Wei Zhan
Peter Vajda
Kurt Keutzer
Masayoshi Tomizuka
3DPC
206
62
0
08 Jun 2021
Decision Transformer: Reinforcement Learning via Sequence Modeling
Decision Transformer: Reinforcement Learning via Sequence ModelingNeural Information Processing Systems (NeurIPS), 2021
Lili Chen
Kevin Lu
Aravind Rajeswaran
Kimin Lee
Aditya Grover
Michael Laskin
Pieter Abbeel
A. Srinivas
Igor Mordatch
OffRL
563
2,023
0
02 Jun 2021
Choose a Transformer: Fourier or Galerkin
Choose a Transformer: Fourier or GalerkinNeural Information Processing Systems (NeurIPS), 2021
Shuhao Cao
444
349
0
31 May 2021
Cross-Attention is All You Need: Adapting Pretrained Transformers for
  Machine Translation
Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine TranslationConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Mozhdeh Gheini
Xiang Ren
Jonathan May
LRM
319
165
0
18 Apr 2021
Generalization over different cellular automata rules learned by a deep
  feed-forward neural network
Generalization over different cellular automata rules learned by a deep feed-forward neural network
Marcel Aach
J. Goebbert
J. Jitsev
AI4CE
145
6
0
27 Mar 2021
Previous
1234
Next
Page 3 of 4