ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.05247
  4. Cited By
Pretrained Transformers as Universal Computation Engines
v1v2 (latest)

Pretrained Transformers as Universal Computation Engines

9 March 2021
Kevin Lu
Aditya Grover
Pieter Abbeel
Igor Mordatch
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)

Papers citing "Pretrained Transformers as Universal Computation Engines"

50 / 151 papers shown
Training Transitive and Commutative Multimodal Transformers with LoReTTa
Training Transitive and Commutative Multimodal Transformers with LoReTTaNeural Information Processing Systems (NeurIPS), 2023
Manuel Tran
Yashin Dicente Cid
Amal Lahiani
Fabian J. Theis
Tingying Peng
Eldad Klaiman
328
4
0
23 May 2023
Introspective Tips: Large Language Model for In-Context Decision Making
Introspective Tips: Large Language Model for In-Context Decision Making
Liting Chen
Lu Wang
Hang Dong
Yali Du
Jie Yan
...
Lu Wang
Si Qin
Saravan Rajmohan
Qingwei Lin
Dongmei Zhang
LLMAGLRM
268
32
0
19 May 2023
Semantic Composition in Visually Grounded Language Models
Semantic Composition in Visually Grounded Language Models
Rohan Pandey
CoGe
206
1
0
15 May 2023
Efficient Feature Distillation for Zero-shot Annotation Object Detection
Efficient Feature Distillation for Zero-shot Annotation Object DetectionIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2023
Zhuoming Liu
Xuefeng Hu
Ram Nevatia
VLMObjD
259
3
0
21 Mar 2023
Contrastive Alignment of Vision to Language Through Parameter-Efficient
  Transfer Learning
Contrastive Alignment of Vision to Language Through Parameter-Efficient Transfer LearningInternational Conference on Learning Representations (ICLR), 2023
Zaid Khan
Yun Fu
VLM
182
20
0
21 Mar 2023
Merging Decision Transformers: Weight Averaging for Forming Multi-Task
  Policies
Merging Decision Transformers: Weight Averaging for Forming Multi-Task PoliciesIEEE International Conference on Robotics and Automation (ICRA), 2023
Daniel Lawson
A. H. Qureshi
MoMeOffRL
379
14
0
14 Mar 2023
Foundation Models for Decision Making: Problems, Methods, and
  Opportunities
Foundation Models for Decision Making: Problems, Methods, and Opportunities
Sherry Yang
Ofir Nachum
Yilun Du
Jason W. Wei
Pieter Abbeel
Dale Schuurmans
LM&RoOffRLLRMAI4CE
397
216
0
07 Mar 2023
PaLM-E: An Embodied Multimodal Language Model
PaLM-E: An Embodied Multimodal Language ModelInternational Conference on Machine Learning (ICML), 2023
Danny Driess
F. Xia
Mehdi S. M. Sajjadi
Corey Lynch
Aakanksha Chowdhery
...
Marc Toussaint
Klaus Greff
Andy Zeng
Igor Mordatch
Peter R. Florence
LM&Ro
479
2,258
0
06 Mar 2023
Optical Transformers
Optical Transformers
Maxwell G. Anderson
Shifan Ma
Tianyu Wang
Logan G. Wright
Peter L. McMahon
163
36
0
20 Feb 2023
Efficiency 360: Efficient Vision Transformers
Efficiency 360: Efficient Vision Transformers
Badri N. Patro
Vijay Srinivas Agneeswaran
409
7
0
16 Feb 2023
Knowledge from Large-Scale Protein Contact Prediction Models Can Be
  Transferred to the Data-Scarce RNA Contact Prediction Task
Knowledge from Large-Scale Protein Contact Prediction Models Can Be Transferred to the Data-Scarce RNA Contact Prediction TaskInternational Conference on Pattern Recognition (ICPR), 2023
Yiren Jian
Chongyang Gao
Chen Zeng
Yunjie Zhao
Soroush Vosoughi
352
1
0
13 Feb 2023
Language Quantized AutoEncoders: Towards Unsupervised Text-Image
  Alignment
Language Quantized AutoEncoders: Towards Unsupervised Text-Image AlignmentNeural Information Processing Systems (NeurIPS), 2023
Hao Liu
Wilson Yan
Pieter Abbeel
254
34
0
02 Feb 2023
Grounding Language Models to Images for Multimodal Inputs and Outputs
Grounding Language Models to Images for Multimodal Inputs and OutputsInternational Conference on Machine Learning (ICML), 2023
Jing Yu Koh
Ruslan Salakhutdinov
Daniel Fried
MLLM
448
151
0
31 Jan 2023
Continuous Spatiotemporal Transformers
Continuous Spatiotemporal TransformersInternational Conference on Machine Learning (ICML), 2023
Antonio H. O. Fonseca
E. Zappala
J. O. Caro
David van Dijk
176
10
0
31 Jan 2023
ClimaX: A foundation model for weather and climate
ClimaX: A foundation model for weather and climateInternational Conference on Machine Learning (ICML), 2023
Tung Nguyen
Johannes Brandstetter
Ashish Kapoor
Jayesh K. Gupta
Aditya Grover
AI4ClAI4CE
584
374
0
24 Jan 2023
A Survey on Transformers in Reinforcement Learning
A Survey on Transformers in Reinforcement Learning
Wenzhe Li
Hao Luo
Zichuan Lin
Chongjie Zhang
Zongqing Lu
Deheng Ye
OffRLMUAI4CE
547
72
0
08 Jan 2023
Evaluating Step-by-Step Reasoning through Symbolic Verification
Evaluating Step-by-Step Reasoning through Symbolic Verification
Yi-Fan Zhang
Hanlin Zhang
Li Erran Li
Eric P. Xing
ReLMLRM
285
10
0
16 Dec 2022
Vision Transformers are Parameter-Efficient Audio-Visual Learners
Vision Transformers are Parameter-Efficient Audio-Visual LearnersComputer Vision and Pattern Recognition (CVPR), 2022
Yan-Bo Lin
Yi-Lin Sung
Jie Lei
Joey Tianyi Zhou
Gedas Bertasius
324
110
0
15 Dec 2022
Deep representation learning: Fundamentals, Perspectives, Applications,
  and Open Challenges
Deep representation learning: Fundamentals, Perspectives, Applications, and Open Challenges
K. T. Baghaei
Amirreza Payandeh
Pooya Fayyazsanavi
Shahram Rahimi
Zhiqian Chen
Somayeh Bakhtiari Ramezani
FaMLAI4TS
226
10
0
27 Nov 2022
I Can't Believe There's No Images! Learning Visual Tasks Using only
  Language Supervision
I Can't Believe There's No Images! Learning Visual Tasks Using only Language SupervisionIEEE International Conference on Computer Vision (ICCV), 2022
Sophia Gu
Christopher Clark
Aniruddha Kembhavi
VLM
354
36
0
17 Nov 2022
On the Effect of Pre-training for Transformer in Different Modality on
  Offline Reinforcement Learning
On the Effect of Pre-training for Transformer in Different Modality on Offline Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2022
S. Takagi
OffRL
187
8
0
17 Nov 2022
Metaphors We Learn By
Metaphors We Learn By
Roland Memisevic
204
0
0
11 Nov 2022
What is Wrong with Language Models that Can Not Tell a Story?
What is Wrong with Language Models that Can Not Tell a Story?
Ivan P. Yamshchikov
Alexey Tikhonov
246
10
0
09 Nov 2022
Pretraining in Deep Reinforcement Learning: A Survey
Pretraining in Deep Reinforcement Learning: A Survey
Zhihui Xie
Zichuan Lin
Junyou Li
Shuai Li
Deheng Ye
OffRLOnRLAI4CE
242
30
0
08 Nov 2022
LMPriors: Pre-Trained Language Models as Task-Specific Priors
LMPriors: Pre-Trained Language Models as Task-Specific Priors
Kristy Choi
Chris Cundy
Sanjari Srivastava
Stefano Ermon
BDL
237
57
0
22 Oct 2022
Equi-Tuning: Group Equivariant Fine-Tuning of Pretrained Models
Equi-Tuning: Group Equivariant Fine-Tuning of Pretrained ModelsAAAI Conference on Artificial Intelligence (AAAI), 2022
Sourya Basu
P. Sattigeri
Karthikeyan N. Ramamurthy
Vijil Chenthamarakshan
Kush R. Varshney
Lav Varshney
Payel Das
254
24
0
13 Oct 2022
Reliable Conditioning of Behavioral Cloning for Offline Reinforcement
  Learning
Reliable Conditioning of Behavioral Cloning for Offline Reinforcement Learning
Tung Nguyen
Qinqing Zheng
Aditya Grover
OffRL
306
7
0
11 Oct 2022
Generating Executable Action Plans with Environmentally-Aware Language
  Models
Generating Executable Action Plans with Environmentally-Aware Language ModelsIEEE/RJS International Conference on Intelligent RObots and Systems (IROS), 2022
Maitrey Gramopadhye
D. Szafir
LM&RoLLMAG
325
38
0
10 Oct 2022
Understanding HTML with Large Language Models
Understanding HTML with Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Izzeddin Gur
Ofir Nachum
Yingjie Miao
Mustafa Safdari
Austin Huang
Aakanksha Chowdhery
Sharan Narang
Noah Fiedel
Aleksandra Faust
AI4CE
505
83
0
08 Oct 2022
Linearly Mapping from Image to Text Space
Linearly Mapping from Image to Text SpaceInternational Conference on Learning Representations (ICLR), 2022
Jack Merullo
Louis Castricato
Carsten Eickhoff
Ellie Pavlick
VLM
1.2K
145
0
30 Sep 2022
Downstream Datasets Make Surprisingly Good Pretraining Corpora
Downstream Datasets Make Surprisingly Good Pretraining CorporaAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Kundan Krishna
Saurabh Garg
Jeffrey P. Bigham
Zachary Chase Lipton
228
37
0
28 Sep 2022
Disentangling Transfer in Continual Reinforcement Learning
Disentangling Transfer in Continual Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2022
Maciej Wołczyk
Michal Zajkac
Razvan Pascanu
Lukasz Kuciñski
Piotr Milo's
CLL
269
36
0
28 Sep 2022
MonoByte: A Pool of Monolingual Byte-level Language Models
MonoByte: A Pool of Monolingual Byte-level Language ModelsInternational Conference on Computational Linguistics (COLING), 2022
Hugo Queiroz Abonizio
Leandro Rodrigues de Souza
R. Lotufo
Rodrigo Nogueira
181
2
0
22 Sep 2022
Non-Linguistic Supervision for Contrastive Learning of Sentence
  Embeddings
Non-Linguistic Supervision for Contrastive Learning of Sentence EmbeddingsNeural Information Processing Systems (NeurIPS), 2022
Yiren Jian
Chongyang Gao
Soroush Vosoughi
SSL
253
16
0
20 Sep 2022
OmniVL:One Foundation Model for Image-Language and Video-Language Tasks
OmniVL:One Foundation Model for Image-Language and Video-Language TasksNeural Information Processing Systems (NeurIPS), 2022
Junke Wang
Dongdong Chen
Zuxuan Wu
Chong Luo
Luowei Zhou
Yucheng Zhao
Yujia Xie
Ce Liu
Yu-Gang Jiang
Lu Yuan
MLLMVLM
294
178
0
15 Sep 2022
Foundations and Trends in Multimodal Machine Learning: Principles,
  Challenges, and Open Questions
Foundations and Trends in Multimodal Machine Learning: Principles, Challenges, and Open QuestionsACM Computing Surveys (ACM CSUR), 2022
Paul Pu Liang
Amir Zadeh
Louis-Philippe Morency
315
169
0
07 Sep 2022
Exploring and Evaluating Personalized Models for Code Generation
Exploring and Evaluating Personalized Models for Code Generation
Andrei Zlotchevski
Dawn Drain
Alexey Svyatkovskiy
Colin B. Clement
Neel Sundaresan
Michele Tufano
188
14
0
29 Aug 2022
What Can Transformers Learn In-Context? A Case Study of Simple Function
  Classes
What Can Transformers Learn In-Context? A Case Study of Simple Function ClassesNeural Information Processing Systems (NeurIPS), 2022
Shivam Garg
Dimitris Tsipras
Abigail Z. Jacobs
Gregory Valiant
658
676
0
01 Aug 2022
Unsupervised Domain Adaptation for Video Transformers in Action
  Recognition
Unsupervised Domain Adaptation for Video Transformers in Action RecognitionInternational Conference on Pattern Recognition (ICPR), 2022
Victor G. Turrisi da Costa
Giacomo Zara
Paolo Rota
Thiago Oliveira-Santos
Andrii Zadaianchuk
Vittorio Murino
Elisa Ricci
187
17
0
26 Jul 2022
Learning Visual Representation from Modality-Shared Contrastive
  Language-Image Pre-training
Learning Visual Representation from Modality-Shared Contrastive Language-Image Pre-trainingEuropean Conference on Computer Vision (ECCV), 2022
Haoxuan You
Luowei Zhou
Bin Xiao
Noel Codella
Yu Cheng
Ruochen Xu
Shih-Fu Chang
Lu Yuan
CLIPVLM
221
56
0
26 Jul 2022
Hidden Progress in Deep Learning: SGD Learns Parities Near the
  Computational Limit
Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational LimitNeural Information Processing Systems (NeurIPS), 2022
Boaz Barak
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Eran Malach
Cyril Zhang
382
157
0
18 Jul 2022
Transformer Neural Processes: Uncertainty-Aware Meta Learning Via
  Sequence Modeling
Transformer Neural Processes: Uncertainty-Aware Meta Learning Via Sequence ModelingInternational Conference on Machine Learning (ICML), 2022
Tung Nguyen
Aditya Grover
BDLUQCV
298
136
0
09 Jul 2022
Big Learning
Big Learning
Yulai Cong
Miaoyun Zhao
AI4CE
394
0
0
08 Jul 2022
CASHformer: Cognition Aware SHape Transformer for Longitudinal Analysis
CASHformer: Cognition Aware SHape Transformer for Longitudinal AnalysisInternational Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2022
Ignacio Sarasua
Sebastian Polsterl
Christian Wachinger
MedIm
209
2
0
05 Jul 2022
TTS-CGAN: A Transformer Time-Series Conditional GAN for Biosignal Data
  Augmentation
TTS-CGAN: A Transformer Time-Series Conditional GAN for Biosignal Data Augmentation
Xiaomin Li
Anne H. H. Ngu
V. Metsis
MedIm
175
39
0
28 Jun 2022
ProGen2: Exploring the Boundaries of Protein Language Models
ProGen2: Exploring the Boundaries of Protein Language ModelsCell Systems (Cell Syst.), 2022
Erik Nijkamp
Jeffrey A. Ruffolo
Eli N. Weinstein
Nikhil Naik
Ali Madani
AI4TS
190
427
0
27 Jun 2022
LIFT: Language-Interfaced Fine-Tuning for Non-Language Machine Learning
  Tasks
LIFT: Language-Interfaced Fine-Tuning for Non-Language Machine Learning TasksNeural Information Processing Systems (NeurIPS), 2022
Tuan Dinh
Yuchen Zeng
Ruisu Zhang
Ziqian Lin
Michael Gira
Shashank Rajput
Jy-yong Sohn
Dimitris Papailiopoulos
Kangwook Lee
LMTD
576
172
0
14 Jun 2022
CyCLIP: Cyclic Contrastive Language-Image Pretraining
CyCLIP: Cyclic Contrastive Language-Image PretrainingNeural Information Processing Systems (NeurIPS), 2022
Shashank Goel
Hritik Bansal
S. Bhatia
Ryan Rossi
Vishwa Vinay
Aditya Grover
CLIPVLM
522
166
0
28 May 2022
History Compression via Language Models in Reinforcement Learning
History Compression via Language Models in Reinforcement LearningInternational Conference on Machine Learning (ICML), 2022
Fabian Paischer
Thomas Adler
Vihang Patil
Angela Bitto-Nemling
Markus Holzleitner
Sebastian Lehner
Hamid Eghbalzadeh
Sepp Hochreiter
OffRLAI4TS
518
51
0
24 May 2022
Housekeep: Tidying Virtual Households using Commonsense Reasoning
Housekeep: Tidying Virtual Households using Commonsense ReasoningEuropean Conference on Computer Vision (ECCV), 2022
Yash Kant
Arun Ramachandran
Sriram Yenamandra
Igor Gilitschenski
Dhruv Batra
Andrew Szot
Harsh Agrawal
LM&RoLRM
416
85
0
22 May 2022
Previous
1234
Next
Page 2 of 4