ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1312.4314
  4. Cited By
Learning Factored Representations in a Deep Mixture of Experts

Learning Factored Representations in a Deep Mixture of Experts

16 December 2013
David Eigen
MarcÁurelio Ranzato
Ilya Sutskever
    MoE
ArXivPDFHTML

Papers citing "Learning Factored Representations in a Deep Mixture of Experts"

32 / 82 papers shown
Title
COUCH: Towards Controllable Human-Chair Interactions
COUCH: Towards Controllable Human-Chair Interactions
Xiaohan Zhang
Bharat Lal Bhatnagar
V. Guzov
Sebastian Starke
Gerard Pons-Moll
52
97
0
01 May 2022
Residual Mixture of Experts
Residual Mixture of Experts
Lemeng Wu
Mengchen Liu
Yinpeng Chen
Dongdong Chen
Xiyang Dai
Lu Yuan
MoE
22
36
0
20 Apr 2022
Beyond Fixation: Dynamic Window Visual Transformer
Beyond Fixation: Dynamic Window Visual Transformer
Pengzhen Ren
Changlin Li
Guangrun Wang
Yun Xiao
Qing Du
Xiaodan Liang
Qing Du Xiaodan Liang Xiaojun Chang
ViT
28
32
0
24 Mar 2022
A Unified Framework for Campaign Performance Forecasting in Online
  Display Advertising
A Unified Framework for Campaign Performance Forecasting in Online Display Advertising
Jun Chen
Cheng Chen
Huayue Zhang
Qing Tan
26
2
0
24 Feb 2022
ST-MoE: Designing Stable and Transferable Sparse Expert Models
ST-MoE: Designing Stable and Transferable Sparse Expert Models
Barret Zoph
Irwan Bello
Sameer Kumar
Nan Du
Yanping Huang
J. Dean
Noam M. Shazeer
W. Fedus
MoE
24
182
0
17 Feb 2022
Unified Scaling Laws for Routed Language Models
Unified Scaling Laws for Routed Language Models
Aidan Clark
Diego de Las Casas
Aurelia Guy
A. Mensch
Michela Paganini
...
Oriol Vinyals
Jack W. Rae
Erich Elsen
Koray Kavukcuoglu
Karen Simonyan
MoE
27
177
0
02 Feb 2022
EvoMoE: An Evolutional Mixture-of-Experts Training Framework via
  Dense-To-Sparse Gate
EvoMoE: An Evolutional Mixture-of-Experts Training Framework via Dense-To-Sparse Gate
Xiaonan Nie
Xupeng Miao
Shijie Cao
Lingxiao Ma
Qibin Liu
Jilong Xue
Youshan Miao
Yi Liu
Zhi-Xin Yang
Bin Cui
MoMe
MoE
29
22
0
29 Dec 2021
Universal Simultaneous Machine Translation with Mixture-of-Experts
  Wait-k Policy
Universal Simultaneous Machine Translation with Mixture-of-Experts Wait-k Policy
Shaolei Zhang
Yang Feng
MoE
28
39
0
11 Sep 2021
Mixed SIGNals: Sign Language Production via a Mixture of Motion
  Primitives
Mixed SIGNals: Sign Language Production via a Mixture of Motion Primitives
Ben Saunders
Necati Cihan Camgöz
Richard Bowden
SLR
27
50
0
23 Jul 2021
Zoo-Tuning: Adaptive Transfer from a Zoo of Models
Zoo-Tuning: Adaptive Transfer from a Zoo of Models
Yang Shu
Zhi Kou
Zhangjie Cao
Jianmin Wang
Mingsheng Long
29
44
0
29 Jun 2021
Scaling Vision with Sparse Mixture of Experts
Scaling Vision with Sparse Mixture of Experts
C. Riquelme
J. Puigcerver
Basil Mustafa
Maxim Neumann
Rodolphe Jenatton
André Susano Pinto
Daniel Keysers
N. Houlsby
MoE
17
575
0
10 Jun 2021
Graph Classification by Mixture of Diverse Experts
Graph Classification by Mixture of Diverse Experts
Fenyu Hu
Liping Wang
Shu Wu
Liang Wang
Tieniu Tan
42
10
0
29 Mar 2021
Switch Transformers: Scaling to Trillion Parameter Models with Simple
  and Efficient Sparsity
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
W. Fedus
Barret Zoph
Noam M. Shazeer
MoE
11
2,075
0
11 Jan 2021
Efficient Continual Learning with Modular Networks and Task-Driven
  Priors
Efficient Continual Learning with Modular Networks and Task-Driven Priors
Tom Véniat
Ludovic Denoyer
MarcÁurelio Ranzato
CLL
30
97
0
23 Dec 2020
THIN: THrowable Information Networks and Application for Facial
  Expression Recognition In The Wild
THIN: THrowable Information Networks and Application for Facial Expression Recognition In The Wild
Estèphe Arnaud
Arnaud Dapogny
Kévin Bailly
CVBM
29
23
0
15 Oct 2020
Scalable Transfer Learning with Expert Models
Scalable Transfer Learning with Expert Models
J. Puigcerver
C. Riquelme
Basil Mustafa
Cédric Renggli
André Susano Pinto
Sylvain Gelly
Daniel Keysers
N. Houlsby
34
62
0
28 Sep 2020
Multi-modal Experts Network for Autonomous Driving
Multi-modal Experts Network for Autonomous Driving
Shihong Fang
A. Choromańska
MoE
28
5
0
18 Sep 2020
Towards Crowdsourced Training of Large Neural Networks using
  Decentralized Mixture-of-Experts
Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts
Max Ryabinin
Anton I. Gusev
FedML
27
48
0
10 Feb 2020
Tree-gated Deep Mixture-of-Experts For Pose-robust Face Alignment
Tree-gated Deep Mixture-of-Experts For Pose-robust Face Alignment
Estèphe Arnaud
Arnaud Dapogny
Kévin Bailly
CVBM
29
10
0
21 Oct 2019
Convergence Rates for Gaussian Mixtures of Experts
Convergence Rates for Gaussian Mixtures of Experts
Nhat Ho
Chiao-Yu Yang
Michael I. Jordan
21
40
0
09 Jul 2019
Task-Driven Modular Networks for Zero-Shot Compositional Learning
Task-Driven Modular Networks for Zero-Shot Compositional Learning
Senthil Purushwalkam
Maximilian Nickel
Abhinav Gupta
MarcÁurelio Ranzato
35
170
0
15 May 2019
Question Guided Modular Routing Networks for Visual Question Answering
Question Guided Modular Routing Networks for Visual Question Answering
Yanze Wu
Qiang Sun
Jianqi Ma
Bin Li
Yanwei Fu
Yao Peng
Xiangyang Xue
23
1
0
17 Apr 2019
Deep Multi-modal Object Detection and Semantic Segmentation for
  Autonomous Driving: Datasets, Methods, and Challenges
Deep Multi-modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges
Di Feng
Christian Haase-Schuetz
Lars Rosenbaum
Heinz Hertlein
Claudius Gläser
Fabian Duffhauss
W. Wiesbeck
Klaus C. J. Dietmayer
3DPC
41
987
0
21 Feb 2019
Mixture Models for Diverse Machine Translation: Tricks of the Trade
Mixture Models for Diverse Machine Translation: Tricks of the Trade
T. Shen
Myle Ott
Michael Auli
MarcÁurelio Ranzato
MoE
33
148
0
20 Feb 2019
Mode Normalization
Mode Normalization
Lucas Deecke
Iain Murray
Hakan Bilen
OOD
29
33
0
12 Oct 2018
Rank of Experts: Detection Network Ensemble
Rank of Experts: Detection Network Ensemble
Seung-Hwan Bae
Youngwan Lee
Y. Jo
Yuseok Bae
Joong-won Hwang
ObjD
28
5
0
01 Dec 2017
FiLM: Visual Reasoning with a General Conditioning Layer
FiLM: Visual Reasoning with a General Conditioning Layer
Ethan Perez
Florian Strub
H. D. Vries
Vincent Dumoulin
Aaron Courville
FAtt
AIMat
OffRL
AI4CE
107
2,151
0
22 Sep 2017
Deep Mixture of Diverse Experts for Large-Scale Visual Recognition
Deep Mixture of Diverse Experts for Large-Scale Visual Recognition
Tianyi Zhao
Jun-chen Yu
Zhenzhong Kuang
Wei Zhang
Jianping Fan
MoE
32
13
0
24 Jun 2017
Outrageously Large Neural Networks: The Sparsely-Gated
  Mixture-of-Experts Layer
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer
Noam M. Shazeer
Azalia Mirhoseini
Krzysztof Maziarz
Andy Davis
Quoc V. Le
Geoffrey E. Hinton
J. Dean
MoE
55
2,513
0
23 Jan 2017
Dynamic Deep Neural Networks: Optimizing Accuracy-Efficiency Trade-offs
  by Selective Execution
Dynamic Deep Neural Networks: Optimizing Accuracy-Efficiency Trade-offs by Selective Execution
Lanlan Liu
Jia Deng
24
200
0
02 Jan 2017
A Survey of Inductive Biases for Factorial Representation-Learning
A Survey of Inductive Biases for Factorial Representation-Learning
Karl Ridgeway
DRL
CML
29
76
0
15 Dec 2016
Opponent Modeling in Deep Reinforcement Learning
Opponent Modeling in Deep Reinforcement Learning
He He
Jordan L. Boyd-Graber
Kevin Kwok
Hal Daumé III
BDL
29
324
0
18 Sep 2016
Previous
12