ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.05032
  4. Cited By
Decoupling Gating from Linearity

Decoupling Gating from Linearity

12 June 2019
Jonathan Fiat
Eran Malach
Shai Shalev-Shwartz
ArXivPDFHTML

Papers citing "Decoupling Gating from Linearity"

17 / 17 papers shown
Title
Implicit Bias in Matrix Factorization and its Explicit Realization in a New Architecture
Yikun Hou
Suvrit Sra
A. Yurtsever
34
0
0
28 Jan 2025
Improving the Expressive Power of Deep Neural Networks through Integral
  Activation Transform
Improving the Expressive Power of Deep Neural Networks through Integral Activation Transform
Zezhong Zhang
Feng Bao
Guannan Zhang
22
0
0
19 Dec 2023
Optimal Sets and Solution Paths of ReLU Networks
Optimal Sets and Solution Paths of ReLU Networks
Aaron Mishkin
Mert Pilanci
49
3
0
31 May 2023
Globally Gated Deep Linear Networks
Globally Gated Deep Linear Networks
Qianyi Li
H. Sompolinsky
AI4CE
27
10
0
31 Oct 2022
The Influence of Learning Rule on Representation Dynamics in Wide Neural
  Networks
The Influence of Learning Rule on Representation Dynamics in Wide Neural Networks
Blake Bordelon
Cengiz Pehlevan
41
22
0
05 Oct 2022
Unraveling Attention via Convex Duality: Analysis and Interpretations of
  Vision Transformers
Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers
Arda Sahiner
Tolga Ergen
Batu Mehmet Ozturkler
John M. Pauly
Morteza Mardani
Mert Pilanci
43
33
0
17 May 2022
Explicitising The Implicit Intrepretability of Deep Neural Networks Via
  Duality
Explicitising The Implicit Intrepretability of Deep Neural Networks Via Duality
Chandrashekar Lakshminarayanan
Ashutosh Kumar Singh
A. Rajkumar
AI4CE
31
1
0
01 Mar 2022
Fast Convex Optimization for Two-Layer ReLU Networks: Equivalent Model Classes and Cone Decompositions
Fast Convex Optimization for Two-Layer ReLU Networks: Equivalent Model Classes and Cone Decompositions
Aaron Mishkin
Arda Sahiner
Mert Pilanci
OffRL
77
30
0
02 Feb 2022
Provable Repair of Deep Neural Networks
Provable Repair of Deep Neural Networks
Matthew Sotoudeh
Aditya V. Thakur
AAML
21
70
0
09 Apr 2021
On the Generalization Power of Overfitted Two-Layer Neural Tangent
  Kernel Models
On the Generalization Power of Overfitted Two-Layer Neural Tangent Kernel Models
Peizhong Ju
Xiaojun Lin
Ness B. Shroff
MLT
29
10
0
09 Mar 2021
Towards Understanding Learning in Neural Networks with Linear Teachers
Towards Understanding Learning in Neural Networks with Linear Teachers
Roei Sarussi
Alon Brutzkus
Amir Globerson
FedML
MLT
57
21
0
07 Jan 2021
Understanding Boolean Function Learnability on Deep Neural Networks: PAC
  Learning Meets Neurosymbolic Models
Understanding Boolean Function Learnability on Deep Neural Networks: PAC Learning Meets Neurosymbolic Models
Márcio Nicolau
Anderson R. Tavares
Zhiwei Zhang
Pedro H. C. Avelar
J. Flach
Luís C. Lamb
Moshe Y. Vardi
NAI
33
2
0
13 Sep 2020
Neural Path Features and Neural Path Kernel : Understanding the role of
  gates in deep learning
Neural Path Features and Neural Path Kernel : Understanding the role of gates in deep learning
Chandrashekar Lakshminarayanan
Amit Singh
AI4CE
11
10
0
11 Jun 2020
An Optimization and Generalization Analysis for Max-Pooling Networks
An Optimization and Generalization Analysis for Max-Pooling Networks
Alon Brutzkus
Amir Globerson
MLT
AI4CE
16
4
0
22 Feb 2020
Learning Parities with Neural Networks
Learning Parities with Neural Networks
Amit Daniely
Eran Malach
24
76
0
18 Feb 2020
Deep Gated Networks: A framework to understand training and
  generalisation in deep learning
Deep Gated Networks: A framework to understand training and generalisation in deep learning
Chandrashekar Lakshminarayanan
Amit Singh
AI4CE
12
1
0
10 Feb 2020
Neural Networks Learning and Memorization with (almost) no
  Over-Parameterization
Neural Networks Learning and Memorization with (almost) no Over-Parameterization
Amit Daniely
18
33
0
22 Nov 2019
1