ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.07897
  4. Cited By
Entropic gradient descent algorithms and wide flat minima

Entropic gradient descent algorithms and wide flat minima

14 June 2020
Fabrizio Pittorino
C. Lucibello
Christoph Feinauer
Gabriele Perugini
Carlo Baldassi
Elizaveta Demyanenko
R. Zecchina
    ODL
    MLT
ArXivPDFHTML

Papers citing "Entropic gradient descent algorithms and wide flat minima"

8 / 8 papers shown
Title
Revisiting Confidence Estimation: Towards Reliable Failure Prediction
Revisiting Confidence Estimation: Towards Reliable Failure Prediction
Fei Zhu
Xu-Yao Zhang
Zhen Cheng
Cheng-Lin Liu
UQCV
49
10
0
05 Mar 2024
FlatNAS: optimizing Flatness in Neural Architecture Search for
  Out-of-Distribution Robustness
FlatNAS: optimizing Flatness in Neural Architecture Search for Out-of-Distribution Robustness
Matteo Gambella
Fabrizio Pittorino
Manuel Roveri
OOD
42
3
0
29 Feb 2024
Phase transitions in the mini-batch size for sparse and dense two-layer
  neural networks
Phase transitions in the mini-batch size for sparse and dense two-layer neural networks
Raffaele Marino
F. Ricci-Tersenghi
30
14
0
10 May 2023
Rethinking Confidence Calibration for Failure Prediction
Rethinking Confidence Calibration for Failure Prediction
Fei Zhu
Zhen Cheng
Xu-Yao Zhang
Cheng-Lin Liu
UQCV
14
39
0
06 Mar 2023
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
Keller Jordan
Hanie Sedghi
O. Saukh
R. Entezari
Behnam Neyshabur
MoMe
46
94
0
15 Nov 2022
Understanding the effect of sparsity on neural networks robustness
Understanding the effect of sparsity on neural networks robustness
Lukas Timpl
R. Entezari
Hanie Sedghi
Behnam Neyshabur
O. Saukh
31
11
0
22 Jun 2022
Learning through atypical "phase transitions" in overparameterized
  neural networks
Learning through atypical "phase transitions" in overparameterized neural networks
Carlo Baldassi
Clarissa Lauditi
Enrico M. Malatesta
R. Pacelli
Gabriele Perugini
R. Zecchina
26
26
0
01 Oct 2021
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp
  Minima
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
281
2,889
0
15 Sep 2016
1