ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.07344
  4. Cited By
Temperature check: theory and practice for training models with
  softmax-cross-entropy losses

Temperature check: theory and practice for training models with softmax-cross-entropy losses

14 October 2020
Atish Agarwala
Jeffrey Pennington
Yann N. Dauphin
S. Schoenholz
    UQCV
ArXivPDFHTML

Papers citing "Temperature check: theory and practice for training models with softmax-cross-entropy losses"

7 / 7 papers shown
Title
Comparative sentiment analysis of public perception: Monkeypox vs. COVID-19 behavioral insights
Comparative sentiment analysis of public perception: Monkeypox vs. COVID-19 behavioral insights
Mostafa Mohaimen Akand Faisal
Rabeya Amin Jhuma
18
0
0
12 May 2025
The Optimization Landscape of SGD Across the Feature Learning Strength
The Optimization Landscape of SGD Across the Feature Learning Strength
Alexander B. Atanasov
Alexandru Meterez
James B. Simon
C. Pehlevan
43
2
0
06 Oct 2024
SAM operates far from home: eigenvalue regularization as a dynamical
  phenomenon
SAM operates far from home: eigenvalue regularization as a dynamical phenomenon
Atish Agarwala
Yann N. Dauphin
17
20
0
17 Feb 2023
Lazy vs hasty: linearization in deep networks impacts learning schedule
  based on example difficulty
Lazy vs hasty: linearization in deep networks impacts learning schedule based on example difficulty
Thomas George
Guillaume Lajoie
A. Baratin
8
5
0
19 Sep 2022
Grassmannian learning mutual subspace method for image set recognition
Grassmannian learning mutual subspace method for image set recognition
L. S. Souza
Naoya Sogi
B. Gatto
Takumi Kobayashi
Kazuhiro Fukui
11
11
0
08 Nov 2021
The large learning rate phase of deep learning: the catapult mechanism
The large learning rate phase of deep learning: the catapult mechanism
Aitor Lewkowycz
Yasaman Bahri
Ethan Dyer
Jascha Narain Sohl-Dickstein
Guy Gur-Ari
ODL
153
232
0
04 Mar 2020
Why bigger is not always better: on finite and infinite neural networks
Why bigger is not always better: on finite and infinite neural networks
Laurence Aitchison
173
51
0
17 Oct 2019
1