ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1708.03888
  4. Cited By
Large Batch Training of Convolutional Networks

Large Batch Training of Convolutional Networks

13 August 2017
Yang You
Igor Gitman
Boris Ginsburg
    ODL
ArXivPDFHTML

Papers citing "Large Batch Training of Convolutional Networks"

19 / 19 papers shown
Title
Task-Optimized Convolutional Recurrent Networks Align with Tactile Processing in the Rodent Brain
Task-Optimized Convolutional Recurrent Networks Align with Tactile Processing in the Rodent Brain
Trinity Chung
Yuchen Shen
Nathan C. L. Kong
Aran Nayebi
75
0
0
23 May 2025
Implicit Contrastive Representation Learning with Guided Stop-gradient
Byeongchan Lee
Sehyun Lee
SSL
150
2
0
12 Mar 2025
USP: Unified Self-Supervised Pretraining for Image Generation and Understanding
Xiangxiang Chu
Renda Li
Yong Wang
108
0
0
08 Mar 2025
Why does my medical AI look at pictures of birds? Exploring the efficacy of transfer learning across domain boundaries
Why does my medical AI look at pictures of birds? Exploring the efficacy of transfer learning across domain boundaries
F. Jonske
M. Kim
Enrico Nasca
J. Evers
Johannes Haubold
...
F. Nensa
Michael Kamp
C. Seibold
Jan Egger
Jens Kleesiek
118
1
0
17 Feb 2025
A Hessian-informed hyperparameter optimization for differential learning rate
A Hessian-informed hyperparameter optimization for differential learning rate
Shiyun Xu
Zhiqi Bu
Yiliang Zhang
Ian Barnett
60
1
0
12 Jan 2025
How Does Critical Batch Size Scale in Pre-training?
How Does Critical Batch Size Scale in Pre-training?
Hanlin Zhang
Depen Morwani
Nikhil Vyas
Jingfeng Wu
Difan Zou
Udaya Ghai
Dean Phillips Foster
Sham Kakade
99
15
0
29 Oct 2024
Rethinking Positive Pairs in Contrastive Learning
Rethinking Positive Pairs in Contrastive Learning
Jiantao Wu
Shentong Mo
Zhenhua Feng
Sara Atito
Josef Kitler
Muhammad Awais
SSL
VLM
72
3
0
23 Oct 2024
Efficient Distribution Matching of Representations via Noise-Injected Deep InfoMax
Efficient Distribution Matching of Representations via Noise-Injected Deep InfoMax
I. Butakov
Alexander Sememenko
Alexander Tolmachev
Andrey Gladkov
Marina Munkhoeva
Alexey Frolov
84
0
0
09 Oct 2024
EEG-Language Modeling for Pathology Detection
EEG-Language Modeling for Pathology Detection
Sam Gijsen
Kerstin Ritter
89
0
0
02 Sep 2024
AdaFisher: Adaptive Second Order Optimization via Fisher Information
AdaFisher: Adaptive Second Order Optimization via Fisher Information
Damien Martins Gomes
Yanlei Zhang
Eugene Belilovsky
Guy Wolf
Mahdi S. Hosseini
ODL
103
2
0
26 May 2024
Towards Calibrated Deep Clustering Network
Towards Calibrated Deep Clustering Network
Yuheng Jia
Jianhong Cheng
Hui Liu
Junhui Hou
UQCV
79
1
0
04 Mar 2024
MIO : Mutual Information Optimization using Self-Supervised Binary Contrastive Learning
MIO : Mutual Information Optimization using Self-Supervised Binary Contrastive Learning
Siladittya Manna
Umapada Pal
Saumik Bhattacharya
SSL
62
1
0
24 Nov 2021
Mime: Mimicking Centralized Stochastic Algorithms in Federated Learning
Mime: Mimicking Centralized Stochastic Algorithms in Federated Learning
Sai Praneeth Karimireddy
Martin Jaggi
Satyen Kale
M. Mohri
Sashank J. Reddi
Sebastian U. Stich
A. Suresh
FedML
56
216
0
08 Aug 2020
Unsupervised Learning of Visual Features by Contrasting Cluster
  Assignments
Unsupervised Learning of Visual Features by Contrasting Cluster Assignments
Mathilde Caron
Ishan Misra
Julien Mairal
Priya Goyal
Piotr Bojanowski
Armand Joulin
OCL
SSL
114
4,032
0
17 Jun 2020
Extrapolation for Large-batch Training in Deep Learning
Extrapolation for Large-batch Training in Deep Learning
Tao R. Lin
Lingjing Kong
Sebastian U. Stich
Martin Jaggi
47
36
0
10 Jun 2020
Integrated Model, Batch and Domain Parallelism in Training Neural
  Networks
Integrated Model, Batch and Domain Parallelism in Training Neural Networks
A. Gholami
A. Azad
Peter H. Jin
Kurt Keutzer
A. Buluç
50
83
0
12 Dec 2017
Train longer, generalize better: closing the generalization gap in large
  batch training of neural networks
Train longer, generalize better: closing the generalization gap in large batch training of neural networks
Elad Hoffer
Itay Hubara
Daniel Soudry
ODL
134
798
0
24 May 2017
Revisiting Distributed Synchronous SGD
Revisiting Distributed Synchronous SGD
Jianmin Chen
Xinghao Pan
R. Monga
Samy Bengio
Rafal Jozefowicz
53
799
0
04 Apr 2016
One weird trick for parallelizing convolutional neural networks
One weird trick for parallelizing convolutional neural networks
A. Krizhevsky
GNN
58
1,297
0
23 Apr 2014
1