ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.12690
38
0

Growth strategies for arbitrary DAG neural architectures

17 February 2025
Stella Douka
Manon Verbockhaven
Théo Rudkiewicz
Stéphane Rivaud
François P. Landes
Sylvain Chevallier
Guillaume Charpiat
    AI4CE
ArXivPDFHTML
Abstract

Deep learning has shown impressive results obtained at the cost of training huge neural networks. However, the larger the architecture, the higher the computational, financial, and environmental costs during training and inference. We aim at reducing both training and inference durations. We focus on Neural Architecture Growth, which can increase the size of a small model when needed, directly during training using information from the backpropagation. We expand existing work and freely grow neural networks in the form of any Directed Acyclic Graph by reducing expressivity bottlenecks in the architecture. We explore strategies to reduce excessive computations and steer network growth toward more parameter-efficient architectures.

View on arXiv
@article{douka2025_2501.12690,
  title={ Growth strategies for arbitrary DAG neural architectures },
  author={ Stella Douka and Manon Verbockhaven and Théo Rudkiewicz and Stéphane Rivaud and François P. Landes and Sylvain Chevallier and Guillaume Charpiat },
  journal={arXiv preprint arXiv:2501.12690},
  year={ 2025 }
}
Comments on this paper