ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1709.06080
29
9

Feedforward and Recurrent Neural Networks Backward Propagation and Hessian in Matrix Form

16 September 2017
Maxim Naumov
ArXivPDFHTML
Abstract

In this paper we focus on the linear algebra theory behind feedforward (FNN) and recurrent (RNN) neural networks. We review backward propagation, including backward propagation through time (BPTT). Also, we obtain a new exact expression for Hessian, which represents second order effects. We show that for ttt time steps the weight gradient can be expressed as a rank-ttt matrix, while the weight Hessian is as a sum of t2t^{2}t2 Kronecker products of rank-111 and WTAWW^{T}AWWTAW matrices, for some matrix AAA and weight matrix WWW. Also, we show that for a mini-batch of size rrr, the weight update can be expressed as a rank-rtrtrt matrix. Finally, we briefly comment on the eigenvalues of the Hessian matrix.

View on arXiv
Comments on this paper