You signed in with another tab or window.Reload to refresh your session.You signed out in another tab or window.Reload to refresh your session.You switched accounts on another tab or window.Reload to refresh your session.Dismiss alert
Now, for each gate and the fully connected layer, we define a function to calculate the gradient of the loss with respect to the input passed and the parameters used. To understand the mathematics behind how the derivatives were calculated we suggest you to follow this helpful[blog](https://christinakouridi.blog/2019/06/19/backpropagation-lstm/) by Christina Kouridi.
655
+
Now, for each gate and the fully connected layer, we define a function to calculate the gradient of the loss with respect to the input passed and the parameters used. To understand the mathematics behind how the derivatives were calculated we suggest you to follow this helpful[blog](https://christinakouridi.github.io/posts/backprop-lstm/) by Christina Kouridi.
656
656
657
657
658
658
Define a function to calculate the gradients in the**Forget Gate**: