[Google_Bootcamp_Day7]

Updated:

Normalizing inputs

1

  • Use same mean, and variance to normalize test set

Why normalize inputs

2

  • Easy and fast to optimize

Vanishing/Exploding gradients

3

  • Single neuron example
    • Partial solution for vanishing gradient problem is careful choice of random initialization 4

Numerical approximation of gradients

5

Gradient Checking

6 7

  • Don’t use in training, only use to debug (slow computation)
  • If algorithm fails grad check, look at components to try to identify bug
  • Remember regularization
  • Doesn’t work with dropout (Turn off dropout, then grad check)
  • Run at random initialization, perhaps again after some training

[Source] https://www.coursera.org/learn/deep-neural-network

Categories:

Updated:

Leave a comment