[Google_Bootcamp_Day7]
Updated:
Normalizing inputs

- Use same mean, and variance to normalize test set
Why normalize inputs

- Easy and fast to optimize
Vanishing/Exploding gradients

- Single neuron example
- Partial solution for vanishing gradient problem is careful choice of random initialization

- Partial solution for vanishing gradient problem is careful choice of random initialization
Numerical approximation of gradients

Gradient Checking

- Don’t use in training, only use to debug (slow computation)
- If algorithm fails grad check, look at components to try to identify bug
- Remember regularization
- Doesn’t work with dropout (Turn off dropout, then grad check)
- Run at random initialization, perhaps again after some training
[Source] https://www.coursera.org/learn/deep-neural-network
Leave a comment