Coursera Deep Learning Course 1 Week 4 notes: Deep neural networks

2017-10-16

Deep Neural Network
Deep L-layer neural network

The word ‘deep’ refers to the number of layer of a neural network (not count the input layer).

Some notations:

  • = number of layer
  • = number of units in layer l (
  • = activations in layer l ().
Forward Propagation in a Deep Network

Vectorization for whole training set (stack lowercase matrices in column to obtain capital matrices):

We can’t avoid having a for loop iterating over all layers.

Getting your matrix dimensions right

No notes.

Why deep representing

No notes.

Building blocks of deep neural networks

Figure 1. Forward and backward implementation.
Source: Coursera Deep Learning course

Forward and Backward Propagation

Forward propagation for layer l

Input

Output , cache()

  • .

  • .

Backward propagation for layer l

Input

Output

  • .

  • .

  • .

  • .

Parameters vs Hyperameters

Parameters: .

Hyperparameters: control parameters.

  • Learning rate .
  • Number of iterations.
  • Number of hidden layers L.
  • Number of hidden units.
  • Choice of activation functions.
  • Many more…

Therefore, applying deep learning is a very empirical process.

What does this have to do with the brain?