# Deep Learning From Scratch - Theory and Implementation

### Open Source Your Knowledge, Become a Contributor

Technology knowledge has to be shared and made accessible for free. Join the movement.

## Gradient descent

Generally, if we want to find the minimum of a function, we set the derivative to zero and solve for the parameters. It turns out, however, that it is impossible to obtain a closed-form solution for W$W$ and b$b$. Instead, we iteratively search for a minimum using a method called **gradient descent**.

As a visual analogy, imagine yourself standing on a mountain and trying to find the way down. At every step, you walk into the steepest direction, since this direction is the most promising to lead you towards the bottom.

If taking steep steps seems a little dangerous to you, imagine that you are a mountain goat (which are amazing rock climbers).

Gradient descent operates in a similar way when trying to find the minimum of a function: It starts at a random location in parameter space and then iteratively reduces the error J$J$ until it reaches a local minimum. At each step of the iteration, it determines the direction of steepest descent and takes a step along that direction. This process is depicted for the 1-dimensional case in the following image.

As you might remember, the direction of steepest ascent of a function at a certain point is given by the gradient at that point. Therefore, the direction of steepest descent is given by the negative of the gradient. So now we have a rough idea how to minimize J$J$:

- Start with random values for W$W$ and b$b$
- Compute the gradients of J$J$ with respect to W$W$ and b$b$
- Take a small step along the direction of the negative gradient
- Go back to 2

Let's implement an operation that minimizes the value of a node using gradient descent. We require the user to specify the magnitude of the step along the gradient as a parameter called `learning_rate`

.

The following image depicts an example iteration of gradient descent. We start out with a random separating line (marked as 1), take a step, arrive at a slightly better line (marked as 2), take another step, and another step, and so on until we arrive at a good separating line.

## Backpropagation

In our implementation of gradient descent, we have used a function `compute_gradient(loss)`

that computes the gradient of a loss$loss$ operation in our computational graph with respect to the output of every other node n$n$ (i.e. the direction of change for n$n$ along which the loss increases the most). We now need to figure out how to compute gradients.

Consider the following computational graph:

By the chain rule, we have

As we can see, in order to compute the gradient of e$e$ with respect to a$a$, we can start at e$e$ an go backwards towards a$a$, computing the gradient of every node's output with respect to its input along the way until we reach a$a$. Then, we multiply them all together.

Now consider the following scenario:

In this case, a$a$ contributes to e$e$ along two paths: The path a$a$, b$b$, d$d$, e$e$ and the path a$a$, b$b$, c$c$, e$e$. Hence, the total derivative of e$e$ with respect to a$a$ is given by:

This gives us an intuition for the general algorithm that computes the gradient of the loss with respect to another node: We perform a backwards breadth-first search starting from the loss node. At each node n$n$ that we visit, we compute the gradient of the loss with respect do n$n$'s output by doing the following for each of n$n$'s consumers c$c$:

- retrieve the gradient G$G$ of the loss with respect to the output of c$c$
- multiply G$G$ by the gradient of c$c$'s output with respect to n$n$'s output

And then we sum those gradients over all consumers.

As a prerequisite to implementing backpropagation, we need to specify a function for each operation that computes the gradients with respect to the inputs of that operation, given the gradients with respect to the output. Let's define a decorator `@RegisterGradient(operation_name)`

for this purpose:

Now assume that our `_gradient_registry`

dictionary is already filled with gradient computation functions for all of our operations. We can now implement backpropagation:

### Gradient of each operation

For each of our operations, we now need to define a function that turns a gradient of the loss with respect to the operation's output into a list of gradients of the loss with respect to each of the operation's inputs. Computing a gradient with respect to a matrix can be somewhat tedious. Therefore, the details have been omitted and I just present the results. You may skip this section and still understand the overall picture. The implementations for `add`

, `reduce_sum`

and `softmax`

are a little more involved, and I'd recommend not spending too much time trying to understand them.

If you want to comprehend how to arrive at the results, the general approach is as follows:

- Find the partial derivative of each output value with respect to each input value. This can be a tensor of a rank greater than 2, i.e. neither scalar nor vector nor matrix, involving a lot of summations.
- Compute the gradient of the loss with respect to the node's inputs given a gradient with respect to the node's output by applying the chain rule. This is now a tensor of the same shape as the input tensor, so if the input is a matrix, the result is also a matrix
- Rewrite this result as a sequence of matrix operations in order to compute it efficiently. This step can be somewhat tricky.

#### Gradient for `negative`

Given a gradient G$G$ with respect to âˆ’x$\xe2\u02c6\u2019x$, the gradient with respect to x$x$ is given by âˆ’G$\xe2\u02c6\u2019G$.

#### Gradient for `log`

Given a gradient G$G$ with respect to log(x)$log(x)$, the gradient with respect to x$x$ is given by Gx$\frac{G}{x}$.

#### Gradient for `sigmoid`

Given a gradient G$G$ with respect to Ïƒ(a)$\mathrm{\xcf\u0192}(a)$, the gradient with respect to a$a$ is given by Gâ‹…Ïƒ(a)â‹…Ïƒ(1âˆ’a)$G\xe2\u2039\dots \mathrm{\xcf\u0192}(a)\xe2\u2039\dots \mathrm{\xcf\u0192}(1\xe2\u02c6\u2019a)$.

#### Gradient for `multiply`

Given a gradient G$G$ with respect to AâŠ™B$A\xe2\u0160\u2122B$, the gradient with respect to A$A$ is given by GâŠ™B$G\xe2\u0160\u2122B$ and the gradient with respect to B$B$ is given by GâŠ™A$G\xe2\u0160\u2122A$.

#### Gradient for `matmul`

Given a gradient G$G$ with respect to AB$AB$, the gradient with respect to A$A$ is given by GBT$G{B}^{T}$ and the gradient with respect to B$B$ is given by ATG${A}^{T}G$.

#### Gradient for `add`

Given a gradient G$G$ with respect to a+b$a+b$, the gradient with respect to a$a$ is given by G$G$ and the gradient with respect to b$b$ is also given by G$G$, provided that a$a$ and b$b$ are of the same shape. If a$a$ and b$b$ are of different shapes, e.g. one matrix a$a$ with 100 rows and one row vector b$b$, we assume that b$b$ is added to each row of a$a$. In this case, the gradient computation is a little more involved, but I will not spell out the details here.

#### Gradient for `reduce_sum`

Given a gradient G$G$ with respect to the output of `reduce_sum`

, the gradient with respect to the input A$A$ is given by repeating G$G$ along the specified axis.

#### Gradient for `softmax`

### Example

Let's now test our implementation to determine the optimal weights for our perceptron.

Notice that we started out with a rather high loss and incrementally reduced it. Let's plot the final line to check that it is a good separator: