๐ Challenge: Backpropagation - 3 Layered Neural Network
As a challenge, code the backpropagation pass for a 3 layered neural network.
We'll cover the following
Problem statement
Code the backpropagation operation for the three layered neural network to compute the gradient of the cross-entropy loss function with respect to weights and biases of the neural network.
The cross entropy loss function is as follows:
Where, is the target output, is the predicted output, and is the number of classes.
Sample input
- The target output:
y
- The predicted output:
out_y
- The output at hidden layer 1:
out_h1
- The output at hidden layer 2:
out_h2
- The weights on the connections between hidden layer 2 and the output layer:
w3
- The weights on the connections between hidden layer 1 and the hidden layer 2:
w2
- The input values:
x
Sample output
The change of weights and bias at the respective layers. For example:
- The gradient of loss w.r.t weights at layer 3:
dW3
- The gradient of loss w.r.t bias at layer 3:
db3
- The gradient of loss w.r.t weights at layer 2:
dW2
- The gradient of loss w.r.t bias at layer 2:
db2
- The gradient of loss w.r.t weights at layer 1:
dW1
- The gradient of loss w.r.t bias at layer 1:
db1
Coding exercise
Write your code below. It is recommendedโ to solve the exercise before viewing the solution.
๐ There is a
backpropagation
function given in the code for testing purposes. Do not modify the function signature.
Good luck!๐ค
Get hands-on with 1400+ tech skills courses.