Introduction to Softmax

Learn briefly about activation functions, especially softmax.

We'll cover the following...

Softmax

Check out the activation functions in our neural network. So far, we took it for granted that both of those functions are sigmoids. However, most neural networks replace the last sigmoid, the one right before the output layer, with another function called the softmax.

Let’s see what the softmax looks like and why it is useful. Like the sigmoid, the softmax takes an array of numbers, that in this case are called the logits. And it returns an array with the same size as the input. Here is the formula of the softmax, in case we want to understand the math behind it:​

softmax(li)=eliel\text{softmax}(l_i)=\frac{e^{l_i}}{\sum{e^l}} ...

Access this course and 1400+ top-rated courses and projects.