Tanh is the rescaled version of the sigmoid.
Deep learning today uses different activation functions. In this short blog post I will show you the relationship between the tanh
and the sigmoid
activation function. In fact I will show that the tanh
is just a rescaled version of the sigmoid
.
Now look
\[\begin{align*} 2 \text{sigmoid}(2x) -1& = 2\frac{e^{2x}}{1+e^{2x}}-1\\ &= \frac{e^{2x}}{1+e^{2x}}+\frac{e^{2x}}{1+e^{2x}}-\frac{1+e^{2x}}{1+e^{2x}}\\ &= \frac{e^{2x}+e^{2x}-1-e^{2x}}{1+e^{2x}}\\ &= \frac{e^{2x}-1}{1+e^{2x}} = \frac{e^{2x}-1}{e^{2x}+1} = \text{tanh}(x) \end{align*}\]Hence we can safely say the sigmoid
is rescaled version of the tanh
.
Written on September 28, 2021