News

Many of the autoencoder examples I see online use relu() activation for interior layers. The relu() function was designed for use with very deep neural architectures. For autoencoders, which are ...
Behind the scenes, the autoencoder uses tanh() activation on the hidden nodes and tanh() activation on the output nodes. The result of the tanh() function is always between -1 and +1. Therefore, the ...