TechRxiv
Activation Function Absolute Function.docx (552.43 kB)
Download file

Activation Function: Absolute Function, One Function Behaves more Individualized

Download (552.43 kB)
preprint
posted on 28.01.2022, 21:14 authored by Jinxin WeiJinxin Wei, Zhe Hou

Inspire by nature world mode, a activation function is proposed. It is absolute function.Through test on mnist dataset and fully-connected neural network and convolutional neural network, some conclusions are put forward. The line of accuracy of absolute function is shaked around the training accuracy that is different from the line of accuracy of relu and leaky relu. The absolute function can keep the negative parts as equal as the positive parts, so the individualization is more active than relu and leaky relu function. The absolute function is less likely to be over-fitting. Through teat on mnist and autoencoder, It is that the leaky relu function can do classification task well, while the absolute function can do generation task well. Because the classification task need more universality and generation task need more individualization. The pleasure irritation and painful irritation is not only the magnitude differences, but also the sign differences, so the negative parts should keep as a part.Stimulation which happens frequently is low value, it is showed around zero in figure 1 . Stimulation which happens accidentally is high value, it is showed far away from zero in figure 1. So the high value is the big stimulation, which is individualization.

History

Email Address of Submitting Author

wjxabai@163.com

Submitting Author's Institution

vocational school of Juancheng

Submitting Author's Country

China