Publication Date

4-1-2021

Comments

Technical Report: UTEP-CS-21-41

Abstract

Since in the physical world, most dependencies are smooth (differentiable), traditionally, smooth functions were used to approximate these dependencies. In particular, neural networks used smooth activation functions such as the sigmoid function. However, the successes of deep learning showed that in many cases, non-smooth activation functions like max(0,z) work much better. In this paper, we explain why in many cases, non-smooth approximating functions often work better -- even when the approximated dependence is smooth.

Share

COinS