Journal of Big Data | |
Evaluation of maxout activations in deep learning across several big data domains | |
Taghi M. Khoshgoftaar1  Paul Morris1  Gabriel Castaneda1  | |
[1] Florida Atlantic University; | |
关键词: Maxout networks; Activation functions; Big data; Deep learning; | |
DOI : 10.1186/s40537-019-0233-0 | |
来源: DOAJ |
【 摘 要 】
Abstract This study investigates the effectiveness of multiple maxout activation function variants on 18 datasets using Convolutional Neural Networks. A network with maxout activation has a higher number of trainable parameters compared to networks with traditional activation functions. However, it is not clear if the activation function itself or the increase in the number of trainable parameters is responsible in yielding the best performance for different entity recognition tasks. This paper investigates if an increase in the number of convolutional filters on traditional activation functions performs equal-to or better-than maxout networks. Our experiments compare the Rectified Linear Unit, Leaky Rectified Linear Unit, Scaled Exponential Linear Unit, and Hyperbolic Tangent activations to four maxout function variants. We observe that maxout networks train relatively slower than networks with traditional activation functions, e.g. Rectified Linear Unit. In addition, we found that on average, across all datasets, the Rectified Linear Unit activation function performs better than any maxout activation when the number of convolutional filters is increased. Furthermore, adding more filters enhances the classification accuracy of the Rectified Linear Unit networks, without adversely affecting their advantage over maxout activations with respect to network-training speed.
【 授权许可】
Unknown