Abstract
In deep learning architectures, rectified linear unit based functions are widely used as activation functions of hidden layers, and the softmax is used for the output layers. Two critical problems of the softmax are introduced, and an improved softmax method to resolve the problems is proposed. The proposed method minimises instability of the softmax while reducing its losses. Moreover, this method is straightforward so its computation complexity is low, but it is substantially reasonable and operates robustly. Therefore, the proposed method can replace the softmax functions.
Original language | English |
---|---|
Pages (from-to) | 1504-1506 |
Number of pages | 3 |
Journal | Electronics Letters |
Volume | 53 |
Issue number | 23 |
DOIs | |
Publication status | Published - 9 Nov 2017 |
Bibliographical note
Publisher Copyright:© The Institution of Engineering and Technology 2017.