|
| | ELU () |
| | Create the ELU object. More...
|
| |
| | ELU (const double alpha) |
| | Create the ELU object using the specified parameter. More...
|
| |
| double const & | Alpha () const |
| | Get the non zero gradient. More...
|
| |
| double & | Alpha () |
| | Modify the non zero gradient. More...
|
| |
| template<typename DataType > |
| void | Backward (const DataType &&input, DataType &&gy, DataType &&g) |
| | Ordinary feed backward pass of a neural network, calculating the function f(x) by propagating x backwards through f. More...
|
| |
| OutputDataType const & | Delta () const |
| | Get the delta. More...
|
| |
| OutputDataType & | Delta () |
| | Modify the delta. More...
|
| |
| template<typename InputType , typename OutputType > |
| void | Forward (const InputType &&input, OutputType &&output) |
| | Ordinary feed forward pass of a neural network, evaluating the function f(x) by propagating the activity forward through f. More...
|
| |
| double const & | Lambda () const |
| | Get the lambda parameter. More...
|
| |
| OutputDataType const & | OutputParameter () const |
| | Get the output parameter. More...
|
| |
| OutputDataType & | OutputParameter () |
| | Modify the output parameter. More...
|
| |
| template<typename Archive > |
| void | serialize (Archive &ar, const unsigned int) |
| | Serialize the layer. More...
|
| |
template<typename InputDataType = arma::mat, typename OutputDataType = arma::mat>
class mlpack::ann::ELU< InputDataType, OutputDataType >
The ELU activation function, defined by.
For more information, read the following paper:
* @article{Clevert2015,
* author = {Djork{-}Arn{\'{e}} Clevert and Thomas Unterthiner and
* Sepp Hochreiter},
* title = {Fast and Accurate Deep Network Learning by Exponential Linear
* Units (ELUs)},
* journal = {CoRR},
* year = {2015}
* }
*
The SELU activation function is defined by
For more information, read the following paper:
* @article{Klambauer2017,
* author = {Gunter Klambauer and Thomas Unterthiner and
* Andreas Mayr},
* title = {Self-Normalizing Neural Networks},
* journal = {Advances in Neural Information Processing Systems},
* year = {2017}
* }
*
In the deterministic mode, there is no computation of the derivative.
- Note
- During training deterministic should be set to false and during testing/inference deterministic should be set to true.
-
Make sure to use SELU activation function with normalized inputs and weights initialized with Lecun Normal Initialization.
- Template Parameters
-
| InputDataType | Type of the input data (arma::colvec, arma::mat, arma::sp_mat or arma::cube). |
| OutputDataType | Type of the output data (arma::colvec, arma::mat, arma::sp_mat or arma::cube). |
Definition at line 109 of file elu.hpp.