Features
Activation Functions
| Activation function | Layer usage | JavaScript | C++ |
|---|---|---|---|
| Sigmoid | Hidden and Output | sigmoid | Sigmoid() |
| Softmax | Output | softmax | Softmax() |
| ReLU | Hidden | relu | ReLU() |
| LeakyReLU | Hidden | leakyrelu | LeakyReLU() |
| ELU | Hidden | elu | ELU() |
| Tanh | Hidden | tanh | Tanh() |
Loss Functions
| Loss function | Output layer activation function | JavaScript | C++ |
|---|---|---|---|
| Mean Squared Error | Sigmoid | mean-squared-error | MeanSquaredError() |
| Sigmoid Cross Entropy | Sigmoid | sigmoid-cross-entropy | SigmoidCrossEntropy() |
| Softmax Cross Entropy | Softmax | softmax-cross-entropy | SoftmaxCrossEntropy() |
Weight Initialization
| Weight initializer | Layer Position | JavaScript | C++ |
|---|---|---|---|
| Xavier Uniform | Hidden | xavier_uniform | XavierUniform() |
| Xavier Normal | Hidden | xavier_normal | XavierNormal() |
| LeCun Uniform | Hidden and Output | lecun_uniform | LeCunUniform() |
| LeCun Normal | Hidden and Output | lecun_normal | LeCunNormal() |
| Gaussian | Hidden and Output | gaussian | Gaussian() |
| Uniform | Hidden and Output | uniform | Uniform() |
| Constant | Hidden and Output | constant | Constant() |
Weights Optimizer
| Weights Optimizer |
|---|
| Stochastic Gradient Descent |
Regularization
| Regularization | Scope |
|---|---|
| Dropout | Input layer and Hidden layers |
| L1 | All weights and bias values |
| L2 | All weights and bias values |