GeLU.java
package neureka.backend.main.operations.functions;
import neureka.backend.main.implementations.fun.api.ScalarFun;
/**
* The GELU activation function is based on the standard Gaussian cumulative distribution function
* and is defined as {@code x Φ( x )} and implemented as {@code x * sigmoid(x * 1.702)}.
* The GELU non-linearity weighs inputs by their percentile,
* rather than gates inputs by their sign as in ReLUs.
* Consequently, the GELU can be thought of as a smoother ReLU.
*/
public class GeLU extends AbstractActivationOperation
{
public GeLU() {
super(ScalarFun.GELU);
}
}