[source] # GELU Gaussian Error Linear Units (GELUs) are rectifiers that are gated by the magnitude of their input rather than the sign of their input as with ReLU variants. Their output can be interpreted as the expected value of a neuron with random dropout regularization applied. ## Parameters This activation function does not have any parameters. ## Example ```php use Rubix\ML\NeuralNet\ActivationFunctions\GELU; $activationFunction = new GELU(); ``` ### References >- D. Hendrycks et al. (2018). Gaussian Error Linear Units (GELUs).