Aussie AI
SiLUSigmoid Activation Function
-
Book Excerpt from "Generative AI in C++"
-
by David Spuler, Ph.D.
SiLU/Sigmoid Activation Function
The SiLU activation function uses the product of x and the sigmoid function of x. This is also equivalent to the Swish activation function, with its parameter beta set to 1. Here is the basic SiLU activation function in C++:
float aussie_SiLU_basic(float x) // Basic SiLU (inefficient)
{
// Sigmoid = 1 + e^(-x)
// SiLU = x * (1 + e^(-x) )
// = x * 1.0 / (1.0 + expf(-x));
return x / (1.0f + expf(-x));
}
The SiLU function is inefficient by default. Its speed can be improved via table lookups and approximations.
|
• Next: • Up: Table of Contents |
|
The new AI programming book by Aussie AI co-founders:
Get your copy from Amazon: Generative AI in C++ |