This module defines a set of activation functions, or method that may be used to determine the sensitivity of neurons in a network layer. To support both forward and backward propagation each method contains the activation function and it’s derivative. This Wikipedia page has a good overview of a number of activation functions.
value
value
value
See also Parallelism with Futures in The Racket Guide In general it is preferable to use the flonum-activator? structure and the corresponding flonum-activation/c form as this reduces the numeric conversions and allows optimization such as futures to work efficiently.
f is the activation function, \phi(v_i)
df is the activation derivative function, \phi^\prime(v_i) – sometimes shown as \phi^{-1}(v_i)
α is an optional stochastic variable sampled from a uniform distribution at training time and fixed to the expectation value of the distribution at test time
procedure
( make-activator namefdf[α])→activator?
name:symbol?procedure
( make-flonum-activator namefdf[α])→flonum-activator?
name:symbol?
Each of the activator? structures below will be defined by it’s activation function (the derivative is not shown). A sample plot shows the shape of the activation function in red and it’s derivative in turquoise.
value
value
value
value
value
value
value
value
value
value
value
procedure
∂:flonum?
Note that the fixed form of this activator uses a delta value \delta=0.01.
value
value
value
value
value
value
value
value
value
value