For hᵢ = σ(Wᵢx + b) z = Σ wᵢxᵢ + b n i=1 Dendrites ⓘ (Input Vector) Synapses ⓘ (Weights) Axon Hillock ⓘ (Activation Function) Axon ⓘ (Output) X1 X2 X3 X4 W1 W2 W3 W4 Soma ⓘ (Summation) Bias ⓘ (Activation Threshold) f Sigmoid ⓘ h = σ(z) = 1 1 + e⁻ᶻ ReLU ⓘ ReLU(z) = max(0, z) Tanh ⓘ tanh(z) = ez - e-z ez + e-z Softmax ⓘ softmax(z) = ezi j ezj for j = 1 ... n 0