Click a step button to begin visualization.
Configuration
- Dataset: Iris
- Architecture: 4 → 10 → 8 → 3
- Optimizer: Adam
- Learning Rate: 0.01
- Epochs: 50
- Hidden Activation: ReLU
- Output Activation: Softmax
Input features are fed into the network.
[−0.90, 1.03, −1.34, −1.32]
Normalized values of original [5.1, 3.5, 1.4, 0.2]
Each input is weighted, summed, and passed through ReLU.
z = Σ(wᵢ·xᵢ) + b → a = max(0, z)
Grey neurons = ReLU zeroed this unit
Same weighted-sum + ReLU process on layer 1 outputs.
z = Σ(wᵢ·aᵢ) + b → a = max(0, z)
Raw scores converted to class probabilities via Softmax.
softmax(zᵢ) = e^zᵢ / Σ e^zⱼ
Highest probability = predicted class
How it works — per neuron
1. ReLU checks the weighted sum (z) already computed in the forward pass:
if z > 0 → neuron remains active (passes z forward)
if z ≤ 0 → neuron becomes inactive (output is exactly 0)
Result in this network
Active (z > 0): Output remains their forward pass value
Inactive (z ≤ 0): Output becomes 0
Important Note
Activation is applied directly on the current neuron values shown after backpropagation. ReLU turns off neurons with z ≤ 0 by setting their output to 0, while keeping positive values unchanged.