Tag: genz-to-ai-enz
All the articles with the tag "genz-to-ai-enz".
-
Optimizers: SGD, Momentum, Adam, and AdamW
Why plain gradient descent isn't enough, and how SGD, momentum, Adam, and AdamW each fix a problem the previous one had.
-
Gradient Descent and Backpropagation: How a Network Actually Learns
How gradient descent uses the loss to update weights, and how backpropagation computes the gradients that make it possible.
-
Loss Functions: How a Neural Network Knows It's Wrong
What loss functions are, how MSE and cross-entropy work, and why picking the wrong one breaks your model even if everything else is right.
-
Activation Functions: Why ReLU, GELU, and SiLU Exist
Why stacking linear layers isn't enough, and how activation functions like ReLU, GELU, and SiLU give neural networks their power.