MuzzammilShah/people-names
Viewer • Updated • 32k • 10
This repository explores manual backpropagation through a 2-layer MLP (with BatchNorm) without using PyTorch autograd's loss.backward(). It involves manually backpropagating through the cross entropy loss, 2nd linear layer, tanh, batchnorm, 1st linear layer, and the embedding table.
For a better reading experience and detailed notes, visit my Road to GPT Documentation Site.
Notes and implementations inspired by the Makemore - Part 4 video by Andrej Karpathy.
For more of my projects, visit my Portfolio Site.