Technical deep dives.
Articles on transformer architectures, model quantization, inference optimization, and the systems behind production AI.
Attention
Frameworks
Fundamentals
Cross Entropy Loss
A comprehensive guide to understanding and implementing cross entropy in machine learning — from information theory to PyTorch.
Elements of Neural Nets: Building Blocks of AI
An essential guide to the foundational components of modern neural networks — activations, cost functions, optimizers, and regularization.
Quantization
Quick Books
Transformers
Building the Transformer from Scratch
An object-oriented implementation of the complete Transformer architecture in PyTorch — from input embeddings to the full encoder-decoder model.
Implementing Transformers
A practical guide to coding the Transformer architecture from scratch — turning the 'Attention is All You Need' paper into working code.