- Essence of linear algebra (linear transformations; matrix multiplication)
- Essence of calculus (derivatives; chain rule)
- Neural Networks (chapter 1 - chapter 4) (animated introduction to neural networks and backpropagation)
- But what is a convolution? (convolution example; convolutions in image processing; convolutions and polynomial multiplication; FFT)
- Neural Networks (chapter 5 - chapter 7) (GPT; visual explanation of attention; LLMs)
A good list of videos to present some classic papers up to SOTA methods in deep learning which help to understand how things work. We recommend to watch them in order, as they are somewhat correlated.
- [Classic] ImageNet Classification with Deep Convolutional Neural Networks
- [Classic] Deep Residual Learning for Image Recognition
- [Classic] Generative Adversarial Networks
- [Classic] Playing Atari with Deep Reinforcement Learning
- Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
- Attention Is All You Need
- An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
- Image GPT: Generative Pretraining from Pixels
- DINO: Emerging Properties in Self-Supervised Vision Transformers
- Perceiver: General Perception with Iterative Attention
- Mamba: Linear-Time Sequence Modeling with Selective State Spaces
- xLSTM: Extended Long Short-Term Memory