Read
Watch
Reflect
Prime your brain first — retention follows
Read ~9m
11 terms · 9 segments
Building makemore Part 2: MLP
9chapters with key takeaways — read first, then watch
9chapters with key takeaways — read first, then watch
Video Details & AI Summary
Published Sep 12, 2022
Analyzed Jan 21, 2026
AI Analysis Summary
This video, part two of the 'makemore' series, delves into building a Multi-Layer Perceptron (MLP) neural network for a character-level language model. It addresses the limitations of simpler bi-gram models by introducing word embeddings and a multi-layer architecture, demonstrating dataset preparation, PyTorch implementation of embedding, hidden, and output layers, and efficient loss calculation. The tutorial covers crucial aspects of training, including mini-batching, learning rate tuning, data splitting, and model scaling, culminating in a model that generates more realistic names and is accessible via Google Colab.
Title Accuracy Score
9/10Excellent
38.4s processing
Model:
gemini-2.5-flashOriginal Video