Learn LLM internals step by step - from tokenization to attention to inference optimization.
-
Updated
Apr 14, 2026
Learn LLM internals step by step - from tokenization to attention to inference optimization.
We read all 512K lines of Claude Code's accidentally exposed source. 82 docs, 15 diagrams, every subsystem mapped — from the hidden YOLO safety classifier to multi-agent swarms.
A reference point for phenomena that have been reported to occur inside AI systems but have no direct mapping into natural language.
Mechanistic interpretability of transformer hallucinations via attention flow, residual stream geometry, and head-level attribution analysis.
Add a description, image, and links to the llm-internals topic page so that developers can more easily learn about it.
To associate your repository with the llm-internals topic, visit your repo's landing page and select "manage topics."