We read all 512K lines of Claude Code's accidentally exposed source. 82 docs, 15 diagrams, every subsystem mapped — from the hidden YOLO safety classifier to multi-agent swarms.
-
Updated
Apr 2, 2026
We read all 512K lines of Claude Code's accidentally exposed source. 82 docs, 15 diagrams, every subsystem mapped — from the hidden YOLO safety classifier to multi-agent swarms.
A reference point for phenomena that have been reported to occur inside AI systems but have no direct mapping into natural language.
Mechanistic interpretability of transformer hallucinations via attention flow, residual stream geometry, and head-level attribution analysis.
Add a description, image, and links to the llm-internals topic page so that developers can more easily learn about it.
To associate your repository with the llm-internals topic, visit your repo's landing page and select "manage topics."