简体中文 | English
Breaking the wall between Academic Math and Engineering Implementation.
This handbook is a "Rosetta Stone" for Deep Learning practitioners. It provides a direct, side-by-side mapping between rigorous mathematical definitions (LaTeX) and their actual implementations in modern frameworks (PyTorch).
- Rosetta Stone Layout: Side-by-side view of math formulas and PyTorch code.
- SOTA Architectures: Detailed mathematical breakdown of Transformer, Mamba (SSM), LoRA, and Diffusion Models.
- High-Quality Visuals: Standardized TikZ diagrams with professional aesthetics (shadows, semantic colors, gradient flow paths).
- Mathematical Depth: Covers everything from Tensor basics and Linear Algebra to Stochastic Processes and Information Theory.
- Academic Typography: Springer-level layout using Times New Roman and asymmetric wide margins for annotations.
The handbook is organized into thematic "Parts":
- Foundations (基石篇): Tensors, Advanced Linear Algebra (SVD, QR), Probability Foundations.
- Anatomy (解剖篇): Activation Functions, Standard Layers (Linear, Conv), Sequence Layers (RNN, LSTM), Normalization.
- Objectives (目标篇): Distance Metrics, Information Theory & Probabilistic Losses.
- Dynamics (动力篇): Autograd (VJP/JVP), Optimization Algorithms, Stochastic Processes (Wiener, SDE).
- Architectures (架构篇): Multi-Head Attention, Generative Models (VAE, GAN).
- Foundation Models (大模型纪元): LLM Components, Mamba (SSM), PEFT (LoRA).
- The Frontiers (前沿探索): Adversarial Training, Graph Neural Networks, Quantization, Next-Gen Generative (Flow Matching).
We provide automation scripts for a seamless build:
Windows:
build.batmacOS / Linux:
makeTo clean intermediate files:
build.bat clean # Windows
make clean # UnixWe welcome contributions to expand the "Atlas"! If you want to add a new SOTA model (e.g., DeepSeek MLA, FlashAttention) or fix a typo, please see CONTRIBUTING.md.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Designed with ❤️ by Antigravity & Sisyphus.