Comprehensive documentation for 292+ state-of-the-art AI algorithms From foundational concepts to cutting-edge research (2023-2025)
Welcome to the Nexus learning system! This documentation provides in-depth tutorials, theory, implementation details, and practical guidance for every algorithm implemented in the Nexus AI framework.
New to Nexus? Start here:
- Browse the Learning Paths to find a structured curriculum
- Check the Category Overview to explore specific domains
- Dive into individual algorithm documentation for deep understanding
Looking for something specific?
- Use the Quick Reference for at-a-glance comparisons
- Search by research area in the Documentation Index
- Check Implementation Status for the latest coverage
Duration: 4-6 weeks
-
Week 1-2: Value-Based Methods
- DQN - Start here!
- Double DQN
- Rainbow DQN
-
Week 3-4: Policy Gradient Methods
-
Week 5-6: Advanced Topics
Duration: 3-4 weeks
-
Week 1: Attention Mechanisms
-
Week 2: State Space Models
-
Week 3: Hybrid Architectures
-
Week 4: Positional Encodings
Duration: 4-5 weeks
-
Week 1-2: Vision Transformers
-
Week 3: Object Detection & Segmentation
-
Week 4-5: 3D Vision
Duration: 2-3 weeks
-
Week 1: Model Compression
-
Week 2: Inference Optimization
-
Week 3: Training Infrastructure
Duration: 4-5 weeks
-
Week 1-2: Diffusion Models
-
Week 3: Video & Audio Generation
-
Week 4-5: Advanced Techniques
92 algorithms | 15,000+ lines of documentation
- Value-Based Methods - DQN, Rainbow, QR-DQN
- Policy Gradient - PPO, SAC, TRPO
- Offline RL - IQL, CQL, TD3+BC
- LLM Alignment - DPO, KTO, SimPO
- Multi-Agent RL - MAPPO, QMIX, MADDPG
- Model-Based RL - DreamerV3, MBPO
- Exploration - ICM, RND
- Sequence-Based RL - Decision Transformer
- Reward Modeling - PRM, ORM
- Planning - AlphaZero, MCTS
18 mechanisms | Comprehensive efficiency analysis
- Core Attention - Flash, GQA, Sparse, Linear
- Advanced Variants - Ring, Differential, MLA
- Specialized - FlashAttention-3, PagedAttention
14 SSMs | From S4 to Mamba-2
- S4 Family - S4, S4D, S5
- Mamba Family - Mamba, Mamba-2
- RWKV Family - RWKV-6, RWKV-7
- Advanced SSMs - RetNet, DeltaNet, HGRN
9 architectures | Attention + SSM combinations
- Attention-Recurrence Hybrids - Griffin, Hawk
- Attention-SSM Hybrids - Jamba, Zamba
- Convolutional Hybrids - Hyena, StripedHyena
- High-Efficiency - Based, GoldFinch
13 encodings | From sinusoidal to 2M+ tokens
- Classic Encodings - Sinusoidal, Learned, RoPE
- Attention Bias Methods - ALiBi, Relative Bias
- Context Extension - YaRN, LongRoPE, CLEX
- Advanced Methods - CoPE, Multiscale RoPE
16 components | Building blocks of modern models
- Mixture of Experts - DeepSeek MoE, Switch, MoD
- Normalization - RMSNorm, QK-Norm
- Activations - SwiGLU, GeGLU
9 techniques | Up to 100x speedup
- Memory Optimizations - KV Cache, PagedAttention
- Speculative Decoding - EAGLE-3, Medusa
- Batching Strategies - Continuous Batching
27 methods | Classification, Detection, 3D
- Vision Transformers - ViT, Swin, DINOv2
- Object Detection - YOLO-World, SAM, RT-DETR
- Segmentation - SAM, SAM 2, MedSAM
- NeRF & 3D - NeRF, Gaussian Splatting, LRM
22 models | Images, Audio, Video
- Diffusion Models - DiT, MMDiT, Flow Matching
- Audio Generation - VALL-E, MusicGen
- Video Generation - CogVideoX, VideoPoet
- Classic Methods - VAE, GANs
38 methods | Reasoning, RAG, Compression
- Reasoning - Chain-of-Thought, Tree of Thoughts, ReAct
- RAG - Self-RAG, GraphRAG, RAPTOR
- PEFT - LoRA, QLoRA, DoRA, LISA
- Quantization - GPTQ, AWQ, QuIP#
- Pruning - SparseGPT, Wanda, SliceGPT
- Embeddings - Matryoshka, BGE-M3
- Tokenization - BLT, MambaByte
20 techniques | Optimizers, Distributed, Mixed Precision
- Optimizers - Lion, SOAP, Sophia
- LR Schedules - WSD, SGDR
- Mixed Precision - FP8, MXFP8, FP4
- Distributed Training - FSDP2, ZeRO++
- Loss Functions - InfoNCE, VICReg
7 methods | Learning without labels
- Vision SSL - DINOv2, MAE, I-JEPA
- Video SSL - V-JEPA 2
- Multimodal SSL - data2vec 2.0
9 models | Vision-Language Integration
- Vision-Language Models - LLaVA, Qwen2-VL, Molmo
- Specialized VLMs - BiomedCLIP, NVLM
5 architectures | Graph representation learning
- Graph Transformers - GPS, Exphormer
- Classic GNNs - GATv2, GraphSAGE
4 models | Dynamics learning
- Model-Based RL - DreamerV3
- Self-Supervised World Models - I-JEPA, V-JEPA 2, Genie
4 methods | Learning without forgetting
- Continual Learning - EVCL, EWC, Prompt-Based CL
3 systems | End-to-end driving
- Autonomous Systems - UniAD, VAD, DriveTransformer
4 methods | Learning from demonstrations
- Imitation Learning - GAIL, DAgger, AIRL
3 techniques | Inference-time scaling
- Test-Time Methods - TTT Layers, Best-of-N, Compute-Optimal Scaling
Deep Learning Foundations
Sequence Modeling
Decision Making
Perception
Generation
Efficiency
Building a Production LLM
Computer Vision Application
Robotics & Control
Research & Experimentation
Every algorithm documentation includes:
- What problem does this solve?
- Key innovations and contributions
- When to use this method
- Mathematical foundations
- Historical context
- Core concepts
- Precise equations
- Loss functions
- Update rules
- Complexity analysis
- Conceptual explanations
- Visual diagrams
- Analogies
- Network architectures
- Hyperparameters
- Training procedures
- Line-by-line explanation
- References to actual Nexus code
- Usage examples
- Best practices
- Performance improvements
- Training stability techniques
- Benchmark performance
- Ablation studies
- Scaling behavior
- What can go wrong
- Symptoms and solutions
- Debugging tips
- Original papers (with arXiv links)
- Implementations
- Related work
- Blog posts and tutorials
- Follow a Learning Path - Structured curriculum from basics to advanced
- Start with READMEs - Category overviews provide context
- Deep dive into methods - Comprehensive 10-section guides
- Run the code - All examples reference actual Nexus implementations
- Compare methods - Comprehensive comparison tables
- Understand theory - Mathematical formulations and proofs
- Reproduce results - Hyperparameters and training details
- Extend algorithms - Clear implementation patterns
- Quick start examples - Get running immediately
- Code walkthroughs - Understand each component
- Optimization tricks - Production-grade techniques
- Troubleshooting - Common issues and solutions
- Search by category - Find specific algorithms quickly
- Comparison tables - Choose the right method
- Benchmarks - Performance metrics
- Paper links - Access original research
- Total Algorithms: 292+
- Documentation Files: 200+
- Lines of Documentation: 150,000+
- Research Papers Referenced: 500+
- Code Examples: 1,000+
- Benchmark Results: 300+
Reinforcement Learning: 92 algorithms (31%)
Computer Vision: 27 algorithms (9%)
NLP & LLMs: 38 algorithms (13%)
Generative Models: 22 algorithms (8%)
Attention Mechanisms: 18 algorithms (6%)
State Space Models: 14 algorithms (5%)
Training Infrastructure: 20 techniques (7%)
Other Categories: 61 algorithms (21%)
- β All algorithms: 10-section comprehensive structure
- β Category READMEs: Overview, comparison, navigation
- β Code references: Actual Nexus implementation files
- β Mathematical rigor: Equations, proofs, algorithms
- β Practical focus: Hyperparameters, tricks, troubleshooting
# Install Nexus
git clone https://github.com/yourusername/Nexus.git
cd Nexus
pip install -e .New to AI? Start with:
Familiar with basics? Jump to:
Advanced user? Explore:
Want to improve the documentation?
- Fix errors - Submit PRs for corrections
- Add examples - Share your use cases
- Improve clarity - Better explanations welcome
- Add visualizations - Diagrams and charts
- Share benchmarks - Your experimental results
See CONTRIBUTING.md for guidelines.
Documentation is licensed under CC BY 4.0 Code examples follow the main Nexus license
- Implementation Status - What's implemented
- Nexus GitHub - Source code
- Paper Collection - Organized research papers
- Benchmarks - Performance comparisons
- Start small - Master fundamentals before advanced topics
- Run the code - Hands-on experience beats reading
- Compare methods - Understand tradeoffs
- Read papers - Original sources provide depth
- Experiment - Modify examples and observe results
- Take notes - Document your insights
- Join discussions - Community learning accelerates progress
Ready to dive in? Pick a Learning Path or explore the Categories!
For questions or feedback, open an issue on GitHub.
Happy learning! π