mirror of
https://github.com/MLSysBook/TinyTorch.git
synced 2026-04-28 11:14:27 -05:00
- Added progressive complexity guidelines (Foundation/Intermediate/Advanced) - Added measurement function consolidation to prevent information overload - Fixed all diagnostic issues in losses_dev.py - Fixed markdown formatting across all modules - Consolidated redundant analysis functions in foundation modules - Fixed syntax errors and unused variables - Ensured all educational content is in proper markdown cells for Jupyter
Module 15: Profiling - Performance Detective Work
Overview
Become a performance detective! You just built MLPs, CNNs, and Transformers - but why is your transformer 100x slower than PyTorch? Build professional profiling infrastructure to reveal bottlenecks and guide optimization decisions.
What You'll Build
- Timer Class: Statistical timing with warmup runs and percentile reporting
- Memory Profiler: Track allocations, peak usage, and memory patterns
- FLOP Counter: Count operations and analyze computational complexity
- Profiler Context: Comprehensive profiling manager combining all tools
- Performance Analysis: Complete bottleneck detection and optimization guidance
Learning Objectives
- Statistical Timing: Build robust timing infrastructure with confidence intervals
- Memory Analysis: Track allocations and identify memory bottlenecks
- Computational Complexity: Count FLOPs and understand scaling behavior
- Bottleneck Detection: Use Amdahl's Law to identify optimization targets
- Systems Thinking: Connect profiling insights to production decisions
Prerequisites
- Module 14: Transformers (need models to profile)
- Understanding of basic complexity analysis (O(n), O(n²))
Key Concepts
Professional Timing Infrastructure
timer = Timer()
stats = timer.measure(model.forward, warmup=3, runs=100)
# Returns: mean, std, p50, p95, p99 with confidence intervals
Memory Profiling with tracemalloc
profiler = MemoryProfiler()
stats = profiler.profile(expensive_operation)
# Tracks: baseline, peak, allocated, memory patterns
FLOP Analysis for Architecture Comparison
counter = FLOPCounter()
flops = counter.count_attention(seq_len=128, d_model=512)
# Reveals: O(n²) scaling, computational bottlenecks
Comprehensive Profiling Context
with ProfilerContext("MyModel") as profiler:
result = profiler.profile_function(model.forward, args=(input,))
# Automatic report: timing + memory + FLOPs + insights
Performance Insights
- MLPs: Linear scaling, memory efficient, excellent for classification
- CNNs: Moderate speed, vectorizable, great for spatial data
- Transformers: O(n²) attention scaling, memory hungry, powerful but expensive
Real-World Applications
- Bottleneck Identification: Find the 20% of code using 80% of time
- Hardware Selection: Use profiling data to choose CPU vs GPU
- Cost Prediction: Estimate infrastructure costs from FLOP counts
- Optimization ROI: Amdahl's Law guides where to optimize first
Module Structure
- Timer Class: Statistical timing with warmup and confidence intervals
- Memory Profiler: Allocation tracking and peak usage analysis
- FLOP Counter: Operation counting for different layer types
- Profiler Context: Integrated profiling with automatic reporting
- Architecture Comparison: MLP vs CNN vs Transformer analysis
- Bottleneck Detection: Complete model profiling and optimization guidance
- Systems Analysis: Connect profiling insights to production decisions
Hands-On Detective Work
# Reveal the transformer bottleneck
with ProfilerContext("Transformer Analysis") as profiler:
output = profiler.profile_function(transformer.forward, args=(tokens,))
# Result: Attention consumes 73% of compute time!
# Next: Optimize attention in Module 16 (Acceleration)
Success Criteria
- ✅ Build timer with statistical rigor (warmup, percentiles, confidence intervals)
- ✅ Implement memory profiler tracking allocations and peak usage
- ✅ Create FLOP counter analyzing computational complexity
- ✅ Develop integrated profiling context for comprehensive analysis
- ✅ Identify bottlenecks using data-driven analysis
Systems Insights
- Attention is O(n²): 2x sequence length = 4x computation
- Memory bandwidth matters: Large models are memory-bound, not compute-bound
- Amdahl's Law rules: Optimize the bottleneck first for maximum impact
- Profiling drives decisions: Every major ML optimization started with profiling
ML Systems Focus
This module teaches performance analysis as the foundation of all optimization work. You'll build the same profiling tools used to optimize GPT, BERT, and every production ML system. Understanding performance through measurement is the first step toward building efficient ML systems.
The detective work you do here reveals the bottlenecks that Module 16 (Acceleration) will fix!