mirror of
https://github.com/MLSysBook/TinyTorch.git
synced 2026-03-25 09:24:52 -05:00
Moved memoization (KV-cache) after compression to align with optimization tier milestones. Changes: - Module 15: Quantization (was 16) - Module 16: Compression (was 17) - Module 17: Memoization (was 15) Pedagogical Rationale: This creates clear alignment with the optimization milestone structure: - M06 (Profiling): Module 14 - M07 (Compression): Modules 15-16 (Quantization + Compression) - M08 (Acceleration): Modules 17-18 (Memoization/KV-cache + Acceleration) Before: Students learned KV-cache before understanding why models are slow After: Students profile → compress → then optimize with KV-cache Updated milestone reference in profile_kv_cache.py: Module 15 → Module 17
128 KiB
128 KiB