Files
cs249r_book/instructors/customization.qmd
Vijay Janapa Reddi 954f09b34d feat(instructors): add The Blueprint instructor site — course in a box
Complete instructor-facing Quarto site with 10 pages:
- Landing page with hero, semester cards, and component overview
- 8-step getting started checklist for adopting instructors
- Week-by-week syllabi for both semesters with direct links
- Course component map with SVG diagrams
- Pedagogy guide with facilitation strategies and cheat sheet
- Assessment rubrics with sample student work
- TA training guide with grading workflows
- Customization guide for quarters, seminars, and emphases
- FAQ addressing common adoption questions

Includes 4 SVG diagrams, ecosystem-aligned SCSS (light + dark),
and navbar structure matching the book sites.
2026-03-15 16:28:50 -04:00

121 lines
5.7 KiB
Plaintext
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
---
title: "Customization Guide"
subtitle: "Adapt the curriculum to your program's format and emphasis"
---
The 16-week syllabi are designed as complete, ready-to-use courses. But not every program has 16 weeks, and not every audience has the same emphasis. This guide shows how to adapt.
---
## 10-Week Quarter Version (Foundations)
For quarter systems, compress the 16-week Foundations syllabus:
| Quarter Week | Content (from 16-week) | What Changes |
|:---|:---|:---|
| 1 | Weeks 12 | Introduction + ML Systems combined; Module 01 only |
| 2 | Week 3 | ML Workflow; Module 02 starts |
| 3 | Weeks 45 | Data Engineering + Neural Computation; Module 02 + 03 |
| 4 | Week 6 | NN Architectures; Module 04 |
| 5 | Weeks 78 | Frameworks + Training; Module 05 + 06 |
| 6 | Week 9 | Data Selection; Module 07 |
| 7 | Week 10 | Model Compression + Lab 09 (Quantization); Module 08 |
| 8 | Week 11 | HW Acceleration + Lab 10 (Roofline) |
| 9 | Weeks 1314 | Serving + Operations; Labs 1213 |
| 10 | Week 16 | Capstone (AI Olympics, reduced scope) |
**What gets dropped:** Benchmarking (Week 12), Responsible Engineering (Week 15) — assign as optional reading. Integrate key responsibility points into the capstone rubric.
**What gets compressed:** TinyTorch Modules 01+02 doubled up in Weeks 1+3. Labs 00, 01, and 03 become optional.
::: {.callout-caution}
## Quarter Tradeoff
The 10-week version sacrifices breathing room. Consider reducing Decision Logs to 100 words and assigning only 2 Design Challenges instead of 4.
:::
---
## 3-Day Workshop Version
For short workshops with experienced practitioners:
| Day | Focus | Materials |
|:---|:---|:---|
| **Day 1** | The Physics of Inference | Iron Law introduction + Labs 01, 05, 09 (Magnitude Gap, Architecture Tradeoffs, Quantization) |
| **Day 2** | The Optimization Frontier | Labs 10, 11 (Roofline, Benchmarking) + TinyTorch Module 08 speed-run |
| **Day 3** | Production Deployment | Labs 12, 13 (Tail Latency, Drift Detection) + mini Design Challenge |
Focus exclusively on the **Iron Law** and **Interactive Labs**. Skip TinyTorch (except as demo). No formal assessment — use labs for hands-on discovery.
---
## For Software Engineers
If your audience is experienced developers, lean into **TinyTorch**:
| Weeks | Focus | Modules |
|:---|:---|:---|
| 14 | Building the Autograd Engine | Modules 0106 (Tensor → Autograd) |
| 58 | From CNNs to Transformers | Modules 0913 (Conv → Transformer) |
| 912 | Production Optimization | Modules 1419 (Profiling → Benchmarking) |
| 1316 | Capstone: Torch Olympics | Module 20 + competition |
Use textbook chapters as background reading, not lecture material. Labs serve as validation checkpoints, not primary pedagogy.
---
## For Computer Architects
Shift the focus toward **Hardware Acceleration** and **mlsysim**:
- Use the hardware Zoo in `mlsysim` to compare architectures (H100, B200, edge devices)
- Spend 2 weeks on the Roofline model — have students plot multiple workloads
- Extend model compression to 2 weeks (quantization + pruning as hardware-aware optimizations)
- Use hardware kits extensively — make them mandatory, not optional
- Reduce TinyTorch to Modules 0103 (enough to understand what frameworks do)
---
## Graduate Seminar Version
For a graduate-level seminar (assumes strong systems background):
| Week | Topic | Textbook | Paper |
|:---|:---|:---|:---|
| 1 | The Iron Law | Vol I: Intro + ML Systems | Hennessy & Patterson, "A New Golden Age" (2019) |
| 2 | Memory Hierarchy | Vol I: HW Acceleration | Williams et al., "Roofline" (2009) |
| 3 | Quantization | Vol I: Model Compression | Dettmers et al., "LLM.int8()" (2022) |
| 4 | Serving Systems | Vol I: Model Serving | Yu et al., "Orca" (2022) |
| 5 | Distributed Training | Vol II: Distributed Training | Shoeybi et al., "Megatron-LM" (2020) |
| 6 | 3D Parallelism | Vol II: Distributed Training | Narayanan et al., "Efficient Large-Scale Training" (2021) |
| 7 | Collective Comms | Vol II: Collective Comm. | Patarasuk & Yuan, "Bandwidth Optimal All-Reduce" (2009) |
| 8 | Fault Tolerance | Vol II: Fault Tolerance | Jeon et al., "Large-Scale GPU Clusters" (2019) |
| 9 | Inference at Scale | Vol II: Inference | Kwon et al., "vLLM/PagedAttention" (2023) |
| 10 | KV-Cache Optimization | Vol II: Inference | Ainslie et al., "GQA" (2023) |
| 11 | Edge Intelligence | Vol II: Edge Intelligence | Lin et al., "MCUNet" (2020) |
| 12 | Fleet Operations | Vol II: Ops at Scale | Zhao et al., "ATC'24 Fleet Analysis" (2024) |
| 13 | Sustainability | Vol II: Sustainable AI | Patterson et al., "Carbon Emissions and AI" (2021) |
| 14 | Student presentations | — | — |
**Assessment:** 40% paper presentations, 30% lab Decision Logs (selected labs only), 30% semester project (original system design or benchmarking study).
---
## Mixing and Matching Components
Each component is independently adoptable:
| Pattern | Components Used | Typical Context |
|:---|:---|:---|
| **Textbook Only** | Vol I or II as required reading | Supplement for existing ML course |
| **Textbook + Labs** | Readings + interactive labs | Active learning without coding assignments |
| **TinyTorch Only** | 20 modules as programming assignments | Systems programming course |
| **Labs Only** | Interactive labs as in-class activities | Active learning supplement for any course |
| **Hardware Kits Only** | Edge deployment labs | Embedded systems course |
| **Full Stack** | All components integrated | Dedicated ML Systems course |
::: {.callout-tip}
## Start Small, Layer Up
If adopting for the first time, start with **Textbook + Labs** for one semester. Add TinyTorch the second time you teach it. Add hardware kits the third. Each component is valuable on its own.
:::