mirror of
https://github.com/harvard-edge/cs249r_book.git
synced 2026-05-07 02:03:55 -05:00
[PR #1549] [MERGED] fix(book): swap inverted MobileNet quantization values + align caption with figure #8257
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
📋 Pull Request Information
Original PR: https://github.com/harvard-edge/cs249r_book/pull/1549
Author: @profvjreddi
Created: 4/26/2026
Status: ✅ Merged
Merged: 4/26/2026
Merged by: @profvjreddi
Base:
dev← Head:fix/issue-1318-quantization-figure📝 Commits (1)
b970393fix(book): swap inverted MobileNet quantization values + align caption with figure📊 Changes
1 file changed (+6 additions, -6 deletions)
View changed files
📝
book/quarto/contents/vol1/optimizations/model_compression.qmd(+6 -6)📄 Description
Closes #1318.
Summary
Reporter observed that the "Quantization Impact" figure in the Model Compression chapter (@fig-quantization_impact) doesn't match its caption. Inspection revealed two compounding issues.
Issue 1: Inverted MobileNet data
In the original TikZ source, MobileNet_v1's FP32 and INT8 values were swapped in both panels:
The original direction was physically impossible — quantization cannot produce a slower, larger model. The other two models (Inception_v3, ResNet_v2) had the correct direction.
Issue 2: Caption claim ≠ figure data
Even after the swap, the gains vary substantially by architecture (Inception ~1.5–2×, ResNet ~1.85–4.3×, MobileNet ~10–23×), so the original "up to 4 times" claim under-described the figure and contradicted what readers actually see.
Caption rewritten to acknowledge per-architecture variability and the role of mobile-class INT8 hardware in MobileNet's outlier behavior. Body prose updated for consistency.
Test plan
🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.