mirror of
https://github.com/harvard-edge/cs249r_book.git
synced 2026-05-08 02:28:25 -05:00
Two new pieces close the generation→validation→saturation feedback loop: 1. gemini_cli_llm_judge.py — multi-criteria validator. For each draft, judges math correctness, cell-fit (does it actually target the declared track/zone/level?), scenario realism, uniqueness vs canonical questions, and visual-asset alignment. Returns PASS/NEEDS_FIX/DROP per item. Batched (default 15 per call) for budget efficiency. 2. iterate_coverage_loop.py — drives the full loop: analyze → plan → generate → render → judge → apply → re-analyze. Self-paced: stops when (a) top priority gap drops below threshold, (b) DROP rate exceeds the saturation/hallucination threshold, (c) total API calls exceed budget, or (d) the same cell is top priority for two iterations in a row (convergence). The user no longer specifies "how many questions" — the loop generates until the corpus reaches a measurable steady state. Plus 25 round-1 visual questions generated by the new batched generator (5 batched calls × 5 cells each, zero failures). The loop is the answer to "we need balance, not just volume": every iteration's plan derives from a fresh analysis of where coverage is weakest, so generation can never over-fill an already-saturated cell.
13 lines
522 B
Python
13 lines
522 B
Python
import os
|
|
import matplotlib.pyplot as plt
|
|
fig, ax = plt.subplots(figsize=(6, 4))
|
|
tiers = ['CPU DRAM\n(PCIe Gen5)', 'Effective Target', 'HBM3 (H100)']
|
|
bw = [64, 1600, 3200]
|
|
ax.bar(tiers, bw, color=['#cfe2f3', '#fdebd0', '#d4edda'], edgecolor='black')
|
|
ax.set_ylabel('Bandwidth (GB/s)')
|
|
ax.set_title('Memory Tier Bandwidth Comparison')
|
|
ax.set_yscale('log')
|
|
plt.axhline(y=1600, color='r', linestyle='--')
|
|
plt.tight_layout()
|
|
out = os.environ.get("VISUAL_OUT_PATH", "out.svg")
|
|
plt.savefig(out, format="svg", bbox_inches="tight") |