1 Commits

Author SHA1 Message Date
tchopra
78095b8213 Add Headroom Context Optimization to LLM Optimization Tools
Headroom reduces LLM API costs by 50-90% through intelligent context
compression for AI agents. Tool outputs are 70-95% redundant boilerplate
that LLMs pay for but don't need - Headroom compresses that away while
preserving accuracy.

Features:
- SmartCrusher: Statistical compression of JSON tool outputs
- CacheAligner: Provider-side cache optimization
- Memory: Persistent memory across conversations
- MCP support: Tool output compression for Claude
- Framework integrations: LangChain, Agno, any OpenAI client
2026-01-20 00:13:53 -08:00