1 link tagged with all of: optimization + llm-efficiency + agent-architecture + token-management + context-tax
Links
This article explains the concept of the "Context Tax" in large language models (LLMs) and offers strategies to minimize token usage and improve performance. It covers techniques like stable prefixes, append-only context, and using precise tools to enhance cache hits and reduce costs.
context-tax ✓
optimization ✓
token-management ✓
llm-efficiency ✓
agent-architecture ✓