[{"data":1,"prerenderedAt":70},["ShallowReactive",2],{"term-c\u002Fcontext-budget":3,"related-c\u002Fcontext-budget":58},{"id":4,"title":5,"acronym":6,"body":7,"category":40,"description":41,"difficulty":42,"extension":43,"letter":44,"meta":45,"navigation":46,"path":47,"related":48,"seo":52,"sitemap":53,"stem":56,"subcategory":6,"__hash__":57},"terms\u002Fterms\u002Fc\u002Fcontext-budget.md","Context Budget",null,{"type":8,"value":9,"toc":33},"minimark",[10,15,19,23,26,30],[11,12,14],"h2",{"id":13},"eli5-the-vibe-check","ELI5 — The Vibe Check",[16,17,18],"p",{},"Context budget is how you allocate tokens across system prompt, memory, tools, and conversation — every token is a dollar at scale. You've got 200k tokens to spend per call. Your system prompt eats 5k, tool definitions another 10k, conversation history is 80k. What's left for actual user content? Context budget is the discipline of making those numbers work before you hit the limit or the invoice.",[11,20,22],{"id":21},"real-talk","Real Talk",[16,24,25],{},"Context budget management involves optimizing the distribution of a model's finite context window across competing inputs: system instructions, retrieved documents, tool schemas, chat history, and the current user message. At scale, poor context budgeting means expensive calls with irrelevant content, or truncated history causing the model to forget earlier conversation. Techniques include dynamic summarization, RAG for selective retrieval, and tool schema minimization.",[11,27,29],{"id":28},"when-youll-hear-this","When You'll Hear This",[16,31,32],{},"\"We blew the context budget with full chat history — switching to rolling summarization.\" \u002F \"Model routing helps, but the real win was fixing the context budget allocation.\"",{"title":34,"searchDepth":35,"depth":35,"links":36},"",2,[37,38,39],{"id":13,"depth":35,"text":14},{"id":21,"depth":35,"text":22},{"id":28,"depth":35,"text":29},"ai","Context budget is how you allocate tokens across system prompt, memory, tools, and conversation — every token is a dollar at scale.","intermediate","md","c",{},true,"\u002Fterms\u002Fc\u002Fcontext-budget",[49,50,51],"Context Window","Token","Prompt Engineering",{"title":5,"description":41},{"changefreq":54,"priority":55},"weekly",0.7,"terms\u002Fc\u002Fcontext-budget","1QLJ6VWEnSN7rjHWVu5LBDikQMyiR9fbggZgHy5vNsY",[59,63,66],{"title":49,"path":60,"acronym":6,"category":61,"difficulty":42,"description":62},"\u002Fterms\u002Fc\u002Fcontext-window","vibecoding","A context window is how much text an AI can 'see' at once — its working memory.",{"title":51,"path":64,"acronym":6,"category":61,"difficulty":42,"description":65},"\u002Fterms\u002Fp\u002Fprompt-engineering","Prompt engineering is the art of talking to AI so it actually does what you want.",{"title":50,"path":67,"acronym":6,"category":61,"difficulty":68,"description":69},"\u002Fterms\u002Ft\u002Ftoken","beginner","In AI-land, a token is a chunk of text — roughly 3\u002F4 of a word.",1775560888758]