[{"data":1,"prerenderedAt":70},["ShallowReactive",2],{"term-c\u002Fcontext-compaction":3,"related-c\u002Fcontext-compaction":59},{"id":4,"title":5,"acronym":6,"body":7,"category":40,"description":41,"difficulty":42,"extension":43,"letter":44,"meta":45,"navigation":46,"path":47,"related":48,"seo":53,"sitemap":54,"stem":57,"subcategory":6,"__hash__":58},"terms\u002Fterms\u002Fc\u002Fcontext-compaction.md","Context Compaction",null,{"type":8,"value":9,"toc":33},"minimark",[10,15,19,23,26,30],[11,12,14],"h2",{"id":13},"eli5-the-vibe-check","ELI5 — The Vibe Check",[16,17,18],"p",{},"Context compaction is summarizing a long AI conversation down to just the important bits so the model can keep going without hitting context limits. Like making a cliff-notes version of your chat history so the AI doesn't lose the plot.",[11,20,22],{"id":21},"real-talk","Real Talk",[16,24,25],{},"Context compaction is the automated or manual summarization of prior conversation context to reduce token count while preserving essential state. Implementations range from naive (keep last N turns) to sophisticated (semantic summarization with preserved facts, decisions, and file references). Claude Code, Cursor, and most agent frameworks implement some form of compaction.",[11,27,29],{"id":28},"when-youll-hear-this","When You'll Hear This",[16,31,32],{},"\"Hit 900k tokens — time for context compaction.\" \u002F \"Our agent loops automatically compact every 50k tokens.\"",{"title":34,"searchDepth":35,"depth":35,"links":36},"",2,[37,38,39],{"id":13,"depth":35,"text":14},{"id":21,"depth":35,"text":22},{"id":28,"depth":35,"text":29},"ai","Context compaction is summarizing a long AI conversation down to just the important bits so the model can keep going without hitting context limits.","intermediate","md","c",{},true,"\u002Fterms\u002Fc\u002Fcontext-compaction",[49,50,51,52],"Context Window","Context Rot","Prompt Compression","Long Context",{"title":5,"description":41},{"changefreq":55,"priority":56},"weekly",0.7,"terms\u002Fc\u002Fcontext-compaction","Ngwsy7-mtjb2lQguvKT8r8vhbH0C6672J9BzBizMj1c",[60,63,67],{"title":50,"path":61,"acronym":6,"category":40,"difficulty":42,"description":62},"\u002Fterms\u002Fc\u002Fcontext-rot","Context rot is when a long AI conversation has been running so long that the model is confusing old instructions with new ones, forgetting what it already...",{"title":49,"path":64,"acronym":6,"category":65,"difficulty":42,"description":66},"\u002Fterms\u002Fc\u002Fcontext-window","vibecoding","A context window is how much text an AI can 'see' at once — its working memory.",{"title":51,"path":68,"acronym":6,"category":40,"difficulty":42,"description":69},"\u002Fterms\u002Fp\u002Fprompt-compression","Prompt compression is shrinking a prompt so it fits more context or costs less, without losing meaning.",1776518269540]