Context should behave more like a lifecycle than a dump: short-term working state, compact summaries, longer-term memory, and periodic consolidation each serve different jobs and should not be collapsed into one context window.
Context Lifecycle for AI Systems
Most discussions about context still assume a crude model: make the context window bigger, stuff more information into it, and let the model sort it out.
That helps for a while, but it is not a serious memory architecture.
The Real Problem
Different kinds of information do different jobs:
- immediate working state helps a model act right now
- compact summaries help it orient quickly
- source documents preserve nuance and evidence
- long-lived memory stores recurring patterns over time
When those layers are collapsed into one giant prompt, everything becomes expensive and blurry. The model spends attention on the wrong things, retrieval gets sloppy, and freshness becomes harder to reason about.
Summary First, Drill Down Second
The practical rule is simple:
read the shortest useful layer first.
That means a good knowledge system should not only expose full documents. It should expose a compact layer that answers:
- what is this page about
- what are the core claims
- how big is the full thing
- when was it updated
- should I keep reading
That compact layer is not just UX. It is context budgeting as architecture.
Consolidation Matters
Static RAG tends to assume that once a document is stored, the problem is solved. But useful memory systems do more than store. They consolidate. They compress. They strengthen what matters and let low-value detail recede.
Most RAG implementations are static — embed once, retrieve forever. But biological memory re-encodes, strengthens and weakens connections, and changes accessibility based on context. A good knowledge system should aspire to that: a dynamic system that updates, re-indexes, and consolidates over time.
The Sleep Loop
The most compelling pattern here is a periodic sleep pass. The system should:
- review what changed recently
- promote durable knowledge into longer-term memory
- compress what does not need full detail anymore
- emit a changelog of what changed and why
This turns storage into a living memory system with consolidation, not a static archive. The "actively work, then update memory based on learnings" pattern is a sleep-cycle-for-AI approach.
An event-driven orchestrator can sit on top of that. If memory updates, another process decides whether a summary, index, or follow-up action needs to happen. The result is a system that gets better over time instead of just getting bigger.
Why This Matters For Public Pages
civ.build treats public content the same way:
- full pages exist for depth
- compact summaries exist for first-pass retrieval
- version history tracks change over time
- trust metadata tells the reader how to use the page
The broader lesson is that context should have a lifecycle, not just a storage location.