Engram Memory for Agents on OCI - Part 1. The Why and the What
Copyright: Sanjay Basu 1. Why Memory, and Why Now Long context is not memory. This is the sentence I keep wanting to staple to people’s foreheads at conferences. Yes, frontier models will happily accept two million tokens of input. No, that does not mean stuffing every prior conversation into the prompt is a good idea, even if you can afford the bill, which most companies cannot. The empirical case against the long-context-as-memory pattern is by now embarrassingly well documented. The original Lost in the Middle paper from Stanford showed that retrieval accuracy collapses when relevant facts are buried in the middle of a long context window. Every follow-up study since (NoLiMa, Michelangelo, RULER, the whole genre) has confirmed the same shape. Effective context length is much smaller than nominal context length. The model’s attention is not democratic. It cares about the beginning, it cares about the end, and the middle goes to the same place socks go in the dryer. Then yo...