Logo
Back to podcasts

Why Your Agent Needs Memory, Not Just Context

with Richmond Alake

Transcript

Chapters

Trailer
[00:00:00]
Introduction and guest welcome
[00:01:03]
Defining agent memory
[00:04:02]
From prompt to context engineering
[00:05:14]
Skills as SOPs for agents
[00:09:18]
What agents need to succeed
[00:13:25]
Getting agents to the right data
[00:16:06]
Security and data privacy
[00:20:18]
Context lifecycle and forgetting
[00:26:10]
Multi-team context sharing
[00:33:13]
File system vs database storage
[00:35:25]
Future of context engineering
[00:38:33]

In this episode

Not onboarding your agent is on you.


Richmond Alake, Director of AI Developer Experience at Oracle, joins Simon Maple to make the case that most agent failures come down to one thing: memory. Not the model, not the infrastructure. Memory.


On the docket:

  • why skills are just SOPs your organisation already has written down
  • the job title that is replacing prompt engineers
  • file systems vs databases for agent memory (and why one gets you hacked)
  • the memory trick that makes agents feel actually intelligent
  • why the agent loop and training loop are about to become one

The developers who figure this out first are going to be very hard to compete with.


Connect with us here:

  • Richmond Alake: https://www.linkedin.com/in/richmondalake/
  • Simon Maple: https://www.linkedin.com/in/simonmaple/
  • Tessl: https://www.linkedin.com/company/tesslio/

Memory Engineering: The Infrastructure Layer of Context Engineering

Context engineering has become the defining practice of agentic development, but there is a layer beneath it that determines whether agents can actually retrieve and apply the right information at the right time. In a recent episode of the AI Native Dev podcast, Simon Maple sat down with Richmond Alake, Director of AI Developer Experience at Oracle, to explore how memory systems, retrieval pipelines, and data infrastructure shape agent capabilities.

The conversation offered a framework for thinking about context that goes beyond prompt construction, grounding the discussion in how enterprise organisations can build agents that remember, adapt, and learn continuously.

What Is Memory Engineering?

Memory engineering is the discipline of optimising how agents store, retrieve, and forget information. It sits at the intersection of database engineering, search optimisation, software engineering, and AI engineering. Unlike prompt engineering, which focuses on how to phrase requests, or context engineering, which addresses what information to provide, memory engineering concerns itself with the infrastructure that makes retrieval efficient and reliable.

Richmond offered a useful framing: "Memory engineering is just a cross-section between a couple of disciplines. If you just cross-pollinate some of the techniques and disciplines within database engineers, search optimisation engineers, software engineers, and a bit of AI engineering, you're going to go into memory engineering."

This matters for development teams because the systems that store and retrieve context determine whether agents can access the right information quickly enough to be useful in practice. A well-designed memory system handles the encoding, storage, retrieval, and crucially, the forgetting of information as circumstances change.

Skills as Procedural Memory for Agents

One of the more instructive parts of the conversation addressed where skills fit within the broader memory architecture. Richmond drew a direct parallel to human cognition: just as the cerebellum stores procedural knowledge like how to perform physical tasks, agents need dedicated systems for storing and retrieving procedural knowledge about how to accomplish specific workflows.

"Skills are SOPs for agents," Richmond explained, referencing Standard Operating Procedures. "It's just a way of telling an agent, 'There's a task, and I'm going to give an arbitrary name to the task, and then I'm going to describe the task. Then I'm going to give you step-by-step instructions and maybe the locations of tools and scripts or APIs that allow you to achieve the outcome.'"

This framing connects skill development to established organisational practices. Most enterprises already have documented procedures for how work gets done. The challenge is translating those procedures into formats that agents can consume and apply. This aligns with the broader shift toward spec-driven development (https://claude.ai/blog/spec-driven-development-guide), where explicit specifications guide agent behaviour rather than relying on implicit assumptions.

The distinction between different types of memory proves useful here. Entity memory stores information about users and their preferences. Procedural memory, where skills reside, stores information about how to accomplish tasks. Separating these concerns allows teams to optimise each independently and reason about what information needs to be retrieved for different types of agent interactions.

The Memory Lifecycle: Ingest, Encode, Store, Retrieve, Forget

The conversation surfaced a lifecycle model for thinking about how context flows through agent systems. It starts with ingestion, where raw data enters the system. Then encoding, where embedding models transform that data into representations suitable for semantic search. Storage follows, typically in a database optimised for the relevant data types. Retrieval mechanisms determine how that stored information gets surfaced when needed.

But the critical addition to traditional data pipelines is forgetting. "When you work with context, forgetting or suppressing information doesn't come naturally," Richmond noted. "But when I speak to developers and I say, if you think about things from the perspective of memory, you immediately start to know that you need to implement a way of forgetting information."

This has practical implications for teams building agent systems. Context that was accurate six months ago may now be outdated or actively misleading. Without mechanisms to deprecate stale information, agents continue to retrieve and apply knowledge that no longer reflects reality. The research on generative agents from Stanford suggests approaches like weighted scoring that can affect how likely information is to be recalled, allowing systems to naturally phase out outdated context.

File Systems Versus Databases for Agent Memory

The conversation addressed a tension familiar to anyone building agent systems: the ease of file-based context versus the robustness of database-backed storage. File systems offer speed and simplicity. Agents have natural affinity for working with files given their training data. Development moves faster when teams do not have to think about infrastructure.

But that speed comes with costs. "Don't experiment with your tool selection," Richmond advised. As teams add requirements for concurrency, transactions, security, and data privacy to file-based systems, they often find themselves rebuilding database functionality from scratch.

The recommendation was clear: for production systems, particularly in enterprise environments, purpose-built databases provide security, scalability, and retrieval mechanisms that file systems cannot match. The experimentation phase benefits from file-based approaches, but production deployments need infrastructure designed for the retrieval patterns agents require.

Agents as Omnipresent Knowledge Capturers

One forward-looking observation from the conversation concerned how agents might solve the tribal knowledge problem that plagues large organisations. When experienced developers leave, their undocumented knowledge often leaves with them. Agents, by contrast, can observe workflows across an organisation and document successful patterns automatically.

"Agents today are omnipresent," Richmond observed. "They can go through different systems and understand different systems, and they can see interactions within a secure environment. Therefore, we don't have the problem of loss of knowledge."

The implication is that memory systems need to support not just retrieval but continuous learning. When an agent observes a successful workflow that was not previously documented, it can generate new procedural knowledge, perhaps a new skill file, and store that for future use. This creates a feedback loop where agent usage improves the context available for future agent interactions.

Toward Continuous Learning

The conversation pointed toward a future where agent loops and training loops merge. Rather than treating context as static information that agents consume, the trajectory appears to be toward systems that continuously learn from agent interactions. Cursor's approach of using agent traces to fine-tune embedding models represents an early example of this pattern.

For development teams working with agents today, the practical takeaway is to build memory systems that can evolve. Context will change. Requirements will shift. The teams that succeed will be those whose infrastructure supports iteration rather than requiring wholesale replacement when circumstances change.

The full conversation covers additional ground on entity memory for personalisation, security considerations for enterprise deployments, and the parallels between human and agent cognition. Worth a listen for teams thinking about the infrastructure layer of their agentic systems.

Chapters

Trailer
[00:00:00]
Introduction and guest welcome
[00:01:03]
Defining agent memory
[00:04:02]
From prompt to context engineering
[00:05:14]
Skills as SOPs for agents
[00:09:18]
What agents need to succeed
[00:13:25]
Getting agents to the right data
[00:16:06]
Security and data privacy
[00:20:18]
Context lifecycle and forgetting
[00:26:10]
Multi-team context sharing
[00:33:13]
File system vs database storage
[00:35:25]
Future of context engineering
[00:38:33]