Modern AI applications often fail not because models are weak, but because memory is poorly designed. In this talk, I’ll share how to design reliable, time-aware memory layers in Python and what breaks when AI systems remember the wrong things.
Large Language Models don’t actually remember anything yet we keep building AI applications as if they do.
In real-world systems, memory is not a model feature but an architectural responsibility. As soon as we add user context, conversation history, or long-term state, our AI systems start behaving in unexpected ways: recalling outdated information, mixing contexts, or confidently acting on stale data.
In this talk, I’ll walk through how memory really works in modern AI applications, and why naive approaches-like endlessly growing prompts or unstructured context injection, eventually fail.
Using practical Python-based examples, we’ll explore different types of memory (short-term, long-term, and time-aware memory), common failure modes, and design patterns that keep AI behavior predictable and safe over time. I’ll also show how forgetting, summarization, and time-based decisions are just as important as remembering.
This talk is aimed at engineers building real AI products who want their systems to stay coherent, reliable, and maintainable.
Hello World, I’m Özge Çinko! 👋
I’m a computer engineer who finds inspiration at the intersection of curiosity and technology. Currently building the future as an AI Engineer at ING.
For me, engineering is a creative craft - turning data into narratives and emotions into visual experiences. I am passionate about making technology more human-centric and purposeful.
When I’m not coding, I’m usually writing, traveling, or chasing the thrill of learning something new.