Discussion about this post

User's avatar
Martin Davidson's avatar

Maybe. Even with today’s tech we could do a better job with memory. Store previous conversations in a RAG database, then search that and pull in relevant previous context for future chats. That’s likely how the current ChatGPT “memory” works - and probably what Mustafa Suleyman is referring to when he talks about “near infinite” memory coming next year.

It also raises some interesting questions:

Is near infinite memory universally better? Sometimes historical context is useful, sometimes it’s not. The human ability to forget inconsequential things may turn out to be a thing AI models need to replicate. Actually, I’ll go further - I reckon AI models _will_ need the ability to prune memory :).

Today models struggle to retain consistency within long chats - Gemini often forgets the original question, ChatGPT has a tendency to hit guard rails in long conversations and refuse to continue. If models struggle within the context of a single conversation, how will they cope with all the history as well? This one is probably soluable - Claude is very good at remaining consistent within a chat - maybe Anthropic have solved it.

Is memory the same as learning? In-context learning (provide lots of examples in the context) works. But I don’t want to have to train an AI model like I train my dog by repeating the same things over and over again.

Memory feels like it brings a new set of human-esque problems: things you want the model to always remember, things that are secret, things you don’t want the model ever to mention again - and the things that just don’t matter!

Piers Finlayson's avatar

With increasing context window sizes, do you see the current as a temporary situation, where eventually the whole of the chat history will be context?

No posts

Ready for more?