ChatGPT will now remember things about you

It appears users will have the feature turned on by default
By Mike Pearl  on 
a graphic representation of personal data stored in a computer system
For all we know, an AI's memory of your personal data might look like this Credit: Vertigo3d via Getty Images

OpenAI announced the release of a new feature late yesterday for a limited number of ChatGPT users that allows the chatbot to retain information gleaned from human-AI interactions. This "memory" capability is meant to save users the trouble of repeating information, though it will no doubt sound to many reasonable observers like yet another piece of tech gathering details about us.

And yes, OpenAI does appear to be turning the memory feature for ChatGPT on by default. "You can turn off memory at any time," the official blog post about memory notes.  

Users who leave memory on are encouraged to manage the feature, much in the way the Men in Black manage the memories of hapless bystanders after encounters with aliens: by forcing it to forget. But no neuralizer is required; instead you can apparently just "tell it to forget conversationally," OpenAI says, which evidently means you can include something like, "don't store this in memory, but..." in a prompt. If only your gossipy barber could be thwarted so easily.  

Exactly what an AI "memory" consists of is not yet clear — and may never be — but an OpenAI video shows a fictional user managing memories in their user settings, and the bulleted list of memories is revealing. Memories outwardly appear to be pithy little snippets of text about preferences and biographical information, similar to what a movie cop would write down in a notebook while interviewing a witness. "Daughter, Lina, loves jellyfish," reads one. "Prefers assistance with writing and blog posts to be more concise, straightforward, and less emotive," says another. "Safe full of valuables is near unlocked side door," reads another. Just kidding about that third one. 

But the type of information the feature retains is, nonetheless, a little concerning, particularly since it's easy to imagine heavy ChatGPT users inadvertently revealing the contours of their workplace, family, and medical situations — not to mention hints as to their their innermost feelings about those situations — to a machine that will remember them conceivably forever. Even more concerning, OpenAI has already has a history of accidentally leaking stored conversations.

In an effort to allay such concerns, OpenAI says it's allowing users with the feature enabled to switch on a "temporary chat" option for memory-free conversations, a feature seemingly inspired by incognito mode in modern web browsers. And OpenAI also claims that it will prevent the proactive memorization of sensitive data, unless "explicitly" requested by the user. This hints at a sub-feature in ChatGPT memory that, when it detects you've just told it, say, your family history of cancer, will say something like, "looks like some pretty sensitive data you've got there. Want me to remember that?"

For now, OpenAI says the feature is in testing, and that it will be "rolling out to a small portion of ChatGPT free and Plus users this week," and that it's still being evaluated for usefulness. If you use ChatGPT don't forget to check and see if it's on.


Recommended For You
Yes, ChatGPT got lazier. But OpenAI finally has a fix.

We tested Roborock's most advanced (and expensive) robot vacuum


The best free dating apps for budget-conscious singles
By Leah Stodart, Miller Kern, Bethany Allard, Tabitha Britt, and Stacia Datskovska

Unblock ChatGPT for free with this quick hack

Trending on Mashable
NYT Connections today: See hints and answers for February 21

Wordle today: Here's the answer and hints for February 21

NYT Connections today: See hints and answers for February 20


How to try Sora, OpenAI's AI video generator
The biggest stories of the day delivered to your inbox.
This newsletter may contain advertising, deals, or affiliate links. Subscribing to a newsletter indicates your consent to our Terms of Use and Privacy Policy. You may unsubscribe from the newsletters at any time.
Thanks for signing up. See you at your inbox!