No menu items!
EletiofeOpenAI Gives ChatGPT a Memory

OpenAI Gives ChatGPT a Memory

-

- Advertisment -

OpenAI says ChatGPT’s Memory is opt-in by default, which means a user has to actively turn it off. The Memory can be wiped at any point, either in settings or by simply instructing the bot to wipe it. Once the Memory setting is cleared, that information won’t be used to train its AI model. It’s unclear exactly how much of that personal data is used to train the AI while someone is chatting with the chatbot. And toggling off Memory does not mean you’ve totally opted out of having your chats train OpenAI’s model; that’s a separate opt-out.

The company also claims that it won’t store certain sensitive information in Memory. If you tell ChatGPT your password (don’t do this) or Social Security number (or this), the app’s Memory is thankfully forgetful. Jang also says OpenAI is still soliciting feedback on whether other personally identifiable information, like a user’s ethnicity, is too sensitive for the company to auto-capture.

“We think there are a lot of useful cases for that example, but for now we have trained the model to steer away from proactively remembering that information,” Jang says.

It’s easy to see how ChatGPT’s Memory function could go awry—instances where a user might have forgotten they once asked the chatbot about a kink, or an abortion clinic, or a nonviolent way to deal with a mother-in-law, only to be reminded of it or have others see it in a future chat. How ChatGPT’s Memory handles health data is also something of an open question. “We steer ChatGPT away from remembering certain health details but this is still a work in progress,” says OpenAI spokesperson Niko Felix. In this way ChatGPT is the same song, just in a new era, about the internet’s permanence: Look at this great new Memory feature, until it’s a bug.

OpenAI is also not the first entity to toy with memory in generative AI. Google has emphasized “multi-turn” technology in Gemini 1.0, its own LLM. This means you can interact with Gemini Pro using a single-turn prompt—one back-and-forth between the user and the chatbot—or have a multi-turn, continuous conversation in which the bot “remembers” the context from previous messages.

An AI framework company called LangChain has been developing a Memory module that helps large language models recall previous interactions between an end user and the model. Giving LLMs a long-term memory “can be very powerful in creating unique LLM experiences—a chatbot can begin to tailor its responses toward you as an individual based on what it knows about you,” says Harrison Chase, cofounder and CEO of LangChain. “The lack of long-term memory can also create a grating experience. No one wants to have to tell a restaurant-recommendation chatbot over and over that they are vegetarian.”

This technology is sometimes referred to as “context retention” or “persistent context” rather than “memory,” but the end goal is the same: for the human-computer interaction to feel so fluid, so natural, that the user can easily forget what the chatbot might remember. This is also a potential boon for businesses deploying these chatbots that might want to maintain an ongoing relationship with the customer on the other end.

“You can think of these as just a number of tokens that are getting prepended to your conversations,” says Liam Fedus, an OpenAI research scientist. “The bot has some intelligence, and behind the scenes it’s looking at the memories and saying, ‘These look like they’re related; let me merge them.’ And that then goes on your token budget.”

Fedus and Jang say that ChatGPT’s memory is nowhere near the capacity of the human brain. And yet, in almost the same breath, Fedus explains that with ChatGPT’s memory, you’re limited to “a few thousand tokens.” If only.

Is this the hypervigilant virtual assistant that tech consumers have been promised for the past decade, or just another data-capture scheme that uses your likes, preferences, and personal data to better serve a tech company than its users? Possibly both, though OpenAI might not put it that way. “I think the assistants of the past just didn’t have the intelligence,” Fedus said, “and now we’re getting there.”

Will Knight contributed to this story.

Latest news

7 Best Handheld Gaming Consoles (2024): Switch, Steam Deck, and More

It feels like a distant memory by now, but right before the Nintendo Switch launched in 2017, it seemed...

The Boeing Starliner Astronauts Will Come Home on SpaceX’s Dragon Next Year

NASA has announced that astronauts Barry Wilmore and Sunita Williams will return to Earth next February aboard SpaceX’s Dragon...

How to Switch From iPhone to Android (2024)

Ignore the arguments about which is better, because iPhones and Android phones have far more in common than some...

12 Best Tablets (2024): iPads, Androids, and More Tested and Compared

Tablets often don't come with kickstands or enough ports, so it's a good idea to snag a few accessories...
- Advertisement -

Will the ‘Car-Free’ Los Angeles Olympics Work?

THIS ARTICLE IS republished from The Conversation under a Creative Commons license.With the Olympic torch extinguished in Paris, all...

Lionel Messi will return before MLS playoffs, says Inter Miami coach Tata Martino

Inter Miami head coach Tata Martino said on Friday that Lionel Messi will return to the team's lineup before...

Must read

7 Best Handheld Gaming Consoles (2024): Switch, Steam Deck, and More

It feels like a distant memory by now, but...

The Boeing Starliner Astronauts Will Come Home on SpaceX’s Dragon Next Year

NASA has announced that astronauts Barry Wilmore and Sunita...
- Advertisement -

You might also likeRELATED
Recommended to you