It may be easy, even comforting, to imagine that using AI tools involves interacting with a purely objective, stoic, independent machine that knows nothing about you. But between cookies, device identifiers, login and account requirements, and even the occasional human reviewer, the voracious appetite online services have for your data seems insatiable.
Privacy is a major concern that both consumers and governments have about the pervasive spread of AI. Across the board, platforms highlight their privacy featuresâeven if they’re hard to find. (Paid and business plans often exclude training on submitted data entirely.) But any time a chatbot âremembersâ something can still feel intrusive.
In this article, we will explain how to tighten your AI privacy settings by deleting your previous chats and conversations and by turning off settings in ChatGPT, Gemini (formerly Bard), Claude, Copilot, and Meta AI that allow developers to train their systems on your data. These instructions are for the desktop, browser-based interface for each.
ChatGPT
Still the flagship of the generative AI movement, OpenAIâs ChatGPT has several features to improve privacy and alleviate concerns about user prompts being used to train the chatbot.
In April, OpenAI announced that ChatGPT could be used without an account. By default, prompts shared through the free, account-less version are not saved. But, if a user does not want their chats used to train ChatGPT, they still need to toggle the âTemporary chatâ settings in the ChatGPT dropdown menu at the top left of the screen.
If you have an account and subscription to ChatGPT Plus, however, how do you keep your prompts from being used? GPT-4 gives users the ability to delete all chats under its general settings. Again, to make sure chats are also not used to train the AI model, look lower to âData controlsâ and click the arrow to the right of âImprove the model for everyone.â
A separate âModel improvementâ section will appear, allowing you to toggle it off and select “Done.” This will remove the ability of OpenAI to use your chats to train ChatGPT.
There are still caveats, however.
“While history is disabled, new conversations wonât be used to train and improve our models, and wonât appear in the history sidebar,” an OpenAI spokesperson told Decrypt. “To monitor for abuseâand reviewed only when we need toâwe will retain all conversations for 30 days before permanently deleting.”
Claude
âWe do not train our models on user-submitted data by default,â an Anthropic spokesperson told Decrypt. âThus far, we have not used any customer or user-submitted data to train our generative models, and weâve expressly stated so in the model card for our Claude 3 model family,â
âWe may use user prompts and outputs to train Claude where the user gives us express permission to do so, such as clicking a thumbs up or down signal, on a specific Claude output to provide us feedback,â the company added, noting that it helps the AI model âlearn the patterns and connections between words.â
Deleting archived chats in Claude will also keep them out of reach. âI do not retain or have access to any previously deleted chats or conversations,â the Claude AI agent helpfully answers in the first person. âMy responses are generated based on the current conversation only.â
Like ChatGPT, Claude does hold on to some information as required by law.
âWe also retain data in our backend systems for the amount of time specified in our Privacy Policy unless required to enforce our Acceptable Use Policy, address Terms of Service or policy violations, or as required by law,â Anthropic explains.
As for Claude’s collection of information across the web, an Anthropic spokesperson told Decrypt that the AI developerâs web crawler respects industry-standard technical signals like robots.txt that site owners can use to opt-out of data collection, and that Anthropic does not access password-protected pages or bypass CAPTCHA controls.
Gemini
By default, Google tells Gemini users that âyour conversations are processed by human reviewers to improve the technologies powering Gemini Apps. Don’t enter anything that you wouldn’t want to be reviewed or used.”
But Gemini AI users can delete their chatbot history and opt out of having their data used to train its model going forward.
To accomplish both, navigate to the bottom left of the Gemini homepage and locate âActivity.â
Once on the activity screen, users can then turn off âGemini Apps Activity.â
A Google representative explained to Decrypt what the âGemini Apps Activityâ setting does.
âIf you turn it off, your future conversations wonât be used to improve our generative machine-learning models by default,â the company representative said. âIn this instance, your conversations will be saved for up to 72 hours to allow us to provide the service and process any feedback you may choose to provide. In those 72 hours, unless a user chooses to give feedback in Gemini Apps, it wonât be used to improve Googleâs products, including our machine learning technology.â
There is also a separate setting to clear out your Google-connected YouTube history.
Copilot
In September, Microsoft added its Copilot generative AI model to its Microsoft 365 suite of business tools, its Microsoft Edge browser, and Bing search engine. Microsoft also included a preview version of the chatbot in Windows 11. In December, Copilot was added to the Android and Apple app stores.
Microsoft does not provide the option to opt out of having user data used to train its AI models, but like Google Gemini, Copilot users can delete their history. The process is not as intuitive on Copilot, however, as previous chats still show on the desktop versionâs home screen even after being deleted.
To find the option to delete Copilot history, open your user profile at the top right of your screen (you must be signed in) and select âMy Microsoft Account.â On the left, select âPrivacy,â and scroll down to the bottom of the screen to find the Copilot section.
Because Copilot is integrated into Bingâs search engine, clearing activity will also clear search history, Microsoft said.
A Microsoft spokesperson told Decrypt that the tech giant protects consumersâ data through various techniques, including encryption, deidentification, and only storing and retaining information associated with the user for as long as is necessary.
âA portion of the total number of user prompts in Copilot and Copilot Pro responses are used to fine-tune the experience,â the spokesperson added. âMicrosoft takes steps to de-identify data before it is used, helping to protect consumer identity,â adding that Microsoft does not use any content created in Microsoft 365 (Word, Excel, PowerPoint, Outlook, Teams) to train underlying âfoundational models.â
Meta AI
In April, Metaâthe parent company of Facebook, Instagram, and WhatsAppârolled out Meta AI to users.
âWe are releasing the new version of Meta AI, our assistant, that you can ask any question across our apps and glasses,â Zuckerberg said in an Instagram video. âOur goal is to build the world’s leading AI and make it available to everyone.â
Meta AI does not provide users the option to opt out of having their inputs used to train the AI model. Meta does give the option to delete past chats with its AI agent.
To do so from a desktop computer, click the Facebook settings tab at the bottom left of your screen, located above your Facebook profile image. Once in settings, users have the option to delete conversations with Meta AI.
Meta does explain that deleting conversations here will not delete chats with other people in Messenger, Instagram, or WhatsApp.
A Meta spokesperson declined to comment on whether or how users could exclude their information from being used in Meta AI model training, instead pointing Decrypt to a September statement by the company about its privacy safeguards and the Meta settings page on deleting history.
âPublicly shared posts from Instagram and Facebookâincluding photos and textâwere part of the data used to train the generative AI models,â the company explains. âWe didnât train these models using peopleâs private posts. We also do not use the content of your private messages with friends and family to train our AIs.â
But anything you send to Meta AI will be used for model trainingâand beyond.
âWe use the information people share when interacting with our generative AI features, such as Meta AI or businesses who use generative AI, to improve our products and for other purposes,â Meta adds.
Conclusion
Of the major AI models we included above, OpenAIâs ChatGPT provided the easiest way to delete history and opt-out of having chatbot prompts used to train its AI model. Meta’s privacy practices appear to be the most opaque.
Many of these companies also provide mobile versions of their powerful apps, which provide similar controls. The individual steps may be differentâand privacy and history settings may function differently across platforms.
Unfortunately, even cranking all privacy settings to their tightest levels may not be enough to safeguard your information, according to Venice AI founder and CEO Erik Voorhees, who told Decrypt that it would be naive to assume your data has been erased.
âOnce a company has your information, you can never trust itâs gone, ever,â he said. âPeople should assume that everything they write to OpenAI is going to them and that they have it forever.â
âThe only way to resolve that is by using a service where the information does not go to a central repository at all in the first place,â Voorhees addedâa service like his own.
Edited by Ryan Ozawa.
Be the first to comment