OpenAI considers encrypting ChatGPT temporary chats
18/08/2025 | Axios
OpenAI is considering adding encryption to temporary chats in ChatGPT to enhance user privacy. Currently, temporary chats do not appear in a user's history or train AI models, but OpenAI may keep copies for up to 30 days for safety purposes.
While this move would make temporary chats more secure, achieving full end-to-end encryption is complicated because OpenAI would be a party to the conversation. This means the company could still access sensitive information and potentially share it with law enforcement. Full encryption for all ChatGPT services would also complicate its operations, as features such as long-term memory require OpenAI to access user data.
Training Announcement: Freevacy offers a range of independently recognised professional AI governance qualifications and AI Literacy short courses that enable specialist teams to implement robust oversight, benchmark AI governance maturity, and establish a responsible-by-design approach across the entire AI lifecycle. Find out more.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 6,250 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.