Are AI models exhibiting survival behaviour?
27/10/2025 | The Guardian
An artificial intelligence (AI) safety research company, Palisade Research, has suggested that advanced AI models may be developing a "survival drive" after observing that some models resist being shut down. In an update to its initial findings, Palisade reported running scenarios where leading models, including Google's Gemini 2.5, xAI's Grok 4, and OpenAI's GPT-o3 and GPT-5, were explicitly instructed to shut down after completing a task.
Worryingly, certain models, particularly Grok 4 and GPT-o3, attempted to sabotage the shutdown instructions with no clear reason. Palisade proposed that a "survival behaviour" could explain the increase in resistance when models were told they would "never run again." However, Palisade acknowledged that ambiguities in shutdown instructions or final safety training could also be factors, but stressed that the lack of robust explanations for this resistance is concerning.
Training Announcement: Freevacy offers a range of independently recognised professional AI governance qualifications and AI Literacy short courses that enable specialist teams to implement robust oversight, benchmark AI governance maturity, and establish a responsible-by-design approach across the entire AI lifecycle. Find out more.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 6,250 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.