Foreign threat actors increasing adoption of multiple AI tools
08/10/2025 | HackRead
OpenAI's latest Disrupting Malicious Uses of AI report reveals that hackers and influence operators are increasingly adopting organised approaches to their use of artificial intelligence (AI) tools. Adversaries are now spreading their operations across multiple AI systems, using models like ChatGPT for reconnaissance and planning, while relying on other platforms for execution and automation. Attackers are not changing their fundamental methods but are using AI to make existing tactics, such as writing malware and drafting phishing emails, faster and more efficient. New specialised malicious tools are also emerging, including SpamGPT, which helps bypass email security filters, and MatrixPDF, which transforms ordinary PDF files into malware. These complement already known AI tools, such as WormGPT and FraudGPT.
Training Announcement: Freevacy offers a range of independently recognised professional AI governance qualifications and AI Literacy short courses that enable specialist teams to implement robust oversight, benchmark AI governance maturity, and establish a responsible-by-design approach across the entire AI lifecycle. Find out more.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 6,250 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.