The UK Department for Science, Innovation and Technology (DSIT) has announced a new voluntary AI cybersecurity Code of Practice and implementation guide, which it claims will set a new global standard for AI security via the European Telecommunications Standards Institute (ETSI).
Developed in conjunction with the National Cyber Security Centre (NCSC), the code outlines 13 principles for secure AI design, development, deployment, maintenance, and end-of-life management. The principles apply to software vendors developing, using, or offering AI, as well as organisations creating or utilising AI services and components. The code covers areas like staff training on AI security risks, secure system design, threat modelling, human oversight of AI systems, asset tracking and protection, infrastructure security, supply chain integrity, documentation, testing, secure deployment, ongoing security maintenance, system monitoring, and data/model disposal.
However, the code does not apply to AI vendors offering models and components; a separate Software and Cyber Governance Code will address these organisations.

What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.