A recent article in the Financial Times (£) highlights the urgent need for transparency in artificial intelligence (AI) research. Despite the growing number of proposals on how to regulate AI, there is little consensus on what needs to be done. Even technologists have serious questions about the behaviour of large language models (LLMs), which they claim often behave unpredictably. Furthermore, the companies developing AI models continue to shield data and algorithmic settings as trademark-protected proprietary information, making it difficult for researchers and regulators to assess foundational models for privacy, discrimination and bias, or security threats. The recent Foundation Model Transparency Index, published by Stanford’s Institute for Human-Centered Artificial Intelligence, reveals just how much we don’t know about proprietary AI and emphasises the risks embedded in technologies for healthcare or law enforcement.
£ - This article requires a subscription.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 4,350 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.