The UK's Information Commission’s Office reminds organizations that data protection laws still apply to unfiltered data used to train large language models. Credit: Thinkstock The UK’s data regulator has issued a warning to tech companies about protecting personal information when developing and deploying large language, generative AI models.Less than a week after Italy’s data privacy regulator banned ChatGPT over alleged privacy violations, the Information Commission’s Office (ICO) published a blog post reminding organizations that data protection laws still apply when the personal information being processed comes from publicly accessible sources.“Organisations developing or using generative AI should be considering their data protection obligations from the outset, taking a data protection by design and by default approach,” said Stephen Almond, the ICO’s director of technology and innovation, in the post. Almond also said that, for organizations processing personal data for the purpose of developing generative AI, there are various questions they should ask themselves, centering on: what their lawful basis for processing personal data is; how they can mitigate security risks; and how they will respond to individual rights requests. “There really can be no excuse for getting the privacy implications of generative AI wrong,” Almond said, adding that ChatGPT itself recently told him that “generative AI, like any other technology, has the potential to pose risks to data privacy if not used responsibly.”“We’ll be working hard to make sure that organisations get it right,” Almond said. The ICO and the Italian data regulator are not the only ones to have recently raised concerns about the potential risk to the public that could be caused by generative AI.Last month, Apple co-founder Steve Wozniak, Twitter owner Elon Musk, and a group of 1,100 technology leaders and scientists called for a six-month pause in developing systems more powerful than OpenAI’s newly launched GPT-4.In an open letter, the signatories depicted a dystopian future and questioned whether advanced AI could lead to a “loss of control of our civilization,” while also warning of the potential threat to democracy if chatbots pretending to be humans could flood social media platforms with propaganda and “fake news.” The group also voiced a concern that AI could “automate away all the jobs, including the fulfilling ones.”Why AI regulation is a challengeWhen it comes to regulating AI, the biggest challenge is that innovation is moving so fast that regulations have a hard time keeping up, said Frank Buytendijk, an analyst at Gartner, noting that if regulations are too specific, they lose effectiveness the moment technology moves on.“If they are too high level, then they have a hard time being effective as they are not clear,” he said. However, Buytendijk added that it’s not regulation that could ultimately stifle AI innovation but instead, a loss of trust and social acceptance because of too many costly mistakes..“AI regulation, demanding models to be checked for bias, and demanding algorithms to be more transparent, triggers a lot of innovation too, in making sure bias can be detected and transparency and explainability can be achieved,” Buytendijk said. Related content news CISA, FBI urge developers to patch path traversal bugs before shipping The advisory highlights how developers can follow best practices to fix these vulnerabilities during production. By Shweta Sharma May 03, 2024 3 mins Vulnerabilities news Microsoft continues to add, shuffle security execs in the wake of security incidents The company has appointed new product security chiefs as well as a customer-facing CISO as it continues to respond to high-profile attacks on its products and own network. By Elizabeth Montalbano May 03, 2024 4 mins CSO and CISO feature Malware explained: How to prevent, detect and recover from it What are the types of malware? How does malware spread? How do you know if you’re infected? We've got answers. By Josh Fruhlinger May 03, 2024 18 mins Ransomware Phishing Malware brandpost Sponsored by Cyber NewsWire LayerX Security Raises $26M for its Browser Security Platform, Enabling Employees to Work Securely from Any Browser, Anywhere Early adoption by Fortune 100 companies worldwide, LayerX already secures more users than any other browser security solution and enables unmatched security, performance and experience By Cyber NewsWire May 02, 2024 4 mins Cyberattacks Security PODCASTS VIDEOS RESOURCES EVENTS SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe