Sorry, OpenAI. The European Union is making life for the leaders of artificial intelligence (AI) much less private. A newly agreed draft of the region's upcoming AI Act will force ChatGPT maker OpenAI and other companies to share previously hidden details about how they build their products. The legislation will still rely on companies to audit themselves, but it's nonetheless a promising development as corporate giants race to launch powerful AI systems with almost no oversight from regulators.
The law, which would come into force in 2025 after approval from EU member states, forces more clarity about the ingredients of powerful, “general purpose” AI systems like ChatGPT that can conjure images and text. Their developers will have to report a detailed summary of their training data to EU regulators, according to a copy of the draft seen by Bloomberg Opinion.
“Training data… who cares?” you might be wondering. As it happens, AI companies do. Two of the top AI companies in Europe lobbied hard to tone down those transparency requirements, and for the last few years, leading firms like OpenAI have become more secretive about the reams of data they've scraped from the Internet to train AI tools like ChatGPT and Google's Bard and Gemini.
OpenAI, for instance, has only given vague outlines of the data it used to create ChatGPT, which included books, websites and other texts. That helped the company avoid more public scrutiny over its use of copyrighted works or the biased data sets it may have used to train its models.
Biased data is a chronic problem in AI that demands regulatory intervention. An October study by Stanford University showed that ChatGPT and another AI model generated employment letters for hypothetical
Read more on tech.hindustantimes.com