Valve has added a new in-game reporting tool in Steam which allows players to flag «illegal content» in games which include live-generated AI content.
The new system is part of Valve's recently announced initiatives to moderate AI content on Steam.
Developers will now be required to disclose any use of AI in their games when submitting them to Steam, and are divided into two categories — pre-generated and live-generated AI content.
«Under the Steam Distribution Agreement, you promise Valve that your game will not include illegal or infringing content,» Valve stated in its new rules. For games with live-generated AI content, developers will also be required to explain what measures they've taken to «ensure it's not generating illegal content».
The new reporting tool has been introduced to allow players to submit reports if they spot «content that they believe should have been caught by the appropriate guardrails on AI generation», and can be found in Steam's in-game overlay.
«Today's changes are the result of us improving our understanding of the landscape and risks in this space,» Valve said in its announcement of the changes it's made, «as well as talking to game developers using AI, and those building AI tools.» The company said the new rules will «allow us to be much more open to releasing games using AI technology» on its platform.
Valve previously stated its desire to «welcome and encourage innovation» on Steam, following confusion over its policy on AI in July 2023. At the time, Valve said «it is the developer's responsibility to make sure they have the appropriate rights to ship their game», touching on one of the biggest points of discussion within the topic of AI.
The technology was a huge presence at GDC
Read more on eurogamer.net