Toxic gamers who enjoy harassing others via in-game comms are set for an equally rude awakening.
As creators and publishers of some of the best co-op games out there, Ubisoft and Riot Games are no strangers to having to deal with toxic players. Fortunately, both companies have joined forces for Zero Harm In Comms(opens in new tab), a research partnership intended to combat bad behavior online. The project will use AI technology to track, tackle, and teach bots how to identify cases of online harassment to “foster more rewarding social experiences and avoid harmful interactions”.
Badly behaved players won’t be able to hide behind the anonymity of their keyboards anymore. The research partnership intends to create a system where information about offending players will be stored and then shared with the entire industry in a database.
You’ve probably heard the term “toxic” being used to describe certain kinds of online gamers. Whether you’re referring to specific gameplay techniques that are frowned upon by the wider playerbase or a case of genuine abuse or bullying, it’s hard to put a face to a screen name in the age of internet anonymity.
Although anti-cheat software has been introduced across many big online PvP games, these don’t tend to make note of when players are actively harming other players with their words. The narrow space between smack-talk and bullying is frequently occupied by the kind of trolls you’d hoped to only find in fairy tales.
In the announcement from Ubisoft Le Forge, executive director Yves Jacquier expresses his empathy(opens in new tab) for players who find themselves in these horrible positions. “Disruptive player behaviors is an issue that we take very seriously but also one that is very
Read more on techradar.com