Xbox is continuing to clamp down on toxic and abusive players on its systems, revealing that it has seen a huge increase in accounts reprimanded for posting "vulgar" content. Since improving its moderation system, the company says there's been a 450 percent increase in "enforcements" over such content, presumably ranging from taking down the post to suspending players altogether.

This also comes as Xbox expands the definition of what it classes as vulgar content, now including "offensive gestures, sexualized content, and crude humour". Xbox defends these increased moderation tactics, saying that such content was "detracting from the core gaming experience" for other users, and that most of the players who posted this material were just given a warning, not banned from their accounts altogether.

Related: The Abuse Of Ada Wong’s Actor Is The Latest Episode Of Gamers Being Awful

This information comes from Xbox's latest transparency report into its moderation practices. Here, the company discusses how it is trying to take a more "proactive" role in enforcing its terms and conditions, using AI to spot material and language that's banned on its services.

"As the needs of players continue to evolve, so do our tools," reads the report. "The safety of our players is a top priority – and to advance safe online experiences, we will continue to invest in innovation, work in close collaboration with industry partners and regulators, and collect feedback from the community."

Master Chief and Fernando Esparza in Halo Infinite

80 percent of content reports were dealt with through "proactive moderation", meaning they could have been decided by either an AI or a human. Xbox doesn't discuss the possibilities of false reports, despite much of this work possibly being conducted by an AI. However, if an AI does flag content as vulgar, then the next step would be for a human to look at it anyway, just to ensure that it wasn't an error. It's just not clear how long the player would have to wait for this to be decided.

While we don't have the info on how reliable it has proven, it's clear that Xbox is pleased with the results. One of the automated systems it uses, Community Sift, went through "20 billion human interactions" last year, so many reports would have gone through this programme.

It's not surprising to see Xbox clamp down on inappropriate content on its platforms, given the increased pressure on the industry to root out racism, hate speech and extremism from the gaming community. Over the weekend, new research suggested that 50 percent of gamers are exposed to extremism while playing games online. Racist language and iconography were also found to be slipping through filters on multiple games, ranging from Call of Duty to Roblox. All in all, the industry was called on to do more to stop extremists from using games as a recruitment tool.

Next: I Can’t Stay Up Playing Games Like I Used To