Content moderation gets automated
OpenAI, one of the pioneers in artificial intelligence, is testing the content moderation capabilities of the advanced GPT-4 model. The firm uses GPT-4 to create a scalable, consistent and customizable content moderation system, aiming for the model not only to help make content moderation decisions, but also to develop policies. Thus, targeted policy changes and development of new policies can take from months to hours.
Anyone with OpenAI API access can already implement this approach to build their own AI-powered moderation system. However, OpenAI claims that its GPT-4 audit tools can help companies do about six months of work in one day.
Important for human health
It is known that the manual review of traumatic content, especially on social media, has significant effects on the mental health of human moderators. Meta, for example, agreed to pay more than 11,000 moderators in 2020 at least $1,000 each for mental health issues that could result from reviewing material posted on Facebook. Using artificial intelligence to remove some of the burden on human moderators can be of great benefit.
However, AI models are miles away from being perfect. It’s a known fact that these tools are prone to making wrong decisions, so OpenAI acknowledges that people still need to be involved.