The company is also developing a tool to give researchers an “effective way to evaluate our content moderation systems and examine existing content available on our platform,” Pappas added. Researchers will also soon be able to upload their own posts to see how different types of content are either permitted, rejected or passed to moderators for further evaluation.
Beyond that, TikTok pledged to expand its transparency reports with information about “countering covert influence operations,” according to Pappas. This information will be included in the company’s quarterly Community Guidelines Enforcement Reports.
“We know that just saying ‘trust us’ is not enough,” Pappas said. “That’s why long ago we made an important commitment to transparency, particularly when it comes to how we moderate and recommend content.”