Known for their ability to process and produce human-like text, Masters are trained on a massive content dataset. This also includes content flagged by Reddit moderators and internal law enforcement for violating the platform’s terms of use. This training helps the AI identify patterns and nuances in language often associated with alleged harassment.
Reddit confirmed the development by updating its support page to acknowledge LLM’s role in the harassment filter. The page clarifies that “moderator actions and content removed by Reddit’s internal tools” form the basis of the AI’s training data.
Although the effectiveness of this new tool has not yet been evaluated, it provides moderators with an additional layer of protection against online harassment. But there are still some questions. Can AI effectively distinguish between real bullying and passionate arguments?
Only time will tell how well this AI model can handle the complexities of online human interaction.
Source: Ferra
I am a professional journalist and content creator with extensive experience writing for news websites. I currently work as an author at Gadget Onus, where I specialize in covering hot news topics. My written pieces have been published on some of the biggest media outlets around the world, including The Guardian and BBC News.