AI Content Moderation
The cluster focuses on using AI, LLMs, and machine learning for moderating online content, such as flagging suspicious, abusive, or AI-generated posts, with discussions on hybrid human-AI approaches, training data challenges, and potential risks like bias or mis-training.
Activity Over Time
Top Contributors
Keywords
Sample Comments
you may be able yo use AI to flag suspicious conent. There probably wouldn't be enough readory posts yet do train it on yet, you may be able to train it on similar data...
Clearly the answer is to throw an LLM to filter AI posts.
It's telling that the AI doesn't just flag something, without banning it, and let a human do a proper review.
Can AI not moderate posts? asking for a friend.
Oh boy, we need AI powered blockers that filter out this stuff.
Interesting. https://openai.com/blog/new-and-improved-content-moderation-...
Wouldn't it be best for them to strip that out of the training data for moderation reasons?
Haha, I call this the "Clbuttic" problem. But nowadays it can be solved with machine learning fairly easy https://moderationapi.com/blog/moderate-text-automatically-u...
Might be a job for an LLM personal moderator? Give it a prompt, what kind of content you want to see, and what to filter out?
Seems like the more likely option. They could be used to live scan every post to work out what it's about and it's sentiment. Similar to how ChatGPT can work out if you are asking for something it won't answer, they could be used to work out if you are saying something not allowed.