I currently moderate 2 subs (down from 5). I'm looking to offload those 2 without having them get banned for being unmoderated (again).
To be fair, that's still possible (and sometimes the case) with human moderators, though I find that experience to be exceptionally rare, and the remedy (just don't use that subreddit) easier to perform. At least if the mods of some sub go haywire, users can always create another sub. With widespread use of AI moderation, there'd be nowhere to go to escape.
THat's how their whole "Appeal" process is now. We had accounts suspended permanently for daring to question the mods at a large subreddit, and they nuked every account that logged in from our home IP. Wife's account, which was innocent, was permanently suspended. Same with housemate behind us that used our WiFi.
The appeal process? Within 5 minutes we got a useless and canned message that was along the lines of "After careful review, it was been determined that... blah blah blah."
We've since made new accounts, but it was 100+ days later, and all of our reddit usage is way down as a result. My account alone was 8+ years old. Gone, no appeal, nothing.
Reports in /r/help were that they used AI to handle appeals. Not sure if it's true, but it feels like it is.
Because they make all their content mods sign one that also absolves them of all responsibility.
There's no scenario where the AI moderators don't become superior to human moderators in a broad sense (only in select, very narrow cases will humans still be better at it). That's one of the easier areas for AI to wipe out in terms of human activity.
This isn't then. The game has changed, permanently and dramatically. Anybody calculating what I'm saying based on GPT 3.5 or 4 is doing it very wrong. That's a comical failure of not looking ahead at all. Look at how far OpenAI has come in just a few years. Progress in that realm is not going to stop anytime soon.
Nvidia is unleashing some extraordinary GPU systems at the datacenter level that will quite clearly enable further leaps in LLMs, and they'll mostly trivially handle moderation tasks.