#AutoMod

Raccoon🏳️‍🌈Raccoon@techhub.social
2024-11-26

You know what would be a good #AutoMod type feature for #Mastodon, or any other instance software for that matter? Detecting if a new user is posting more than a certain number of posts in a brief period of time.

This thing was posting like five messages per minute. I will say that it's good that 127 of the people who got those messages apparently hit the report button, that really helps us as #moderation.

Maho Pacheco 🦝🍻mapache@hachyderm.io
2024-02-27
Cragsand :catjam:cragsand
2023-12-13

Thank you everyone who helped bring attention to this!
But looks like it was shut down by Twitch.

They misinterpreted the whole thing, perhaps intentionally. AI is being used to flag/ban without considering context and I hope that this is reconsidered.

"Twitch Global AI AutoMod does not understand context."

Response from Twitch misinterpreting the article userpost:

Hi, thanks for taking the time to share your feedback with us. We wanted to clarify a couple of things. AutoMod can’t ban people from Twitch. Mods can use the tool to identify chat messages that break their channel’s rules. But AutoMod can’t be used to timeout, ban, or mute users from any channel, or from Twitch – those aren’t AutoMod capabilities.

We also wanted to share a bit more about how we enforce our Community Guidelines. These are the policies that apply to all of Twitch, and help make Twitch safer. We have human teams that review our content moderation decisions. We call this “human in the loop.” These reviewers help ensure that our policies are being applied accurately. We don’t use a “global AI” system to enforce our guidelines, and we manually review suspension decisions.

We recognize that, while the majority of enforcements we issue are correct, sometimes we may get it wrong... (cont)
Cragsand :catjam:cragsand
2023-11-26

Here it is: "Spirit AI" using "proactive" technology.

I fear this has backfired to instead mean presumed guilt until proven innocent. Makes me think of Minority Report. The road to hell is paved with good intentions and all that I guess.

From a 2022 Twitch blog post:
safety.twitch.tv/s/article/An-

From Twitch a 2022 blog Post:
"Fortifying the technology that detects harmful text of all kinds on Twitch.

We recently completed the acquisition of Spirit AI, a leading natural language processing company who will help us continue to refine AutoMod and other proactive detection for catching harmful text or phrases sent on Twitch."
2023-11-26

@cragsand@mastodon.social
Discussion regarding Twitch moderation AI spread to Reddit where I clarified some questions that arose:

Since this global AI AutoMod remains an undocumented "feature" of Twitch chat from a while back a lot of the conclusions I've listed in the thread are based on deduction from watching active chatters get suspended and tell their stories on Discord and social media.

Most can luckily get their account reinstated after appealing but it relies on having an actual human look at the timestamp of the VOD and take their time to figure out what actually happened as well as get the complete context of what was going on on stream when it occurred. I've seen many apologies from Twitch moderation sent in emails after appealing, but if you get unbanned, an apology or stay banned seems mostly random.

Being banned like this will also make it much less likely that you want to participate and joke around in chat in the future, leading to a much worse chatting experience.

I see some discussions are arguing that all AI flagged moderation events are actually reviewed by humans (but poorly) and this is a possibility. Because of the lack of transparency from Twitch regarding how this works it's very difficult to know for sure how these reviews are done. A manual report in combination with an AI flag is almost certainly a ban. One thing is sure though, and that is that too much power is given to AI to judge in these cases.

Seeing as permanent suspensions from accounts who have had active paying subscriptions for YEARS on Twitch can be dished out in seconds, either those reviewing are doing a lousy job, or its mostly done by AI. Even worse, if those reviewing are underpaid workers who get paid by "number of cases solved per hour" there is little incentive for them to take their time to gather more context when reviewing.

It's likely that if Twitch gets called out for doing this, they have little incentive to admit it as it may even be in violation of consumer regulations in some countries. Getting a response that they "Will oversee their internal protocol for reviewing" may be enough of a win which results in them actually turning this off. Since there is no transparency we can't really know for sure.

A similar thing happened on YouTube at the start of 2023, where they went through all old videos speech-to-text transcripts and issues strikes retroactively. It got a lot of old channels to disappear, especially those with hours of VOD content where something could get picked up and flagged by AI. For the communities I'm engaged in, it meant relying less on YouTube for saving Twitch VODs. It was brought up by MoistCritical about a year ago since it also affected monetization of old videos.

#Twitch #Moderation #BadAI #AI #Enshittification #AutoMod #AIAutoMod #ModMeta #ModerationMeta

Cragsand :catjam:cragsand
2023-11-17

AI based is causing issues for streamers on .

When any single line of text can be taken to mean something horrific out of context, streamers are seeing their most active chatters getting permanently suspended without any human intervention.

I wrote an article to hopefully change this. Maybe it can at least lead to further discussion regarding the matter.

Outsourcing to , while it may save money, is NOT a good idea.

Full article: twitch.uservoice.com/forums/95

📡 Daan Bergdb@bla.daanberg.net
2023-09-05

Interesting: I haven't even made a post or reaction on #Facebook today and my account is limited for an hour due to a violation of the Community Guidelines.

Nothing strange visible in the activity log.

I wonder what triggered this 🤔

#socialmedia #algorithm #zuck #moderation #automod

Dutch language warning in the Facebook app, indicating to the user that their account has been locked for an hour due to a Community Guidelines violation.

The literal text reads:

"Je account is beperkt gedurende 1 uur

Je accountactiviteit is in strijd met onze richtlijnen voor de community. Daarom kun je een of meer dingen die je normaal doet nu niet meer doen."
Vyr Cossont 🏳️‍🌈🏳️‍⚧️vyr@universeodon.com
2023-05-03

@supernovae following up: here's the moderation stuff i was working on. shelving it to get back to Feditext. might be useful to others.

github.com/VyrCossont/m1000 implements a webhook listener automod bot in Rust which automatically creates reports when posts or users match on configurable words, regexes, or link domains. should be able to handle another #thegx situation, or a small link-spam campaign. proof of concept, not quite prod-ready, but i'm already running it on my home instance.

github.com/VyrCossont/mastodon removes the `if local?` check from Mastodon's `status.*` and `account.*` webhooks, allowing `m1000` to moderate remote posts and users.

github.com/VyrCossont/mastodon adds a `report.updated` webhook to Mastodon, useful for integration with external ticketing systems.

github.com/VyrCossont/mastodon creates an API for Mastodon's audit log, which currently doesn't have one. this would be useful for admin apps.

#MastoAdmin #MastoDev #MastoMod #moderation #automod #webhook #webhooks

Vyr Cossont 🏳️‍🌈🏳️‍⚧️vyr@demon.social
2023-03-19

so this isn't a documented feature yet, but Mastodon has webhooks for a handful of events. (you can find them in your instance's web UI, in the admin prefs area.)

up until now, they could be set up to trigger when an account on the instance is created or approved, or when a report is created. as of PR #24133 getting merged, they can also trigger when an account on the instance is updated, and when a post on the instance is created or updated. they can also now run on the same machine or in the same cluster as Mastodon.

this means that when this change hits a release version, you can build external applications that react to posts! think Reddit, Discord, or Twitch automod systems. in conjunction with existing admin API methods, webhooks can be used to match posts or account bios against filters, and if spam, hate speech, etc. is detected, automatically DM a mod, report a post, suspend the problem account, or even block a related IP range from signing up.

in the future, i hope to see both a mature admin/moderation API (for example, i don't think it's possible to, through the public API, delete posts you didn't create, even with admin credentials) and eventually more automod features in the server itself. meanwhile, being able to iterate outside the Mastodon codebase with the existing admin API and webhooks will let instance admins experiment, and hopefully share tooling.

#MastoDev #MastoAdmin #spam #moderation #automod

Norobiik @Norobiik@noc.socialNorobiik@noc.social
2023-03-09

If this experiment works out I expect the tech behind tools like #AutoMod to spread.

"Where the power of #AI could get interesting for Discord admins is with an upgrade to AutoMod. Discord rolled out its autonomous moderation tool widely last year to fight spam and slurs, and now, it’s experimenting with an AI-powered version of AutoMod"

#Discord is using #OpenAI’s technology to improve its #ClydeBot, #ModerationTools, and platform features | #ChatGPT #GenerativeAI
theverge.com/2023/3/9/23631930

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst