Hello Lemmings!
I am thinking of making a community moderation bot for Lemmy. This new bot will have faster response times with the help of Lemmy webhooks, an amazing plugin for Lemmy instances by @[email protected] to add webhook support. With this, there is no need to frequently call the API at a fixed interval to fetch new data. Any new data will be sent via the webhook directly to the bot backend. This allows for actions within seconds, thus making it an effective auto moderation tool.
I have a few features I thought of doing:
- Welcome messages
- Auto commenting on new posts
- Scheduled posts
Punish content authors or take action onAuto report content via word blacklist/regex- Ban members of communities by their usernames via word blacklist or regex
- Auto community lockdown during spam
What other features do you think are possible? Please let me know. Any questions are also welcome.
Community requested features:
- Strike system
Strikes are added to a certain member of the community and the member will be temporarily banned within a time period if their strike count reaches a certain threshold
- Post creation restriction by account age
If an account’s age is lower than X, remove the post.
I beg of you, please don’t. The worst thing to happen to Reddit was their Automod. Please reconsider.
Why? Automod is just a tool, the issues people have with it is how overzealous the mods using it are. If you’re moderating a community with 10,000+ people you can’t expect to filter and manage everything yourself, so a bot scheduling posts and filtering potential spam/low effort content is necessary.
Exactly.
Automod is just a tool, indeed, but how a tool is designed dictates or at least encourages its usage.
deleted by creator
Trying to automate things and decrease mod burden is great, so I don’t oppose OP’s idea on general grounds. My issues are with two specific points:
Instead of those two I think that a better use of regex would be an automated reporting system, bringing potentially problematic users/pieces of content to the attention of human mods.
Alright. Sounds fair. Instead of taking dangerous actions, I’ll make it create a report instead. Though I’ll probably keep the feature to punish members by their usernames via regex or word blacklist.
This right here is the attitude that I have a problem with. I can think of one user who would get blacklisted right away because of their username alone. And that does not sit right with me.
Thank you! Frankly, if done this way I’d be excited to use it ASAP.
It’s to easen the work of community moderators. And you can’t just catch every comment that needs to be removed. Or posts, etc. This is where an automated moderation bot comes in. No matter how much you hate it, it is a must to have some automated system in growing platforms such as Lemmy.
It’s also not like the bot instantly bans everyone. I honestly don’t get the hate
Banning members on their username. Locking down an entire community because of a small group of people spamming. Deleting posts because an account isn’t old enough?
Why not throw in the system to have to approve posts before they get published? Really make the community welcoming.
It was said in another comment above that this tool is easily abused by “overzealous mods”, but I believe the real problem are overzealous programmers.
Reddit failed for reasons, and I believe automod was one of them. But you’ll do you, and nothing I say can change that.
I am merely trying to give community mods options. This feature and the other features are optional. Direct your complaints to the community owners if they use some regex that matches usernames that you think shouldn’t be banned.
The bot just locks it down to stop the spam, otherwise everyone’s feed will just be filled with spam. I haven’t seen such a spam yet, but that does not mean there won’t be any in the future. Just trying to be prepared for it.
Again, I am just giving the mods options. If they enable the feature and use it, direct your complaints to them.
That is possible with post locking and with a dashboard. I’ll look into it.
Again, I’m only giving them options.
Every tool can be used both in good and bad purposes. Why is it that it is the fault of the tool or its creator?
OP I agree with you, it’s a great idea imo.
I’ve been a moderator before on a Discord server with +1000 members, for one of my FOSS projects,
and maintenance against scam / spam bots grew so bad,
that I had to get a team of moderators + an auto moderation bot + wrote an additional moderation bot myself!..
Here is the source to that bot, might be usable for inspiration or just plain usable some other users:
https://github.com/Rikj000/Discord-Auto-Ban
I think it will only be a matter of time before the spam / scam bots catch up to Lemmy,
so it’s good to be ahead of the curve with auto-moderation.
However I also partially agree with @dohpaz42, auto-moderation on Reddit is very, uhm, present.
Imo auto moderation should not really be visible to non-offenders.