139
How do we get "normies" to adopt the Fediverse?
(lemmy.zip)
A community to talk about the Fediverse and all it's related services using ActivityPub (Mastodon, Lemmy, KBin, etc).
If you wanted to get help with moderating your own community then head over to !moderators@lemmy.world!
Learn more at these websites: Join The Fediverse Wiki, Fediverse.info, Wikipedia Page, The Federation Info (Stats), FediDB (Stats), Sub Rehab (Reddit Migration), Search Lemmy
@dch82 first, "normies" have to not get harassed when they come here.
Unfortunately the biggest Fedi software refuses to add automated reporting of offensive posts so if it's not reported, the admins won't even see it.
People coming from corporate social media are used to ignoring the report button because in their experience, it either doesn't work, or gets ignored by admins anyway.
We need automated reporting.
@fediverse
I'm fine with auto REPORTING, but the actual moderation needs to be a human. Auto moderation is bad. It gets things wrong. It's how I got banned from both twitter (calm down, this was back in 2018 before it was an elon owned nazi cesspool), and reddit.
On twitter I saw a funny video that was posted, and I replied "Aw man, that killed me".
I was banned for "inciting death threats"
@Lost_My_Mind yeah, just reporting.
I want to do the actual judgement, but if I don't know the post exists, I can't judge anything and it makes me so mad that possible racist stuff can exist on my instance without my knowledge because I havent "seen" it.
@fediverse
That's the thing about automation and training models.
First, they implement some sort of auto-reporting bot that requires a human to review them. In the beginning, it only about 50% accurate, but as they give it more and more examples of good and bad results through the human reviews, it moves to 80%, then 90%, then 99%, then 99.99% accuracy.
After a while, the humans on the other end are so numb to the 9999 entries they have to mark as approved that they can barely tell what's a rejection themselves, and the moderation team is asking itself just what this human review is actually doing. If it's 99.99% accurate, why not let the bot decide?
Then, the model moves on from auto-reporting to auto-moderation.
Federated reporting would help too
@BeAware@social.beaware.live @dch82@lemmy.zip Maybe im a little lost. Isn't there a block and report button on Mastodon? I'm using Misskey and both buttons seem to work. I mean im reporting to myself, but the button seems to work. What kind of automated blocking are you trying to do here?
@AterNox @dch82 blocking and reporting work fine.
However, people from corporate social media won't report posts because in their experience, it either doesn't get taken seriously or the admins ignore it. Corporate social media sites don't exactly act on reports in a timely manner.
I'm on my own instance, I moderate for myself. I don't want slurs to exist on my instance at all. However, if I don't see them with my own eyes, I cannot ban the user.
PS. I'm talking about banning users that are harassing others on the instance level. These are user actions. I am an admin. I run my own instance.
@fediverse
I'm confused, do you mean like automated enforcement rules/algorithms like big SM has? I.e. if user gets reported for breaking Y rule X amount of times ban user for Z amount of time and forward to admin for further action?
@cm0002 no, I want automated reports.
A user using the n word, full on with the hard R, isn't gonna be a good post. It should be automatically reported to me so that I can judge context and take action.
If a user doesn't report it, I won't see it.
I'm on my own instance, I am the user.
If I don't report it, nobody sees it.
That's dumb.
@fediverse
Ah, makes sense now, that is dumb. I can totally see why they would have issues with automated enforcement, but what you described I don't see why anyone would be against it lol
@BeAware@social.beaware.live @dch82@lemmy.zip So Mastodon not have a wordlist you can populate that "removes" posts with the keywords you provide? It took me a while to find it in Misskey, works like a charm,
@AterNox @dch82 doesn't exist for admins. It works on a "user" level. But that won't remove the post or data from the instance, it just "hides" it so the single user can't see.
@fediverse
We have instancewide admin blocks, so the accounts that would be automatically reported can be blocked preemptively, no report needed. That can be both good and bad... but pick a sheltered instance and you shouldn't get harassed. How would automatic reporting even work? I don't recall, but doesn't the admin interface let you specify keywords that alert the admins in a post? Is that what you mean?
CC: @dch82@lemmy.zip @fediverse@lemmy.world
@cy
Unfortunately not. Mastodon has no such thing. It does have filtered words for normal users. However, that doesn't do anything besides hide posts that contain the filtered words, nothing more.😬
@dch82 @fediverse
I unironically think it would be easier to train users that the report button works now than it would to get automated reporting that was worth a damn implemented.
By automated reporting do you mean something like filters on the backend to flag offensive posts per some custom settings?
Definitely. Back when I used FB and Twitter I learned that reporting is entirely useless. You just end up with some automated message about how they reviewed it and it "didn't violate their community standards" with some lame verbiage like "we realize this isn't the outcome you were looking for", regardless of how ridiculously blatant whatever you reported was. On the flip side, I was banned for clearly misinterpreted or brigaded comments, and then an appeal just gives you the inverse where they reviewed it and whatever you posted was definitely terrible and they "realize this isn't the outcome you were looking for".