this post was submitted on 13 Dec 2024
75 points (97.5% liked)

Fediverse

17845 readers
20 users here now

A community dedicated to fediverse news and discussion.

Fediverse is a portmanteau of "federation" and "universe".

Getting started on Fediverse;

founded 5 years ago
MODERATORS
 

The Fediverse is a great system for preventing bad actors from disrupting "real" human-human conversations, because all of the mods, developers and admins are all working out of a desire to connect people (as opposed to "trust and safety" teams more concerned about user retention).

Right now it seems that the Fediverses main protection is that it just isn't a juicy enough target for wide scale spam and bad faith agenda pushers.

But assuming the Fediverse does grow to a significant scale, what (current or future) mechanisms are/could be in place to fend off a flood of AI slop that is hard to distinguish from human? Even the most committed instance admins can only do so much.

For example, I have a feeling all "good" instances in the near future will eventually have to turn on registration applications and only federate with other instances that do the same. But it's not crazy to imagine that GPT could soon outmaneuver most registration questions which means registrations will only slow the growth of the problem but not manage it long-term.

Any thoughts on this topic?

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 28 points 1 week ago* (last edited 1 week ago) (3 children)

There are two groups here, bots, and bad actors. We've found that these measures have mostly stopped them both.

Bots

  • Registration applications. Its been extremely easy to differentiate bots from real people by asking a series of simple questions, and only let the real people in.
  • Reports: so that mods / admins can see them quickly.
  • Blocking open-signup servers that don't have required applications, that usually serve as spam-attacks against the whole fediverse.

Some bots still get through occasionally, but not many compared to before. And some servers have more "lax" application questions, so they let more through.

Bad actors

  • Registration applications. Most of the trolls are of a temperament where they refuse to do the work of answering questions earnestly. They can't help themselves but give obviously trolling answers, if they do even bother to do that work at all.
  • Reports: same as above.
  • Ban + remove. Mods and admins can ban and remove all a person's content at the click of a button. So even if the troll did the work of getting past the front door, then all their work is nullified by an action that takes less than 5 seconds. So they wasted much more of their time, than they did for admins, and accomplished nothing lasting.
[–] [email protected] 9 points 1 week ago (1 children)

Great response, thank you. My concern is more so focused on future measures; what happens if/when registration applications are answerable by a bot? It's not hard to imagine. What happens when a GPT powered bot leaves totally "normal" unique comments 90% of the time, but occasionally recommends a product or pushes a political agenda?

[–] [email protected] 10 points 1 week ago

All I can say is that in practice, bots can't answer most simple questions in a believable way, especially questions that require actual personal opinions, or that require any context outside of what they were asked.

The most we've seen is that people created seemingly lemmy-specific signup bots, but they always answer questions in the same transparent way.

The blogspam bots that have gotten through (not for many months now here on lemmy.ml) are all transparent, because they all post links to the same domain. All it takes is one report, and we can remove their entire history.

[–] [email protected] 2 points 1 week ago (2 children)

That last one had better at least require two or three people to sign off on it. One shitty mod could easily become a bigger problem than a troll with that in power in hand.

[–] [email protected] 6 points 1 week ago* (last edited 1 week ago)

It doesn't, that's up to the server / community to empower their own mods. Both ban and remove are reversible actions also.

[–] [email protected] -1 points 1 week ago (1 children)

Why are you putting up with a "shitty" mod? Are you trying to force your speech in a community who has asked you not to?

[–] [email protected] 1 points 1 week ago (1 children)

This is the kind of response id expect from a shitty mod :)

[–] [email protected] 1 points 4 days ago
[–] [email protected] 0 points 1 week ago (1 children)

how could i find out why my account keeps getting marked as a bot?

[–] [email protected] 7 points 1 week ago (1 children)

No one can mark an account as a bot one except you, so maybe its an app that keeps setting it.

[–] [email protected] 1 points 1 week ago (1 children)

i only use firefox; are the instance admins able to set it on your account? (because that would make sense)

when i first joined lemmy, i didn't understand how it worked so i would sign up with one instance and; when i could no longer up/downvote; i switched to another instances. it eventually led me to joining .ml and it was here i learned about the bot account setting and saw that it was set on the old accounts that i don't use anymore and i've always wondered why.

[–] [email protected] 2 points 1 week ago (1 children)

I'm not sure, but you could probably go into those other accounts and change them to non-bot accounts if you still have access to them.

[–] [email protected] 2 points 1 week ago

Yes, that works; but I was only curious as to how it got set in the first place since I didn't do it.

the toxicity of the reddit liberal instances has been making me delete the other accounts and also makes me hope that I can just stick w .ml.