cross-posted from: https://lemm.ee/post/177673

Cross posting this here for visibility since lemmy.ml federation has been very hit or miss the last week. Original post from @sunaurus@lemm.ee

Today, a bunch of new instances appeared in the top of the user count list. It appears that these instances are all being bombarded by bot sign-ups.

For now, it seems that the bots are especially targeting instances that have:

  • Open sign-ups
  • No captcha
  • No e-mail verification

I have put together a spreadsheet of some of the most suspicious cases here.

If this is affecting you, I would highly recommend considering one of the following options:

  1. Close sign-ups entirely
  2. Only allow sign-ups with applications
  3. Enable e-mail verification + captcha for sign-ups

Additionally, I would recommend pre-emptively banning as many bot accounts as possible, before they start posting spam!

Please comment below if you have any questions or anything useful to add.

  • AFK BRB Chocolate@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    Why wouldn’t an instance require an email and a captcha? Seems like such an easy add that’s totally worth if it even just reduces bot traffic.

  • minus5m@fedia.io
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    This seems to be an issue on reddit aswell currently. Someone was guessing it may have to do with the presidential elections next year?

  • nucleative@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I think there is a gold rush going on here - a race to register usernames and community names in the event that a mass migration starts.

  • Wander@yiffit.net
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Well, since we own the instances and every user, even if a remote user, is also user of every instance they interact with, we should be able to run some analysis tools and weed out the bots and astroturfers.

  • cerevant@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    3
    ·
    1 year ago

    Sadly, this may signal the end of open federation. I can’t see how trust by default is going to work long term.

    • Ulu-Mulu-no-die@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      It depends on how federated platforms react, it’s necessary to control who signs up in some way, if that’s “globally” accepted trust can hold well long term.

      • cerevant@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        Right now the instances are working to stifle this by restricting account creation, but we’re just a step away from spammers creating instances on demand, flooding networks with stuff like the crypto spam on Reddit. I’m thinking major instances are going to have to go whitelist federation as a result.

        • marsokod@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          I would not be surprised if at some point you have a few whitelists, or some kind of reputation management for instances. One could even say they can have a karma number associated to them.

          In the end, this is the same problem as for emails and the landscape will probably structure itself around several big actors and countless of smaller actors who will have to be really careful in order to not be defederated.

          • Zak@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            1 year ago

            A reputation system is really important. Open federation may not be viable long term without one.

          • WalterLatrans@yiffit.net
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            1 year ago

            That’s an interesting idea. For each instance give users the ability to mark as spam comments/posts, then make it so each instance keeps track of what the ratio of spam vs not-spam is coming from peer instances and block any that exceed a certain ratio. It could easily be made automatic with manual intervention for edge cases.

            One issue I could see is that it could be used as a way of blacklisting smaller instances from larger instances by using bot accounts on the larger instances to mark the smaller instance’s legitimate traffic as spam. It would likely be necessary to implement a limit on how young/active an account can be to mark comments/posts as spam, as well as rate-limiting for situations where a given smaller community that is a subset of the larger one decides to dogpile on a smaller instance in an attempt to block them from the entire community.

    • Cinner@lemmy.worldB
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      This is one place where 2-of-3 multisig crypto could truly excel. If posters were required to have, say, $5 in XYZ per account per site, but that $5 will get you access to every site on the fediverse AND you can withdrawal it whenever you want, but it will close your accounts. Like a $5 participation escrow. Spammers could still spam, but they’d need to have $5 per account created to do so. I’m sure there are pros and cons to this, but it is technologically feasible.

      • Zetaphor@zemmy.ccOP
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 year ago

        You just described token gating, not multisig. Also you can do multisig without involving any money (or tokens)

        • Cinner@lemmy.worldB
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          My comment says deleted by creator (I didn’t delete it) so I don’t remember exactly what I said, but the point is to temporarily (as long as the account exists) put money on the books, which can be taken if you spam but you can withdrawal it when you close your account if you don’t. There would need to be a trusted signing authority so instance admins couldn’t just take it.

  • fax@sffa.community
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I’m an admin from https://sffa.community

    We did notice a large amount of bot sign-ups early on. We already ha email verification on and they never made it past that stage.

    We have since turned on CAPTCHA and now they aren’t signing up any longer.

    We’ve purged the bot accounts from our database.

    We are working on generating organic content but we’ve only been open for a couple weeks so I just ask that other instances bear with us. We’ve a very niche community of sci-fi and fantasy geeks. We’d love to be federated again.

  • tyfi@wirebase.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 year ago

    You should make the spreadsheet private, otherwise botters will find it. Don’t make their job too easy.

    • Zetaphor@zemmy.ccOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Not my spreadsheet, this is a crosspost. Also it looks like the content is being auto-generated and the sheet is read-only. There’s nothing they could really do by having this information.

    • NetHandle@kbin.social
      link
      fedilink
      arrow-up
      9
      ·
      1 year ago

      It’s the new form of marketing really. Ads aren’t super effective anymore as they’ve reached saturation and are showing severely diminished returns, so the next thing you can do is to create bot posts in the form of ‘product testimonials’ similar to what a person would see on amazon reviews, but from a source they’re more likely to trust such as a subreddit comment chain.
      Ie. Initial: I use [product], going on 15 years etc.
      Reponse: I use it too works great.
      Alt response: I use [diff. product], its cheaper but works the same.

      The “buy it for life” subreddit was full of them.

      Of course it gets a little more inisidious when it’s not just used for pushing products, but instead pushing ideas or propaganda. It’s commonly referred to as psyops, and they try to maintain a steady presence on any popular online forum. It was a big problem on Tumblr for a long time before Reddit.

      Also news post bots are super common, they want to generate traffic to their news sites, as that is their source of ad revenue. Lots of ad revenue related bots making posts to generate site traffic. Some of it’s not the worst thing to have, creates something of a newsfeed and a lot of it is already present here or on Mastodon.

      It becomes problematic when you have very biased news organizations and they’re allowed to use bots to upvote their news articles with impunity so all you see is biased news, it circles right back around to psyops.

      Also ad revenue fraud is something that happens a lot. Ad engagement with a bot that is meant to simulate a user browsing the site and clicking on an ad so that the person hosting the ads gets paid by the person who put the ad up. Sometimes it happens on entirely fake websites with entirely fake traffic, its much easier when you dont have to fake all the traffic and just the ad engaging traffic, as it adds legitimacy to the website. I wouldnt be surprised if a good portion of Reddits revenue is from ad fraud. It would go a long way to explaining why traffic was down only 6% during blackouts, if a large portion of the 94% remaining traffic was ad engagement bots.

    • Mereo@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      It comes to the territory. When you’re progressively getting bigger, you’re bound to get attacked.