I’m really enjoying lemmy. I think we’ve got some growing pains in UI/UX and we’re missing some key features (like community migration and actual redundancy). But how are we going to collectively pay for this? I saw an (unverified) post that Reddit received 400M dollars from ads last year. Lemmy isn’t going to be free. Can someone with actual server experience chime in with some back of the napkin math on how expensive it would be if everyone migrated from Reddit?

  • honk@feddit.de
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I don’t believe this is how it works though.

    Let’s say your tiny 3 person instance is connected to a big one. I believe it only pulls in content from the communities somebody from the small instance is subscribed too. Correct me if I’m wrong.

    • panoptic@fedia.io
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      That’s what they’re saying.

      Essentially - if someone from the small instance subscribes to a community that has a ton of data (huge post volume, images, whatever), the small instance needs to pull data over from the larger instance. At some point there may be communities that are so large small instances can’t pull them in without tanking.

      • Silviecat44@vlemmy.net
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 year ago

        I wonder if there is a way to get around this? maybe smaller instances will have to be text-only?

        • panoptic@fedia.io
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          If I’m reading the protocol right, it’s probably larger instances that will avoid more duplication, since:

          1. There’s a higher chance they’re going to have more communities shared among users (for really tiny instances you’re probably going to get a lot of overlap since those people likely have interconnected interests, but I expect that would fall off quickly, but then converge at scale).
          2. The larger number of users will mean they ‘use’ more of the content they’re pulling down (I can’t read all of a highly active community in a day, but 1000 people together checking through the day might ‘use’ it all).

          I’m not sure I see where you see caching fitting in.
          I am surprised I don’t see some kind of lower resolution digest concept in the protocol (which might be what you’re looking for)

    • flambonkscious@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      That’s what I’ve gathered, but I don’t believe there’s a way for instance owners to limit what’s fetched - a user crafts the query and the server does the needful.

      I imagine this could amount to a denial of service attack of sorts, if some high-churn communities are imported into tiny instances. How bad that could be, I have no idea - I’m speaking pretty theoretically, here. Text is tiny, after all, so it’s probably not much of a concern, since most of the media is actually handled elsewhere…

      • honk@feddit.de
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        I’m not a web developer. I’m sort of a sysadmin so i have some experiences maintaining machines for web apps for other people. And you are right…text will not create massive amounts of data. But a lot of tiny transactions can bring down machines surprisingly fast even if the total amount of data is relatively small.

        I guess we are here to experience it first hand. I don’t think anybody…not even the developers have a clear idea of how well this will scale. There is only one way to find out lol