Source: https://front-end.social/@fox/110846484782705013

Text in the screenshot from Grammarly says:

We develop data sets to train our algorithms so that we can improve the services we provide to customers like you. We have devoted significant time and resources to developing methods to ensure that these data sets are anonymized and de-identified.

To develop these data sets, we sample snippets of text at random, disassociate them from a user’s account, and then use a variety of different methods to strip the text of identifying information (such as identifiers, contact details, addresses, etc.). Only then do we use the snippets to train our algorithms-and the original text is deleted. In other words, we don’t store any text in a manner that can be associated with your account or used to identify you or anyone else.

We currently offer a feature that permits customers to opt out of this use for Grammarly Business teams of 500 users or more. Please let me know if you might be interested in a license of this size, and I’II forward your request to the corresponding team.

  • harmonea@kbin.social
    link
    fedilink
    arrow-up
    17
    ·
    1 year ago

    Framing this solely as fear is extremely disingenuous. Speaking only for myself: I’m not against the development of AI or LLMs in general. I’m against the trained models being used for profit with no credit or cut given to the humans who trained it, willing or unwilling.

    It’s not even a matter of “if you aren’t the paying customer, you’re the product” - massive swaths of text used to train AIs were scraped without permission from sources whose platforms never sought to profit from users’ submissions, like AO3. Until this is righted (which is likely never, I admit, because the LLM owners have no incentive whatsoever to change this behavior), I refuse to work with any site that intends to use my work to train LLMs.

    • Jaded@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      3
      arrow-down
      3
      ·
      edit-2
      1 year ago

      Models need vast amounts of data. Paying individual users isnt feasible, and like you said most of it can be scraped.

      The only way I see this working is if scraped content is a no go and then you pay the website, publishing house, record company, etc which kills any open source solution and doesn’t really help any of the users or creators that much. It also paves the way for certain companies owning a lot of our economy as we move towards an AI driven society.

      It’s definitely a hot mess but the way I see it, the more restrictive we are with it, the more gross monopolies we create for no real gains.

      • harmonea@kbin.social
        link
        fedilink
        arrow-up
        12
        arrow-down
        1
        ·
        1 year ago

        Paying individual users isnt feasible

        Sounds like their problem to solve, not mine.

      • kibiz0r@midwest.social
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I don’t see why those are the only two options.

        We could update GPL, CC, etc. licensing so that it specifies whether the author intends to allow their work to be used for LLM training. And you could still put a non-commercial or share-alike constraint on it.

        Hooray, open source is saved while greedy grubby hands are thwarted.

    • valveman@lemmy.eco.br
      link
      fedilink
      arrow-up
      2
      arrow-down
      4
      ·
      1 year ago

      I’m against the trained models being used for profit with no credit or cut given to the humans who trained it.

      Sorry mate, hell’s gonna get cold before this happens. We’re talking about the biggest moth******ers on earth since always. Do you think Meta/[insert big tech company name here] will start to behave all of the sudden? These people literally KILL people everyday for a profit (looking at you Instagram).

      The only way to get something from these scumbags is fining them something like 100k per hour, until they start respecting people’s privacy

      • harmonea@kbin.social
        link
        fedilink
        arrow-up
        6
        ·
        edit-2
        1 year ago

        I did already say I don’t expect this to ever change, so “sorry mate,” but you’re not exactly telling me anything I don’t know here.

        But I suspect this was a knee-jerk rant typed before bothering to read past what you quoted. Oh well. Good thing I can still stand against something even if I don’t expect it to change much.

        • valveman@lemmy.eco.br
          link
          fedilink
          arrow-up
          3
          arrow-down
          1
          ·
          1 year ago

          Sorry if it sounded rude (and yeah, it was kind of a rant, sorry). What I’m trying to say is: these people do much worse things and don’t bother to say “sorry” publicly. The only way to make them behave is to fine them by a huge amount, just like Norway did.

          • harmonea@kbin.social
            link
            fedilink
            arrow-up
            4
            ·
            1 year ago

            Well, we can agree on that! Make paying contributors the cheaper option.

            I won’t hold my breath though. :')