• Grimy@lemmy.world
    link
    fedilink
    arrow-up
    26
    ·
    7 months ago

    Each time this pops up, there is a rush of people saying to delete or edit your comments.

    They have a database of your comments and all your edits. Its easy to see when you mass delete or edit them. Anything done past a certain point in time, especially all at once, is automatically reverted.

    By deleting and editing, you are taking the data away from scrappers making the dataset they are selling actually unique and more valuable.

    • CosmoNova@lemmy.world
      link
      fedilink
      arrow-up
      11
      ·
      edit-2
      7 months ago

      I mean that’s completely illegal at least in places like Germany where people have the right to be forgotten, but unfortunately you’re still right. They already commited the biggest heist in human history and got away with it. I guess NFT grifters only got punished because they dared to also steal from some rich people while Altman and his cronies are smart enough to only steal from the other 99.9%. When they have your data once, you can’t request it back anymore. Because the worst that can happen to them is a slap on the wrist and the cost of being in the fastest growing business of our times. In other words: World’s fucked and shit sucks.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        7 months ago

        I just did a bit of poking around on the subject of the “right to be forgotten” and it’s legally complex. Data without personally identifying information, and data that’s been anonymized through statistical analysis (which LLM training is a form of) aren’t covered.

        • Barbarian@sh.itjust.works
          link
          fedilink
          arrow-up
          9
          ·
          edit-2
          7 months ago

          Yup. As someone who’s worked a little bit on GDPR compliance, it’s not some magic wand you wave at your data. Any data they receive after the request is also not covered by that request. Also, only EU citizens and residents are legally entitled to make a request. A company may choose to comply with non-EU users, but that’s purely their choice.

          Comments that contain any info about where you live, your ethnicity, disabilities (cognitive or physical), gender, where you work, etc must be deleted as part of a forget request, so that might impact LLM training data.

          Personally identifying information can be somewhat of a grey area in some situations as well. If I were to say I’m from New York, that’d be personally identifying. If I were to say I’m a fan of a sports team in New York, that’s not (even if that implies my location). If I were to say I’m a fan of a New York sports team, my favourite pizza place is in New York, my favourite park is in New York, etc etc, that might arguably be identifying, even if each of the pieces by itself is not.

          EDIT: Oh, and I forgot one of the most important parts: it’s not like there are any spot checks or anything. You’d need someone to actually lodge a formal complaint, with some kind of evidence they haven’t done what they’re supposed to, and the procedures are different for every EU country. They are normally very involved and complex. Essentially, you’d need to lawyer up and care enough to slowly and painfully shove it through the legal system.

    • gila@lemm.ee
      link
      fedilink
      English
      arrow-up
      4
      ·
      7 months ago

      Surely the use of user-deleted content as training data carries the same liabilities as reinstating it on the live site? I’ve checked my old content and it hasn’t been reinstated. I’d assume such a dataset would inherently contain personal data protected by the right to erasure under GDPR, otherwise they’d use it for both purposes. If that is correct, regardless of how they filtered it, the data would be risky to use.

      Perhaps the cumulative action of disenfranchised users could serve toward the result of both the devaluation of a dataset based on a future checkpoint, or reduction in average post quality leading to decreased popularity over time (if we assume content that is user-deleted en masse was useful, which I think is fair).

      • Grimy@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        7 months ago

        I think you need to make a special request to get that level of deletion that comes with gdpr. I’m not certain, I just remember other users specifically talking about how you need to send them an email so they have to comply.

        I also wouldn’t be surprised if their dataset is mostly stripped of user names to get around GDPR though I’m no expert.

        All that to say I’d be very very surprised if they deleted comments in their dataset.

        Very valid point of devaluating the user experience thought, especially when you take into account google searches. I’m sure they have already fallen off compared to a year ago where reddit would pop up half the time no matter what you searched.

        • gila@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          7 months ago

          Well, that’d be the mechanism of how GDPR protections are actioned, yes; but leaving themselves open to these ramifications broadly would be risky. I don’t think it’d satisfy ‘compliance’ to ignore GDPR except upon request. Perhaps the issues with it are even more significant when using it as training data, given they’re investing compute and potentially needing to re-train down the track.

          Based on my understanding; de-identifying the dataset wouldn’t be sufficient to be in compliance. That’s actually how it worked prior to it for the most part, but I know companies largely ended up just re-identifying data by cross-referencing multiple de-identified datasets. That nullification forming part of the basis for GDPR protections being as comprehensive as they are.

          There’d almost certainly be actors who previously deleted their content that later seek to verify whether it was later used to train any public AI.

          Definitely fair to say I’m making some assumptions, but essentially I think at a certain point trying to use user-deleted content as a value add just becomes riskier than it’s worth for a public company

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        1
        ·
        7 months ago

        Surely the use of user-deleted content as training data carries the same liabilities as reinstating it on the live site?

        Why would that be? It’s not the same.

        And what liabilities would there be for reinstating it on the live site, for that matter? Have there been any lawsuits?

  • SirSamuel@lemmy.world
    link
    fedilink
    arrow-up
    6
    ·
    7 months ago

    Your posts, maybe. Apparently I’ve been shadow banned

    I don’t even know what the offensive post was, there’s zero communication. I only post on my local city subreddit and one directly related to my trade, so whatever i guess. It was a good site for a while

  • ALostInquirer@lemm.ee
    link
    fedilink
    arrow-up
    4
    ·
    7 months ago

    Asking similarly as I did with a Twitter post, because I think it’s worth discussing (and people should want others to leave the corporate enclosures so info on the internet may move more freely):

    How might we help and encourage people to leave Reddit?

  • vamp07@lemm.ee
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    6
    ·
    7 months ago

    We all need to move on from this, including myself. We know Reddit’s business practice. If we don’t like it help make the fediverse better. Having said that your posts will be scrapped over here to the benefit of nobody so is that better? At least on Reddit you have the option of owning their stock.

    • haui@lemmy.giftedmc.com
      link
      fedilink
      arrow-up
      5
      ·
      7 months ago

      I disagree very strongly.

      Owning any stock has nothing to do with your content being held hostage. When you and I made the posts on there (I have to assume) we made this for the benefit of everyone. We didnt know that LLMs would come to pass and gobble up our stuff but you can technically make your own LLM and gobble it up yourself. The only problem I have with this one and all other „reddit sells your data to…“ is that its not reddit‘s to sell (legal bla bla, I mean I never agreed to it being sold) and if its reddits to sell, its also mine to sell.

      I know thats not how the corporate owned legal shitshow works but it is why I dont mind anyone scraping lemmy or mastodon. Everyone can do it which makes it worthless for one party. The control to access is what creates worth, which is why we need to abandon all proprietary media asap.

      • vamp07@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        7 months ago

        I don’t really care if content I create is used to train LLMs, but I do object to Reddit monetizing the content I placed on their platform. Without somehow rewarding me for that content. I’d rather just post it here and let the LLMs train on it and let the value accumulate to the LLMs. I do think there’s a greater good for humanity. And just because I say something intelligent or unintelligent, I see that as a gift to the greater good of society. I did not create that gift as a benefit to Reddit shareholders but having said all this I’m realistic, and the best thing I can do is try to keep the value I add within open platforms like this one.

        • haui@lemmy.giftedmc.com
          link
          fedilink
          arrow-up
          1
          ·
          7 months ago

          Great that we are on the same side here.

          Now I‘d like to add that your idea is much better suited for the fediverse than reddit because nobody can control which LLM reads your data and therefore gives the „greater good for humanity“ to exactly that: humanity, instead of some corporation making money off of it.