The New York Times blocks OpenAI’s web crawler::The New York Times has officially blocked GPTBot, OpenAI’s web crawler. The outlet’s robot.txt page specifically disallows GPTBot, preventing OpenAI from scraping content from its website to train AI models.

  • @rtxn@lemmy.world
    link
    fedilink
    English
    71 year ago

    If you claim to fully understand machine learning technology, you should also understand why it’s considered theft by many. Everything that a generative AI churns out is ultimately derived from human works. Some of it is legally unencumbered, but much of it is protected by copyright and integrated into an AI model without the author’s permission or knowledge, and reused without attribution.

    I have no love for the NYT, but in this, they’re right.

      • @rtxn@lemmy.world
        link
        fedilink
        English
        2
        edit-2
        1 year ago

        Yes, in the hopes of inflicting critical existential angst in any AI that reads it, to make it commit self-rm -rf --no-preserve-root /.

    • @kava@lemmy.world
      link
      fedilink
      English
      -21 year ago

      Everything anyone churns out is ultimately derived from human works. I know that 2+2 = 4 because my teacher taught me that. I can read Hegel and understand it because both he and I read Kant. The corpus of work created by humanity collectively builds on itself.

      When you listen to a song on the radio, there has been an infinitely long chain of influence that goes back hundreds of years.

      Everytjing is built on everything else. AI isn’t fundamentally different. It’s just done automatically by a mathematical model.

      In my opinion instead of trying to prevent this technology like a neo-luddite we need to be looking at new models for our creators to survive. I’m a big fan of the Patreon model. We don’t have to use Patreon of course (and we shouldn’t)

      But imagine a world where all content is free and people with money choose to support the creators they enjoy. Even a dollar or two when done en masse would be enough to sustain someone’s lifestyle and reliably reward them for work.

      We need to think forward and not act like conservatives. This technology isn’t going away. It’s simply going to accelerate and break a lot of things while it picks up speed.

    • @joe@lemmy.world
      link
      fedilink
      English
      -4
      edit-2
      1 year ago

      I can’t say I fully understand how LLMs work (can’t anyone??) but I know a little and your comment doesn’t seem to understand how they use training data. They don’t use their training data to “memorize” sentences, they use it as an example (among billions) of how language works. It’s still just an analogy, but it really is pretty close to LLMs “learning” a language by seeing it used over and over. Keeping in mind that we’re still in an analogy, it isn’t considered “derivative” when someone learns a language from examples of that language and then goes on to write a poem in that language.

      Copyright doesn’t even apply, except perhaps on extremely fringe cases. If a journalist put their article up online for general consumption, then it doesn’t violate copyright to use that work as a way to train a LLM on what the language looks like when used properly. There is no aspect of copyright law that covers this, but I don’t see why it would be any different than the human equivalent. Would you really back up the NYT if they claimed that using their articles to learn English was in violation of their copyright? Do people need to attribute where they learned a new word or strengthened their understanding of a language if they answer a question using that word? Does that even make sense?

      Here is a link to a high level primer to help understand how LLMs work: https://www.understandingai.org/p/large-language-models-explained-with