AI companies have all kinds of arguments against paying for copyrighted content::The companies building generative AI tools like ChatGPT say updated copyright laws could interfere with their ability to train capable AI models. Here are comments from OpenAI, StabilityAI, Meta, Google, Microsoft and more.

    • Sibbo@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      31
      arrow-down
      1
      ·
      11 months ago

      This. If the model and its parameters are open source and under an unrestricted license, they can scrape anything they want in my opinion. But if they make money with someone’s years of work writing a book, then please give that author some money as well.

      • abhibeckert@lemmy.world
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        11
        ·
        edit-2
        11 months ago

        But if they make money with someone’s years of work writing a book, then please give that author some money as well.

        Why? I’ve read many books on programming, and now I work as a programmer. The authors of those books don’t get a percentage of my income just because they spent years writing the book. I’ve also read (and written) plenty of open source code over the years, and learned from that code. That doesn’t mean I have to give money to all the people who contributed to those projects.

          • LufyCZ@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            2
            ·
            11 months ago

            He might’ve borrowed them from a library.

            OpenAI could’ve trained on borrowed ebooks as well

        • Davin@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          1
          ·
          11 months ago

          Like with most things, consent and intent matter. I went out on Halloween when I was a kid and got free candy, so why is it bad if I break in and steal other people’s candy?

    • makyo@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      11 months ago

      I will never be totally happy with this situation until they’re required to offer a free version of all the models that were created with unlicensed content.

  • sunbeam60@lemmy.one
    link
    fedilink
    English
    arrow-up
    39
    arrow-down
    2
    ·
    11 months ago

    As I’ve said before: Sure, go ahead, train on copyrighted material, but anything AI generates shouldn’t be copyrightable either - it wasn’t created by a human.

      • sunbeam60@lemmy.one
        link
        fedilink
        English
        arrow-up
        15
        ·
        11 months ago

        Not where I live. In the U.K. copyright for AI generated works is owned by the person that caused the AI to create the work. See s178 and s9(3) of the Copyright, Design and Patent Act of 1987.

  • FireTower@lemmy.world
    link
    fedilink
    English
    arrow-up
    23
    ·
    11 months ago

    Stock image companies have probably the strongest CR claim here IMO. An AI trained off their images without paying for licence could act as a market replacement for their service.

  • Fisk400@feddit.nu
    link
    fedilink
    English
    arrow-up
    22
    arrow-down
    3
    ·
    11 months ago

    Tough tits. Imagine all the books, movies and games that could have been made if copyright didn’t exist. Nobody else gets to ignore the rules just because it’s inconvenient.

    • CosmicTurtle@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      ·
      11 months ago

      Honestly, if tech companies have to battle it out in court with Disney, imma grab some popcorn and watch the money go brrrrr.

    • Evotech@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      11 months ago

      And if it’s ok. Then what’s the limit on what an AI is, do you have to prove an AI made it? Or can you just write some repetitive work and say it’s made by AI and dodge copyright?

      • Fisk400@feddit.nu
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        11 months ago

        Because they cant. They do not have the ability to be creative or be inspired.

        • fuckwit_mcbumcrumble@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          Why can’t they? Sure if you ask it to make a pixel perfect replica of the Mona Lisa it will probably do it. But if you’re asking it to make something new it will make something new.

          • Fisk400@feddit.nu
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            11 months ago

            Then you don’t need to use copyrighted material as training data. If the AI can be creative and make new things it doesn’t need the copyrighted material. Just tell it to make new things and use the new things to make other new things.

  • Uriel238 [all pronouns]@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    2
    ·
    11 months ago

    Here is what deliberative experts way smarter and more knowledgeable than I am are saying ( on TechDirt )

    TLDR: Letting AI be freely trained on human-made artistic content may be dangerous. We may decide to stop it so long as capitalists control who eats and lives. But copyright is not the means to legally stop it. This is a separate issue to how IP law is way, way broken. And precedents stopping software from training on copyrighted work will be used to stop humans from training on copyrighted work. And that’s bad.

    • ryannathans@aussie.zone
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      5
      ·
      11 months ago

      Agree, it’s not much different from a human learning from all these sources and then applying said knowledge

      • realharo@lemm.ee
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        1
        ·
        edit-2
        11 months ago

        Scale matters. For example

        • A bunch of random shops having security cameras, where their employees can review footage

        • Every business in a country having a camera connected to a central surveillance network with facial recognition and search capabilities

        Those two things are not the same, even though you could say they’re “not much different” - it’s just a bunch of cameras after all.

        Also, the similarity between human learning and AI training is highly debatable.

        • 𝙲𝚑𝚊𝚒𝚛𝚖𝚊𝚗 𝙼𝚎𝚘𝚠@programming.dev
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          11 months ago

          Both of your examples are governed by the same set of privacy laws, which talk about consent, purpose and necessity, but not about scale. Legislating around scale open up the inevitable legal quagmires of “what scale is acceptable” and “should activity x be counted the same as activity y to meet the scale-level defined in the law”.

          Scale makes a difference, but it shouldn’t make a legal difference w.r.t. the legality of the activity.

          • lollow88@lemmy.ml
            link
            fedilink
            English
            arrow-up
            3
            ·
            11 months ago

            Scale makes a difference, but it shouldn’t make a legal difference w.r.t. the legality of the activity.

            What do you think the difference between normal internet traffic and a ddos attack is?

      • topinambour_rex@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        2
        ·
        edit-2
        11 months ago

        When Google trained their playing neural network, they trained it to starcraft2 . It was better at it than professional gamer. It trained by watching 100years of play. Or 36500 days of play. Or 876000 hours of play.

        Does a human can do that ? We both know it’s impossible. As the other person said, the issue is scale.

  • Mahlzeit@feddit.de
    link
    fedilink
    English
    arrow-up
    11
    ·
    11 months ago

    This thread is interesting reading. Normally, people here complain about capitalism left and right. But when an actual policy choice comes up, the opinions become firmly pro-capitalist. I wonder how that works.

    • YⓄ乙 @aussie.zone
      link
      fedilink
      English
      arrow-up
      5
      ·
      11 months ago

      Human beings are funny characters. They only care when it starts to affect them personally otherwise they say all kinda shit.

    • ThatWeirdGuy1001@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      11 months ago

      Everyone’s always up for changing things until it comes to making the actual sacrifices necessary to enact the changes

      • Mahlzeit@feddit.de
        link
        fedilink
        English
        arrow-up
        5
        ·
        11 months ago

        That’s the thing. I don’t see how there is sacrifice involved in this. I would guess that the average user here has personally more to lose than to gain from expanded copyrights.

  • eldrichhydralisk@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    9
    ·
    11 months ago

    Most of these companies are just arguing that they shouldn’t have to license the works they’re using because that would be hard and inconvenient, which isn’t terribly compelling to me. But Adobe actually has a novel take I hadn’t heard before: they equate AI development to reverse engineering software, which also involves copying things you don’t own in order to create a compatible thing you do own. They even cited a related legal case, which is unusual in this pile of sour grapes. I don’t know that I’m convinced by Adobe’s argument, I still think the artists should have a say in whether their works go into an AI and a chance to get paid for it, but it’s the first argument I’ve seen for a long while that’s actually given me something to think about.

    • PhlubbaDubba@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      11 months ago

      Yeah the difference is that a software company still makes money if you use their working code to improve your own shit. You’re not allowed to just copy paste Oracle’s entire repo and then sell it as your own original code.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    11 months ago

    This is the best summary I could come up with:


    The US Copyright Office is taking public comment on potential new rules around generative AI’s use of copyrighted materials, and the biggest AI companies in the world had plenty to say.

    We’ve collected the arguments from Meta, Google, Microsoft, Adobe, Hugging Face, StabilityAI, and Anthropic below, as well as a response from Apple that focused on copyrighting AI-written code.

    There are some differences in their approaches, but the overall message for most is the same: They don’t think they should have to pay to train AI models on copyrighted work.

    The Copyright Office opened the comment period on August 30th, with an October 18th due date for written comments regarding changes it was considering around the use of copyrighted data for AI model training, whether AI-generated material can be copyrighted without human involvement, and AI copyright liability.

    There’s been no shortage of copyright lawsuits in the last year, with artists, authors, developers, and companies alike alleging violations in different cases.

    Here are some snippets from each company’s response.


    The original article contains 168 words, the summary contains 168 words. Saved 0%. I’m a bot and I’m open source!

  • Mahlzeit@feddit.de
    link
    fedilink
    English
    arrow-up
    1
    ·
    11 months ago

    So this has been going around in my head for the last couple days. Why are opinions here, on this topic, so decidedly right-wing?

    I’ll have to pick this apart.

    Copyrights are a form of property. Where such intellectual property is used to make money, it is intangible capital. License payments are capital income. Property is distributed very unequally. Most of it is owned by rich people. Those who demand license payments here are literally demanding that more money should go to rich capitalists.

    People who create copyrighted materials for their employers do not own the copyrights thereto. They are just like factory workers who do not own the product either. The people who worked on animations in the pre-CGI era were basically factory workers. When these jobs disappeared due to computers, where was the hand-wringing?

    A brush-wielding artist has as much to do with the copyright industry as a pitchfork-wielding farmer with the agro-industry.

    This isn’t even normal capitalism but the absolutely worst kind. The copyrighted material was uploaded to the net for many reasons, including making a profit. Some people used this publically available resource to train AIs. The owners contributed no labor. They were affected so little that they mostly seem to have been unaware that anything was going on.

    The sole argument for paying seems to be mainly “muh property rights!”. I am not seeing any consideration of the good of society, public benefit, the general welfare, or anything of the sort. Those who say the trained AI models should be released for free, seem to imply that they should not be able to profit, because they looked at someone else’s property.

    This is far more capitalistic than even US capitalism.

    Consider patents. To get a patent, a new invention has to be registered, which involves publishing how it works in enough detail so that others can copy it. Then the government will enforce a monopoly on that invention for 20 years. During that time, the inventor can demand license fees. But also, other people can learn from it and maybe find other solutions. After those 20 years, the knowledge becomes public domain. This is often framed as a social contract: temporary monopoly in exchange for advancing knowledge. Scientific discoveries don’t get anything at all.

    Compared to how copyright is treated by so many here, actual US capitalism looks almost like socialism!

    US copyright used to work exactly like patents, with the same duration. Today, copyrights last until 70 years after the death of the creator. It’s just FUBAR. The US Constitution, far-left manifesto that it is, still it limits to the purpose of promoting intellectual output (to put it in modern terms). It is supposed to help society and not to enable capitalist grift.

    People like to blame corrupt politicians or lobbyists for what is going wrong in the US but perhaps US politics is delivering exactly what people want. They may not like the necessary and predictable outcome of their choices, but it’s still what they want.

    Americans left and right curse those evil corporations. Of course, Americans side with the individuals when some faceless corporation tries to bully money out of them. Well, a union is just such a corporation. Look up the definition of corporation if you don’t believe me.

  • sugarfree@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    11
    ·
    11 months ago

    The billion dollar companies will win, and we’ll be better off for it. AI models need training, the idea that the open internet shouldn’t be used to train it is asinine. AI is the future.