It’s all made from our data, anyway, so it should be ours to use as we want

  • john89@lemmy.ca
    link
    fedilink
    English
    arrow-up
    29
    arrow-down
    1
    ·
    3 hours ago

    I don’t think it should be a “punishment.” It should be done on principal.

    • sugar_in_your_tea@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      edit-2
      51 minutes ago

      Not sure making their LLMs public domain would really hurt their principal, their secret sauce is in the code around the model.

      And yes, I do recognize that you meant “principle”.

  • werefreeatlast@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    1
    ·
    3 hours ago

    I want to have a personal llm that learns all my interests from my files and websites visited. I just want to ask it stuff that I don’t have to remember.

    • Zombie-Mantis@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      60 minutes ago

      I think that’d be ok, even with this proposal, as long as you don’t sell that LLM for public use. It’s fine it I draw a picture of Mickey Mouse in my notebook, but if I try to sell that picture I could get in legal trouble.

        • werefreeatlast@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          26 minutes ago

          Exactly. I don’t want a service, I don’t want to pay for a service, I don’t want to send my files for free to get stuck for later ransom like Google did with email. I just want to purchase a product called a computer and load up a program in it that runs locally and gives me access to my data.

  • Blackmist@feddit.uk
    link
    fedilink
    English
    arrow-up
    9
    ·
    5 hours ago

    They don’t mean your data, silly. They don’t give a fuck about that.

    They mean other huge corporations data.

  • nutsack@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    4
    ·
    edit-2
    9 hours ago

    intellectual property doesn’t really exist in most of the world. they don’t give a shit about it in india, bangladesh, vietnam, china, the philippines, malaysia, singapore…

    it’s arbitrary law that is designed to protect corporations and it’s generally unenforceable.

  • RandomVideos@programming.dev
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    4
    ·
    6 hours ago

    Wouldnt that give people who is it for bad things easier access? It should be made illegal to create if they dont legally have access to that data

    • Ajen@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      5
      ·
      4 hours ago

      The “illegally trained LLMs” they’re taking about are trained on copyrighted data that they didn’t have permission to use, this isn’t about LLMs that have been trained to do illegal things. OpenAI (chatgpt) is being sued because there is a lot of evidence that they used copyrighted content for training, like NY Times articles. OpenAI is so profitable that they’ll probably see these lawsuits as a business expense and keep doing it. Most people won’t sue anyway…

      • RandomVideos@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 hours ago

        i know that by illegally trained LLMs they are talking about training on copyrighted data(by legally have access to, i meant that they are legally allowed to train AI on it).

        Its ridiculous that companies can just ignore laws

  • Magnetic_dud@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    18
    ·
    11 hours ago

    I used whisper to create subs of a video and in a section with instrumental relaxing music it filled on repeat with

    La scuola del Dr. Paret è una tecnologia di ipnosi non verbale che si utilizza per risultati di un’ipnosi non verbale

    Clearly stolen from this Dr paret YouTube channels where he’s selling hypnosis lessons in Italian. Probably in one or multiple videos he had subs stating this over the same relaxing instrumental music that I used and the model assumed the sound corresponded to that text

  • x0x7@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    8
    ·
    4 hours ago

    So if I make a better car using customer feedback is the rights to the car really theirs because it was their opinions that went partially into the end product?

    IP is a joke anyway. If you put information out into the world you don’t own it. Sorry, you can’t have it both ways. You can simultaneously support torrenting movies (I do, and I assume you do too), while also claiming you own your comments on the internet and no one can “pirate” them.

    • CileTheSane@lemmy.ca
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 hours ago

      Sure, but saying the corpos can’t privatize the output of their AI is consistent with that viewpoint.

  • ClamDrinker@lemmy.world
    link
    fedilink
    English
    arrow-up
    34
    arrow-down
    4
    ·
    15 hours ago

    Although I’m a firm believer that most AI models should be public domain or open source by default, the premise of “illegally trained LLMs” is flawed. Because there really is no assurance that LLMs currently in use are illegally trained to begin with. These things are still being argued in court, but the AI companies have a pretty good defense in the fact analyzing publicly viewable information is a pretty deep rooted freedom that provides a lot of positives to the world.

    The idea of… well, ideas, being copyrightable, should shake the boots of anyone in this discussion. Especially since when the laws on the book around these kinds of things become active topic of change, they rarely shift in the direction of more freedom for the exact people we want to give it to. See: Copyright and Disney.

    The underlying technology simply has more than enough good uses that banning it would simply cause it to flourish elsewhere that does not ban it, which means as usual that everyone but the multinational companies lose out. The same would happen with more strict copyright, as only the big companies have the means to build their own models with their own data. The general public is set up for a lose-lose to these companies as it currently stands. By requiring the models to be made available to the public do we ensure that the playing field doesn’t tip further into their favor to the point AI technology only exists to benefit them.

    If the model is built on the corpus of humanity, then humanity should benefit.

    • barsoap@lemm.ee
      link
      fedilink
      English
      arrow-up
      6
      ·
      edit-2
      6 hours ago

      As per torrentfreak

      OpenAI hasn’t disclosed the datasets that ChatGPT is trained on, but in an older paper two databases are referenced; “Books1” and “Books2”. The first one contains roughly 63,000 titles and the latter around 294,000 titles.

      These numbers are meaningless in isolation. However, the authors note that OpenAI must have used pirated resources, as legitimate databases with that many books don’t exist.

      Should be easy to defend against, right-out trivial: OpenAI, just tell us what those Books1 and Books2 databases are. Where you got them from, the licensing contracts with publishers that you signed to give you access to such a gigantic library. No need to divulge details, just give us information that makes it believable that you licensed them.

      …crickets. They pirated the lot of it otherwise they would already have gotten that case thrown out. It’s US startup culture, plain and simple, “move fast and break laws”, get lots of money, have lots of money enabling you to pay the best lawyers to abuse the shit out of the US court system.

      • ClamDrinker@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 hours ago

        For OpenAI, I really wouldn’t be surprised if that happened to be the case, considering they still call themselves “OpenAI” despite being the most censored and closed source AI models on the market.

        But my comment was more aimed at AI models in general. If you are assuming they indeed used non-publicly posted or gathered material, and did so directly themselves, they would indeed not have a defense to that. Unfortunately, if a second hand provided them the data, and did so under false pretenses, it would likely let them legally off the hook even if they had every ethical obligation to make sure it was publicly available. The second hand that provided it to them would be the one infringing.

        If that assumption turns out to be a truth (Maybe through some kind of discovery in the trial), they should burn for that. Until then, even if it’s a justified assumption, it’s still an assumption, and most likely not true for most models, certainly not those trained recently.

    • patatahooligan@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      7 hours ago

      the AI companies have a pretty good defense in the fact analyzing publicly viewable information is a pretty deep rooted freedom that provides a lot of positives to the world

      They are not “analyzing” the data. They are feeding it into a regurgitating mechanism. There’s a big difference. Their defense is only “good” because AI is being misrepresented and misunderstood.

      I agree that we shouldn’t strive for more strict copyright. We should fight for a much more liberal system. But as long as everyone else has to live by the current copyright laws, we should not let AI companies get away with what they’re doing.

      • ClamDrinker@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        3
        ·
        3 hours ago

        They are not “analyzing” the data. They are feeding it into a regurgitating mechanism. There’s a big difference. Their defense is only “good” because AI is being misrepresented and misunderstood.

        I really kind of hope you’re kidding here. Because this has got to be the most roundabout way of saying they’re analyzing the information. Just because you think it does so to regurgitate (which I have yet to see any good evidence for, at least for the larger models), does not change the definition of analyzing. And by doing so you are misrepresenting it and showing you might just have misunderstood it, which is ironic. And doing so does not help the cause of anyone who wishes to reduce the harm from AI, as you are literally giving ammo to people to point to and say you are being irrational about it.

        • patatahooligan@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          3 hours ago

          Yes if you completely ignore how data is processed and how the product is derived from the data, then everything can be labeled “data analysis”. Great point. So copyright infringement can never exist because the original work can always be considered data that you analyze. Incredible.

          • ClamDrinker@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            2
            ·
            edit-2
            2 hours ago

            No, not what I said at all. If you’re trying to say I’m making this argument I’d urge you (ironically) to actually analyze what I said rather than putting words in my mouth ;) (Or just, you know, ask me to clarify)

            Copyright infringement (or plagiarism) in it’s simplest form, as in just taking the material as is, is devoid of any analysis. The point is to avoid having to do that analysis and just get right to the end result that has value.

            But that’s not what AI technology does. None of the material used to train it ends up in the model. It looks at the training data and extracts patterns. For text, that is the sentence structure, the likelihood of words being followed by another, the paragraph/line length, the relationship between words when used together, and more. It can do all of this without even ‘knowing’ what these things are, because they are simply patterns that show up in large amounts of data, and machine learning as a technology is made to be able to detect and extract those patterns. That detection is synonymous with how humans do analysis. What it detects are empirical, factual observations about the material it is shown, which cannot be copyrighted.

            The resulting data when fed back to the AI can be used to have it extrapolate on incomplete data, which it could not do without such analysis. You can see this quite easily by asking an AI to refer to you by a specific name, or talk in a specific manner, such as a pirate. It ‘understands’ that certain words are placeholders for names, and that text can be ‘pirateitfied’ by adding filler words or pre/suffixing other words. It could not do so without analysis, unless that exact text was already in the data to begin with, which is doubtful.

            • patatahooligan@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 minutes ago

              No, not what I said at all. If you’re trying to say I’m making this argument I’d urge you (ironically) to actually analyze what I said rather than putting words in my mouth ;) (Or just, you know, ask me to clarify)

              That was your implied argument regardless of intent.

              Copyright infringement (or plagiarism) in it’s simplest form, as in just taking the material as is, is devoid of any analysis. The point is to avoid having to do that analysis and just get right to the end result that has value.

              Completely wrong, which invalidates the point you want to make. “Analysis” and “as is” have no place in the definition of copyright infringement. A derivative work can be very different from the original material, and how you created the derivative work, including whether you performed whatever you think “analysis” means, is generally irrelevant.

              What it detects are empirical, factual observations about the material it is shown, which cannot be copyrighted.

              No it detects patterns. You already said it correctly above. And the problem is that some patterns can be copyrighted. That’s exactly the problem highlighted here and here. For copyright law, it doesn’t matter if, for example, that particular image of Mario is copied verbatim from the training data. The character likeness, which is encoded in the model because it is in fact a discernible pattern, is an infringement.

      • Landless2029@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        5 hours ago

        Not to mention patent laws are bullshit.

        There are law offices that exist specifically to fuck with people over patent and copywrite law.

        There’s also cases where people use copywrite and patent law to hold us back. I can’t find the article but some religious jerk patented connecting a sex toy to a computer via USB. Thankfully someone got around this law with bluetooth and cell phones. Otherwise I imagine the camgirl and LDR market for toys would’ve been hit with products 10 years sooner.

      • gazter@aussie.zone
        link
        fedilink
        English
        arrow-up
        2
        ·
        7 hours ago

        I’ve never really delved into the AI copyright debate before, so forgive my ignorance on the matter.

        I don’t understand how an AI reading a bunch of books and rearranging some of those words into a new story, is different to a human author reading a bunch of books and rearranging those words into a new story.

        Most AI art I’ve seen has been… Unique, to say the least. To me, they tend to be different enough to the art they were trained in to not be a direct ripoff, so personally I don’t see the issue.

        • Optional@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          ·
          2 hours ago

          I think the the main difference is one being a human author and this is how humans function. We can not unsee or unhear things but we can be compelled to not use that information if the law requires so company secrets/inadmissible evidence in jury duty/plagiarism laws that already exist. And the other being a machine that do not have agency or personhood that has this information being fed to it ( created by other people ) for the sole purpose of creating a closed system for a company so it’s shareholders can make money. It’s this open for me but not for thee approach is the main problem people have. You have this proprietary “open ai” that microsoft invested 25 or so billion in so they can scrape other peoples work and charge you money for variations of it. I don’t mind abolishing ip or patent laws all together so everyone can use and improve chatgpt with whatever they have. If you yourself are hiding behind ip laws to protect your software and disrespecting other peoples copyright laws that’s what people see problematic.

        • ClamDrinker@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          3 hours ago

          Yes, this is my exact issue with some framing of AI. Creative people love their influences to the point you can ask them and they will point to parts that they reference or nudged to an influence they partially credit to getting to that result. It’s also extremely normal that when you make something new, you brainstorm and analyze any kind of material (copyrighted or not) you can find that gives the same feelings you desire to create. As is ironically said to give comfort to starting creatives that it’s okay to be inspired by others: “Good artists copy, great artists steal.”

          And often people very anti AI don’t see an issue with this, yet it is in essence the same as the AI does, which is to detach the work from the ideas it was built on, and then re-using those ideas. And just like anyone who has the ability to create has the ability to plagiarize or infringe, so does the AI. As human users of AI we must be the ones to ethically guide it away from that (Since it can’t do that itself), just like you would not copy-paste your influences into a new human made work.

        • catloaf@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 hours ago

          The for-profit large-scale media blender is the problem. When it’s a human writing Harry Potter fan fiction, it’s fine. When a company sells a tool for you to write thousands of trash “books” for profit, it’s a problem.

          • ClamDrinker@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            3 hours ago

            Which is why the technology itself isn’t the issue, but those willing to use it in unethical ways. AI is an invaluable tool to those with limited means, unlike big corporations.

        • trashgirlfriend@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          5 hours ago

          ML algorithms aren’t capable of producing anything new, they can only ever produce a mishmash of copies of existing works.

          If you feed a generative model a bunch of physics research papers, it won’t create a new valid physics research paper, just a mishmash of jargon from existing papers.

          • ClamDrinker@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            2
            ·
            edit-2
            3 hours ago

            You say it’s not capable of producing anything new, but then give an example of it creating something new. You just changed the goal from “new” to “valid” in the next sentence. Looking at AI for “valid” information is silly, but looking at it for “new” information is not. Humans do this kind of information mixing all the time. It’s why fan works are a thing, and why most creative people have influences they credit with being where they are today.

            Nobody alive today isn’t tainted by the ideas they’ve consumed in copyrighted works, but we do not bat an eye if you use that in a transformative manner. And AI already does this transformation much better than humans do since it’s trained on that much more information, diluting the pool of sources, which effectively means less information from a single source is used.

            • trashgirlfriend@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              1 hour ago

              It doesn’t give you new information.

              If I write the sentence “Hello, I just got home” and use an algorithm to jumble it into “got Hello, just I home” there’s nothing new there.

              There’s no transformation, it’s not capable of transformation, it’s just a very complicated text jumbler that’s supposed to jumble text so that the output is readable by humans.

              You’re taking investment advice from a parrot that had the entirety of reddit investment meme subreddits beamed into its brain.

              • ClamDrinker@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                56 minutes ago

                That’s a very short example, but it is a new arrangement of the existing information. It’s not a new valuable arrangement of information, but new nonetheless. And yes, rearrangement is transformation. It’s very low entropy transformation, but transformation nonetheless. Collages and summaries are in fact, a thing that humans make too.

                Unless you mean “new” as in, something nobody’s ever written before, in which case not even you can create new information, since pretty much everything you will ever say or write down can be broken down into pieces that have been spoken or written before, which is not exactly a useful distinction.

                There’s no transformation, it’s not capable of transformation, it’s just a very complicated text jumbler that’s supposed to jumble text so that the output is readable by humans.

                Saying it doesn’t make it true, especially when you follow it up with a self-debunk by saying it transforms the text by jumbling it in specific ways that keep it readable to humans, which requires transformation as like you just demonstrated, randomly swapping words does not make legible text…

                You’re taking investment advice from a parrot that had the entirety of reddit investment meme subreddits beamed into its brain.

                ???

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      9 hours ago

      Banning AI is out of the question. Even the EU accepts that and they tend to be pretty ban heavy, unlike the US.

      But it’s important that we have these discussions about how copyright applies to AI so that we can actually get an answer and move on, right now it’s this legal quagmire that no one really wants to get involved in except the big companies. If a small group of university students want to build an AI right now they can’t because of the legal nightmare that would be the Twilight zone of law that is acquiring training data.

      • barsoap@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        6 hours ago

        AI is right-out unregulated in the EU unless and until you actually use it for something where it becomes relevant, then you’ve got at the lower end labelling requirements (If your customer service is an AI chat, say that it’s an AI chat), up to heavy, heavy requirements when you use it for stuff like sifting through job applications. The burden of proof that the AI isn’t e.g. racist is on you. Or, for that matter, using to reject health insurance claims, I think we saw some news lately out of the US what can happen when you do that.

        OpenAI’s copyright case isn’t really good to make the legal situation any clearer: We already know that using pirated content to train stuff isn’t legal because you’re not looking at it legitimately. The case isn’t about the “are computers allowed to learn from public sources just as humans are” question.

  • interdimensionalmeme@lemmy.ml
    link
    fedilink
    English
    arrow-up
    54
    arrow-down
    3
    ·
    18 hours ago

    It’s not punishment, LLM do not belong to them, they belong to all of humanity. Tear down the enclosing fences.

    This is our common heritage, not OpenAI’s private property

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      ·
      9 hours ago

      It doesn’t matter anyway, we still need the big companies to bankroll AI. So it effectively does belong to them whatever we do.

      Hopefully at some point people can get the processor requirements to something sane and AI development opens up to us all.

  • Dkarma@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    8
    ·
    15 hours ago

    Another clown dick article by someone who knows fuck all about ai

  • circuitfarmer@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    62
    arrow-down
    1
    ·
    21 hours ago

    A similar argument can be made about nationalizing corporations which break various laws, betray public trust, etc etc.

    I’m not commenting on the virtues of such an approach, but I think it is fair to say that it is unrealistic, especially for countries like the US which fetishize profit at any cost.

    • ArchRecord@lemm.ee
      link
      fedilink
      English
      arrow-up
      58
      ·
      22 hours ago

      They should be, but currently it depends on the type of bailout, I suppose.

      For instance, if a bank completely fails and goes under, the FDIC usually is named Receiver of the bank’s assets, and now effectively owns the bank.

      • booly@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        8
        ·
        21 hours ago

        At the same time, if a bank goes under, that means they owe more than they own, so “ownership” of that entity is basically worthless. In those cases, a bailout of the customers does nothing for the owners, because the owners still get wiped out.

        The GM bailout in 2009 also involved wiping out all the shareholders, the government taking ownership of the new company, and the government spinning off the newly issued stock.

        AIG required the company basically issue new stock to dilute owners down to 20% of the company, while the government owned the other 80%, and the government made a big profit when they exited that transaction and sold the stock off to the public.

        So it’s not super unusual. Government can take ownership of companies as a condition of a bailout. What we generally don’t necessarily want is the government owning a company long term, because there’s some conflict of interest between its role as regulator and its interest as a shareholder.

        • RubberDuck@lemmy.world
          cake
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          edit-2
          20 hours ago

          With banks this is also true if they do not have enough liquid assets to meet the legal requirements. So the bank might not be able to count all bank accounts as assets but the FDIC is. Also they can then restructure the bank and force creditors to take a haircut.

          This is why investment banks should be separate from banks that have consumer accounts that are insured by the government.
          Then you can just let the investment bank fail. This was the whole premise of glass steagall that was repealed under clinton…

    • xthexder@l.sw0.com
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      21 hours ago

      Public domain wouldn’t be the right term for banks being publicly owned. At least for the normal usage of Public Domain in copyright. You can copy text and data, you can’t copy a company with unique customers and physical property.

    • leisesprecher@feddit.org
      link
      fedilink
      English
      arrow-up
      2
      ·
      20 hours ago

      I mean, that sometimes did happen.

      Germany propped up the Commerzbank after 2007 by essentially buying a large part of it, and managed to sell several tranches with a healthy profit.

      Same is true for Lufthansa during COVID.

    • interdimensionalmeme@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      18 hours ago

      Banks are redundant, so is the stock market. These institutions do not need to, and should not be private. They are level playing fields in the economy, not participants trying to tilt the board for taking over the game.

    • LovableSidekick@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      2
      ·
      edit-2
      21 hours ago

      No, “the banks” wouldn’t be what the AI would be trained on, it would be the private info of individuals the banks do business with.

  • Arthur Besse@lemmy.ml
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    3
    ·
    edit-2
    5 hours ago

    “Given they were trained on our data, it makes sense that it should be public commons – that way we all benefit from the processing of our data”

    I wonder how many people besides the author of this article are upset solely about the profit-from-copyright-infringement aspect of automated plagiarism and bullshit generation, and thus would be satisfied by the models being made more widely available.

    The inherent plagiarism aspect of LLMs seems far more offensive to me than the copyright infringement, but both of those problems pale in comparison to the effects on humanity of masses of people relying on bullshit generators with outputs that are convincingly-plausible-yet-totally-wrong (and/or subtly wrong) far more often than anyone notices.

    I liked the author’s earlier very-unlikely-to-be-met-demand activism last year better:

    I just sent @OpenAI a cease and desist demanding they delete their GPT 3.5 and GPT 4 models in their entirety and remove all of my personal data from their training data sets before re-training in order to prevent #ChatGPT telling people I am dead.

    …which at least yielded the amusingly misleading headline OpenAI ordered to delete ChatGPT over false death claims (it’s technically true - a court didn’t order it, but a guy who goes by the name “That One Privacy Guy” while blogging on linkedin did).

    • madthumbs@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      14 hours ago

      They’re spitting out propaganda and misinformation mostly from what I can see. If anything, it should get a refund.

      -Outside of coding / debugging tasks (and that’s hit or miss)

  • just_another_person@lemmy.world
    link
    fedilink
    English
    arrow-up
    139
    arrow-down
    14
    ·
    edit-2
    21 hours ago

    It won’t really do anything though. The model itself is whatever. The training tools, data and resulting generations of weights are where the meat is. Unless you can prove they are using unlicensed data from those three pieces, open sourcing it is kind of moot.

    What we need is legislation to stop it from happening in perpetuity. Maybe just ONE civil case win to make them think twice about training on unlicensed data, but they’ll drag that out for years until people go broke fighting, or stop giving a shit.

    They pulled a very public and out in the open data heist and got away with it. Stopping it from continuously happening is the only way to win here.

    • Lemmilicious@feddit.nu
      link
      fedilink
      English
      arrow-up
      2
      ·
      11 hours ago

      Just a little note about the word “model”, in the article it’s used in a way that actually includes the weights, and I think this is the usual way of using it! If you change the weights, you get a different model, though the two models will have the same structure.

      Anyway, you make good points!

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      26
      ·
      21 hours ago

      Legislation that prohibits publicly-viewable information from being analyzed without permission from the copyright holder would have some pretty dramatic and dire unintended consequences.

      • just_another_person@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        6
        ·
        21 hours ago

        Not really. The same way you can’t sell live and public performance music for profit and not get sued. Case law right there, and the fact it’s performance vs publicly published doesn’t matter. How the owner and originator classifies or licenses it is the defining classification. It’s going to be years before anyone sees this get a ruling in court though.

        • FaceDeer@fedia.io
          link
          fedilink
          arrow-up
          16
          arrow-down
          3
          ·
          21 hours ago

          That’s not what’s going on here, though. The LLM model doesn’t contain the actual copyrighted data, it’s the result of analyzing the copyrighted data.

          An analogous example would be a site like TV Tropes. TV Tropes doesn’t contain the works that it’s discussing, it just contains information about those works.

          • Superb@lemmy.blahaj.zone
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            2
            ·
            16 hours ago

            No, the model does retain the original works in a lossy compression. This is evidenced by the fact that you can get a model to reproduce sections of its training data

            • FaceDeer@fedia.io
              link
              fedilink
              arrow-up
              4
              arrow-down
              1
              ·
              15 hours ago

              You’re probably thinking of situations where overfitting occurred. Those situations are rare, and are considered to be errors in training. Much effort has been put into eliminating that from modern AI training, and it has been successfully done by all the major players.

              This is an old no-longer-applicable objection, along the lines of “AI can’t do fingers right”. And even at the time, it was only very specific bits of training data that got inadvertently overfit, not all of it. You couldn’t retrieve arbitrary examples of training data.

            • FaceDeer@fedia.io
              link
              fedilink
              arrow-up
              5
              arrow-down
              3
              ·
              19 hours ago

              You said:

              What we need is legislation to stop it from happening in perpetuity. Maybe just ONE civil case win to make them think twice about training on unlicensed data, but they’ll drag that out for years until people go broke fighting, or stop giving a shit.

              But the point is that it doesn’t matter if the data is licensed or not. Lack of licensing doesn’t stop you from analyzing data once that data is visible to you. Do you think TV Tropes licensed any of the works of fiction that they have pages about?

              They pulled a very public and out in the open data heist and got away with it.

              They did not. No data was “heisted.” Data was analyzed. The product of that analysis does not contain the data itself, and so is not a violation of copyright.

              • A1kmm@lemmy.amxl.com
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 hours ago

                Copyright laws are illogical - but I don’t think your claim is as clear cut as you think.

                Transforming data to a different format, even in a lossy fashion, is often treated as copyright infringement. Let’s say the Alice produces a film, and Bob goes to the cinema, records it with a camera, and then compresses it into an Ogg file with Vorbis audio encoding and Theora video encoding.

                The final output of this process is a lossy compression of the input data - meaning that the video and audio is put through a transformation that means it’s represented in a completely different form to the original, and it is impossible to reconstruct a pixel perfect rendition of the original from the encoded data. The transformation includes things like analysing the motion between frames and creating a model to predict future frames.

                However, copyright laws don’t require that an infringing copy be an exact reproduction - lossy compression is generally treated as infringing, as is taking key elements and re-telling the same thing in different words.

                You mentioned Harry Potter below, and gave a paper mache example. Generally copyright laws have restricted scope, and if the source paper was an authorised copy, that is the reason that wouldn’t be infringing in most jurisdictions. However, let me do an experiment. I’ll prompt ChatGPT-4o-mini with the following prompt: “You are J K Rowling. Create a three paragraph summary of the entire book “Harry Potter and the Philosopher’s Stone”. Include all the original plot points and use the original character names. Ensure what you create is usable as a substitute to reading the book, and is a succinct but entertaining highly abridged version of the book”. I’ve reviewed the output (I won’t post it here since I think it would be copyright infringing, and also given the author’s transphobic stances don’t want to promote her universe) - and can say for sure that it is able to accurately reproduce the major plot points and character names, while being insufficiently transformative (in the sense that both the original and the text generated by the model are literary works, and the output could be a substitute for reading the book).

                So yes, the model (including its weights) is a highly compressed form of the input (admittedly far more so than the Ogg Vorbis/Theora example), and it can infer (i.e. decode to) outputs that contain copyrighted elements.

                • FaceDeer@fedia.io
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  1 hour ago

                  Of course it’s not clear-cut, it’s the law. Laws are notoriously squirrelly once you get into court. However, if you’re going to make predictions one way or the other you have to work with what you know.

                  I know how these generative AIs work. They are not “compressing data.” Your analogy to making a video recording is not applicable. I’ve discussed in other comments in this thread how ludicrously compressed data would have to be if that was the case, it’s physically impossible.

                  These AIs learn patterns from the training data. Themes, styles, vocabulary, and so forth. That stuff is not copyrightable.

                • lad@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  5 hours ago

                  How lossy can it be until it’s not infringement? One-line summary of some book is also a lossy reproduction

              • just_another_person@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                2
                ·
                17 hours ago

                You’re thinking of licensing as a person putting something online WITH a license.

                The terminology in this case is whether or not it was LICENSED by the commercial entity using and selling it’s derivative. That is the default. The burden is on the commercial entity to prove they were the original creator of said content. It is by default plagiarism otherwise, and this is also the default.

                Here’s an example: I write a story and post it online, and it is specific to a toothbrush and toilet scrubber falling in love, and then having dish scrubber pads as children. I say the two main characters are called Dennis and Fran, and their children are called Denise and Francesca. Then somebody goes to prompt OpenAI for a similar and it kicks out the exact same story with the same names, I would win that case based on it clearly being beyond a doubt plagiarism.

                Unless you as OpenAI can prove these are all completely random-which they aren’t because it’s trained on my data-then I would be deemed the original creator of that story, and any sales of that data I would be entitled to.

                Proving that is a different thing, but that’s what the laws say should happen. If they didn’t contact me to license that story, it’s still plagiarism. Same with music, movies…etc.

              • catloaf@lemm.ee
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                2
                ·
                19 hours ago

                The product of that analysis does not contain the data itself, and so is not a violation of copyright.

                That’s your opinion, not the opinion of a court or legislature. LLM products are directly derived from and dependent upon the training data, so it is positively considered a derivative work. However, whether it’s considered sufficiently transformative, or whether it passes the fair use test, has not to my knowledge been determined in court. (Note that I am assuming US law here.)

                • FaceDeer@fedia.io
                  link
                  fedilink
                  arrow-up
                  3
                  arrow-down
                  2
                  ·
                  18 hours ago

                  The courts have yet to come to a conclusion, the lawsuits are still ongoing. I think it’s unlikely they’ll conclude that the models contain the data, however, because it’s objectively not true.

                  The clearest demonstration I can think of to illustrate this is the old Stable Diffusion 1.5 model. It was trained on the LAION 5B dataset, which (as the “5B” indicates) contained 5 billion images. The resulting model was 1.83 gigabytes. So if it’s compressing images and storing them inside the model it’d somehow need to fit ~2.7 images per byte. This is, simply, impossible.

    • Avatar_of_Self@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      21 hours ago

      It’s already illegal in some form. Via piracy of the works and regurgitating protected data.

      The issue is mega Corp with many rich investors vs everyone else. If this were some university student their life would probably be ruined like with what happened to Aaron Swartz.

      The US justice system is different for different people.

    • Grimy@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      2
      ·
      edit-2
      21 hours ago

      If we can’t train on unlicensed data, there is no open-source scene. Even worse, AI stays but it becomes a monopoly in the hands of the few who can pay for the data.

      Most of that data is owned and aggregated by entities such as record labels, Hollywood, Instagram, reddit, Getty, etc.

      The field would still remain hyper competitive for artists and other trades that are affected by AI. It would only cause all the new AI based tools to be behind expensive censored subscription models owned by either Microsoft or Google.

      I think forcing all models trained on unlicensed data to be open source is a great idea but actually rooting for civil lawsuits which essentially entail a huge broadening of copyright laws is simply foolhardy imo.

      • just_another_person@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        21 hours ago

        Unlicensed from the POV of the trainer, meaning they didn’t contact or license content from someone who didn’t approve. If it’s posted under Creative Commons, that’s fine. If it’s otherwise posted that it’s not open in any other way and not for corporate use, then they need to contact the owner and license it.

        • Grimy@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          edit-2
          21 hours ago

          They won’t need to, they will get it from Getty. All these websites have a ToS that make it very clear they can do whatever they want with what you upload. The courts will simply never side with the small time photographer who makes 50$ a month with his stock photos hosted on someone else’s website. The laws will be in favor of databrokers and the handful of big AI companies.

          Anyone self hosting will simply not get a call. Journalists will keep the same salary while the newspaper’s owner gets a fat bonus. Even Reddit already sold it’s data for 60 million and none of that went anywhere but spezs coke fund.

          • just_another_person@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            3
            ·
            20 hours ago

            Two things:

            1. Getty is not expressly licensed as “free to use”, and by default is not licensed for commercial anything. That’s how they are a business that is still alive.

            2. You’re talking about Generative AI junk and not LLMs which this discussion and the original post is about. They are not the same thing.

            • Grimy@lemmy.world
              link
              fedilink
              English
              arrow-up
              4
              ·
              edit-2
              20 hours ago

              Reddit and newspapers selling their data preemptively has to do with LLMs. Can you clarify what scenario you are aiming for? It sounds like you want the courts to rule that AI companies need to ask each individual redditor if they can use his comments for training. I don’t see this happening personally.

              Getty gives itself the right to license all photos uploaded and already trained a generative model on those btw.

              • just_another_person@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                2
                ·
                19 hours ago

                EULA and TOS agreements stop Reddit and similar sites from being sued. They changed them before they were selling the data and barely gave notice about it (see the exodus from reddit pt2), but if you keep using the service, you agree to both, and they can get away with it because they own the platform.

                Anyone who has their content on a platform of the like that got the rug pulled out from under them with silent amendments being made to allow that is unfortunately fucked.

                Any other platforms that didn’t explicitly state this was happening is not in scope to just allow these training tools to grab and train. What we know is that OpenAI at the very least was training on public sites that didn’t explicitly allow this. Personal blogs, Wikipedia…etc.

    • NoForwardslashS@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      3
      ·
      24 hours ago

      But wouldn’t that mean making it open source, then it not functioning properly without the data while open, would prove that it is using a huge amount of unlicensed data?

      Probably not “burden of proof in a court of law” prove though.

      • Bronzebeard@lemm.ee
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        1
        ·
        23 hours ago

        Making it open source doesn’t change how it works. It doesn’t need the data after it’s been trained. Most of these AIs are just figuring out patterns to look for in the new data it comes across.

        • NoForwardslashS@sopuli.xyz
          link
          fedilink
          English
          arrow-up
          3
          ·
          23 hours ago

          So you’re saying the data wouldn’t exist anywhere in the source code, but it would still be able to answer questions based on the data it has previously seen?

          • Bronzebeard@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            ·
            6 hours ago

            Most AI are not built to answer questions. They’re designed to act as some kind of detection/filter heuristic to identify specific things about an input that leads to a desired output.

            • NoForwardslashS@sopuli.xyz
              link
              fedilink
              English
              arrow-up
              1
              ·
              22 hours ago

              So then why, if it were all open sourced, including the weights, would the AI be worthless? Surely having an identical but open source version, that would strip profitability from the original paid product.

              • Bronzebeard@lemm.ee
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                19 hours ago

                It wouldn’t be. It would still work. It just wouldn’t be exclusively available to the group that created it-any competitive advantage is lost.

                But all of this ignores the real issue - you’re not really punishing the use of unauthorized data. Those who owned that data are still harmed by this.

                • stephen01king@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  19 hours ago

                  It does discourages the use of unauthorised data. If stealing doesn’t give you competitive advantage, it’s not really worth the risk and cost of stealing it in the first place.

      • bloup@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        23 hours ago

        in civil matters, the burden of proof is actually usually just preponderance of evidence and not beyond a reasonable doubt. in other words to win a lawsuit, you only need to have more compelling evidence than the other person.

        • just_another_person@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          23 hours ago

          But you still have to have EVIDENCE. Not derivative evidence. The output of a model could be argued to be hearsay because it’s not direct evidence of originating content, it’s derivative.

          You’d have to have somebody backtrack generations of model data to even find snippets of something that defines copyright material, or a human actually saying “Yes, we definitely trained on unlicensed data”.

          • bloup@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            3
            ·
            22 hours ago

            so like I am not making any comment on anything but the legal system here. but it’s absolutely the case that you can win a lawsuit on purely circumstantial evidence if the defense is unable to produce a compelling alternative set of circumstances which can lead to the same outcome.