164
submitted 1 year ago by floofloof@lemmy.ca to c/technology@lemmy.ml
top 50 comments
sorted by: hot top controversial new old
[-] Moobythegoldensock@lemm.ee 27 points 1 year ago

At the crux of the author's lawsuit is the argument that OpenAI is ruthlessly mining their material to create "derivative works" that will "replace the very writings it copied."

The authors shoot down OpenAI's excuse that "substantial similarity is a mandatory feature of all copyright-infringement claims," calling it "flat wrong."

Goodbye Star Wars, Avatar, Tarantino’s entire filmography, every slasher film since 1974…

[-] taanegl@beehaw.org 17 points 1 year ago

Uh, yeah, a massive corporation sucking up all intellectual property to milk it is not the own you think it is.

[-] Umbrias@beehaw.org 11 points 1 year ago

But this is literally people trying to strengthen copyright and its scope. The corporation is, out of pure convenience, using copyright as it exists currently with the current freedoms applied to artists.

load more comments (6 replies)
[-] Even_Adder@lemmy.dbzer0.com 9 points 1 year ago* (last edited 1 year ago)

AI training isn’t only for mega-corporations. We can already train our own open source models, so we shouldn't applaud someone trying to erode our rights and let people put up barriers that will keep out all but the ultra-wealthy. We need to be careful not weaken fair use and hand corporations a monopoly of a public technology by making it prohibitively expensive to for regular people to keep developing our own models. Mega corporations already have their own datasets, and the money to buy more. They can also make users sign predatory ToS allowing them exclusive access to user data, effectively selling our own data back to us. Regular people, who could have had access to a corporate-independent tool for creativity, education, entertainment, and social mobility, would instead be left worse off with fewer rights than where they started.

[-] HawlSera@lemm.ee 15 points 1 year ago* (last edited 1 year ago)

Is actually reminds me of a Sci-Fi I read where in the future, they use an ai to scan any new work in order to see what intellectual property is the big Corporation Zone that may have been used as an influence in order to Halt the production of any new media not tied to a pre-existing IP including 100% of independent and fan-made works.

Which is one of the contributing factors towards the apocalypse. So 500 years later after the apocalypse has been reversed and human colonies are enjoying post scarcity, one of the biggest fads is rediscovering the 20th century, now that all the copyrights expired in people can datamine the ruins of Earth to find all the media that couldn't be properly preserved heading into Armageddon thanks to copyright trolling.

It's referred to in universe as "Twencen"

The series is called FreeRIDErs if anyone is curious, unfortunately the series may never have a conclusion, (untimely death of co creator) most of its story arcs were finished so there's still a good chunk of meat to chew through and I highly recommend it.

[-] anachronist@midwest.social 10 points 1 year ago

OpenAI is trying to argue that the whole work has to be similar to infringe, but that's never been true. You can write a novel and infringe on page 302 and that's a copyright infringement. OpenAI is trying to change the meaning of copyright otherwise, the output of their model is oozing with various infringements.

load more comments (4 replies)
load more comments (2 replies)
[-] mindbleach@sh.itjust.works 21 points 1 year ago

I don't care what works a neural network gets trained on. How else are we supposed to make one?

Should I care more about modern eternal copyright bullshit? I'd feel more nuance if everything a few decades old was public-domain, like it's fucking supposed to be. Then there'd be plenty of slightly-outdated content to shovel into these statistical analysis engines. But there's not. So fuck it: show the model absolutely everything, and the impact of each work becomes vanishingly small.

Models don't get bigger as you add more stuff. Training only twiddles the numbers in each layer. There are two-gigabyte networks that have been trained on hundreds of millions of images. If you tried to store those image, verbatim, they would each weigh barely a dozen bytes. And the network gets better as that number goes down.

The entire point is to force the distillation of high-level concepts from raw data. We've tried doing it the smart way and we suck at it. "AI winter" and "good old-fashioned AI" were half a century of fumbling toward the acceptance that we don't understand how intelligence works. This brute-force approach isn't chosen for cost or ease or simplicity. This is the only approach that works.

[-] anachronist@midwest.social 3 points 1 year ago

Models don’t get bigger as you add more stuff.

They will get less coherent and/or "forget" the earlier data if you don't increase the parameters with the training set.

There are two-gigabyte networks that have been trained on hundreds of millions of images

You can take a huge tiff of an image, put it through JPEG with the quality cranked all the way down and get a tiny file out the other side, which is still a recognizable derivative of the original. LLMs are extremely lossy compression of their training set.

[-] mindbleach@sh.itjust.works 4 points 1 year ago

which is still a recognizable derivative of the original

Not in twelve bytes.

Deep models are a statistical distillation of a metric shitload of data. Smaller models with more training on more data don't get worse, they get more abstract - and in adversarial uses they often kick big networks' asses.

load more comments (7 replies)
[-] archomrade@midwest.social 19 points 1 year ago

Copyright is already just a band-aid for what is really an issue of resource allocation.

If writers and artists weren't at risk of loosing their means of living, we wouldn't need to concern ourselves with the threat of an advanced tool supplanting them. Nevermind how the tool is created, it is clearly very valuable (otherwise it would not represent such a large threat to writers) and should be made as broadly available (and jointly-owned and controlled) as possible. By expanding copyright like this, all we're doing is gatekeeping the creation of AI models to the largest of tech companies, and making them prohibitively expensive to train for smaller applications.

If LLM's are truly the start of a "fourth industrial revolution" as some have claimed, then we need to consider the possibility that our economic arrangement is ill-suited for the kind of productivity it is said AI will bring. Private ownership (over creative works, and over AI models, and over data) is getting in the way of what could be a beautiful technological advancement that benefits everyone.

Instead, we're left squabbling over who gets to own what and how.

[-] Franzia@lemmy.blahaj.zone 7 points 1 year ago

fourth industrial revolution" as some have claimed

The people claiming this are often the shareholders themselves.

prohibitively expensive to train for smaller applications.

There is so much work out there for free, with no copyright. The biggest cost in training is most likely the hardware, and I see no added value in having AI train on Stephen King ☠️

Copyright is already just a band-aid for what is really an issue of resource allocation.

God damn right but I want our government to put a band aid on capitalists just stealing whatever the fuck they want "move fast and break things". It's yet another test for my confidence in the state. Every issue, a litmus test for how our society deals with the problems that arise.

[-] archomrade@midwest.social 3 points 1 year ago

There is so much work out there for free, with no copyright

There's actually a lot less than you'd think (since copyright lasts for so long), but even less now that any online and digitized sources are being locked down and charged for by the domain owners. But even if it were abundant, it would likely not satisfy the true concern here. If there was enough data to produce an LLM of similar quality without using copyrighted data, it would still threaten the security of those writers. What is to say a user couldn't provide a sample of Stephen King's writing to the LLM and have it still produce derivative work without having trained it on copyrighted data? If the user had paid for that work, are they allowed to use the LLM in the same way? If they aren't who is really at fault, the user or the owner of the LLM?

The law can't address the complaints of these writers because interpreting the law to that standard is simply too restrictive and sets an impossible standard. The best way to address the complaint is to simply reform copyright law (or regulate LLM's through some other mechanism). Frankly, I do not buy that the LLM's are a competing product to the copyrighted works.

The biggest cost in training is most likely the hardware

That's right for large models like the ones owned by OpenAI and Google, but with the amount of data needed to effectively train and fine-tune these models, if that data suddenly became scarce and expensive it could easily overtake hardware cost. To say nothing for small consumer models that are run on consumer hardware.

capitalists just stealing whatever the fuck they want “move fast and break things”

I understand this sentiment, but keep in mind that copyright ownership is just another form of capital.

load more comments (1 replies)
[-] mrcleanup@lemmy.world 18 points 1 year ago

I think the place we haven't quite gotten to yet is that copyright is probably the wrong law for this. What the AI is doing is reverse engineering the authors magic formula for creating new works, which would likely be patent law.

In the past this hasn't really been possible for a person to do reliably, and it isn't really quantifiable as far as filling a patent for your process, yet the AI does it anyway, leaving us in a weird spot.

[-] Mandarbmax@lemmy.world 18 points 1 year ago

US patent professional here

Ya, saying it isn't possible to do under patent law is no understatement. Even making the patent applications possible to allow would require changes to 35 U. S. C. 112 (A, and probably also B), 35 U. S. C. 101. This all assumes that all authors would have the time and money and energy to file a patent, which even with a good attorney is analogous to is many many hours of work and filing pro se would be like writing a whole new book. After the patent is allowed the costs of continuation applications to account for changes in the process as the author learns and grows would be a hellish burden. After this comes the 20 year lifespan of a patent (assuming all maintenance fees are paid, which is quite the assumption, those are not cheap) at which point the patent protections are dead and the author needs to invent a new process to be protected. Don't even get me started on enforcing a patent.

Patent law is fundamentally flawed to be sure but even if every author gets infinite money and time to file patents with then the changes needed to patent law to let them do so would leave patent law utterly broken for other purposes.

Using patent law for this is a good idea to bring up but for the above reasons I don't think it is viable at all. It would be better and more realistic to have congress change copyright law than to change patent law I think. Sadly, I don't think that is particularly likely either. :(

[-] TheGreenGolem@lemm.ee 6 points 1 year ago* (last edited 1 year ago)

Can't they just create a brand new law, specifically to cover these use-cases we have here?

Edit: And thank you for your detailed answer! It was educational.

[-] Mandarbmax@lemmy.world 4 points 1 year ago

They could and imho (I'm not an expert on this) they probably should. This would fall under unfucking copyright though, or perhaps under a new thing along side copyright and patent law (though that sounds like more work than updating copyright law). Amending it into patent law would be the toughest option. The simple answer as to why I think that is that the vibes are off.

As a rough analogy it would be like combating public flashers by changing the rules for the department of transportation rather than the criminal justice system (ignoring how fucked the criminal justice system is).

[-] tias@discuss.tchncs.de 11 points 1 year ago

And patent law is even more broken than copyright law.

[-] Mandarbmax@lemmy.world 10 points 1 year ago

I don't know if I would say more broken, at least patents have limits on how long they can exist for, putting an upper bound on how much damage they can cause. The again, limiting the production of vaccines during a pandemic is a lot more urgent than letting people do micky mouse cartoons so the standard for what broken is has to be a lot more stringent. It is more important for patent law to not be broken than it is for copyright law so the same amount of brokenness feels worse with patents.

load more comments (1 replies)
[-] autotldr 15 points 1 year ago

This is the best summary I could come up with:


ChatGPT creator OpenAI has been on the receiving end of two high profile lawsuits by authors who are absolutely livid that the AI startup used their writing to train its large language models, which they say amounts to flaunting copyright laws without any form of compensation.

One of the lawsuits, led by comedian and memoirist Sarah Silverman, is playing out in a California federal court, where the plaintiffs recently delivered a scolding on ChatGPT's underlying technology.

At the crux of the author's lawsuit is the argument that OpenAI is ruthlessly mining their material to create "derivative works" that will "replace the very writings it copied."

The authors shoot down OpenAI's excuse that "substantial similarity is a mandatory feature of all copyright-infringement claims," calling it "flat wrong."

It can brag that it's a leader in a booming AI industry, but in doing so it's also painted a bigger target on its back, making enemies of practically every creative pursuit.

High profile literary luminaries behind that suit include George R. R. Martin, Jonathan Franzen, David Baldacci, and legal thriller maestro John Grisham.


The original article contains 369 words, the summary contains 180 words. Saved 51%. I'm a bot and I'm open source!

[-] beejjorgensen@lemmy.sdf.org 6 points 1 year ago

"substantial similarity is a mandatory feature of all copyright-infringement claims"

Is that not a requirement? Time for me to start suing people!

[-] NounsAndWords@lemmy.world 5 points 1 year ago

I take it we don't use the phrase "good writers borrow, great writers steal" in this day and age...

[-] AphoticDev@lemmy.dbzer0.com 10 points 1 year ago

Wait till they find out photographers spend their whole careers trying to emulate the style of previous generations. Or that Adobe has been implementing AI-driven content creation into Photoshop and Lightroom for years now, and we've been pretending we don't notice because it makes our jobs easier.

load more comments
view more: next ›
this post was submitted on 02 Oct 2023
164 points (89.4% liked)

Technology

34975 readers
103 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS