I haven't been very worried about AI, even though I'm a writer.
Why?
Because it takes a while for the law teams employed by the titans of old media to rumble to action, but it was always clear they were coming. These are the teams that don't sue other companies unless they're certain of winning.
And today, the New York Times sued OpenAI for several billion dollars.
https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html
BTW: this isn't even the lawsuit OpenAI is *really* scared of.
The House of Mouse has yet to tee up to the plate against them or Midjourney or the like.
Modern copyright laws are terrifying, y'all, and the courts have found definitively and repeatedly that AI products are derivative material, and cannot be copyrighted. Unless copyright law is completely rewritten (BTW, it needs to be) the *only* thing AI can be used for is to build better search.
Just another tech hustle, like crypto & NFTs.
@Impossible_PhD what I'm curious about is: how much effort is it to re-train such a generative model?
They're gonna get sued again and again, and each time it's gonna end with "remove our stuff from your model and pay us damages, or keep it in and pay us damages and licensing fees". And as far as I understand, the way these models work, it's impossible to "remove" anything because the training data isn't stored inside as discrete units, it all contributes to the biasing of these artificial neurons.
So will they have to re-train their models after each lawsuit with one source fewer? Or how is that gonna work.
@amberage@eldritch.cafe @Impossible_PhD@hachyderm.io
@Impossible_PhD@hachyderm.io @simontoth@hachyderm.io
they could train models on only copyright-free/copyleft/their own sources but 1) the quality would be lower and 2) they still couldn't copyright the results so they couldn't use it for certain things, just like, lazy stock art generators
Edit: copyleft/other licenses still won't be enough for them to use. The main point was that they'd be limited significantly by being held to legal uses
@rachel @amberage @simontoth The problem with that is the 92-year threshold for copyright expiry. Anything trained on such old material would spit out prose that's *hopelessly* and unrecognizably weird.
@Impossible_PhD@hachyderm.io @amberage@eldritch.cafe @simontoth@hachyderm.io true. There is art/writing/etc that people produce today that they publish with explicit copyright-free/copyleft licences. Some realms have more of it than others and it is absolutely in lower volume that what is otherwise obtainable for training.
Overall my biggest fears is how it'll be used to accelerate the crumbling of the web as we know it, and the torrent of shit news articles.
@Impossible_PhD@hachyderm.io @amberage@eldritch.cafe @simontoth@hachyderm.io mostly thinking of things like open source code, social media posts used for training by those social media companies and such and that is still full of legal landmines
@rachel @Impossible_PhD @amberage @simontoth
If the business case gets made well enough, I’ve figured for a while we’d end up in an uncomfortably dystopian scenario where content farming by AI companies becomes normal.
IE, people are hired to just produce stuff for which they’ve signed away the rights.
Historically high rates of education/skill plus capitalism hell scape leads pretty neatly to such a state IMO.
These “content cattle” will be doing what they’re passionate about anyway.
@rachel @Impossible_PhD @amberage @simontoth
Like, “are you a philosopher but need to pay rent? Come, debate ideas and publish for pay!”
“Music career not panning out and sick of the YouTube algorithm? … just jam everyday with us!”
Seems to tie too well into the education industrial complex that’s developed around the middle class and their “passion coddling”.