upvote
> but because AI surprised the copyright industry, it's now too late to enforce copyright like that.

I think I've got whiplash from the way a lot of the tech scene has gone from 'IP troll outfits are malicious actors who make everything worse for everyone else' to 'IP troll outfits are an ethical and effective solution to exploitation in the AI industry'.

I'm not a huge fan of much of the generative AI industry, but is IP maximalism really the answer here? Before 2022 most of us would have agreed that DRM is generally a scourge for example, and the 'copyright industry' are a big part of pushing for the end of general-purpose computing in favour of DRM-controlled appliances. Personally I'd rather go in the opposite direction, copyright lasts for exactly thirty years and after that a work enters the public domain without exception, and I'd weaken anti-circumvention laws too.

reply
"Copyright" is, frankly, just an excuse people who hate AI latch onto.

Many of the people who rally against AI now used to rally against Napster being prosecuted by RIAA and the Big Mouse renewing copyright expiration dates once again.

It's not that they suddenly gained an appreciation for the copyright law. It's that they found something they hate more than the big record label megacorps - and copyright became a tool they think they can leverage against it. Very stupid, IMO.

reply
Even in a counterfactual world where any data that's not in public domain can't be used in AI training at all, ever, AIs would exist. Training on public domain data is a bitch, but it's doable. It's just that it results in worse AIs for more effort. So no one does it other than to flex.

It would still be "commercially viable", mind. I'm not sure how much would it stall the AI development in practice, but all the inputs of making AIs only get cheaper over time. So I struggle to imagine not having something like DALL-E 1 by 2030.

If we extend the counterfactual and allow for licensed media, we compress the timelines and raise the bar. The "best" image generation AIs of 2026 are now made by the likes of Adobe and locked behind some kind of $500 a month per seat Creative Cloud Pro Future subscription. Because Adobe is rich enough to afford big bulk licensing deals, while the likes of academia and smaller startups have to subsist on old public domain data, permissively licensed scraps and small carefully selected batches of licensed data that might block them from sharing the resulting weights with the licensing deals.

In the "counterfactual: licensed media" world, the local AI generation powerhouse of Stable Diffusion ecosystem probably doesn't exist at all. Big companies selling AI do. Their offerings cost a lot more and perform considerably worse than the actual AIs we have today. So you can't just go to a random website and get an image edited for a shitpost for free. But the high end commercial suites exist, they're used by the media and the marketing companies, and they are still way cheaper than hiring artists. The big copyright companies get their pound of flesh, but don't confuse that for the artists getting a win.

reply
> No, a counterfactual world where artists were paid for AI training wouldn't see commercially viable AI at all. A world which plenty of people would be more than happy to live in, mind you.

You recon Disney and Shutterstock don't have enough images to make commercially viable AI?

Or for that matter, Facebook? Even just for photorealistic images from, you know, all the photos people upload.

> AI relies on mass piracy worth Googols of dollars if you count like you would the million dollar iPod, but because AI surprised the copyright industry, it's now too late to enforce copyright like that.

Not that I disagree that people use everything they can get their hands on for marginal improvements, they obviously do, but the copyright industry being "surprised" is the default state of affairs for infringement, and "piracy" is the wrong word because that's a law and the judges so far have ruled that training isn't itself a copyright offence, while also affirming that it is possible to commit a copyright offence by pirating training data.

reply