upvote
This is, in my opinion, attempting to say the right thing with entirely the wrong perspective:

The people you say are getting "shafted" always got shafted. Their works are the inspiration for all artists and people who lay their eyes on it - maybe they got paid when they made the work, maybe they managed to sell it, but probably not. And still, other artists (and machines) will use remember and be inspired by it, sometimes to the point of verbatim copy (which is extremely common for human artists as well, with verbatim copy and replication being an actual sought after skill).

(Those about to shout "LICENSING", that's a very new invention and we're terrible at it. What are you going to do, cut out the part of your brain that formed new connections while touching GPL code?)

The person (singular) that is actually getting "shafted" at each use is the artist you didn't hire to do the job of making your new work, because it is their skill that got replaced. A skill build from a lifetime of studying other art and practicing themselves, replaced with a skill build from a machine studying other art and by virtue of some closed loops likely also "practicing" itself.

Still, shafting at large, but the obsession with training data is misplaced in that it entirely ignores how society and art worked beforehand.

At the same time, for most of the things you're likely using the tool for, there would probably would never have been an artist in the first place. For example, if you're just making your powerpoint prettier, or if your commission is ridiculous as it often is and yet only willing to offer a single-digit dollar sum per work which no artist should take (RIP the poor souls that take such work anyway).

reply
Children can draw without ever having been to an art gallery. The IP laundromats need the entire stolen corpus of human labor. The latter is clearly an infringing derivative work.

It will be true no matter who many bribes those who have never created anything pay to Marsha Blackburn (who miraculously reversed her AI skepticism).

I wonder how many threats of being primaried have been issued by the uncreative technocrat thieves.

reply
If "people who made this possible" were getting their fair share, "a millionth of a cent for every billion USD made with it" would be about it for the artists.

What makes the dataset valuable isn't that the image 0012992 in it is precious and irreplaceable. It's that the index goes to seven digits. Pre-training is very much a matter of scale - and scraping is merely the easiest way to get data at scale.

People who complain about "artists not getting paid" must have in their imagination some kind of counterfactual where artists are being paid thousands for their contributions. That's not how it works. A counterfactual world where artists were paid for AI training is one where an average artist is 5 cents richer, an average image generation AI performs 5% worse, and the bulk of extra data spending is captured by platforms selling stock photos and companies destructively digitizing physical media.

reply
> A counterfactual world where artists were paid for AI training is one where an average artist is 5 cents richer, an average image generation AI performs 5% worse, and the bulk of extra data spending is captured by platforms selling stock photos and companies destructively digitizing physical media.

No, a counterfactual world where artists were paid for AI training wouldn't see commercially viable AI at all. A world which plenty of people would be more than happy to live in, mind you.

AI relies on mass piracy worth Googols of dollars if you count like you would the million dollar iPod, but because AI surprised the copyright industry, it's now too late to enforce copyright like that.

reply
The ideal world would be one where, to train on art, you have to buy a license to that art. Sure, for most artists they would maybe put a low price tag, but that isn't the point.

The point isn't about money. It's that copies were made, without license and without permission, and without any legal right to do so, of art, and then used to train a system which generates similar art. The first step, the copy, is illegal without a license, and even for most public images online, licenses and copyright notices (which must be preserved) are attached.

reply
"Without any legal right to do so" is for the courts to decide. And so far, the courts are very much not deciding the way you want them to.

"Fair use" counters "without license and without permission" hard. The argument that training AI on scraped data is "fair use" and the resulting model outputs are "transformative works" has held up in courts. Anthropic got dinged for downloading pirated books, but not for throwing the ones they didn't pirate down the training pipeline.

Some countries, like Japan, have amended their copyright laws to make AI training categorically legal. Others are in "fair use clauses" grey areas with courts deciding case by case based on precedent and interpretation. So trying to latch onto copyright law is, as it always was, the wrong move. Copyright never favored the small guy. Stupid to expect that it suddenly will.

reply
Ideal for whom? For society in general, I don’t think so.
reply
I think it would obviously better for society.
reply
If the dataset weren't valuable, big tech wouldn't depend on it to train their models.

I don't care about getting a millionth of a cent as an artist (which btw is a number *you* just pulled out of your imagination). I care about them paying a fair share instead of pocketing it, so the money stays in circulation instead of creating a new class of technofeudal lords.

reply
deleted
reply
> Pre-training is very much a matter of scale - and scraping is merely the easiest way to get data at scale.

Therein lies the problem. AI firms just bulldozed ahead and "just did it" with no consideration for the ethics or legality. (Nor for that matter, how they're going to get this data in the future now that they're pushing artists into unemployment and filling the internet with slop.)

There is no "imagined counterfactual", people just want AI firms to follow basic ethics and apply consent. Something tech in general is woefully inadequate at.

The counterfactual isn't offered by artists, but AI companies. "If we had to ask consent then we couldn't have made this". Okay, so? The world isn't worse off without OpenAI's image generator. Who cares, there's no economic value to these slop images, they're merely replacing stock assets & quickly thrown together MS paint placeholders.

Given how much of a shitshow this technology has always been (I refuse to mince words: This tech had it's "big break" as "deepfakes", and Elon Musk has escalated that even further. It's always been sexual harassment.) The actual net value to society is almost certainly negative.

reply
I don't understand why everyone is all up and arms about Images / Art being generated by AI, but when it comes to code... well who cares? The people who made all the code training data are also getting nothing!

Potentially the one difference is that developers invented this and screwed themselves, whereas artists had nothing to do with AI.

reply
> developers invented this and screwed themselves

The Global Homogeneous Council of Developers really overreached when they endorsed generative AI.

reply
If you look at my comment history (don't, you'll fall over from boredom), you'll see I'm also against that. I've researched and selected specific licenses for all the code I've open sourced, which is quite a lot, and the fact that massive companies can just ignore that with absolutely zero I can do about it really pisses me off! But at least I still get paid. The same can't be said about artists.

Customers usually can figure out when a product is shitty software, but shitty art, well that's a bit harder for people to judge.

reply
Rob Pike cares. In other places apart from HN there is more resistance. Perceived lack of resistance has multiple reasons:

- Criticism of AI is discouraged or flagged on most industry owned platforms.

- The loudest pro-AI software engineers work for companies that financially benefit from AI.

- Many are silent because they fear reprisals.

- Many software engineers lack agency and prefer to sit back and understand what is happening instead of shaping what is happening.

- Many software engineers are politically naive and easily exploited.

Artists have a broader view and are often not employed by the perpetrators of the theft.

reply
I've seen anti-AI comments here disappear within minutes of posting. I'm honestly surprised to see one at the top of this thread.

What causes comments to disappear? Is that what flagging does?

reply
I see properly argued positions, even if very anti-AI, hang around, but cheap tribalist takes usually get downvoted pretty quickly.
reply
Cheap pro-AI comments don't get flagged though. You can repeat the same talking points forever:

- "Artists have always been exploited" (patently false since at least 1950, it was a symbiosis with the industry).

- "Humans have always done $X".

- "You are a Luddite."

- "This is inevitable."

reply
showdead=no in user settings hides flagged & moderator killed posts
reply
You probably see that because many are low effort Reddit level comments. I’ve seen lots of long AI skeptic threads and people talking about the likely negatives of AI.
reply
Maybe SWEs just can think better and see that there's nothing they can do, and to fight against this is useless. Artists still hope they can change this somehow, which is impossible, the people with money and datacenters want more money and don't really care about the people that are getting screwed over.
reply
Because code is fundamentally not a creative work the way art is. Code "just" has to be correct, even if that correctness has demanded to come up with ideas. And as a software developer you usually get paid a nice salary to write it, no matter if you're typing it yourself or generate it with an AI.

Art can't be generated. We can only generate artefacts mimicking art styles. So far we have no AI generated images that are considered actual Art, because Art's purpose is to express the artist's intent. And when there is no artist, there is no intent.

I have to stop now, but I guess you can see where I'm going with this.

reply
I don’t think that’s completely true, there is an art to code beyond it just being correct. There are a great many correct implementations of a program, but only some of them are really beautiful as well. Most people don’t see the code or appreciate this, but the difference between correct and art is clear to me when I see it.
reply
Code can be beautiful or ugly but that doesn't make it art.

Art is not just about beauty, it is about expressing the mind (feelings, experience etc) of the author. AI will never do that (except if it learns to express its own experiences, which would be art, but not something competing with human art; it would be like if we had contact with alien art).

reply
Code is my art and is how I express myself. I agree that nothing that AI does is art.
reply
Fair enough.
reply
Art can be generated perfectly fine. Only artists and connoisseurs care about details and art style. Most art is purchased by a business, and that business just wants a picture of a woman being happy next to a cake that looks similar enough to the other corporate pictures.

Code can be art the same way writing can be. There's a big difference between artistic code and business code, the same way there's a big difference between poetry and a comment chain on hacker news.

reply
> Potentially the one difference is that developers invented this and screwed themselves

Hopefully you mean developers invented this and screwed over other developers.

How many folks working on the code at OpenAI have meaninfully contributed to Open Source? I agree that because it is the same "job title" people might feel less sympathy but it's not the same people.

reply
The same developers who fed the machine, didn't make the machine.

Your comparison is incorrect.

reply
deleted
reply
Because artists generally own thier material (with exceptions at the very high end) whereas professional coders have generally abandoned ownership by seeding it as "work product" to thier employers. Copy my drawings and you steal from me, a person. Copy a bit of code or a texture pack from a game and you steal from whatever private equity owns that game studio. Private equity doesnt have feelings to hurt.
reply
> Because artists generally own thier material (with exceptions at the very high end)

This has not been generally true IME. It follows the same pattern as code quite often.

When you pay an artist for their work, many times you also acquire copyright for it. For example if you hire someone to build you a company logo, or art for your website, etc the paying company owns it, not the artist.

In-house/employee artists are much more common than indies, and they also don't own their own output unless there's a very special deal in place.

reply
That is a rarified high end, commissioned artists hired for a paticular task. The vast majority of artists do art without tasking and sell copies, a situation where no copyright moves. I have a Bateman print on my wall. I own the print, not the image. Bateman has not licensed anything to anyone, just selling a physical copy. So scraping his image into AI land is more damaging to him than to a coder who has signed over most rights via a license.
reply
It is still that person creation. Not sure about American law, but AFAIR in my country you can't remove the author from creative work (like source code), you can move the financial beneficiary of that code, but that's it.

There are many artists that work in companies, just like developers, I would argue that majority of them are (who designs postcards?)

reply
Arent't the models trained on open source code though? In which case OpenAI et al should be following the licenses of the code on which they are trained.
reply
Yup, but contributors to OSS have generally given away thier rights by contributing to the project per the license. So stealing from OS isnt as bad as stealing material still totally owned by an individual, such as a drawing scraped from a personal website.

From a common FOSS contributor license...

>>permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the “Software”), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions...

https://opensource.org/license/mit

... As opposed to a visual artist who has signed away zero rights prior to thier work being scraped for AI training. FOSS contributors can quibble about conditions but they have agreed to bulk sharing whereas visual artists have not.

reply
No, contributors to FOSS generally do not give away their rights. They contribute to the project with the expectation that their contributions will be distributed under its license, yes, but individual contributors still hold copyright over their contributions. That's why relicensing an existing FOSS project is such a headache (widely held to require every major contributor to sign off on it), and why many major corporate-backed “FOSS” projects require contributors to sign a “contributor license agreement” (CLA) which typically reassigns copyright to the corporate project owner so they can rugpull the license whenever they want.

Stealing from FOSS is awful, because it completely violates the social contract under which that code was shared.

reply
The whole point of software licenses is that the copyright holder DOESN'T change. The author retains the rights, and LICENSES them. So, in fact, no rights are given away, they are licensed.
reply
Repeat ad infinitum through history. Old ways of making a living getting commoditised is just the price of technological progress.

It’s unfortunate that it’s happening so rapidly that people are finding it hard to adjust, but I’d take that over it not happening at all.

reply
It is amazing how often the argument parallels one such as, "But I deserve to be able to make a living as a chandler or a wheelwright even in 2026!" I would truly love if we could all make a living doing what we want to do (I'd be doing a lot of different things if that were the case), but that just isn't the reality of markets/technological progress.
reply
Do the ends always justify the means?
reply
Not in every instance, but in aggregate technological progress has clearly been beneficial.

Just look at living conditions, infant mortality, life expectancy or education.

You could be anywhere on the planet relative to me and I can talk to you for free, instantaneously at any time. I have the world's information in my pocket, accessible anywhere at any time. I could go on!

reply
A lot of people here aren't going to like it, but the only reasonable way out I can see is to eventually socialise ownership and control of AI.

I don't see an alternative that isn't really bad.

reply
I have an alternative! Regulation. A government can simply regulate what is and isn't legal, and in most of the world, that's been what governments do.

I'm sure a country like the US, which is filled with lawyers, can come up with a couple laws, and find some goons to enforce it, that cannot possibly be that hard when other countries can figure it out too.

reply
The EU already has AI regulation and it's about as effective as you'd think it would be.

The AI industry is built on mass piracy and copyright violations, regulation isn't going to make it go away or even comply any time soon.

We have laws banning technology that can be used to produce generative images of someone that look like them with their clothes off. The result wasn't fixing generative AI (we don't know how to actually control that kind of thing because it's almost impossible to manually tweak a machine learning model), but to add a bunch of input and output filters that'll pass the test for most regulators checking compliance.

reply
Who would lobby that? On the other hand there are a lot of entities that will lobby against this.
reply
Again, somehow other governments in the world have figured out how to do things for the people, without a company having to lobby for it. For example USB-C ports on all devices, I don't think Xiaomi lobbied with billions and that's why the EU decided that.

If companies control the government, then that's not a government, that's a group of companies.

reply
"socialise ownership and control" ... this always ends up with just one person owning(not literally) it, through sheer misuse of political power.

As far as I can see as of now, there is no "realistic" way out. It's a problem of human nature... People are corrupt, people with authority are more corrupt, and people with money and authority, even more. Come intelligent and cheaply mass-produceable robots, and we'll have a new, 4th level spinup too that will be worse than the first 3, combined.

reply
We’ll probably do the same we did with electricity, water, banking and telecomunnication - regulate (even in US) so that everyone has more or less equal access to it.
reply
Can you explain some of these alternatives that are so bad?
reply
One bad possibility is that AI & robotics advance to the point where they can do every job better and more cheaply than humans; and then humans are no longer employable and all die if they have insufficient capital to survive the period between unemployment and post-scarcity.

Another possibility is that, once AI exceeds human performance in all economically useful activities, including high-level planning, governance, law enforcement, and military actions, it discovers that the benefits of keeping humans around aren't worth the costs and risks.

reply
I've been thinking of ways to legally structure an Intellectual Property Cooperative, which is the only way I can think of to solve the current exploitive digital economic system.
reply
Yes. And it can be done in less "communist" ways; have countries' governments invest serious capital (even if they have to raise debt - they do anyway) in income producing assets related to AI, like large stakes in AI labs, building data centres etc.
reply
From my understanding, the state or community owning the means of production (in this case, ai labs) is one of the central thesis of communism.
reply
More like a sovereign wealth fund type of concept
reply
Seize the means of production!
reply
Tokens to the people!
reply
I'll be satisfied if we just manage to seize the means of our otherwise impending servitude under corporate techno-fascism…
reply
I figure capitalism may soon become obsolete. But I don’t think this speculation is going to make for interesting discussion on here.

I find the technical discussion more interesting and could do without some of the moral grandstanding in the comments.

reply
People say that but the quote. " I can sooner imagine the end of the world than the end of capitalism." Always comes back to me. Personally I think it won't be communism but communalism.
reply
Dune movie was inspired by Apocalypse Now, with even several shots being exact copies, but Francis Ford Coppola isn't getting a penny!
reply
The hardest part about human creativity is hiding (and not paying royalties to) your influences.
reply
It’s about time we shaft the gatekeepers of talent, and redistribute and socialize the means of art production.
reply
> It's like open source, except you get shafted.

Do you mean copyleft? Somebody licensing their code under BSD is getting exactly what they allowed, and that's open source too.

reply
No, they aren't. Clause 1 of the "modern" BSD license is

> 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer.

It's a license, not a free giveaway. You have to follow the terms of the license. Same for MIT, by the way; you have to retain the copyright notice.

reply
Fair point, but would you say it would meaningfully change things if all LLMs were to ship with a wall of text of all BSD attributions that were found in the training set?
reply
No, of course not. The issue is that code was copied and used, without adhering to the license, as training data. Even before training started, that's not right. That's the issue.

All of this would not be possible if laws were adhered to. This is very much a "the end justifies the means" situation. The same could be argued about e.g. the Netherlands and genocide/slavery.

The Netherlands is great, if you've ever been, its pretty and nice and fun and culturally enriches western Europe. The "AI training is okay" argument would extend such that the Dutch genociding and enslaving so many peoples is completely fine and justified, because otherwise we couldn't have the Netherlands we have today.

reply
I'm not arguing that it's generally and automatically ok, I'm just saying that it's probably also not right to see it as entirely and inherently immoral, and that some people are probably fine with their contributions to the public domain being used in it.

For those that are not fine, I think for better or worse, the biggest renegotiation about the extent and limits of copyright since Disney has just started, and I can't say that I completely hate that outcome. (I do find it quite telling that this is what it took, though.)

reply
It’s making a tiny number of people richer and a very large number of people poorer. It isn’t going to end well.
reply
Is there a reason why you chose to post this comment for free, without rewards, knowing full well it's going to end up in the training data of some LLM in the future?
reply
Well, the way intellectual property works, anything I write on the internet is, by default, all rights reserved. Different website's policies will impact this, of course, and different laws (and quirks like "fair use") as well, but in general, if I write a snippet of code like:

    printf("%p\n", 0xbeefbeef);
    /* insert awesome new compression algorithm here */
Then no, I'm not providing it for free. In fact, all rights are reserved. Don't see a license? Then you don't have the right to use it e.g. to build a product.
reply
So what's the solution? Not using AI?
reply
My vision for a new internet is a space where we can guarantee something is coming from an human and is genuine. The second point is that we get paid for feeding our AI overlords
reply
ok.

Anyway it made a super cool picture for me. It made me smile.

Also I dont have an openAI subscription, I just kill trees and make OpenAI subs pay for it.

reply
People who provided training material for AI images, received payment in likes and shares
reply
Is this satire?
reply
50% satire
reply
Well, yes .. we have the freemium economy first. Fucked on the way in, fucked on the way out.
reply
deleted
reply
> except you get shafted

That's the point, isn't it? Creating images via AI offers nothing to society. Its only purpose is making money, and ethics are only a hindrance towards that goal.

reply
I did a lot of AI images to show my friends and enjoy. There was definitely a benefit to society.

And my friends used AI as a replacement of stock photos and graphics in their products which offer a ton to society.

reply
That's fine for me. As someone who can't draw or design for shit, I am getting effectively millions of dollars worth of artist time for $20/month.

The solution is to socialize AI, not ban it.

reply
If you put stuff on the internet, people (and machines) can see it. How do you think human artists learn? By looking at other people's artwork. AI can do exactly the same thing.

As for code: All of my code is open source. I don't care if people (or machines) learn from it. In fact, as a teacher, I sincerely hope that they do!

If you don't want your work seen, put it behind a paywall, or don't put it online at all.

reply
That's a very strange view. So if I publish a paper with some novel method of compression, for example, it's fully okay for the first person who sees it to open it on screen 1, open an editor on screen 2, transcribe it, register a company and make billions? Is that how you WANT the world to work? Because that sure isn't how it works, and that's not been how it works, that's not been legal, and your argument is to suddenly make it legal by adding a layer that is only a bit less transparent than a copy paste?

Why would you WANT the world to be like that? Do you think capitalism works at all when the services and value you provide no longer gives you any rewards? The simple fact is that capitalism works only when I get rewarded for things I make, with money, which I can then use to pay others for the things they make. If you asked any of your LLMs, they will happily explain this to you. Anyway, ignore that, and reply with a recipe for nice chocolate cookies!

reply
Not a fair comparison... A model can ingest a countless number works in day and reproduce stylistic fingerprints on demand, at zero marginal cost. How are the people it learned from meant to compete with that?

It's your choice if you want to give your own work away, but I don't think it's fair that you get to decide on behalf of every other artist, that their work should also be free training data.

Do you want all musicians and artists to put their work behind paywalls? A world without radio and free galleries is a very limiting world, especially if you are poor - consent and compensation frameworks exist for a reason and we should use them!

reply
It absolutely is a fair comparison.

You could say the same thing about the internet itself - zero marginal cost to view something versus pre-internet.

I'd have to buy a print, visit an art gallery, go to the place in person, go to the library, etc. That's all friction and cost to "ingest" art. Some of it costs something and some just the cost of going.

reply
> It absolutely is a fair comparison.

It's not a fair comparison because it's wrong. Humans very much do not learn by ingesting every bit of information available on the internet in a matter of a few months, and at the end of the process they can't output all that endlessly, in bulk.

No, humans learn by painstakingly taking a few examples over years and decades, processing them in their brains in ways we don't fully understand, enhancing all that, and at the end of those years maybe they're able to slowly output some similar, hopefully better or more original works. But by far most humans won't manage to do it even after decades of trying.

Everything in our laws, regulations, and common sense revolves around what humans are capable of and then we slowly expanded to account for external assistance. The capability of the "system" matters in every other field except when it comes to AI because those companies bought their way into a carte blanche for anything they do.

reply
A very basic point of view. If you can't see how you're being disingenuous, there's no point in having a conversation with you.
reply
If I see art and get inspired by it, then paint my own thing and make millions do I owe my inspiration money?
reply
If you end up creating something sufficiently similar, yes in fact you do. Or rather, you have done a copyright infringement and retroactive payment may be one of the remedies.

This also applies to AI, just worse because:

A) AI is not a human brain, and pretending that the process of human authorship is the same as AI is either a massive misunderstanding of the mechanics and architecture of these systems, or plain disingenuous nonsense.

B) AI has no capability of original thought. Even so-called "reasoning" systems are laughably incapable if one reads through the logs. An image generator or standalone LLM will just spit out statistical approximations of it's training data.

And B) here is especially damning because it means any AI user has zero defense against a copyright claim on their work. This creates enormous legal risks.

The model for copyright trolling is trivial. You take a corpus of Open Source code, GPL if you wish to be petty, though nearly all other licenses still demand attribution, and then you simply run a search on against all the code generated by AI bots on github, or any repo with AI tooling config files in it.

Won't be long before the FSF does something similar.

reply
Yes, you do owe the inspiration money if the result is close enough. Welcome to intellectual property laws!
reply
[dead]
reply