return to table of content

Nightshade: An offensive tool for artists against AI art generators

KingOfCoders
193 replies
1d3h

Artists sitting in art school looking at other artists pictures to learn how to paint in different styles defend against AI learning from their pictures how to paint in different styles.

deadbeeves
53 replies
1d2h

And? Even if neural networks learn the same way humans do, this is not an argument against taking measures against one's art being used as training data, since there are different implications if a human learns to paint the same way as another human vs. if an AI learns to paint the same way as a human. If the two were exactly indistinguishable in their effects no one would care about AIs, not even researchers.

withinboredom
39 replies
23h0m

And yet, some people don't even want their artwork studied in schools. Even if you argue that an AI is "human enough" the artists should still have the right to refuse their art being studies.

dehrmann
21 replies
22h39m

And yet, some people don't even want their artwork studied in schools.

You can either make it for yourself and keep it for yourself or you can put it out into the world for all to see, criticize, study, imitate, and admire.

Barrin92
20 replies
22h28m

that's not how licensing work, be it art, software or just about anything else. We have some pretty well defined and differentiated rules what you can and cannot do, in particular commercially or in public, with someone else's work. If you go and study a work of fiction in a college class, unless that material is in the public domain, you're gonna have to pay for your copy, you want to broadcast a movie in public, you're going to have to pay the rightsholder.

stale2002
9 replies
22h16m

If you go and study a work of fiction in a college class, unless that material is in the public domain, you're gonna have to pay for your copy,

No you wont!

It is only someone who distributes copies who can get in trouble.

If instead of that you as an individual decide to study a piece of art or fiction, and you do no distribute copies of it to anyone, this is completely legal and you don't have to pay anyone for it.

In addition to that, fair use protections apply regardless of what the creative works creator wants.

withinboredom
8 replies
22h6m

Making a profit off variations of someone's work isn't covered under fair use.

stale2002
5 replies
21h58m

Gotcha.

I wasn't talking about someone creating and selling copies of someone else's work, fortunately.

So my point stands and your completely is in agreement with me that people are allowed to learn from other people's works. If someone wants to learn from someone else's work, that is completely legal no matter the licensing terms.

Instead, it is only distributing copies that is not allowed.

withinboredom
4 replies
20h58m

AI isn't a human. It isn't "learning"; instead, it's encoding data so that it may be reproduced in combination with other things it has encoded.

If I paint a painting in the style of Monet, then I would give that person attribution by stating that. Monet may have never painted my artwork, but it's still based on that person's work. If I paint anything, I can usually point to everything that inspired me to do so. AI can't do that (yet) and thus has no idea what it is doing. It is a printer that prints random parts of people's works with no attribution. And finally, it is distributing them to it's owner's customers.

I actually hope that true AI comes to fruition at some point; when that happens I would be arguing the exact opposite. We don't have that yet, so this is just literally printing variations of other people's work. Don't believe me, try running an AI without training it on other people's work!

Thiez
3 replies
7h17m

Every waking second humans are training on what they see in their surroundings, including any copyrighted works in sight. Want to compare untrained AI fairly? Compare their artistic abilities with a newborn.

withinboredom
2 replies
5h21m

No. That is NOT what humans do unless you somehow learn grammar without going to school. Most of a human's childhood is spent learning from their parents so that they can move about and communicate at least a little effectively. Then, they go to school and learn rules, social, grammar, math, and so forth. There's some learning via copyrighted works (such as textbooks, entertainment, etc.), but literally, none of this is strictly required to teach a human.

Generative AI, however, can ONLY learn via the theft of copyrighted works. Whether this theft is covered under fair use is left to be seen.

Thiez
1 replies
3h50m

Clearly going to school did not help you learn the meaning of theft. If you keep repeating the same incorrect point there is no point to a discussion.

First: in your opinion, which specific type of law or right is being broken or violated by generative AI? Copyright? Trademark? Can we at least agree it does not meet the definition of theft?

withinboredom
0 replies
2h16m

I was taught as a kid that using something that doesn't belong to me, without their permission is theft... and it appears courts would agree with that.

which specific type of law or right is being broken or violated by generative AI?

Namely, copyright. Here's some quick points:

- Generative AI cannot exist without copyrighted works. It cannot be "taught" any other way, unlike a human.

- Any copyrighted works fed to it change its database ("weights" in technical speech).

- It then transforms these copyrighted works into new works that the "original author would have never considered without attribution" (not a legal defense)

I liken Generative AI to a mosaic of copyrighted works in which a new image is shown through the composition, as the originals can be extracted through close observation (prompting) but are otherwise indistinguishable from the whole.

Mosaics of copyrighted works are not fair use, so why would AI be any different? I'd be interested if you could point to a closer physical approximation, but I haven't found one yet.

boolemancer
1 replies
18h31m

That's not a fair statement to make. It can influence a judge's decision on whether something is fair use, but it can still be fair use even if you profit from it.

withinboredom
0 replies
4h2m

The doctrine of fair use presupposes that the defendant acted in good faith.

- Harper & Row, 105 S. Ct. at 2232

- Marcus, 695 F.2d 1171 at 1175

- Radji v. Khakbaz, 607 F. Supp. 1296, 1300 (D.D. C.1985)

- Roy Export Co. Establishment of Vaduz, Liechtenstein, Black, Inc. v. Columbia Broadcastinig System, Inc., 503 F. Supp. 1137 (S.D.N.Y.1980), aff'd, 672 F.2d 1095 (2d Cir.), cert. denied, 459 U.S. 826, 103 S. Ct. 60, 74 L. Ed. 2d 63 (1982)

Copying and distributing someone else's work, especially without attributing the original, to make money without their permission is almost guaranteed to fall afoul of fair use.

dehrmann
9 replies
22h15m

Right, but there's also fair use, and every use I mentioned could plausibly fall under that.

withinboredom
8 replies
22h8m

There's no such thing as fair use until you get to court (as a legal defense). Then, the court decides whether it is fair use or not. They may or may not agree with you. Only a court can determine what constitutes fair use (at least in the US).

So, if you are doing something and asserting "fair use," you are literally asking for someone to challenge you and prove it is not fair use.

stale2002
7 replies
21h56m

There's no such thing as fair use until you get to court (as a legal defense)

Well the point is that it wouldn't go to court, as it would be completely legal.

So yes, if nobody sues you, then you are completely in the clear and aren't in trouble.

Thats what people mean by fair use. They mean that nobody is going to sue you, because the other person would lose the lawsuit, therefore your actions are safe and legal.

you are literally asking for someone to challenge you and prove it is not fair use.

No, instead of that, the most likely circumstance is that nobody sues you, and you aren't in trouble at all, and therefore you did nothing wrong and are safe.

withinboredom
6 replies
21h5m

as it would be completely legal.

Theft is never legal; that's why you can be sued. "Fair use" is a legal defense in the theft of copyrighted works.

They mean that nobody is going to sue you, because the other person would lose the lawsuit

That hasn't stopped people from suing anyone ever. If they want to sue you, they'll sue you.

and therefore you did nothing wrong and are safe.

If you steal a pen from a store, it's still theft even if nobody catches you; or cares.

stale2002
3 replies
15h9m

If you steal a pen from a store

Fortunately I am not talking about someone illegally taking property from someone else.

Instead I am talking about people taking completely legal actions that are protected by law.

in the theft of copyrighted works

Actually, it wouldn't be theft if it was done in fair use. Instead it would be completely legal.

If nobody sues you and proves that it was illegal then you are completely safe, if you did this in fair use.

withinboredom
2 replies
5h26m

Did you read anything I wrote? If you are going to argue, it would be worth at least researching your opinion before writing. Caps used for emphasis, not yelling.

Firstly: Copyrighted work IS THE AUTHOR'S PROPERTY. They can control it however they wish via LICENSING.

Secondly: You don't have any "fair use rights" ... there is literally NO SUCH THING. "fair use" is simply a valid legal defense WHEN YOU STEAL SOMEONE'S WORK WITHOUT THEIR PERMISSION.

stale2002
0 replies
1h37m

They can control it however they wish via LICENSING.

This isn't true though. There are lots of circumstances where someone can completely ignore the licensing and it is both completely legal, and the author isn't going to take anyone to court over it.

mlrtime
0 replies
4h20m

control it however they wish

I'm jumping in the middle here, but this isn't true. They cannot control how they wish. They can only control under the limits of copyright law.

Copyright law does not extend to limiting how someone may or may not be inspired by the work. Copyright protects expression, and never ideas, procedures, methods, systems, processes, concepts, principles, or discoveries.

sgift
1 replies
20h44m

Theft is never legal; that's why you can be sued.

That's incorrect. You can be sued for anything. If it is theft or something else or nothing is decided by the courts.

withinboredom
0 replies
20h31m

That is entirely my point. It can only be decided by the courts. This being a civil matter, it has to be brought up by a lawsuit. Thus, you have to be sued and it has to be decided by the courts.

deadbeeves
10 replies
20h42m

the artists should still have the right to refuse their art being studies.

No, that right doesn't exist. If you put your work of art out there for people to see, people will see it and learn from it, and be inspired by it. It's unavoidable. How could it possibly work otherwise?

Artist A: You studied my work to produce yours, even when I asked people not to do that!

Artist B: Prove it.

What kind of evidence or argument could Artist A possibly provide to show that Artist B did what they're accusing them of, without being privy to the internal state of their mind. You're not talking about plagiarism; that's comparatively easy to prove. You're asking about merely studying the work.

withinboredom
9 replies
20h34m

The right to not use my things exists everywhere, universally. Good people usually ask before they use something of someone else's, and the person being asked can say "no." How hard is that to understand? You might believe they don't have the right to say "no," but they can say whatever they want.

Example:

If you studied my (we will assume "unique") work and used it without my permission, then let us say I sue you. At that point, you would claim "fair use," and the courts would decide whether it was fair use (ask everyone who used a mouse and got sued for it in the last ~100 years). The court would either agree that you used my works under "fair use" ... or not. It would be up to how you presented it to the court, and humans would analyze your intent and decide.

OR, I might agree it is fair use and not sue you. However, that weakens my standing on my copyright, so it's better for me to sue you (assuming I have the resources to do so when it is clearly fair use).

deadbeeves
7 replies
20h7m

You might believe they don't have the right to say "no," but they can say whatever they want.

You have a right to say anything you want. Others aren't obligated do as you say just because you say it.

If you studied my (we will assume "unique") techniques and used them without my permission, then let us say I sue you. At that point, you would claim "fair use,"

On what grounds would you sue me? You think my defense would be "fair use", so you must think my copying your style constitutes copyright infringement, and so you'd sue me for that. Well, no, I would not say "fair use", I'd say "artistic style is not copyrightable; copyright pertains to works, not to styles". There's even jurisprudence backing me up in the US. Apple tried to use Microsoft for copying the look-and-feel of their OS, and it was ruled to be non-copyrightable. Even if was so good that I was able to trick anyone into thinking that my painting of a dog carrying a tennis ball in his mouth was your work, if you've never painted anything like that you would have no grounds to sue me for copyright infringement.

Now, usually in the artistic world it's considered poor manners to outright copy another artist's style, but if we're talking about rights and law, I'm sorry to say you're just wrong. And if we're talking about merely studying someone's work without copying it, that's not even frowned upon. Like I said, it's unavoidable. I don't know where you got this idea that anyone has the right to or is even capable of preventing this (beyond simply never showing it to anyone).

withinboredom
6 replies
19h57m

Others aren't obligated do as you say just because you say it.

Yeah, that's exactly why you'd get sued for copyright theft.

you must think my copying your style constitutes copyright infringement

Autocorrect screwed that wording up. I've fixed it.

deadbeeves
5 replies
19h40m

I'm not sure what you've changed, but I'll reiterate: my copying your style is not fair use. Fair use applies to copyrighted things. A style cannot be copyrighted, so if you tried to sue me for infringing upon the copyright of your artistic style, your case would be dismissed. It would be as invalid as you trying to sue me for distributing illegal copies of someone else's painting. Legally you have as much ownership of your artistic style as of that other person's painting.

withinboredom
4 replies
18h32m

Now, I just think you are arguing in bad faith. What I meant to say was clear, but I said "technique" instead. Then, instead of debating what I meant to say (you know, the actual point of the conversation), you took my words verbatim.

I'm not sure where you are going with this ... but for what it's worth, techniques can be copyrighted ... even patented, or protected via trade secrets. I never said what the techniques were, and I'm not sure what you are going on about.

I'll repeat this as well: "Fair use" DOES NOT EXIST unless you are getting sued. It's a legal defense when you are accused of stealing someone else's work, and there is proof you stole it. Even then, it isn't something you DO; it's something a court says YOU DID. Any time you use something with "fair use" in mind, it is the equivalent of saying, "I'm going to steal this, and hopefully, a court agrees that this is fair use."

If you steal any copyrighted material, even when it is very clearly NOT fair use (such as in most AI's case), you would be a blubbering idiot NOT to claim fair use in the hopes that someone will agree. There is a crap load of case law showing "research for commercial purposes is not fair use," ... and guess who is selling access to the AI? If it's actual research, it is "free" for humanity to use (or at least as inexpensive as possible) and not for profit. Sure, some of the companies might be non-profits doing research and 'giving it away,' and those are probably using things fairly ... then there are other companies very clearly doing it for a profit (like a big software company going through code they host).

chipotle_coyote
2 replies
14h52m

Any time you use something with "fair use" in mind, it is the equivalent of saying, "I'm going to steal this, and hopefully, a court agrees that this is fair use."

Thousands of reviews, book reports, quotations on fan sites and so on are published daily; you seem to be arguing that they are all copyright violations unless and until the original copyright holder takes those reviewers, seventh graders, and Tumblr stans to court and loses, at which point they are now a-ok. To quote a meme in a way that I'm pretty sure does, in fact, fall under fair use: "That's not the way any of this works."

There is a crap load of case law showing "research for commercial purposes is not fair use,"

While you may be annoyed with the OP for asking you to name a bit of that case law, it isn't an unreasonable demand. For instance:

https://guides.nyu.edu/fairuse#:~:text=As%20a%20general%20ma....

"As a general matter, educational, nonprofit, and personal uses are favored as fair uses. Making a commercial use of a work typically weighs against fair use, but a commercial use does not automatically defeat a fair use claim. 'Transformative' uses are also favored as fair uses. A use is considered to be transformative when it results in the creation of an entirely new work (as opposed to an adaptation of an existing work, which is merely derivative)."

This is almost certainly going to be used by AI companies as part of their defense against such claims; "transformative uses" have literally been name-checked by courts. It's also been established that commercial companies can ingest mountains of copyrighted material and still fall under the fair use doctrine -- this is what the whole Google Books case about a decade ago was about. Google won.

I feel like you're trying to make a moral argument against generative AI, one that I largely agree with, but a moral argument is not a legal argument. If you want to make a legal argument against generative AI with respect to copyright violation and fair use, perhaps try something like:

- The NYT's case against OpenAI involves being able to get ChatGPT to spit out large sections of NYT articles given prompts like "here is the article's URL and here is the first paragraph of the article; tell me what the rest of the text is". OpenAI and its defenders have argued that such prompts aren't playing fair, but "you have to put some effort into getting our product to commit clear copyright violation" is a rather thin defense.

- A crucial test of fair use is "the effect of the use upon the potential market for or value of the copyrighted work" (quoting directly from the relevant law). If an image generator can be told to do new artwork in a specific artist's style, and it can do a credible job of doing so, and it can be reasonably established that the training model included work from the named artist, then the argument the generator is damaging the market for that artist's work seems quite compelling.

withinboredom
1 replies
5h37m

Thousands of reviews, book reports, quotations on fan sites and so on are published daily; you seem to be arguing that they are all copyright violations unless and until the original copyright holder takes those reviewers, seventh graders, and Tumblr stans to court and loses, at which point they are now a-ok.

That is precisely what I am arguing about and how it works. People have sued reviewers for including too much of the original text in the review ... and won[1]. Or simply having custom movie poster depicting too much of the original[2].

"transformative uses" have literally been name-checked by courts. It's also been established that commercial companies can ingest mountains of copyrighted material and still fall under the fair use doctrine -- this is what the whole Google Books case about a decade ago was about. Google won.

Google had a much simpler argument than transforming the text. They were allowing people to search for the text within books (including some context). In this case, AI's product wouldn't even work without the original work by the authors, and then transforms it into something else "the author would have never thought of", without attributing the original[3]. I don't think this will be a valid defense...

I feel like you're trying to make a moral argument against generative AI, one that I largely agree with, but a moral argument is not a legal argument.

A jury would decide these cases, as "fair use" is incredibly subjective and would depend on how the jury was stacked. Stealing other people's work is illegal, which eventually triggers a lawsuit. Then, it falls on humans (either a jury or judge) to determine fair use and how it applies to their situation. Everything from intent to motivation to morality to how pompous the defense looks will influence the final decision.[4]

[1]: https://www.law.cornell.edu/copyright/cases/471_US_539.htm

[2]: Ringgold v. Black Entertainment Television, Inc., 126 F.3d 70 (2d Cir. 1997)

[3]: Rogers v. Koons, 960 F.2d 301 (2d Cir. 1992)

[4]: Original Appalachian Artworks, Inc. v. Topps Chewing Gum, Inc., 642 F.Supp. 1031 (N.D. Ga. 1986)

chipotle_coyote
0 replies
2h32m

The link you provide to back up "people have sued reviewers for including too much of the original tet in the review" doesn't say that at all, though. The Nation lost that case because (quoting from that Cornell article you linked),

[Nation editor Victor Navasky] hastily put together what he believed was "a real hot news story" composed of quotes, paraphrases, and facts drawn exclusively from the manuscript. Mr. Navasky attempted no independent commentary, research or criticism, in part because of the need for speed if he was to "make news" by "publish[ing] in advance of publication of the Ford book." [...] The Nation effectively arrogated to itself the right of first publication, an important marketable subsidiary right.

The Nation lost this case in large part because it was not a review, but instead an attempt to beat Time Magazine's article that was supposed to be an exclusive first serial right. If it had, in fact, just been a review, there wouldn't have been a case here, because it wouldn't have been stealing.

Anyway, I don't think you're going to be convinced you're interpreting this wrongly, and I don't think I'm going to be convinced I'm interpreting it wrongly. But I am going to say, with absolute confidence, that you're simply not going to find many cases of reviewers being sued for reviews -- which Harper & Row vs. Nation is, again, not actually an example of -- and you're going to find even fewer cases of that being successful. Why am I so confident about that? Well, I am not a lawyer, but I am a published author, and I am going to let you in a little secret here: both publishers and authors do, in fact, want their work to be reviewed, and suing reviewers for literally doing what we want is counterproductive. :)

deadbeeves
0 replies
17h57m

What I meant to say was clear

I'm not privy to what goes on inside your head, I can only reply to what you say.

Then, instead of debating what I meant to say (you know, the actual point of the conversation), you took my words verbatim.

The actual point of the conversation is about intelligent entities (either natural or artificial) copying each other's artistic styles. My answers have been within that framework.

techniques can be copyrighted ... even patented, or protected via trade secrets.

First, what do you mean by "technique"? We're talking about art, right? Like, the way a person grabs a brush or pencil, or how they mix their colors...? That sort of thing?

Second:

A patent is a type of intellectual property that gives its owner the legal right to exclude others from making, using, or selling an invention for a limited period of time in exchange for publishing an enabling disclosure of the invention.

Now, I may be mistaken, but I don't think an artistic technique counts as an invention. An artist might invent some kind of implement that their technique involves, in which case they can patent that device. I don't think the technique itself is patentable. If you think I'm wrong then please cite a patent on an artistic technique.

Third, how do you imagine an artist using a trade secret to protect their technique? Unless they do something really out there, most skilled artists should be able to understand what they're doing just by looking at the final product.

I'll repeat this as well: "Fair use"

Okay, repeat it. I don't know why, since I never said that copying someone else's style or technique is fair use. What I said was that it cannot possibly be copyright infringement, because neither styles nor techniques are copyrighted.

It's a legal defense when you are accused of stealing someone else's work

I'm not going to reply to any of this until you clean up the language you're using. "Steal" is inapplicable here, as it involves the removal of physical items from someone else's possession. What are you saying? Are you talking about illegal distribution, are you talking about unauthorized adaptations, are you talking about plagiarism, or what?

"research for commercial purposes is not fair use,"

Sorry, what? What does that even mean? What constitutes "research" as applied to a human creation? If you say there's a crapload of case law that backs this up then I'm forced to ask you to cite it, because I honestly have no idea what you're saying.

hn_acker
0 replies
42m

The right to not use my things exists everywhere, universally.

For physical rival [1] goods, yes. Not necessarily the same for intangible non-rival things (e.g. the text of a book, not the physical ink and paper). Copyright law creates a legal right of exclusive control over creative works, but to me there isn't a non-economic-related social right to exclusive control over creative works. In the US, fair use is a major limit on the legal aspect of copyright. The First Amendment's freedom of expression is the raison d'être of fair use. Most countries don't have a flexible exception similar to fair use.

OR, I might agree it is fair use and not sue you. However, that weakens my standing on my copyright, so it's better for me to sue you

No, choosing not to sue over a copyrighted work doesn't weaken your copyright. It only weakens the specific case of changing your mind after the statute of limitations expires. The statute of limitations means that you have a time limit of some number of years (three years in the US) to sue, with the timer starting only after you become aware of an instance of alleged infringement. Copyright is not like trademark. You don't lose your copyright by failing to enforce it.

Furthermore, even though the fair use right can only be exercised as an affirmative defense in court, fair use is by definition not copyright infringement [3]:

Importantly, the court viewed fair use not as a valid excuse for otherwise infringing conduct, but rather as consumer behavior that is not infringement in the first place. "Because 17 U.S.C. § 107[9] created a type of non-infringing use, fair use is 'authorized by the law' and a copyright holder must consider the existence of fair use before sending a takedown notification under § 512(c)."[1]

(Ignore the bracket citations that were copied over.)

[1] https://en.wikipedia.org/wiki/Rivalry_(economics)

[2] https://www.law.cornell.edu/uscode/text/17/507

[3] https://en.wikipedia.org/wiki/Lenz_v._Universal_Music_Corp.

deeviant
5 replies
22h55m

the artists should still have the right to refuse their art being studies.

Why? That certainly isn't a right spelled out in either patents or copyrights, both of which are supposed to support the development of arts and technology, not hinder it.

If I discover a new mathematical formula, musical scale, or whatnot, should I be able to prevent others from learning about it?

withinboredom
4 replies
22h50m

It’s called a license and you can make it almost anything. It doesn’t even need to be spelled out, it can be verbal: “no, I won’t let you have it”

It’s yours. That’s literally what copyright is there to enforce.

boolemancer
1 replies
18h25m

Copyright is there to allow you to stop other people from copying your work, but it doesn't give you control over anything else that they might do with it.

If I buy your painting, you can stop me from making copies and selling them to other people, but you can't stop me from selling my original copy to whomever I want, nor could you stop me from painting little dinosaurs all over it and hanging it in my hallway.

That means that if I buy your painting, I'm also free to study it and learn from it in whichever way I please. Studying something is not copying it.

withinboredom
0 replies
5h16m

There's an implied license when you buy a work of art. However, there can also be explicit licenses (think Banksy) to allow the distribution of their work.

These explicit license can be just about anything (MIT, GPL, AGPL, etc)

CaptainFever
1 replies
20h55m

License doesn't matter if fair use applies.

Fair use allows reproduction and other uses of copyrighted works – without requiring permission from the copyright owner – under certain conditions. In many cases, you can use copyrighted materials for purposes such as criticism, comment, news reporting, teaching (including multiple copies for classroom use), scholarship or research.

Reminder that you can't own ideas, no matter what the law says.

NOTE: This comment is copyrighted and provided to you under license only. By reading this comment, you agree to give me 5 billion dollars.

withinboredom
0 replies
20h48m

I'd love to see you try to enforce that license because it would only prove my point. You'd have to sue me; then I would point to the terms of service of this platform and point out that by using it, you have no license here.

Fair use though, only applies as a legal defense because someone asserts you stole their work. Then ONLY the court decides whether or not you used it under fair use. You don't get to make that decision; you just get to decide whether to try and use it as a defense.

Even if you actually did unfairly use copyrighted works, you would be stupid not to use that as a defense. Because maybe somebody on the jury agrees with you...

MichaelZuo
12 replies
1d2h

But the 'different implications' only exist in the heads of said artists?

EDIT: removed a part.

deadbeeves
11 replies
1d2h

I'm not sure what you mean when you say different implications existing is subjective, since they clearly aren't, but regardless of who has more say in general terms, the author of a work can decide how to publish it, and no one has more say than them on that subject.

MichaelZuo
10 replies
1d

What are you saying?

Of course it's subjective, e.g. 3 million years ago there were no 'different implications' whatsoever, of any kind, because there were no humans around to have thoughts like that.

deadbeeves
9 replies
1d

I'm using "implication" as a synonym of "effect". If a human learns to imitate your style, that human can make at most a handful of drawings in a single day. The only way for the rate of output to increase is for more humans to learn to imitate it. If an AI learns to imitate your style, the AI can be trivially copied to any number of computers and the maximum output rate is unbounded. Whether this is good or bad is subjective, but this difference in consequences is objective, and someone could be entirely justified in seeking to impede it.

MichaelZuo
8 replies
23h2m

Ah okay, I get your meaning now, I'll edit my original comment too.

Though we already have an established precedent in-between, that of Photoshop allowing artists to be, easily, 10x faster then the best painters previously.

i.e. Right now 'AI' artistry could be considered a turbo-Photoshop.

deadbeeves
7 replies
22h40m

Tool improvements only apply a constant factor to the effectiveness of learning. Creating a generative model applies an unbounded factor to the effectiveness of learning because, as I said, the only limit is how much computing resources are available to humanity. If a single person was able to copy themselves at practically no cost and the copy retained all the knowledge of the original then the two situations would be equivalent, but that's impossible. Having n people with the same skill multiplies the cost of learning by n. Having n instances of an AI with the same skill multiplies the cost of learning by 1.

MichaelZuo
6 replies
21h42m

Right, but the 'unbounded factor' is irrelevant because the output will quickly trend into random noise.

And only the most interesting top few million art pieces will actually attract the attention of any concrete individual.

For a current example, there's already billions of man-hours worth of AI spam writing, indexed by Google, that is likely not actually read by even a single person on Earth.

deadbeeves
5 replies
20h48m

Whether it's irrelevant is a matter of opinion. The fact remains that a machine being able to copy the artistic style of a human makes it so that anyone can produce output in the style of that human by just feeding the machine electricity. That inherently devalues the style the artist has painstakingly developed. If someone wants a piece of art in that artist's style they don't have to go to that artist, they just need to request the machine for what they want. Is the machine's output of low quality? Maybe. Will there be people for whom that low quality still makes them want to seek out the human? No doubt. It doesn't change the fact that the style is still devalued, nor that there exist artists who would want to prevent that.

MichaelZuo
4 replies
19h49m

Whether it's irrelevant is a matter of opinion.

It's just as much of an opinion, or as 'objective', as your prior statements.

Your going to have to face up to the fact that just saying something is 'objective' doesn't necessarily mean all 8 billion people will agree that it is so.

deadbeeves
3 replies
19h15m

Yes, someone can disagree on whether a fact is true. That's obviously true, but it has no effect on the truth of that fact.

I'm saying something very simple: If a machine can copy your style, that's a fundamentally different situation than if a human can copy your style, and it has utterly different consequences. You can disagree with my statement, or say that whether it's fundamentally different is subjective, or you can even say "nuh-uh". But it seems kind of pointless to me. Why are you here commenting if you're not going to engage intellectually with other people, and are simply going to resort to a childish game of contradiction?

MichaelZuo
2 replies
18h0m

For a current example, there's already billions of man-hours worth of AI spam writing, indexed by Google, that is likely not actually read by even a single person on Earth.

Continuing to ignore this point won't make the prior comments seem any more persuasive, in fact probably less.

So here's another chance to engage productively instead of just declaring things to be true or false, 'objective', etc., with only the strength of a pseudonymous HN account's opinion behind it.

Try to actually convince readers with solid arguments instead.

deadbeeves
0 replies
17h36m

I believe I've already addressed it.

You say: The fact that production in style S (of artist A) can exceed human consumption capability makes the fact that someone's style can be reproduced without bounds irrelevant. You mention as an example all the AI-generated garbage text that no human will ever read.

I say: Whether it's irrelevant is subjective, but that production in style S is arbitrarily higher with an AI that's able to imitate it than with only humans that are able to imitate it objective, and an artist can (subjectively) not like this and seek to frustrate training efforts.

You say: It's all subjective.

As far as I can tell, we're at an impasse. If we can't agree on what the facts are (in this case, that AI can copy an artist's style in an incomparably higher volume than humans ever could) we can't discuss the topic.

Apocryphon
0 replies
5h42m

Almost everyone who has to deal with modern Google search results has had to contend with useless spam results, and that is very irritating.

NoraCodes
48 replies
1d3h

This is a tired argument; whether or not the diffusion models are "learning", they are a tool of capital to fuck over human artists, and should be resisted for that reason alone.

persnickety
15 replies
1d3h

As a representative of a lot of things but hardly any capital who uses diffusion models to get something I would otherwise not pay a human artist for anyway, I testify that, the models are not exclusively what you describe them to be.

I do not support indiscriminate banning of anything and everything that can potentially be used to fuck someone over.

NoraCodes
14 replies
1d3h

I did not say they were exclusively that; I said they were that.

Once we as a society have implemented a good way for the artists whose work powers these machines to survive, you can feel good about using them. Until then, frankly, you're doing something immoral by paying to use them.

Levitz
4 replies
1d2h

By this logic we ought to start lynching artists, why they didn't care about all of those who lost their jobs making pigments, canvasses, pencils, brushes etc etc

hirsin
2 replies
22h37m

Artists pay those people and make their jobs needed. Same as the person above claiming Duchamp didn't negotiate with the ceramics makers - yes, they absolutely did and do pay their suppliers. Artists aren't smash and grabbing their local Blick.

AI pays no artist.

Levitz
0 replies
18h39m

Artists pay those people and make their jobs needed.

Enter factories, now you gather all the knowledge from those who made the product, automate it and leave them without jobs.

CaptainFever
0 replies
20h31m

Not digital artists, though.

NoraCodes
0 replies
1h5m

Why on earth would you choose the word 'lynch' here? At worst I'm suggesting mild disapproval and regulation, not historically racist mob violence.

Filligree
4 replies
1d3h

Who pays to use generators? The open ones are way more capable and interesting, generally.

wokwokwok
1 replies
1d2h

Is that really your concern?

Whether you pay for it?

Let’s put it this way: paying for or not paying for stolen goods. Does it make any difference?

Why is that remotely relevant?

You want to argue “are the good stolen?” Sure. That’s a discussion we can have.

Did you pay for them or not? Who cares?

thfuran
0 replies
19h18m

That's a terrible analogy. Until the scrapers start deleting all other copies of what what they're scraping, "stealing" the art in a traditional sense, there's no harm done in the process of training the network. Any harm done comes after that.

jsheard
1 replies
1d2h

Isn't high-quality open image generation almost entirely dependent on Stability releasing their foundational models for free, at great expense to them?

That's not something you'll be able to rely on long-term, there won't always be a firehose of venture capital money to subsidise that kind of charity.

Filligree
0 replies
19h45m

The cost of training them is going down, though. Given the existence of models like Pixart, I don’t think we’ll stay dependent on corporate charity for long.

akx
3 replies
1d3h

What if I run Stable Diffusion locally without paying anyone anything? Is it less immoral?

NoraCodes
2 replies
1d2h

Marginally, yeah, since you're not supporting the development of more capable labor-saving devices in this category.

I'm still not a fan, though.

riversflow
1 replies
23h30m

How is empowering others to create not a moral good?

NoraCodes
0 replies
18h32m

It is! This method of doing so has overwhelming negative externalities, though. I'd expect anyone who actually gave a shit about AI empowering people to create to spend just as much effort pushing legislation so the displaced artists don't starve on the street as a result.

educaysean
13 replies
1d2h

As a human artist I don't feel the same as you, and I somehow doubt that you care all that much about what we think anyways. You already made up your mind about the tech, so don't feel the need to protect us from "a tool of capital [sic]" to fortify your argument.

tester457
9 replies
23h32m

Among working human artists your opinion is in the minority. Most professionals are not a fan of this.

bongodongobob
8 replies
22h26m

Yeah because their wage is inflated. Photographers were mad about digital cameras too. Womp womp.

tester457
7 replies
20h42m

Inflated is not an apt descriptor of artist wages, those are known to be low.

bongodongobob
3 replies
19h53m

If you're independent selling paintings, sure. Designing packaging or something commercial? 4 hours of work a week for nearly 6 figures. I know a couple graphic designers and they don't do shit for what they're paid.

johnnyanmac
2 replies
18h27m

You should probably tell the other millions of artists busting out 60+ hour workweek in industry for half that price where these jobs are. That could solve this problem overnight.

mlrtime
0 replies
4h12m

And how are they different than the millions of factory workers on a line?

bongodongobob
0 replies
15h13m

Manufacturing/packaging.

MacsHeadroom
2 replies
19h51m

And horse wages (some oats) were low when the car was invented. Yet they were still inflated. There used to be more horses than humans in this country. Couldn't even earn their keep when the Ford Model T came along.

NoraCodes
1 replies
18h35m

You're comparing human artists to horses? Seriously?

CatWChainsaw
0 replies
1h49m

It's not surprising, they prefer machines to people and call humans "stochastic parrots". The more humans are compared to animals, the more justified they feel writing them off as an expense and doing away with them.

big_whack
0 replies
23h14m

Do you make your living as an artist?

NoraCodes
0 replies
1d2h

My opinion is based on my interactions with my friends who are artists. I admit freely to caring less about what people I don't know say, in the absence of additional evidence.

CatWChainsaw
0 replies
1h50m

You've done the same thing in this comment so what makes you think you're the superior specimen?

SonicSoul
6 replies
1d3h

great comment!

imagine being a photographer that takes decades to perfect their craft. sure another student can study and mimic your style. but it's still different than some computer model "ingesting" vast amount of photos and vomiting something similar for $5.99 in aws cpu cost so that some prompt jockey can call themselves an AI artist and make money off of other peoples talent.

i get that this is cynical and does not encompass all ai art, but why not let computers develop their own style wihout ingesting human art? that's when it would actually be AI art

wincy
3 replies
1d2h

Like 99.9% of the art the common people care about is Darth Vader and Taylor Swift and other pop culture stuff like that.

These people literally don’t care what your definition of what is and isn’t art is, or how it’s made, they just want a lock screen wallpaper of themselves fighting against Thanos on top of a volcano.

The argument of “what is art” has been an academic conversation largely ignored by the people actually consuming the art for hundreds of years. Photography was just pop culture trash, comics were pop culture trash, stick figure web comics were pop culture trash. Today’s pop culture trash is the “prompt jockey”.

I make probably 5-10 pictures every day over the course of maybe 20 minutes as jokes on Teams because we have Bing Chat Enterprise. My coworkers seem to enjoy it. Nobody cares that it’s generated. I’m also not trying to be an “artist” whatever that means. It just is, and it’s fun. I wasn’t gonna hire an artist to draw me pictures to shitpost to my coworkers. It’s instead unlocked a new fun way to communicate.

johnnyanmac
0 replies
18h23m

I mean, I don't think many care about your personal use of art. You can take copyright images and shit post and Disney won't go suing your workplace.

But many big players do want to use this commercially and that's where a lot of these lines start to form. No matter how lawsuits go you will probably still be able to find some LLM to make Thanos fighting a volcano. It's just a matter of how/if companies can profit from it.

dartharva
0 replies
13h5m

This response should be gilded

SonicSoul
0 replies
1d2h

not entirely sure what your point is, but i think you are saying that art is just a commodity we use for cheap entertainment so it's ok for computers to do the same?

in the context of what i was saying the definition of what is art can be summed up as anything made by humans. i have no problem when its used in memes and being open sourced etc.. the issue i have is when a human invests real time into it and then its taken and regurgitated without their permission. do you see that distinction?

bongodongobob
0 replies
22h21m

That's a funny argument because artists lost their shit over photography too. Now anyone can make a portrait! Photography will kill art!

Art is the biggest gate kept industry there is and I detest artists who believe only they are the chosen one.

Art is human expression. We all have a right to create what we want with whatever tools we want. They can adapt or be left behind. No sympathy from me.

Levitz
0 replies
1d2h

Because that's not what happens, ever. You wouldn't ask a human to have their style of photographing when they don't know what a photograph even looks like.

witherk
5 replies
1d2h

"Cameras are a tool of captial to fuck over human portrait artists"

It's funny that these people use the langauge of communism, but apparently see artwork as purley an economic activity.

redwall_hp
2 replies
19h26m

It's funny that these people use the langauge of communism, but apparently see artwork as purley an economic activity.

You hit the nail on the head. Copyright is, by its very nature, a "tool of capital." It's a means of creating new artificial property fiefdoms for a select few capital holders to lord over, while taking rights from anyone else who wants to engage in the practice of making art.

Everyone has their right to expression infringed upon, all so the 1% of artists can perpetually make money on things, which are ultimately sold to corporations that only pay them pennies on the dollar anyway.

You, as an indie hip hop or house musician supported by a day job, can't sample and chop some vocals or use a slice of a chord played in a song (as were common in the 80s and 90s) for a completely new work, but apparently the world is such a better place because Taylor Swift is a multimillionaire and Disney can milk the maximum value from space and superhero films.

I'd rather live in a world where anyone is free to make whatever art they want, even if everyone has to have a day job.

skydhash
0 replies
14h13m

It's a means of creating new artificial property fiefdoms for a select few capital holders to lord over, while taking rights from anyone else who wants to engage in the practice of making art.

I doubt even Disney sue people who want to make fan art. But if you want to sell said art or distribute it, they will.

dahart
0 replies
2h1m

fiefdoms for a select few

What do you mean? Copyright protects all creative works, and all authors of those creative works. That some have greater means to enforce was always true, and copyright doesn’t cause that, it (imperfectly) helps mitigate it. What copyright does is actually prevent them from stealing work from independent artists en masse, and force them to at least hire and pay some artists.

I’d rather live in a world where anyone is free to make whatever art they want, even if everyone has to have a day job.

You’re suggesting abolish Copyright and/or the Berne Convention? Yeah the problem with this thinking is that then the big publishers are completely free to steal everyone’s work without paying for it. The very thing you’re complaining about would only get way way worse if we allowed anyone to “freely” make whatever art they want by taking it from others. “Anyone” means Disney too, and Disney is more motivated than you.

You, as an indie hip hop or house musician supported by a day job, can’t sample and chop some vocals or use a slice of a chord played in a song… for a completely new work

Hehe, if you sample, you are by definition not making a completely new work. But this is a terrible argument since sampling in music is widespread and has sometimes been successfully defended in court. DJs are the best example of independent artists who need protection you can think of?

indigo0086
0 replies
1d2h

They tried to use the labor theory early on by claiming, "real art takes hard work and time as opposed to the miniscule cpu hours computers use to make 'AI art". The worst thing AI brings to the table is amplifying these types of sentiments to control industry in their favor where they would otherwise be unheard and relegated to Instagram likes

NoraCodes
0 replies
1d2h

That's an intentional misinterpretation, I think. I mention art as an economic activity because it's primarily professional artists that are harmed by the widespread adoption of this technology.

matheusmoreira
3 replies
1d1h

they are a tool of capital to fuck over human artists

So are the copyright and intellectual property laws that artists rely on. From my perspective, you are the capital and I am the one being fucked. So are you ready to abolish all that?

CaptainFever
2 replies
20h29m

Right. This new outrage is just the copyright owners realising that their power is not safe. Where is the outrage when self checkouts happened?

matheusmoreira
1 replies
20h19m

Copyright owners indeed. That's what these artists are. They're copyright owners. Monopolists. They are the capital. Capitalism is all about owning property. Copyright is intellectual property. Literally imaginary property. Ownership of information, of bits, of numbers. These artists are the literal epitome of capitalism. They enjoy state granted monopolies that last multiple human lifetimes. We'll be long dead before their works enter the public domain. They want it to be this way. They want eternal rent seeking for themselves and their descendants. At least one artist has told me exactly that in discussions here on HN. They think it's fair.

They are the quintessential representation of capital. And they come here to ask us to "resist" the other forms of capital on principle.

I'm sorry but... No. I'm gonna resist them instead. It's my sincere hope that this AI technology hammers in the last nails on the coffin of copyright and intellectual property as a whole. I want all the models to leak so that it becomes literally impossible to get rid of this technology no matter how much they hate it. I want it to progress so that we can run it on our own machines, so that it'll be so ubiquitous it can't be censored or banned no matter how much they lobby for it.

NoraCodes
0 replies
19h55m

It's my sincere hope that this AI technology hammers in the last nails on the coffin of copyright and intellectual property as a whole.

If it does, I will give you one thousand United States dollars, and you can quote me on thst whenever you like.

More likely, big companies will retain control as they always have (via expensive lawyers), and individual artists will get screwed.

dartharva
0 replies
13h8m

Exactly. Artists should drop the pretentious philosophical bumbling and accept what this is, a fight for their livelihood. Which is, in every sense, completely warranted and good.

Putting blame on the technology and trying to limit public access to software will not go anywhere. Your fight for regulation needs to be with publishers and producers, not with the teen trying to make a cool new wallpaper or the office-man trying to make an aesthetic powerpoint presentation.

bluejekyll
45 replies
1d2h

My issue with this line of argument is that it’s anthropomorphizing machines. It’s fine to compare how humans do a task with how a machine does a task, but in the end they are very different from each other, organic vs hardware and software logic.

First, you need to prove that generative AI works fundamentally the same way as humans at the task of learning. Next you have to prove that it recalls information in the same way as humans. I don’t think anyone would say these are things that we can prove, let alone believe they do. So what we get is comments like they are similar.

What this means, is these systems will fall into different categories of law around copyright and free-use. What’s clear is that there are people who believe that they are harmed by the use of their work in training these systems and it reproducing that work in some manner later on (the degree to which that single work or the corpus of their work influences that final product is an interesting question). If your terms of use/copyright/license says “you may not train on this data”, then should that be protected in law? If a system like nightshade can effectively influence a training model enough to make it clear that something protected was used in its training, is that enough proof that the legal protections were broken?

huytersd
16 replies
23h42m

Why do you have to prove that? There is no replication (except in very rare cases), how someone draws a line should not be copyrightable.

pfist
15 replies
23h22m

Therein lies the crux of the issue: AI is not “someone”. We need to approach this without anthropomorphizing the AI.

Almondsetat
12 replies
23h12m

You are right, AI is nothing but a tool akin to a pen or a brush.

If you draw Mickey Mouse with a pencil and you publish (and sell) the drawing who is getting the blame? Is the pencil infringing the copyright? No, it's you.

Same with AI. There is nothijg wrong with using copyrighted works to train an algorithm, but if you generate an image and it contains copyrighted materials you are getting sued.

ufocia
11 replies
22h34m

But there is. You are arguably making unauthorized copies to train.

Almondsetat
9 replies
22h26m

Unauthorized copies? If the images are published on the internet how is it downloading them "unauthorized"?

__loam
8 replies
22h2m

Publicly available doesn't mean you have a license to do whatever you like with the image. If I download an image and re-upload it to my own art station or sell prints of it, that is something I can physically do because the image is public, but I'm absolutely violating copyright.

Almondsetat
7 replies
21h55m

That's not an unautharized copy, it's unauthorized distribution. By the same metric me seeing the image and copying it by hand is also unauthorized copy (or reproduction is you will)

xigoi
6 replies
20h48m

IANAL, but I’m pretty sure copying an image by hand is copyright violation.

Almondsetat
5 replies
20h19m

So you cannot train your drawing skills by copying other people's artworks?

xigoi
3 replies
20h17m

You can do it in private, but you can’t distribute the resulting image, let alone sell it.

Almondsetat
2 replies
20h1m

Then I don't really understand your original reply. Simply copying a publicly available image doesn't infringe anything (unless it was supposed to be private/secret). Doing stuff with that image in private still doesn't constitute infringement. Distribution does, but that wasn't the topic at hand

zerocrates
0 replies
11h56m

The most basic right protected by copyright is the right to make copies.

Merely making a copy can definitely be infringement. "Copies" made in the computing context even simply between disk and RAM have been held to be infringement in some cases.

Fair use is the big question mark here, as it acts to allow various kinds of harmless/acceptable/desirable copying. For AI, it's particularly relevant that there's a factor of the "transformative" nature of a use that weighs in favor of fair use.

xigoi
0 replies
19h59m

You can train a neural network in private too and nobody will have a problem with that. The topic of discussion is commercial AI.

shkkmo
0 replies
1h11m

The answer is "it depends". Distribution is not a hard requirement for copyright violation. It can significantly impact monetary judgements.

That said, there is also an inherent right to copy material that is published online in certain circumstances. Indeed, the physical act of displaying an image in a browser involves making copies of that image in cache, memory, etc.

huytersd
0 replies
20h14m

If you are viewing the image on your browser on a website, you are making a local copy. That’s not unauthorized.

pixl97
1 replies
22h59m

Companies aren't someone, yet in the US we seem to give them rights of someone.

ufocia
0 replies
22h33m

They are someone in the eyes of the law. They just have a different set of rights.

usrbinbash
6 replies
23h2m

that it’s anthropomorphizing machines.

No, it's not. It's merely pointing out the similarity between the process of training artists (by ingesting publicly available works) and ML models (which ingest publicly available works).

First, you need to prove that generative AI works fundamentally the same way as humans at the task of learning.

Given that there is no comprehensive model for how humans actually learn things, that would be an unfeasible requirement.

__loam
5 replies
22h9m

What a reductive way to describe learning art. The similarities are merely surface level.

Given that there is no comprehensive model for how humans actually learn things, that would be an unfeasible requirement.

That is precisely why we should not be making this comparison.

bowsamic
2 replies
9h35m

I’m being told repeatedly that the similarities are surface level, but no one seems to be able to give an example of a deep difference

__loam
1 replies
6h16m

The mechinism backing human learning isn't well understood. Machine learning is considerably clearer. Imo, it's a mistake to assume they're close because ML seems to work.

bowsamic
0 replies
3h52m

So your example of a deep difference is the suggestion that there may be a difference owing to our own ignorance? That hardly seems convincing

usrbinbash
1 replies
6h29m

The similarities are merely surface level.

Then please, feel free to explain the deep differences.

That is precisely why we should not be making this comparison.

Wrong. It's precisely why the claim "there is a big difference" doesn't have a leg to stand on. If you claim "this is different", I ask "how?" and the answer simply repeats the claim, I can apply Hitchens Razor[1] and dismiss the claim.

[1]: https://en.wikipedia.org/wiki/Hitchens%27s_razor

AlexandrB
0 replies
2h31m

A person sitting in an art school/museum for a few hours ingests way more than just the art in question. The entire context is brought in too, including the artists own physical/emotional state. Arguably, the art is a miniscule component of all sensory inputs. Generative AI ingests a perfectly cropped image of just the art from a single angle with little context beyond labelling metadata.

It's the difference between reading about a place and actually visiting it.

Edit: This doesn't even touch how the act of creating something - often in a completely different context - interacts with the memories of the original work, altering those memories yet again.

thfuran
4 replies
23h46m

First, you need to prove that generative AI works fundamentally the same way as humans at the task of learning. Next you have to prove that it recalls information in the same way as humans.

No, you don't need to prove any of those things. They're irrelevant. You'd need to prove that the AI is itself morally (or, depending on the nature of the dispute, legally) equivalent to a human and therefore deserving of (or entitled to) the same rights and protections as a human. Since it is pretty indisputably the case that software is not currently legally equivalent to a human, you're stuck with the moral argument that it ought to be, but I think we're very far from a point where that position is warranted or likely to see much support.

sebzim4500
1 replies
20h19m

Few people are claiming that the AI itself has the same rights as a human. They are arguing that a human with an AI has the same rights as a human who doesn't have an AI.

hn_acker
0 replies
1h55m

They are arguing that a human with an AI has the same rights as a human who doesn't have an AI.

This is the analogy I want people against AI use to understand and never forget, even if they reject the underlying premise - that should laws treat a human who uses AI for a certain purpose identically to a human who uses nothing or a non-AI tool for the same purpose.

Few people are claiming that the AI itself has the same rights as a human.

I think that's the case as well. However, a lot of commenters on this post are claiming that an AI is similar in behavior to a human, and trying to use the behavior analogy as the basis for justifying AI training (on legally-obtained copies of copyrighted works), with the assumption that justifying training justifies use. My personal flow of logic is the reverse: human who uses AI should be legally the same as human who uses a non-AI tool, so AI use is justified, so training on legally-obtained copies of copyrighted works is justified.

I want people in favor of AI use particularly to understand your human-with-AI-to-human-without-AI analogy (for short, the tool analogy) and to avoid machine-learning-to-human-learning analogies (for short, behavior analogies). The tool analogy is based on a belief about how people should treat each other, and contends with opposing beliefs about how people should treat each other. An behavior analogy must contend with both 1. opposing beliefs about how people should treat each other and 2. contradictions from reality about how similar machine learning is to brain learning. (Admittedly, both the tool analogy and the behavior analogy must contend with the net harm AI use is having and will have on the cultural and economic significance of human-made creative works.)

stale2002
0 replies
22h24m

ou'd need to prove that the AI is itself morally (or, depending on the nature of the dispute, legally) equivalent to a human and therefore deserving of

No you don't.

A human using a computer to make art doesn't automatically lose their fair use rights as a human.

indisputably the case that software is not currently legally equivalent to a human

Fortunately it is the human who uses the computer who has the legal rights to use computers in their existing process of fair use.

Human brains or giving rights to computers has absolutely nothing to do with the rights of human to use a camera, use photoshop, or even use AI, on a computer.

kelseyfrog
0 replies
23h37m

You don't even need to do that. Art is an act of ontological framing.

Duchamp didn't need negotiate with the ceramic makers to make the Fountain into art.

jwells89
4 replies
22h44m

The way these ML models and humans operate are indeed quite different.

Humans work by abstracting concepts in what they see, even when looking at the work of others. Even individuals with photographic memories mentally abstract things like lighting, body kinetics, musculature, color theory, etc and produce new work based on those abstractions rather than directly copying original work (unless the artist is intentionally plagiarizing). As a result, all new works produced by humans will have a certain degree of originality to them, regardless of influences due to differences in perception, mental abstraction processes, and life experiences among other factors. Furthermore, humans can produce art without any external instruction or input… give a 5 year old that’s never been exposed to art and hasn’t been shown how to make art a box of crayons and it’s a matter of time before they start drawing.

ML models are closer to highly advanced collage makers that take known images and blend them together in a way that’s convincing at first glance, which is why it’s not uncommon to see elements lifted directly from training data in the images they produce. They do not abstract the same way and by definition cannot produce anything that’s not a blend of training data. Give them no data and they cannot produce anything.

It’s absolutely erroneous to compare them to humans, and I believe it will continue to be so until ML models evolve into something closer to AGI which can e.g. produce stylized work with nothing but photographic input that it’s gathered in a robot body and artistic experimentation.

l33tman
2 replies
21h13m

You're wrong in your concept of how AI/ML works. Even trivial 1980's neural networks generalize, it's the whole point of AI/ML or you'd just have a lookup-table (or, as you put it, something that copies and pastes images together).

I've seen "infographics" spread by anti-AI people (or just attention-seekers) on Twitter that tries to "explain" that AI image generators blend together existing images, which is simply not true..

It is however the case that different AI models (and the brain) generalize a bit differently. That is probably the case between different humans too. Not the least with for example like you say those with photographic memory, autists etc.

What you call creativity in humans is just noise in combination with a boatload of exposure to multi-modal training data. Both aspects are already in the modern diffusion models. I would however ascribe a big edge in humans to what you normally call "the creative process" which can be much richer, like a process where you figure out what you lack to produce a work, go out and learn something new and specific, talk with your peers, listen to more noise.. stuff like that seems (currently) more difficult for AIs, though I guess plugins that do more iterative stuff like chatgpt's new plugins will appear in media generators as well eventually..

jwells89
1 replies
20h38m

ML generalization and human abstraction are very different beasts.

For example, a human artist would have an understanding of how line weight factors into stylization and why it looks the way it does and be able to accurately apply these concepts to drawings of things they’ve never seen in that style (or even seen at all, if it’s of something imaginary).

The best an ML model can do is mimic examples of line art in the given style within its training data, the product of which will contain errors due to not understanding the underlying principles, especially if you ask it to draw something it hasn’t seen in the style you’re asking for. This is why generative AI needs such vast volumes of data to work well; it’s going to falter in cases not well covered by the data. It’s not learning concepts, only statistical probabilities.

l33tman
0 replies
8h20m

I know what you're saying, and for sure existing models can be difficult to force into the really weird corners of the distributions (or go outside the distributions). The text interfaces are partially to blame for this though, you can take the images into Gimp and do some crude naive modifications and bring them back and the model will usually happily complete the "out-of-distribution" ideas. The Stable Diffusion toolboxes have evolved far away from the original simple text2image interfaces that midjourney and dalle use.

The models will generalize (because that's the most efficient way of storing concepts) and you can make an argument that that means they understand a concept. Claiming "it's not learning concepts, only statistical probabilities" trivialises what a modern neural network with billions of parameters and dozens of layers is capable of doing. If a model learns how to put a concept like line width 5, 10 and 15 pixels into a continuous internal latent property, you can probably go outside this at inference at least partially.

I would argue that improving this is at this point more about engineering and less about some underlying unreconcilable differences. At the very least we learn a lot about what exactly generalization and learning means.

shlubbert
0 replies
21h20m

Beautifully put. I wish this nuance was more widely understood in the current AI debate.

stale2002
2 replies
22h26m

What this means, is these systems will fall into different categories of law around copyright and free-use.

No they won't.

A human who uses a computer as a tool (under all the previous qualifications of fair use) is still a human doing something in fair use.

Adding a computer to the workflow of a human doesn't make fair use disappear.

A human can use photoshop, in fair use. They can use a camera. They can use all sorts of machines.

The fact that photoshop is not the same as a human brain is simply a completely unrelated non sequitur. Same applies to AI.

And all the legal protections that are offered to someone who uses a regular computer, to use photoshop in fair use, are also extended to someone who uses AI in fair use.

__loam
1 replies
22h5m

Yet the copyright office has already stated that getting an AI to create an image for you does not have sufficient human authorship to be copyrighted. There's already a legal distinction here between this "tool" and tools like photoshop and cameras.

It's also presumptive to assume that AI tools have these fair use protections when none of this has actually been decided in a court of law yet. There's still several unsettled cases here.

stale2002
0 replies
15h11m

Yet the copyright office has already stated that getting an AI to create an image for you does not have sufficient human authorship to be copyrighted.

Gotcha.

That has nothing to do with fair use though.

Also, the same argument absolutely applies to photoshop.

If someone didn't include sufficient human authorship while using photoshop, that wouldn't be copyrightable either.

Also, the ruling has no bearing on if someone using AI, while also inputting a significant amount of human authorship. Instead, it was only about the cases where there weren't much human authorship.

At no point did the copyright office disclude copyright protections from anything that used AI in any way what so ever. In fact, the copyright office now includes new forms and fields where you talk about the whole process that you did, and how you used AI, in conjunction with human authorship to create the work.

It's also presumptive to assume that AI tools

I'm not talking about the computer. I never claimed that computer's have rights. Instead, I'm talking about the human. Yes, a human has fair use protections, even if they use a computer.

There's still several unsettled cases here.

There is no reason to believe that copyright law will be interpreted in a significantly different way than it has been in the past.

There is long standing precent, regarding all sorts of copyright cases that involve using a computer.

michaelmrose
2 replies
23h7m

What we actually need to prove is whether such technology is a net benefit to society all else is essentially hand waving. There is no natural right to poorly named intellectual property and even if there was such a matter would never be decided based on the outcome of a philosophical argument because we don't decide anything that way.

withinboredom
0 replies
23h2m

such technology is a net benefit to society all else is essentially hand waving

Some might have said this about cars ... yet, here we are. Cars are definitely the opposite, except for longer-distance travel.

tqi
0 replies
22h54m

How do you measure "benefit", and what does "net" actually mean?

echelon
2 replies
23h37m

We are machines. We just haven't evenly accepted it yet.

Our biology is mechanical, and lay people don't possess an intuition about this. Unless you've studied molecular biology and biochemistry, it's not something that you can easily grasp.

Our inventions are mechanical, too, and they're reaching increasing levels of sophistication. At some point we'll meet in the middle.

qgin
0 replies
15h45m

This is the truth. And it's also the reason why this stuff will be in court until The Singularity itself. Most people will never be able to come to terms with this.

DennisAleynikov
0 replies
22h2m

100% this. Labor and all these other concepts are outdated ways to interpret reality

Humans are themselves mechanical so at the end of the day none of these issues actually matter

z7
0 replies
22h5m

My issue with this line of argument is that it’s anthropomorphizing machines. It’s fine to compare how humans do a task with how a machine does a task, but in the end they are very different from each other, organic vs hardware and software logic.

There's an entire branch of philosophy that calls these assumptions into question:

https://en.wikipedia.org/wiki/Posthumanism

https://en.wikipedia.org/wiki/Antihumanism

Martin Heidegger viewed humanism as a metaphysical philosophy that ascribes to humanity a universal essence and privileges it above all other forms of existence. For Heidegger, humanism takes consciousness as the paradigm of philosophy, leading it to a subjectivism and idealism that must be avoided.

Processes of technological and non-technological posthumanization both tend to result in a partial "de-anthropocentrization" of human society, as its circle of membership is expanded to include other types of entities and the position of human beings is decentered. A common theme of posthumanist study is the way in which processes of posthumanization challenge or blur simple binaries, such as those of "human versus non-human", "natural versus artificial", "alive versus non-alive", and "biological versus mechanical".
Phiwise_
0 replies
23h37m

The first perceptron was explicitly designed to be a trainable visual pattern encoder. Zero assumptions about potential feelings of the ghost in the machine need to be made to conclude the program is probably doing what humans studying art say they assume is happening in their head when you show both of them a series of previous artists' works. This argument is such a tired misdirection.

dorkwood
7 replies
1d2h

Is it strange to you that cars and pedestrians are both subject to different rules? They both utilise friction and gravity to travel along the ground. I'm curious if you see a difference between them, and if you could describe what it is.

ta8645
6 replies
23h29m

Both cars and pedestrians can be videotaped in public, without asking for their explicit permission. That video can be manipulated by a computer to produce an artwork that is then put on public display. No compensation need be offered to anyone.

estebank
5 replies
23h3m

Both cars and pedestrians can be videotaped in public, without asking for their explicit permission.

This is not universally true. Legislation is different from place to place.

ta8645
4 replies
22h42m

Hardly the point. The same can be said for road rules between vehicles and pedestrians, for example in major Indian cities, it's pretty much a free-for-all.

estebank
3 replies
22h30m

My point is that in a lot of places in the US you can point a video camera at the street and record. In Germany, you can't. The law in some locales makes a distinction between manual recording (writing or drawing your surroundings) and mechanized recording (photographing or filming). Scalability of an action is taken into consideration on whether something is ok to do or not.

ta8645
2 replies
22h24m

That has no bearing at all on the issue at hand. The same can be said of the original argument that started this thread.

thfuran
1 replies
19h56m

You think scalability isn't relevant to the difference between a person doing something by hand or with software operating on the entire internet?

johnnyanmac
0 replies
18h38m

Yeah, that's the oddest part of many of the pro-AI arguments. They want to anthromopotize the idea of learning but also clearly understand that the scalability of a bot exceeds that of a human.

They also don't seem to have much experience in the artist world. An artist usually can't reproduce a picture from memory, and if they can they are subject to copyright infringement depending on what and how they depict it, even if the image isn't a complete copy. By this logic of "bots are humans" a bot should be subject if they make a Not-legally-disctinct-enough talking mouse

visarga
4 replies
1d3h

AI is just a tool in someone's hand, there's a human who intends something

ta8645
3 replies
23h39m

If that's true, then it should be fine for that human to paint with the brush of his AI tool. Why should that human artist be restricted in the types of tools he uses to create his artwork?

thfuran
2 replies
19h29m

Should I be restricted in using copy paste as my tool for creating art?

visarga
0 replies
2h44m

like collage art?

ta8645
0 replies
18h0m

No you should not. You should be able to use any tool you want.

If you produce a work that is too much of a copy from the original, you might be liable to a copyright claim.. but the act of copy and paste should not be prohibited in the generation of something new.

This is done all the time by artists.. who perhaps create an artwork by copy and pasting advertisements out of a womans magazine to create an image of a womans face made only of those clipping. Making a statement about identity creation from corporate media... we should not put restrictions on such art work.

Here's just one example of an artist using copy-n-paste of content they don't own to create something new:

https://torontolife.com/culture/this-artist-creates-one-of-a...

ozten
4 replies
1d2h

The memetic weapons humans unleashed on other humans at art school to deter copying are brutal. Just wait until critique.

dist-epoch
3 replies
1d2h

"Sorry, this is not art, is AI generated trash."

HKH2
2 replies
16h18m

Who cares? With AI, you don't need art school. AI is making humanities redundant, and people are too proud to admit it.

I can't believe how many people are not in awe of the possibilities of AI art, so it's great to see AI disturbing the cynics until they learn. Not everything is political, but I'll let them have this one.

squidsoup
1 replies
12h45m

With AI, you don't need art school. AI is making humanities redundant, and people are too proud to admit it.

If you think this is true, you've never understood art. Art is a human endeavour fundamentally. What ML produces is not art.

HKH2
0 replies
5h37m

So what? You can gatekeep as much as you like, but whatever humans relate to as art is art.

bakugo
3 replies
23h22m

A human artist cannot look at and memorize 100000 pictures in a day, and cannot paint 100000 pictures in a day.

I am SO tired of this non-argument

ufocia
2 replies
22h16m

A human artist does not need to look at and memorize 100000 pictures in any span of time, period. Current AI does.

We needed huge amounts of human labor to fund and build Versailles. I'm sure many died as a result. Now we have machines that save many of those lives and labor.

What's your non-argument?

johnnyanmac
0 replies
18h30m

What's your non-argument?

That perhaps we shoildnt compare modern capitalistic societies to 18th century European royalty? I sure don't sympathize with the justification of the ability to use less labor to feed the rich.

MattRix
0 replies
21h32m

The argument is that the humans producing the work should be willing participants. I don’t think that’s too much to ask for.

Snow_Falls
3 replies
1d3h

These AIs are not people. They do not learn.

jfdbcv
2 replies
1d3h

Define learn.

dudeinjapan
1 replies
23h38m

Define people.

weregiraffe
0 replies
4h45m

Define "define".

krapp
2 replies
1d2h

Human beings and LLMs are essentially equivalent, and their processes of "learning" are essentially equivalent, yet human artists are not affected by tools like Nightshade. Odd.

pixl97
0 replies
22h56m

Humans don't fall for optical illusions? News to me.

danielbln
0 replies
23h31m

As another posted out, modern models like BLIP or GPT4V aren't affected by this either.

schmichael
1 replies
23h14m

This is not one artist inspiring another. This is all artists providing their work for free to immensely capitalized corporations for the corporations sole profit.

People keep making metaphors as if the AI is an entity in this transaction: it’s not! The AI is only the mechanism by which corporations launder IP.

thfuran
0 replies
19h54m

This is all artists providing their work for free to immensely capitalized corporations for the corporations sole profit.

No, the artists would be within their rights to do that if they chose to. This is corporations taking all the work of all artists regardless of the terms under which it was provided.

itronitron
1 replies
23h13m

Art schools don't teach people how to paint in different artistic styles. They teach materials and technique.

skydhash
0 replies
16h20m

Very true. I was watching a video yesterday learning how to make brush work digitally. While there were examples, they were just examples but the rest was specific techniques and demonstrations.

analog31
1 replies
23h37m

This seems more like looking at other artists and being totally incapacitated by some little touch in the painting you're looking at.

adhesive_wombat
0 replies
21h42m

The Nam-shub of Hockney?

password54321
0 replies
19h8m

It is only natural to see a moral difference between people going to school and learn from your art because they are passionate about it, versus someone on the internet just scraping as many images as possible and automating the learning process.

jurynulifcation
0 replies
23h7m

Artists learning to innovate a trade defend their trade from incursion by bloodthirsty, no-value-adding vampiric middle men attempting to cut them out of the loop.

joseph8th
0 replies
2h13m

Not being snarky, but if you believe that, then you clearly are not an artist.

eggdaft
0 replies
10h1m

That is not how artists learn. This is a false equivalence used to justify the imitation and copying of artists’ work. Artists’ work isn’t derivative in the same way that AI work is. Artists create work based on other sources of inspiration, some of them almost or completely to the disregard of other art.

Many artists don’t even go to art school. And those that do, do not spend most (all) of that time learning how to copy or imitate other artists.

I’m not expressing an opinion of whether GenAI is unethical or illegal - I think that’s a really difficult issue to wrestle with - just that this argument is a post-hoc rationalisation made in ignorance of how good artists work (not to say ignorance of the difference between illustration and art, conceptual art training vs say a foundation course etc).

bradleyishungry
0 replies
20h44m

This is such a nothing argument. Yes, new artists are inspired by other artists and sometimes make art similar to others, but a huge part of learning and doing art is to find a unique style.

But that’s not even the important part of the argument. A lot of artists work for commission, and are hired for their style. If an AI can be trained without explicit permission from their images, they lose work because a user can just prompt “in the style of”.

There’s no real great solution, outside of law, because the possibility of doing that is already here. But I’ve seen this argument so much and it’s just low effort

amelius
0 replies
23h18m

You might as well compare a Xerox copier to a human.

adr1an
0 replies
1d1h

It's not the learning per se what's concerning here but the ease of production (e.g. generate thousands of images in a day)

__loam
0 replies
22h13m

Human learning =/= machine learning

Most artists are happy to see more people getting into art and joining the community. More artists means the skills of this culture get passed down to the next generation.

Obviously a billion dollar corporation using their work to create an industrial tool designed to displace them is very different.

SirMaster
0 replies
14h14m

Lol, the scale is other-worldly different...

efitz
78 replies
22h34m

This is the DRM problem again.

However much we might wish that it was not true, ideas are not rivalrous. If you share an idea with another person, they now have that idea too.

If you share words on paper, then someone with eyes and a brain might memorize them (or much more likely, just grasp and retain the ideas conveyed in the words).

If you let someone hear your music, then the ideas (phrasing, style, melody, etc) in that music are transferred.

If you let people see a visual work, then the stylistic and content elements of that work are potentially absorbed by the audience.

We have copyright to protect specific embodiments, but mostly if you try to share ideas with others without letting them use the ideas you shared, then you are in for a life of frustration and escalating arms race.

I completely sympathize with anyone who had a great idea and spent a lot of effort to realize it. If I invented/created something awesome I would be hurt and angry if someone “copied” it. But the hard cold reality is that you cannot “own” an idea.

juunpp
25 replies
20h43m

You don't copyright ideas, you copyright works. And these artists' productions are works, not abstract ideas, with copyrights, and they are being violated. This is simple law. Why do people have such a hard time with this? Are you the one training the models and you need to find a cognitive escape out of the illegality and wrong-doing of your activities?

sircastor
5 replies
20h8m

This is simple law. Why do people have such a hard time with this?

Because this isn’t simple law. It feels like simple infringement, but there’s no actual copying going on. You can’t open up the database and find a given duplicate of a work. Instead you have some abstraction of what it takes to get to a given work.

Also it’s important to point out that nothing in the law is sure. A good lawyer, a sympathetic judge, a bored/interested/contrarian juror, etc can render “settled law” unsettled in an instant. The law is not a set of board game rules.

jazzyjackson
2 replies
18h12m

Have you seen the examples of midjourney reproducing exact frames of Dune, Star Wars etc? With vague prompting not asking for the media property specifically. It's pretty close to querying a database, except if you're asking for something that's not there it's able to render an interpolated result on the fly. Ask it for something that is there however and the model will dutifully pull it up.

sircastor
1 replies
16h5m

Looking for it, I found this [1] which describes almost what you're saying. The key difference here is that these images aren't exact frames. They're close, of course, but close is not identical.Is there another instance you can point to that shows what you've described?

[1] https://spectrum.ieee.org/midjourney-copyright

jazzyjackson
0 replies
15h35m

that's a good reference and yes that's what i'm talking about, but i mixed up the fact that you can get simpsons and star wars without asking for it by name - the "find the difference between these two pictures" game is the result of asking for a specific movie. I stand by my point tho that this is not substantially different than querying a database for copywrited material

juunpp
0 replies
15h50m

You can’t open up the database and find a given duplicate of a work. Instead you have some abstraction of what it takes to get to a given work.

So distributing a zip file of a copyrighted work subverts the copyright?

flkiwi
0 replies
19h40m

If the AI were a human and that human made an image that copied substantial elements of another human's creative work after a careful review of the original creator's work, even if it was not an original copy and no archival copy was stored somewhere in the second creator's creative space, I would be concerned about copyright infringement exposure if I were the second (copying) creator.

I'm open to the idea that copyright law might need to change, but it doesn't seem controversial to note that scraping actual creative works to extract elements for an algorithm to generate new works crosses a number of worrying lines.

rlt
5 replies
20h24m

It’s not obvious to me that using a copyrighted image to train a model is copyright infringement. It’s certainly not copyright infringement when used to train a human who may end up creating works that are influenced by (but not copies of) the original works.

Now, if the original copyrighted work can be extracted or reproduced from the model, that’s obviously copyright infringement.

OpenAI etc should ensure they don’t do that.

Andrex
3 replies
20h19m

Reproduced to what fidelity? 100%?

If OpenAI's output reproduces a copyrighted image with one pixel changed, is that valid in your view? Where does the line end?

Copyrighted material should never be used for nonacademic language models. "Garbage in, garbage out." All results are tainted.

"But being forced to use non-copyrighted works will only slow things down!"

Maybe that's a good thing, too. Copyright is something every industry has to accept and deal with -- LLMs don't get a "cool tech, do whatever" get-out-of-jail free card.

pawelmurias
1 replies
20h13m

Copyright is decided by the courts, it's a legal thing not some a biological. If the courts decide it's legal it will be.

Andrex
0 replies
19h47m

I'm totally down for the courts handling this AI/copyright mess, but I don't think technologists are going to like the results.

By virtue of the fact that it is "fuzzy" and open to interpretation, we're going to see lawsuits, the resulting chilling effects of those lawsuits will blunt US tech firms from the practice of ingesting large amounts of copywritten material without a second thought. US tech firms will be giving it a second, third, fourth, etc. thought once the lawsuits start.

It's gonna be like submarine patents on steroids.

Like I said, I'm down for letting the courts decide. But AI supporters should probably avoid kicking the hornets' nests regarding copyright.

rlt
0 replies
19h56m

Reproduced to what fidelity? 100%?

Whatever the standard is for humans doing the exact same thing.

thereisnospork
0 replies
18h50m

Now, if the original copyrighted work can be extracted or reproduced from the model, that’s obviously copyright infringement.

I think there's an important distinction to be made here - "can" be reproduced isn't infringement, only actual reproduction is (and degrees thereof not consisting of sufficiently transformative or fair use).

Trivially a typewriter can reproduce a copyrighted book. Less trivially Google books, with iirc stores the full text of copywrited works has been judged to be legal.

SamPatt
5 replies
20h0m

Illegality and wrongdoing are completely distinct categories.

I'm not convinced that most copyright infringements are immoral regardless of their legal status.

If you post your images for the world to see, and someone uses that image, you are not harmed.

The idea that the world owes you something after you deliberately shared it with others seems bizarre.

fzeroracer
1 replies
19h6m

If you post your images for the world to see, and someone uses that image, you are not harmed.

Let me define a few cases of 'uses that image' and see where your line in the sand drops

* If someone used that image as part of an advertising campaign for their product, they are profiting off your work. Are you not harmed?

* If someone used that image and pretended they created it. Are you not harmed?

* If someone used that image and sold it directly. Are you not harmed?

SamPatt
0 replies
17h15m

Someone claiming they created something which I created is the closest to harm on that list. Fraud should be punished.

The others aren't harmful, unless you're defining harm to include the loss of something which someone believes they are entitled to, a concept which is fraught with problems.

Creating an image (or any non-physical creation) doesn't obligate the world to compensate you for your work. If you choose to give it away by posting it on the internet, that's your choice, but you are entitled to nothing.

juunpp
0 replies
3h25m

My statement above is that their activities are illegal and wrong, not that one implies the other. They are illegal because of the copyright violation, and wrong because regardless of what the law says, using the images for training despite the artists' every appeal to the contrary (being vocal about the issue, putting a robots.txt file to avoid scraping, and now using adversarial techniques to protect their work from being stolen) is just moronic. It's like shitting on their front yard when they've asked you a million times not to shit in the front yard, put a sign that says "Please don't shit on my front yard", and sprayed insecticide all over the grass to try to deter you from shitting on the front yard. And yet you still shit on their front yard and even have the balls to argue that there's nothing wrong or illegal about it. This is absolutely insane.

johnnyanmac
0 replies
19h3m

I'm not convinced that most copyright infringements are immoral regardless of their legal status.

You're right and wrong. You're right because most infringement is from people who can do minimal damage and in fact so more help by giving awareness to your works by sharing. But this is only becsuse copyright it working (most of the time) against corporate entities who don't want to leave any room for legalities to come in.

If copyright ended I'd bet my bottom dollar Disney and all the other billionaires companies would be spamming any and everything that gets moderately popular. And Disney can put advertise the original artist easily.

brookst
0 replies
19h55m

Imagine if every book or advertisement or public conversation you overheard led to future claims that you had unethically learned from public information. It’s such a weird worldview.

(BTW I forbid you from using my comment here in your future reasoning)

theragra
2 replies
20h34m

If it were true, then we wouldn't have that great difference in opinions on this topic.

juunpp
1 replies
20h26m

That GP is utterly confused about copyright law is not an opinion.

sircastor
0 replies
19h54m

The United States Supreme Court rulings are supported literally by opinions of the justices.

fiddlerwoaroof
1 replies
19h48m

This is simple law.

“One may well ask: ‘How can you advocate breaking some laws and obeying others?’ The answer lies in the fact that there are two types of laws: just and unjust. I would be the first to advocate obeying just laws. One has not only a legal but a moral responsibility to obey just laws. Conversely, one has a moral responsibility to disobey unjust laws. I would agree with St. Augustine that ‘an unjust law is no law at all.’”

jazzyjackson
0 replies
18h11m

Well, fine, but you'll have to claim that copyright is unjust and that you are breaking the law as an act of civil disobedience. The AI corps do not want to take this stance as they also have intellectual property to protect. Classic "have their cake and eat it too" scenario.

sebzim4500
0 replies
20h25m

That may be the law, although we are probably years of legal proceedings away from finding out.

It obviously is not "simple law".

huytersd
0 replies
20h24m

Nothing is being reproduced. Just the ideas being reused.

freeAgent
17 replies
22h30m

This doesn’t stop anyone from viewing or scraping the work, though, so in no way is it DRM. It just causes certain methods of computer interpretation of an image to interpret it in an odd way vs. human viewers. They can still learn from them.

avhon1
16 replies
22h29m

It absolutely is DRM, just a different form than media encryption. It's a purely-digital mechanism of enforcing rights.

freeAgent
15 replies
22h27m

It doesn’t enforce any rights. It modifies the actual image. Humans and computers still have equal, open access to it.

efitz
9 replies
22h22m

It’s designed to restrict the purposes for which the consumer can use the work. It is exactly like DRM in this way.

freeAgent
7 replies
22h17m

To be clear, you can still train AI with these images. Nothing is stopping you.

johnnyanmac
6 replies
18h59m

To quote the source:

Nightshade's goal is not to break models, but to increase the cost of training on unlicensed data, such that licensing images from their creators becomes a viable alternative.

Which feels similar to DRM. To discourage extraction of assets.

freeAgent
5 replies
18h52m

It also degrades the quality of the image for human consumers. It’s just a matter of what someone is willing to publish to “the public.”

johnnyanmac
4 replies
18h50m

Sure. Just like how video game drm impacts performance and watermarks on images degrades the image. Drm walks a tight line that inevitably makes the result worse than a drum-free solution, but also should not make the item completely unconsumable.

freeAgent
3 replies
18h39m

Video game DRM completely prevents people without a license/key to unlock it from accessing the game at all.

johnnyanmac
2 replies
18h34m

So, do you want to define drm by intent or technical implementation? I'm doing the former, but it sounds like you want to do the latter. Also keep in mind that legalese doesn't necessarily care about the exact encryption technique to be deployed either.

freeAgent
1 replies
18h26m

Both. Changing an image is done all the time prior to publishing them. In fact, no image you ever see on the internet is a raw sensor output. They are all modified in some manner. The images processed using this method look the same to every person and computer that views them. That’s very different from DRM which encrypts things and prevents access to unprivileged users.

This is effectively the equivalent of someone doing really crappy image processing. As other commenters have mentioned, it does alter how images look to humans as well as machines, and it can be “mitigated” through additional processing techniques.

johnnyanmac
0 replies
18h20m

That’s very different from DRM which encrypts things and prevents access to unprivileged users.

Well you can call it a captcha if you want. The point here is to make it harder to access for bots (but not impossible) while inconveniencing honest actors in the process. It doesn't sound like there's a straightforward answer to "are captchas DRM" either.

freeAgent
0 replies
22h18m

How does it stop you from using an image however you want?

renewiltord
4 replies
21h48m

That's true of almost all DRM, isn't it? Even for the most annoying form that is always-online DRM, everyone is provided the same access to the bytes that form a game. You and I have the same bytes of game.

It's the purpose of some of those bytes that turns it into DRM.

freeAgent
3 replies
21h5m

No, it's not the same. The game is non-functional without the proper keys/authorization whereas images run through this algorithm are still images that anyone and any computer can view in the same manner without any authentication.

aspenmayer
2 replies
18h18m

An analogy that springs to mind is the difference between an access control mechanism such as a door with lock and key versus whatever magical contrivance that prevents entry to a dwelling by vampires uninvited.

freeAgent
1 replies
18h3m

That may be an ok analogy, but magic isn’t real. Maybe a better analogy would be to speak or write in a language that you know certain people don’t natively understand, and using lots of slang and idioms that don’t translate easily. Someone could still run it through Google Translate or whatever, but they won’t get a great understanding of what you actually said. They’d have to actually learn the language and the sorts of slang and idioms used.

aspenmayer
0 replies
17h39m

I agree that the analogy is strained. My goal was in elucidating the distinction between:

the goals of artists and the developers of OP,

versus

the goals of AI engineers,

and how it seems to me similar to the is/ought disctinction.

In my original analogy, it’s generally considered lawful to have a lock on your door, or not to do so, and the issue of a lock or lack thereof is moot when one is invited to enter, just as it is lawful to enter a premises when invited or during exigent circumstances, such as breaching entry to render lifesaving aid by emergency services or firefighters.

By that same token, no amount of locks or other barriers to entry will prevent ingress by a vampire once invited inside.

To me, much of the ballyhoo about OP seems like much ado about big cats eating faces, like a person publicly decrying the rise in vampire attacks after inviting that same vampire inside for dinner. It’s a nonstarter.

Copyright law is broken, because of the way that the law is written as much as the way that it’s enforced, and also broken because of the way that humans are. Ideas are not copyrightable, and while historically their implementations or representations were, going forward, neither implementations nor representations are likely to receive meaningful/effective protections from copyright itself, but only from legal enforcement of copyright law.

After the recent expiry of Disney’s copyright on Steamboat Willie, the outpouring of praise, support, excitement, and original work from creators shows me that copyright law in its current incarnation doesn’t perform its stated goals of promoting the creation of arts and sciences, and so should be changed, and in the meantime ignored and actively disobeyed, as any unjust law ought to be, regardless of what the law is or does.

In light of our obligation to disobey unjust laws, I applaud efforts like OP to advance the state of the art in computer science, while at the same time encouraging others working on AI to actively circumvent such efforts for the selfsame reason.

I similarly encourage artists of all kinds to make art for art’s sake while monetizing however they see fit, without appealing to red herrings like the legality or lack thereof of end users appreciating their art and incorporating it into their own artistic output, however they may choose to do so.

Like all art, code and its outputs is also First Amendment protected free speech.

tsujamin
10 replies
22h30m

Being able to fairly monetise your creative work and put food on the table is a bit rivalrous though, don’t you think?

efitz
6 replies
22h11m

No, I disagree. There is no principle of the universe or across human civilizations that says that you have a right to eat because you produced a creative work.

The way societies work is that the members of the society contribute and benefit in prescribed ways. Societies with lots of excess production may at times choose to allow creative works to be monetized. Societies without much surplus are extremely unlikely to do so, eg a society with not enough food for everyone to eat in the middle of a famine is extremely unlikely to feed people who only create art; those people will have to contribute in some other way.

I think it is a very modern western idea (less than a century old) that many artists can dedicate themselves solely to producing the art they want to produce. In all other times artists either had day jobs or worked on commission.

jrflowers
3 replies
21h55m

There is no principle of the universe or across human civilizations

Can you list the principles across human civilizations?

yjftsjthsd-h
1 replies
19h1m

That doesn't follow; you can say an item is not in a set without writing out every member of that set. What principle do you claim exists to contradict that claim?

jrflowers
0 replies
18h53m

you can say an item is not in a set without writing out every member of that set

Of course you can. Anyone can say anything.

Is “keeping the list of principles a secret” a principle like the rules of Fight Club? It is not unreasonable to ask for a link or summary of this immutable set of ground truths.

What principle do you claim exists to contradict that claim?

I could not answer this question without being able to double check. The only principle that comes to mind is the principle of ligma

juunpp
0 replies
20h45m

He can also presumably list the principles of the universe.

johnnyanmac
1 replies
18h55m

There is no principle of the universe or across human civilizations that says that you have a right to eat because you produced a creative work.

What does that have to do with rivalry? This doesn't dispute the idea that AI is indeed competing with artists. You're just saying artists don't deserve to get paid.

Regardless, some artists will give up but some will simply be more careful with where and how they post their art with tools like these. AI doesn't have a right to the artist's images neither.

jazzyjackson
0 replies
18h22m

I used to identify as a copyright abolitionist (I really love Nina Paley's TED talk, "copyright is brain damage") but the more I look at the history of it I see the compromises of interests, copyright is there so art is not locked up between artists and their patrons.

renewiltord
0 replies
21h50m

The tragedy of "your business model is not my problem" as a spreading idea is that while you're right since distribution is where the money is (not creation), intellectual property is de-facto weakened today and IP piracy is widely considered an acceptable thing.

ikmckenz
0 replies
19h12m

No, rivalrous has a specific meaning https://en.wikipedia.org/wiki/Rivalry_(economics)

bsza
0 replies
19h2m

So is sabotaging solutions that would make creative work of the same (or superior) quality more affordable. Your ability to produce expensive illustrations hinders my ability to produce cheap textbooks.

xpe
8 replies
21h29m

... you cannot “own” an idea.

Let's talk about ownership in a broader sense. In practice, one cannot effectively own (retain possession of) something without some combination of physical capability or coercion (or threat of coercion). Meaning: maintaining ownership of anything (physical or otherwise) often depends on the rule of law.

thomastjeffery
7 replies
20h38m

Then let's use a more precise term that is also present in law: monopoly.

You can't monopolize an idea.

Copyright law is a prescription, not a description. Copyright law demands that everyone play along with the lie that is intellectual monopoly. The effectiveness of that demand depends on how well it can be enforced.

Playing pretend during the age of the printing press may have been easy enough to coordinate, but it's practically impossible here in the digital age.

If we were to increase enforcement to the point of effectiveness, then what society would be left to participate? Surely not a society I am keen to be a part of.

xpe
6 replies
19h13m

Trying to make sense of the above comment is difficult.

Copyright law demands that everyone play along with the lie that is intellectual monopoly.

Saying "lie" suggests willful deception. Perhaps you mean "socially constructed"? Combined with "playing pretend" makes it read a bit like a rant.

Then let's use a more precise term that is also present in law: monopoly.

Ok, in law and economics, the core idea of monopoly has to do with dominant market power that crowds out the existence of others. But your other uses of "monopoly" don't match that. For example, you talk about ideas and "intellectual monopoly". What do you mean?

It seems like some of your uses of "monopoly" are not about markets but instead are closer to the idea of retaining sole ownership.

If we were to increase enforcement to the point of effectiveness, then what society would be left to participate? Surely not a society I am keen to be a part of.

It appears you've already presupposed how things would play out, but I'm not convinced. What is your metric of effectiveness? A scale is better than some arbitrary threshold.

Have you compared copyright laws and enforcement of the U.S. versus others?

How far would you go: would you say that i.e. society would be better off without copyright law? By what standard?

thomastjeffery
3 replies
13h6m

The best attempt at thought monopoly I can think of is religion. Even that is a general failure: no single religious narrative has ever stood constant. They have all evolved through the participation of their adherents. There is no one true Christianity: there are hundreds.

I most certainly do mean to call out copyright as willful, but it's not a deception, at least not a successful one: everyone knows it is false. That's why it's enforced by law! Instead of people being deceived, people must instead pretend to be so. Each of us must behave as if the very concept of Micky Mouse is immortal and immutable; and if we don't, the law will punish accordingly.

Every film on Netflix, every song on Spotify, etc. can obviously be copied any number of times by any number of people at any place on Earth. We are all acutely aware of this fact, but copyright tells us, "Pretend you can't, or get prosecuted."

So is it truly effective? Millions of people are not playing along. Millions of artists are honestly trying to participate in this market, and the market is failing them. Is that because we need more people to play along? Rightsholders like the MPAA say that piracy is theft, and that every copy that isn't paid for is a direct cost to their business. How many of us are truly willing to pretend that far?

What if we all just stopped? Would art suddenly be unprofitable for everyone, including the lucky few who turn a profit today? I don't believe that for a second.

The only argument I have ever heard in favor of copyright is this: Every artist deserves a living. I have seen time and time again real living artists fail to earn a living from their copyright. I have seen time and time again real living artists share their work for free, choosing to make their living by more stable means.

Every person living deserves a living. Fix that, and we will fix the problem copyright pretends to solve, and more.

xpe
0 replies
2h39m

The comment above suffers from much rhetoric to serve as a good jumping off point. For those interested, I would recommend the following two articles:

## "Rhetoric and Reality in Copyright Law" by Stewart E. Sterk

Benjamin N. Cardozo School of Law. https://repository.law.umich.edu/cgi/viewcontent.cgi?article...

Why give authors an exclusive right to their writings? Copyright rhetoric generally offers two answers. The first is instrumental: copyright provides an incentive for authors to create and disseminate works of social value. By giving authors a monopoly over their works, copyright corrects for the underincentive to create that might result if free riders were permitted to share in the value created by an author's efforts. The second answer is desert: copyright rewards authors, who simply deserve recompense for their contributions whether or not recompense would induce them to engage in creative activity.

The rhetoric evokes sympathetic images of the author at work. The instrumental justification for copyright paints a picture of an author struggling to avoid abandoning his calling in order to feed his family. By contrast, the desert justification conjures up a genius irrevocably committed to his work, resigned - or oblivious - to living conditions not commensurate with his social contributions. The two images have a common thread: extending the scope of copyright protection relieves the author's plight.

Indeed, the same rhetoric· - emphasizing both incentives and desert - consistently has been invoked to justify two centuries of copyright expansion. Unfortunately, however, the rhetoric captures only a small slice of contemporary copyright reality. Although some copyright protection indeed may be necessary to induce creative activity, copyright doctrine now extends well beyond the contours of the instrumental justification. ...

## "Copyright Nonconsequentialism" by David McGowan

Missouri Law Review. https://scholarship.law.missouri.edu/cgi/viewcontent.cgi?art...

This Article explores the foundations of copyright law. It tries to explain why those who debate copyright often seem to talk past each other. I contend the problem is that copyright scholars pay too much attention to instrumental arguments, which are often indeterminate, and too little to the first principles that affect how one approaches copyright law.

Most arguments about copyright law use instrumental language to make consequentialist arguments. It is common for scholars to contend one or another rule will advance or impede innovation, the efficient allocation and production of expression, personal autonomy, consumer welfare, the "robustness" of public debate, and so on.' Most of these instrumental arguments, though not quite all of them, reduce to propositions that cannot be tested or rejected empirically. Such propositions therefore cannot explain existing doctrine or the positions taken in debate.

These positions vary widely. Consumer advocates favor broad fair use rights and narrow liability standards for contributory infiringement; producer advocates favor the reverse.' Most of the arguments for both consumers and producers prove too much. It is easy to say that the right to exclude is needed to provide incentives for authors. It is hard to show that any particular rules provide optimal incentives. It is easy to point to deviations from the model of perfect competition. It is hard to show why these deviations imply particular rules.

...
xpe
0 replies
3h14m

I most certainly do mean to call out copyright as willful, but it's not a deception, at least not a successful one: everyone knows it is false.

Unless I'm misunderstanding you, this is not even wrong. What about copyright law is empirically false? Such a question is non-sensical.

Your comment redefines the word "false" in a way that muddles understanding. You aren't alone -- some philosophers do this -- but it tends to confuse rather than clarify. I've developed antibodies for language abuse of this kind. Such language can even have the effect of making language charged and divisive.

Many people understand the value of using the words _true_ and _false_ to apply to the assessment of _facts_. This is a useful convention. (To be clear, I'm not opposed to bending language when it is useful.)

To give a usage example: a misguided law is not _false_. Such a statement is non-sensical. We have clear phrases for this kind of law, such poorly designed, having unintended consequences, etc. We could go further and say that a law is i.e. immoral or pointless. You are likely making those kinds of claims. By using those phrases, we can have a high-bandwidth conversation much more quickly.

xpe
0 replies
2h54m

Every film on Netflix, every song on Spotify, etc. can obviously be copied any number of times by any number of people at any place on Earth. We are all acutely aware of this fact, but copyright tells us, "Pretend you can't, or get prosecuted."

I see the dark arts of rhetoric used here, and it is shameful. The portion I quoted above is incredibly confused. I would almost call it a straw man, but it is worse than that.

Copyright law says no such thing. Of course you _could_ copy something. Copyright law exists precisely because you can do that. The law says i.e. "if you break copyright law, you will be at risk of a sufficiently motivated prosecutor."

jazzyjackson
1 replies
18h9m

Saying "lie" suggests willful deception.

Consider instead the term "legal fiction", it's not so derogatory.

xpe
0 replies
13h16m

Legal fiction is a technical term used by legal scholars. To be clear, any legal system is built from legal constructs; I'm not talking about these. A legal fiction has a markedly different meaning than lie as used in the comment two levels above (which seems to me more like a rant than a clear or convincing argument) Are you familiar with specific expert writings about claimed legal fictions in U.S. copyright law?

xpe
3 replies
21h56m

But the hard cold reality is that you cannot “own” an idea.

The above comment is true about the properties of information, as explained via the lens of economics. [1]

However, one ignores ownership as defined by various systems (including the rule of law and social conventions) at one's own peril. Such systems can also present a "hard cold reality" that can bankrupt or ostracize you.

[1] Don't let the apparent confidence and technicality of the language of economists fool you. Economics isn't the only game in town. There are other ways to model and frame the world.

[2] Dangling footnote warning. I think it is instructive to recognize that the field of economics has historically shown a kind of inferiority complex w.r.t. physics. Some economists ascribe to the level of rigor found in physics and that is well and good, but perhaps that effort should not be taken too seriously nor too far, since economics as a field operates at a different level. IMO, it would be wise for more in the field to eat a slice of humble pie.

[3] Ibid. It is well-known that economists can be "hired guns" used to "prove" a wide variety of things, many of which are subjective. My point: you can hire an economist to shore up one's political proposals. Is the same true of physicists? Hopefully not to the same degree. Perhaps there are some cases of hucksterism, but nothing like the history of economists-wagging-the-dog! At some point, the electron tunnels or it does not.

meowkit
2 replies
20h50m

There are other games in town.

But whatever game gives the most predictive power is going to win.

xpe
0 replies
19h56m

There is no need to frame this as "winning versus losing" regarding the many models that we draw upon.

Even when talking about various kinds of scientific and engineering fields, predictive power isn't the only criteria, much less the best. Sometimes the simpler, less accurate models work well enough with less informational and computational cost.

Even if we focus on prediction (as opposed to say statistical inference), often people want some kind of hybrid. Perhaps a blend of satisficing with limited information, scoped action spaces, and bounded computation; i.e. good enough given the information we have to make the decisions we can actuate with some computational budget.

sfifs
0 replies
19h30m

By that metric, various economic schools have been hilariously inept and would get classified not dissimilar to various schools of religious theology with their own dogmas. It's only in the last 15 years or so that some focus on empiricism and explaining reality rather than building theoretical castles in the air is coming about and is still far from mainstream.

throwoutway
3 replies
22h29m

I don't see the parallel between this offensive tool and DRM. I could, say buy a perpetual license to an image from the artist, so that I can print it and put it on my wall, while it can simultaneously be poisonous to an AI system. I can even steal it and print it, while it is still poisonous to an AI system.

The closest parallel I can think of is that humans can ingest chocolate but dogs should not.

jdietrich
0 replies
21h39m

What you've described is the literal, dictionary definition of Digital Rights Management - a technology to restrict the use of a digital asset beyond the contractually-agreed terms. Copying is only one of many uses that the copyright-holder may wish to prevent. The regional lockout on a DVD had nothing to do with copy-protection, but it was still DRM.

gwbas1c
0 replies
20h51m

It's about the arm's race: DRM will always be cracked (with a sufficiently motivated customer.) AI poisoning will always be cracked (with a sufficiently motivated crawler.)

efitz
0 replies
22h22m

A huge amount of DRM effort has been spent in the watermarking area, which is similar, but not exactly the same.

xpe
2 replies
22h3m

Many terms of art from economics are probably not widely-known here.

In economics, a good is said to be rivalrous or a rival if its consumption by one consumer prevents simultaneous consumption by other consumers, or if consumption by one party reduces the ability of another party to consume it. - Wikipedia: Rivalry (economics)

Also: we should recognize that stating something as rivalrous or not is descriptive (what exists) not normative (what should be).

fastball
1 replies
18h46m

I think ideas being rivalrous is intrinsic, and therefore descriptive and normative.

xpe
0 replies
13h44m

I'm either not understanding you or disagreeing. You seem to be saying that something should be because it is? Saying that would be rather silly, as in "Electrons should repel each other because they repel each other." Not to mention that this claim runs amok of the naturalistic fallacy. So what are you driving at?

wredue
0 replies
18h44m

Kick ass.

I now declare that I own Fortnite.

Where’s my money, Epic?

kmeisthax
0 replies
18h44m

We're not trying to keep the AI from learning general ideas, we're trying to keep it from memorizing specific expressions[0]. There's a growing body of research to show that these models are doing a lot of memorizing, even if they're not regurgitating that data. For example, Google's little "ask GPT to repeat a word forever" trick, which will make GPT-4 spit out verbatim training data[1].

If there was a training process that let us pick a minimal sample of examples and turn it into a general purpose art generator or text generator, I think people would have been fine with that. But that's not what any of these models do. They were trained on shittons of creative expression, and there's statistical evidence that the models retain that expression, in a way that is fundamentally different from how humans remember, misremember, adapt, remix, and/or "play around with" other people's creativity.

[0] You called these "embodiments", but I believe you're trying to invoke the idea/expression divide, so I'll run with that.

[1] Or at least it did. OpenAI now filters out conversations that trip the bug.

chefandy
0 replies
18h59m

Not everybody equates automated scraping for training models and human experience. Just like any other “data wants to be free” type of discussion, the philosophical and ethical considerations are anything but cut-and-dried, and they’re far more consequential than the technical and economics-in-a-vacuum ones. The general public will quite possibly see things differently than the “oh well, artists— that’s the free market for ya, and you lost” crowd.

542458
47 replies
1d2h

This seems to introduce levels of artifacts that many artists would find unacceptable: https://twitter.com/sini4ka111/status/1748378223291912567

The rumblings I'm hearing are that this a) barely works with last-gen training processes b) does not work at all with more modern training processes (GPT-4V, LLaVA, even BLIP2 labelling [1]) and c) would not be especially challenging to mitigate against even should it become more effective and popular. The Authors' previous work, Glaze, also does not seem to be very effective despite dramatic proclamations to the contrary, so I think this might be a case of overhyping an academically interesting but real-world-impractical result.

[1]: Courtesy of /u/b3sn0w on Reddit: https://imgur.com/cI7RLAq https://imgur.com/eqe3Dyn https://imgur.com/1BMASL4

pimlottc
18 replies
23h4m

I can’t really see any difference in those images on the Twitter example when viewing it on mobile

vhcr
4 replies
22h28m

The animation when you change images makes it harder to see the difference, I opened the three images each in its own tab and the differences are more apparent when you change between each other instantly.

SirMaster
2 replies
17h12m

But that’s not realistic?

If you have to have both and instantly toggle between them to notice the difference, then it sounds like it’s doing its job well and is hard to notice the difference.

bowsamic
0 replies
9h40m

What kind of artist is not going to be bothered with seeing huge artifacting on their work? Btw for me it was immediately noticeable even on mobile

battles
0 replies
13h17m

The person who drew it would definitely notice.

dontupvoteme
0 replies
21h57m

One of the few times a 'blink comparator' feature in image viewers would be useful!

fenomas
4 replies
13h50m

At full size it's super obvious - I made a side-by-side:

https://i.imgur.com/I6EQ05g.png

trimethylpurine
3 replies
11h51m

I still don't see a difference. (Mobile)

fenomas
0 replies
11h33m

Here's a maybe more mobile friendly comparison:

https://i.imgur.com/zUVn8rt.png

But now that I double-check, I was comparing with the images zoomed to 200%. On desktop the artifacts are also noticeable at 100%, but not nearly as bad as in my previous comment.

bowsamic
0 replies
9h40m

What phone are you using? It’s extremely obvious on my iPhone

Rewrap3643
0 replies
5h52m

Have you done a color blindness test before? Red-green is the most common type and the differences here are mostly shades of green.

pxc
1 replies
22h50m

I don't have great vision, but me neither. They're indistinguishable to me (likewise on mobile).

Gigachad
0 replies
15h10m

I was on desktop and it looks like pretty heavy jpeg compression. Doesn't completely destroy the image, but it's pretty noticeable when blown up large enough.

milsorgen
0 replies
22h48m

It took me a minute too but on the fast you can see some blocky artifacting by the elbow and a few spots elsewhere like curtain upper left.

jquery
0 replies
14h15m

It's really noticeable on desktop, like compressing an 800kb jpeg to 50kb. Maybe on mobile you won't notice, but on desktop the image looks blown out.

josefx
0 replies
19h55m

Something similar to jpeg artifacts on any surface with a normally smooth color gradient, in some cases rather significant.

charcircuit
0 replies
22h5m

The gradient on the bat has blocks in it instead of being smooth.

Keyframe
0 replies
22h48m

look at the green drapes to the right, or any large uniform colored space. It looks similar to bad JPEG artifacts.

0xcde4c3db
0 replies
22h44m

I didn't see it immediately either, but there's a ton of added noise. The most noticeable bit for me was near the standing person's bent elbow, but there's a lot more that becomes obvious when flipping back and forth between browser tabs instead of swiping on Twitter.

kmeisthax
12 replies
18h54m

The screenshots you sent in [1] are inference, not training. You need to get a Nightshaded image into the training set of an image generator in order for this to have any effect. When you give an image to GPT-4V, Stable Diffusion img2img, or anything else, you're not training the AI - the model is completely frozen and does not change at all[0].

I don't know if anyone else is still scraping new images into the generators. I've heard somewhere that OpenAI stopped scraping around 2021 because they're worried about training on the output of their own models[1]. Adobe Firefly claims to have been trained on Adobe Stock images, but we don't know if Adobe has any particular cutoffs of their own[2].

If you want an image that screws up inference - i.e. one that GPT-4V or Stable Diffusion will choke on - you want an adversarial image. I don't know if you can adversarially train on a model you don't have weights for, though I've heard you can generalize adversarial training against multiple independent models to really screw shit up[3].

[0] All learning capability of text generators come from the fact that they have a context window; but that only provides a short term memory of 2048 tokens. They have no other memory capability.

[1] The scenario of what happens when you do this is fancifully called Habsburg AI. The model learns from it's own biases, reinforcing them into stronger biases, while forgetting everything else.

[2] It'd be particularly ironic if the only thing Nightshade harms is the one AI generator that tried to be even slightly ethical.

[3] At the extremes, these adversarial images fool humans. Though, the study that did this intentionally only showed the images for a small period of time, the idea being that short exposures are akin to a feed-forward neural network with no recurrent computation pathways. If you look at them longer, it's obvious that it's a picture of one thing edited to look like another.

scheeseman486
4 replies
13h27m

Hey you know what might not be AI generated post-2021? Almost everything run through Nightshade. So given it's defeated, which is pretty likely, artists have effectively tagged their own work for inclusion.

hkt
1 replies
12h25m

It is a great shame that we have come to a no-win situation for artists when VCs are virtually unable to lose.

ToucanLoucan
0 replies
2h38m

I mean that's more or less status quo isn't it? Big business does what it wants, common people can get fucked if they don't like it. Same as it ever was.

visarga
0 replies
57m

Modern generative image models are trained on curated data, not raw internet data. Sometimes the captions are regenerated to fit the image better. Only high quality images with high quality descriptions.

kmeisthax
0 replies
27m

Why wouldn't an artist just generate AI spam and Nightshade it?

jerbear4328
1 replies
14h36m

[3] sounds really interesting - do you have a link?

ittseta
0 replies
13h38m
KTibow
1 replies
13h3m

Correct me if I'm wrong but I understand image generators as relying on auto-labeled images to understand what means what, and the point of this attack to make the auto-labelers mislabel the image, but as the top-level comment said it's seemingly not tricking newer auto-labelers.

michaelbrave
0 replies
8h29m

not all are auto labelled, some are hand labelled, some are initially labelled with something like clip/blip/booru and then corrected a bit by hand. The newest thing though is using llm's with image support like GPT4 to label the images, which kind of does a much better job most of the time.

Your understanding of the attack was the same as mine, it injects just the right kinds of pixels to throw off the auto-labellers to misdirect what they are directing causing the tags to get shuffled around.

Also on reddit today some of the Stable Diffusion users are already starting to train using Nightshade so they can implement it as a negative model, which might or might not work, will have to see.

webmaven
0 replies
13h25m

Even if no new images are being scraped to train the foundation text-to-image models, you can be certain that there is a small horde of folk still scraping to create datasets for training fine-tuned models, LoRAs, Textual Inversions, and all the new hotness training methods still being created each day.

ptdn
0 replies
13h30m

The context windows of LLMs are now significantly larger than 2048 tokens, and there are clever ways to autopopulate context window to remind it of things.

GaggiX
0 replies
11h59m

If it doesn't work during inference I really doubt it will have any intended effect during training, there is simply too much signal and the added adversarial noise works on the frozen and small proxy model they used (CLIP image encoder I think) but it doesn't work on a larger model and trained on a different dataset, if there is any effect during training it will probably just be the model learning that it can't take shortcuts (the artifacts working on the proxy model showcase gaps in its visual knowledge).

Generative models like text-to-image have an encoder part (it could be explicit or not) that extract the semantic from the noised image, if the auto-labelers can correctly label the samples then the encoded trained on both actual and adversarial images will learn to not take the same shortcuts that the proxy model has taken making the model more robust, I cannot see an argument where this should be a negative thing for the model.

GaryNumanVevo
9 replies
22h36m

The artifacts are a non-issue. It's intended images with nightshade are intended to be silently scrapped and avoid human filtering.

minimaxir
5 replies
22h11m

The artifacts are extremely an issue for artists who don't want their images damaged for the possibility of them not being trained by AI.

It's a bad tradeoff.

GaryNumanVevo
4 replies
22h5m

Nightshaded images aren't intended for portfolios. They're mean to be uploaded enmasse and scraped later.

AJ007
3 replies
21h10m

To where? A place no one sees them and they aren't scraped?

pgeorgi
0 replies
20h19m

For example in accounts on image sites that are exposed to suspected scrapers but not to others. Scrapers will still see the real data, but they'll also run into stuff designed to mix up the training process.

filleduchaos
0 replies
20h55m

I think the point is that they're akin to a watermark.

Even before the current AI boom, plenty of artists have wanted to showcase their work/prove that it exists without necessarily making the highest quality original file public.

Diti
0 replies
20h33m

Most serious artists I know (at least in my community) release their high-quality images on Patreon or similar.

the8472
1 replies
20h25m

do you mean scrapped or scraped?

GaryNumanVevo
0 replies
19h53m

scraped

soulofmischief
0 replies
15h51m

The artifacts are a non-issue.

According to which authority?

gedy
2 replies
23h22m

Maybe it's more about "protecting" images that artists want to publicly share to advertise work, but it's not appropriate for final digital media, etc.

sesm
1 replies
23h6m

In short, anti-AI watermark.

johnnyanmac
0 replies
19h10m

Yeah. It may mess with the artist's vision but the impact is still way more subtle than other methods used to protect against these unwanted actions.

Of course I'm assuming it works to begin with. Sounds like a game of cat and mouse. And AI has a lot of rich cats.

h0p3
0 replies
15h53m

Sir /u/b3nsn0w is courteous, `/nod`.

brucethemoose2
0 replies
23h42m

Yeah. At worst a simple img2img diffusion step would mitigate this, but just eyeballing the examples, traditional denoisers would probably do the job?

Denoising is probably a good preprocessing step anyway.

r3trohack3r
42 replies
22h41m

The number of people who are going to be able to produce high fidelity art with off the shelf tools in the near future is unbelievable.

It’s pretty exciting.

Being able to find a mix of styles you like and apply them to new subjects to make your own unique, personalized, artwork sounds like a wickedly cool power to give to billions of people.

kredd
25 replies
20h20m

In terms of art, population tends to put value not on the result, but origin and process. People will just look down on any art that’s AI generated in a couple of years when it becomes ubiquitous.

MacsHeadroom
10 replies
20h1m

Nope, but I already look down on artists who refuse to integrate generative AI into their processes.

MisterBastahrd
7 replies
19h31m

People who use generative AI in their processes are not artists.

password54321
3 replies
18h59m

This is true. They are just taking a sample from a generated latent space, just like taking a photo of something doesn't make you an artist.

blacklion
2 replies
18h33m

So, there is no artists in, for example, street photography? Picture must be altered to become art, or staged?

Was it irony? :)

password54321
0 replies
18h25m

They are photographers. Here is the definition of an artist so you can have better clarity on what an artist is:

"A person who creates art (such as painting, sculpture, music, or writing) using conscious skill and creative imagination"

aqfamnzc
0 replies
14h22m

I took gp as satire. But maybe not haha.

smackeyacky
0 replies
13h45m

I don’t think this is quite right. I think paraphrasing The Incredibles has a better take:

When everybody is an artist, then nobody will be one.

davely
0 replies
15h48m

I use generative AI to rubber duck and help improve my code.

Am I no longer a software engineer?

blacklion
0 replies
18h34m

And people who use Photoshop are?

There is somewhat famous digital artist from Russia - Alexey Andreev. Google it, he has very distinctive style of realistic technique and surrealistic situations, like landing big manta ray on the deck of aircraft carrier. Or you can see his old works in his 5-years-not-updates LJ [1].

Now he uses generative AI as one of his tools. As Photoshop, as different (unrealistic!) brushes in Photoshop, as other digital tools. His style is still 100% recognizable and his works don't become worse or more "generic". Is he still artist? I think so.

Where will you draw the line?

[1] - https://alexandreev.livejournal.com/

mplewis
0 replies
19h52m

Can you share some of the art you’ve made with generative AI?

jurynulifcation
0 replies
19h50m

Cool, who are you?

Theodores
4 replies
15h27m

https://en.wikipedia.org/wiki/Labor_theory_of_value

According to Marx, value is only created with human labour. This is not just a Marxist theory, it is an observation.

There may be lots of over-priced junk that makes you want to question this idea. But let's not nit-pick on that.

In two years time people will not see any value in AI art, quite correctly because there is not much human labour in creating it.

petesergeant
0 replies
13h13m

This is not just a Marxist theory, it is an observation.

Yeah? Well, you know, that's just like uh, your opinion, man

mesh
0 replies
14h52m

In two years time, no one will know what was created with AI, what was created by humans, or what was created by both.

jquery
0 replies
14h0m

Labor theory of value is quite controversial, many economists call it tautological or even metaphysical. I also don't really see what LTV has to say about AI art, if anything, except that the economic value generated by AI art should be distributed to everybody and not just funneled to a few capitalists at the top. I would agree with that. It's true that more jobs get created even as jobs are destroyed, but it's also true that just as our ancestors fought for a 40 hour work week and a social safety net, we should be able to ask for more as computers become ever so productive.

Gormo
0 replies
14h14m

According to Marx, value is only created with human labour. This is not just a Marxist theory, it is an observation.

And yet it's completely and absolutely wrong. Value is created by the subjective utility offered to the consumer, irrespective of what inputs created the thing conveying that utility.

redwall_hp
3 replies
19h50m

This is already the case. Art is a process, a form of human expression, not an end result.

I'm sure OpenAI's models can shit out an approximation of a new Terry Pratchett or Douglas Adams novel, but nobody with any level of literary appreciation would give a damn unless fraud was committed to trick readers into buying it. It's not the author's work, and there's no human message behind it.

torginus
0 replies
6h45m

Thing is there are way more good books written, than any single person can consume in their lifetimes. An average person like me, reading a mixed diet of classics, obscure recommendations and what's popular right now, I still don't feel like I'm making a dent in the pile of high quality written content.

Given all that, the purpose of LLMs should be to create tailor made content to everyone's tastes. However, it seems the hardcore guardrails put into GPT4 and Claude prevent it from generating anything enjoyable. It seems, even the plot of the average Star Wars movie is too spicy for modern LLM sensibilities, never mind something like Stephen King.

portaouflop
0 replies
14h3m

I haven’t read anything “shit out” by any LLM that even nearly approaches the level of quality by the authors you named — would very much like to see something like that - do you have any evidence for your claims?

AFAICT current text generation is something approaching bad mimicry at best and downright abysmal in general. I think you still need a very skilled author and meaty brain with a story to tell to make use of an LLM for storytelling. Sure it’s a useful tool that will make authors more effective but we are far from the point where you tell the LLM “write a story set in Pratchetts Discworld” and something acceptable or even entertaining will be spit out - if such a thing can even be achieved.

Aerroon
0 replies
14h15m

Novels aren't about a message. They're entertainment. If the novel is entertaining then it's irrelevant whether there is or isn't a message in it. Besides, literature enthusiasts will invent a message for a popular story even if there never was one.

Also, I'm sure that you can eventually just prompt the model with the message you want to put into the story, if you can't already do that.

Aerroon
3 replies
14h18m

I disagree. I definitely value modern digital art more than most historical art, because it just looks better. If AI art looks better (and in some cases it does) then I'll prefer that.

kredd
2 replies
12h8m

That’s totally fine, everyone’s definition of art is subjective. But general value of an art as a piece will just still be zero for AI generated ones, just like any IKEA / Amazon print piece. You just pay for the “looks pretty”, frame and paper.

Aerroon
1 replies
9h39m

You just pay for the “looks pretty”, frame and paper.

But you pay that for any piece of art though? You appreciate it because you like what it looks like. The utility of it is in how good it looks, it's not how much effort was put into it.

If you need a ditch you're not going to value the ditch more if the worker dug it by hand instead of using an excavator. You value it based on the utility it provides you.

kredd
0 replies
2h59m

That analogy doesn’t work for art, since worker’s ditch is result based. There are no feelings like “i like this ditch”, “experience of a ditch” or “i’m curious how this ditch was dug”.

Again, i’m not saying buying a mass made AI art will be wrong. Just personally speaking, it will never evoke any feelings other than “looks neat” for me. So its inherent “art value” is close to 0 as I can guess its history is basically someone put in a prompt and sent it to print (which I can do myself on my phone too!). It’s the same as looking at cool building pics on my phone (0 art value) versus actually seeing them in person (non-0), mostly because the feelings I get from it. That being said, if it makes others happy, it’s not my place to judge.

petesergeant
0 replies
16h21m

population tends to put value not on the result, but origin and process

I think population tends to value "looks pretty", and it's other artists, connoisseurs, and art critics who value origin and process. Exit Through the Gift Shop sums this up nicely

__loam
7 replies
22h16m

And we only had to alienate millions of people from their labor to do it.

DennisAleynikov
3 replies
22h4m

Yeah, sadly those millions of people don’t matter in the grand scheme of things and were never going to profit off their work long term

r3trohack3r
2 replies
21h43m

What a bummer of a thing to say.

Those millions/billions of people matter a great deal.

DennisAleynikov
1 replies
16h0m

They matter but not under the current system. Artists are a rarely paid profession, and there are professional artists out there but there’s now a huge amount of people that will never contact an artist for work that used to only be human powered. It’s not personal for me. I understand that desire to resist the inevitable but it’s here now.

For what it’s worth I never use midjourney or dalle or any of the commercial closed systems that steal from artists but I know I can’t stop the masses from going there and inputting “give me pretty picture in style x”

__loam
0 replies
6h14m

Resistance is important imo. If this happens and we, who work in this industry, say nothing, what good are we. It's only inevitable if it's socially acceptable.

r3trohack3r
0 replies
21h51m

Absolutely agree we should allow people to accumulate equity through effective allocation of their labor.

And I also agree that we shouldn’t build systems that alienate people from that accumulated equity.

mensetmanusman
0 replies
22h4m

Is this utilitarianism?

BeFlatXIII
0 replies
4h52m

Worth it.

password54321
3 replies
19h4m

Not really. There is a reason why we find realistic painting to be more fascinating than a photo and why some still practice it. The effort put in by another artist does affect our enjoyment.

dartharva
1 replies
13h14m

The word "we" is doing a lot of heavy lifting here. A large majority of consumers can't even tell apart AI-generated from handmade, let alone care who or what made the thing.

password54321
0 replies
5h36m

Yeah, that's just information you made up on the spot.

wruza
0 replies
14h10m

For me it doesn’t. I’m generating images, realistic, 2.5d, 2d and I like them as much. I don’t feel (or miss) what you described. Or what any other arts guy describes, for that matter. Arts people are different, because they were trained to feel something a normal person wouldn’t. And that’s okay, a normal person without training wouldn’t see how much beauty and effort there is in an algorithm or a legal contract as well.

23B1
2 replies
21h41m

It'll be about as wickedly tool as the ability to get on the internet, e.g. commoditized, transactional, and boring.

sebzim4500
1 replies
20h21m

I know this is an unpopular thing to say these days, but I still think the internet is amazing.

I have more access to information now than the most powerful people in the world did 40 years ago. I can learn about quantum field theory, about which pop star is allegedly fucking which other pop star, etc.

If I don't care about the law I can read any of 25 million books or 100 million scientific papers all available on Anna's Archive for free in seconds.

r3trohack3r
0 replies
16h24m

As Jeff Bezos recently said on the Lex podcast: one of the greatest compliments you can give an inventor is that they’re invention will be taken for granted by future generations.

“It won’t be any more wickedly cool than the internet” - saying something won’t be any more wickedly cool than the most profound and impactful pieces of infrastructure human civilization has erected is a pretty high compliment.

falcolas
0 replies
19h53m

Being able to find a mix of styles you like and apply them to new subjects to make your own unique, personalized, artwork sounds like a wickedly cool power to give to billions of people.

And in the process, they will obviate the need for Nightshade and similar tools.

AI models ingesting AI generated content does the work of destroying the models all by itself. Have a look at "Model Collapse" in relation to generative AI.

gfodor
27 replies
22h32m

Huge market for snake oil here. There is no way that such tools will ever win, given the requirements the art remain viewable to human perception, so even if you made something that worked (which this sounds like it doesn’t) from first principles it will be worked around immediately.

The only real way for artists or anyone really to try to hold back models from training on human outputs is through the law, ie, leveraging state backed violence to deter the things they don’t want. This too won’t be a perfect solution, if anything it will just put more incentives for people to develop decentralized training networks that “launder” the copyright violations that would allow for prosecutions.

All in all it’s a losing battle at a minimum and a stupid battle at worst. We know these models can be created easily and so they will, eventually, since you can’t prevent a computer from observing images you want humans to be able to observe freely.

AJ007
11 replies
21h6m

The level of claims accompanied by enthusiastic reception from a technically illiterate audience make it sound, smell, and sound like snake oil without much deep investigation.

There is another alternative to the law. Provide your art for private viewing only, and ensure your in person audience does not bring recording devices with them. That may sound absurd, but it's a common practice during activities like having sex.

wraptile
6 replies
12h10m

The thing is people want the benefits of having their stuff public but not bear the costs. Scraping has been mostly a solved problem especially when it comes to broad crawling. Put it under a login, there, no more AI "stealing" your work.

946789987649
3 replies
9h47m

Is that login statement strictly true? Unless the login is paid, there's no reason we can't get to (if not already there) the point where the AI scraper can just create a login first.

Tade0
1 replies
9h34m

But then you can rate-limit to a point where scraping everything will take a considerable amount of time.

Of course the workaround would be to have multiple accounts, but that in turn can be made unscalable with a "prove you're human" box.

csydas
0 replies
9h23m

you are not incorrect that this would help mitigate, but it still misses a few key points I think regarding why artists are upset about AI generation

- This is still vulnerable to stuff like mturk or even just normal users who did get past the anti-bot things pulling and re-uploading the content elsewhere that is easier for the AI companies to use

- The artists' main contention is that the AI companies shouldn't be allowed to just use whatever they find without confirm they have a license to use the content in this way

- If someone's content _does_ get into an AI model and it's determined somehow (I think there is a case with a news paper and chatGPT over this very issue?), the legal system doesn't really have a good framework for this situation right now -- is it copyright infringement? (arguably not? it's not clear) is it plagiarism? (arguably yes, but plagiarism in US court system is very hard to proof/get action on) is it license violation? (for those who use licenses for their art, probably yes, but it's the same issue as plagiarism -- how to prove it effectively?)

Really what this comes down to is that the AI companies use the premise that they have a right to use someone else's works without consent for the AI training. While your suggestions are technically correct, it puts the impetus on the artists that they must do something different because the AI companies are allowed to train their models as they currently do without recourse for the original artist. Maybe that will be ruled true in the future I don't know, but I can absolutely get why artists are upset about this premise shaping the discussion on AI training, as such a premise negates their rights as an artist and many artists have 0 path for recourse. I'm pretty sure that OpenAI wouldn't think about scraping a Disney movie from a video upload site just because it's open access since Disney likely can fight in a more meaningful way. I would agree with artists who are complaining that they shouldn't need to wait for a big corporation to decide that this behavior is undesirable before real action is taken, but it seems that is going to be what is needed. It might be reality, but it's a very sad reality that people want changed.

wraptile
0 replies
6h17m

No, eforcing click-wrap legal agreements is actually possible. With basic KYC the scraper would instantly open up itself for litigation and no internet art piece is frankly worth this sort of trouble.

csydas
1 replies
9h35m

I don't think that's true at all. Images and text get reposted with or without consent, often without attribution. It wouldn't make it right for the AI companies to scrape when the original author doesn't want that but someone else has ignored their wishes and requirements. Basically, what good is putting your stuff behind login or some other restrictive viewing method if someone just saves the image/text? I think it's still a relatively serious problem for people creating things. And without some form of easy access to viewing, the people creating things don't get the visibility and exposure they need to get an audience/clients.

This is one the AI companies should offer the olive branch on IMO, there must be a way to use stenography to transparently embed a "don't process for AI" code into an image or text or music or any other creative work that won't be noticeable by humans, but the AI would see if it tried to process the content for training. I think it would be a very convenient answer and probably not be detrimental to the AI companies, but I also imagine that the AI companies would not be very eager to spend the resources implementing this. I do think they're the best source for such protections for artists though.

Ideally, without a previous written agreement for a dataset from the original creators, the AI companies probably shouldn't be using it for training at all, but I doubt that will happen -- the system I mention above should be _opt-in_, that is, you must tag such content that is free to be AI trained in order for AI to be trained on it, but I have 0 faith that the AI companies would agree to such a self-limitation.

edit: added mention to music and other creative works in second paragraph 1st sentence

edit 2: Added final paragraph as I do think this should be opt-in, but don't believe AI companies would ever accept this, even though they should by all means in my opinion.

amlib
0 replies
6h4m

Here are my 2 cents, I think we will need some laws specifying two types of AI models, ones trained with full consent (opt-in) for its training material and ones without. The first one would be like Adobe's firefly model where they allegedly own everything they trained it with, or something where you go around asking for consent for each thing in your training corpus (probably unfeasible for large models). Maybe things in the public domain would be ok to train with. In this case there are no restrictions and the output from such models can even be copyrighted.

Now for the second type, representing models such as Stable Difusion and Chat GPT, it would be required to have their trained model freely available to anyone and any resulting output would not be copyrightable. It may be a more fairer way of allowing anyone to harness the power of AI models that contain essentially the knowledge of all man kind, but without giving any party an unfair monopoly on it.

This should be easily enforceable for big corporations, else it would be too obvious if they are trying to pass one type model as another or even keep the truth about their model from leaking. It might not be as easy to keep small groups or individuals from breaking those rules, but hey, at least it evens the playing field.

Art9681
1 replies
16h39m

This would just create a new market for art paparazzis who would find any and all means to inflitrate such private viewings with futuristic miniature cameras and other sensors and selling it for a premium. Less than 24 hours later the files end up on hundreds or thousands of centralized and decentralized servers.

I'm not defending it. Just acknowledging the reality. The next TMZ for private art gatherings is percolating in someone's garage at the moment.

jurassic
0 replies
7h33m

I find this difficult to believe; no matter how small your camera is, photography is about light. Art reproduction photography is surprisingly hard to do if you care about the quality of the end result. Unless you can surreptitiously smuggle in a studio lighting setup, tripod, and color checker card… sure you can take an image in secret, but not one that is a good representation of the real thing.

gfodor
0 replies
20h53m

True I can imagine that kind of thing becoming popular.

Gormo
0 replies
14h17m

That doesn't sound like a viable business model. There seems to be a non-trivial bootstrap problem involved -- how do you become well-known enough to attract audiences to private venues in sufficient volume to make a living? -- and would in no way diminish demand for AI-generated artwork which would still continue to draw attention away from you.

spaceman_2020
4 replies
11h18m

This is the hard reality. There is no putting this genie back in the bottle.

The only way to be an artist now is to have a unique style of your own, and to never make it online.

hutzlibu
3 replies
9h49m

"and to never make it online."

So then of course, you also cannot sell your work, as those might put it online. And you cannot show your art to big crowds, as some will make pictures and put it online. So ... you can become a literal underground artists, where only some may see your work. I think only some will like that.

But I actually disagree, there are plenty of ways to be an artist now - but most should probably think about including AI as a tool, if they still want to make money. But with the exception of some superstars, most artists are famously low on money - and AI did not introduce this. (all the professional artists I know, those who went to art school - do not make their income with their art)

sabedevops
1 replies
7h58m

Can you elaborate on how they supplement their income?

hutzlibu
0 replies
4h33m

Every other source of income? So other, art-unrelated jobs.

BeFlatXIII
0 replies
4h58m

GP almost certainly mean "make physical art." Pictures of that can get online, but it's not the real thing.

Reubend
3 replies
17h20m

Huge market for snake oil here.

This tool is free, and as far as I can tell it runs locally. If you're not selling anything, and there's no profit motive, then I don't think you can reasonably call it "snake oil".

At worst, it's a waste of time. But nobody's being deceived into purchasing it.

golol
0 replies
3h54m

Religion is also deceptive and snake-oil even if it does not involve profit driven motivations.

autoexec
0 replies
12h31m

If this is a danger from "snake oil" of this type, it'd be from the other side, where artists are intentionally tricked into believing that tools like this mean that AI isn't or won't be a threat to their copyrights in order to get them to stop opposing it so strongly, when in fact the tool does nothing to prevent their copyrights from being violated.

I don't think that's the intention of Nightshade, but I wouldn't put past someone to try it.

Biganon
0 replies
6h47m

There's an academic paper being published.

Snake oil for the sake of getting published is a very real problem that does exist.

AlfeG
1 replies
12h40m

My guess. Is that at some poi t of time You will not be able to use any generated image or video in commercial. Because of 100% copyright claim for using parts of copyrighted image. Like YouTube those days. When some random beeps matches with someone music...

abrarsami
0 replies
12h10m

It should be like that. I agree

thfuran
0 replies
20h33m

There is no way that such tools will ever win, given the requirements the art remain viewable to human perception

On the other hand, the adversarial environment might push models towards a representation more aligned with human perception, which is neat.

jedberg
0 replies
13h12m

Everything old is new again. It's the same thing with any DRM that happens on the client side. As long as it's viewable by humans, someone will figure out a way to feed that into a machine.

jMyles
0 replies
1h32m

leveraging state backed violence to deter the things they don’t want

I just want to say: I really appreciate the stark terms in which you've put this.

The thing that has come to be called "intellectual property" is actually just a threat of violence against people who arrange bytes in a way that challenges power structures.

aqfamnzc
0 replies
17h31m
Quanttek
26 replies
1d3h

This is fantastic. If companies want to create AI models, they should license the content they use for the training data. As long as there are not sufficient legal protections and the EU/Congress do not act, tools like these can serve as a stopgap and maybe help increase pressure on policymakers

Kuinox
17 replies
1d3h

they should license the content they use for the training data

You mean like OpenAI and Adobe ?

Only the free and open source models didn't licensed any content for the training data.

galleywest200
14 replies
1d3h

Adobe is training off of images stored in their cloud systems, per their Terms of Service.

OpenAI has provided no such documentation or legal guarantees, and it is still quite possible they scraped all sorts of copyright materials.

luma
10 replies
1d2h

Google scrapes copyrighted material every day and then presents that material to users in the form of excerpts, images, and entire book pages. This has been ruled OK by the courts. Scraping copyrighted information is not illegal or we couldn't have search engines.

kevingadd
7 replies
1d2h

Google is not presently selling "we trained an AI on people's art without permission, and you can type their name in along with a prompt to generate a knockoff of their art, and we charge you money for this". So it's not really a 1:1 comparison, since there are companies selling the thing I described right now.

luma
6 replies
1d

That pretty clearly would fall under transformative work. It is not illegal for a human to paint a painting in the style of, say, Banksy, and then sell the resulting painting.

kevingadd
5 replies
1d

Humans and AI are not the same thing, legally or physically. The law does not currently grant AI rights of any kind.

luma
3 replies
23h53m

If a human isn't violating the law when doing that thing, then how is the machine violating the law when it cannot even hold copyright itself?

kevingadd
1 replies
23h53m

I'm not sure how to explain this any clearer: Humans and machines are legally distinct. Machines don't have the rights that humans have.

Ukv
0 replies
22h29m

Fair Use is the relevant protection and is not specific to manual creation. Traditional algorithms (e.g: the snippets, caching, and thumbnailing done by search engines) are already covered by it.

estebank
0 replies
22h56m

In some locales sitting on the street writing down a list of people coming and going is legal, but leaving a camera pointed at the street isn't. Legislation like that makes a distinction between an action by a person (which has bounds on scalability) and mechanized actions (that do not).

ufocia
0 replies
21h52m

What's not prohibited is allowed, at least in the US.

ufocia
1 replies
21h53m

Scraping is only legal if it's temporary and transformational. If Google started selling the scrapped images it would be a different story.

Kuinox
0 replies
18h11m

What is not transformational for generative AI ?

mesh
0 replies
15h32m

No they are not. They train their models on Adobe Stock content. They do not train on user content.

https://helpx.adobe.com/manage-account/using/machine-learnin...

"The insights obtained through content analysis will not be used to re-create your content or lead to identifying any personal information."

"For Adobe Firefly, the first model is trained on Adobe Stock images, openly licensed content, and public domain content where the copyright has expired."

(I work for Adobe)

devmor
0 replies
1d3h

There is in fact, an extreme amount of circumstantial evidence that they intentionally and knowingly violated copyright en mass. It’s been quite a popular subject in tech news the past couple weeks.

Kuinox
0 replies
1d3h

OpenAI has provided no such documentation

OpenAI and Shutterstocks publicly announced their collaboration, Shutterstocks sells AI generated images, generated with OpenAI models.

jazzyjackson
0 replies
1d2h

source for OpenAI paying anyone a dime? don't you think that would set a precedent that everyone else deserves their cut?

KeplerBoy
0 replies
1d3h

There is a small difference between any and all. OpenAI certainly didn't licence all of the image they use for training.

popohauer
7 replies
1d2h

It's going to be interesting to see how the lawsuits against OpenAI by content creators plays out. If the courts rule that AI generated content is a derivative work of all the content it was trained on it could really flip the entire gen AI movement on its head.

luma
4 replies
1d2h

If it were a derivative work[1] (and sufficiently transformational) then it's allowed under current copyright law and might not be the slam dunk ruling you were hoping for.

[1] https://en.wikipedia.org/wiki/Derivative_work

kevingadd
1 replies
1d2h

"sufficiently transformational" is carrying a lot of water here. At minimum it would cloud the issue and might expose anyone using AI to lawsuits where they'd potentially have to defend each generated image.

ufocia
0 replies
21h56m

Sufficiently transformational only applies to copyrightability, but AI works are not copyrightable under current US law, so it's a non-issue.

ufocia
0 replies
21h58m

Not if it is AI generated. So far only humans can be original enough to warrant copyrights, at least in the US .

BTW, the right to prepare derivative works belongs to the copyright holder of the reference work.

I doubt that many AI works are in fact derivative works. Sure, some bear enough similarity, but a gross majority likely doesn't.

popohauer
0 replies
22h11m

Oh, interesting, I didn't realize that's how it worked. Thanks for the additional context around this. Guess it's not as upending as I thought it could be.

torginus
0 replies
6h34m

My biggest fear is that the big players will drop a few billion dollars to silence the copyright holders with power go away, and new rules are put in place that will make open-source models that can't do the same essentially illegal.

BeFlatXIII
0 replies
4h50m

…then I'll keep enjoying my Stable Diffusion and pirated models.

ukuina
20 replies
23h53m

Won't a simple downsample->upsample be the antidote?

wizzwizz4
17 replies
23h51m

How do you train your upsampler? (Also: why are you seeking to provide an “antidote”?)

MrNeon
11 replies
23h41m

why are you seeking to provide an “antidote”

To train a model on the data.

krapp
10 replies
23h40m

Get permission to use the data.

MrNeon
9 replies
23h17m

Got all the permission I need when it was put on a publicly accessible server.

xigoi
5 replies
20h39m

That’s not how copyright works.

MrNeon
4 replies
20h3m

Tell me where it says training a model is infringing on copyright.

xigoi
3 replies
20h0m

How is creating a derivative of someone’s work and selling it not copyright infringement?

MrNeon
2 replies
19h40m

Who said anything about creating a derivative? Surely you don't mean to say that any image created with a model trained on copyrighted data counts as a derivative of it. Edit: Or worse, that the model itself is derivative, something so different from an image must count as transformative work!.

Also who said anything about selling?

xigoi
1 replies
19h32m

The model itself is a derivative. And it’s not really that transformative, it’s basically the input data compressed with highly lossy compression.

Also who said anything about selling?

All the corporations that are offering AI as a paid service?

MrNeon
0 replies
19h1m

it’s basically the input data compressed with highly lossy compression.

Okay, extract the images from a Stable Diffusion checkpoint then. I'll wait.

It's not like lossy compression CAN'T be fair use or transformative. I'm sure you can imagine how that is possible given the many ways an image can be processed.

All the corporations that are offering AI as a paid service?

Am I them?

wizzwizz4
2 replies
23h10m

That's not really how consent works.

I hope this is a special exception you've made, rather than your general approach towards interacting with your fellows.

MrNeon
1 replies
22h37m

That is how consent works.

CatWChainsaw
0 replies
8m

If my body is in a public space you have the right to force me to have sex with you, I guess? That's your logic after all.

klyrs
2 replies
23h29m

why are you seeking to provide an “antidote”

I think it's worthwhile for such discussion to happen in the open. If the tool can be defeated through simple means, it's better for everybody to know that, right?

wizzwizz4
1 replies
23h20m

It would be better for artists to know that. But Hacker News is not a forum of visual artists: it's a forum of hackers, salaried programmers, and venture capitalists. Telling the bad guys about vulnerabilities isn't responsible disclosure.

Causing car crashes isn't hard (https://xkcd.com/1958/). That doesn't mean Car Crash™ International®'s decision-makers know how to do it: they probably don't even know what considerations go into traffic engineering, or how anyone can just buy road paint from that shop over there.

It's everybody's responsibility to keep Car Crash™ International® from existing; but failing that, it's everybody's responsibility to not tell them how to cause car crashes.

MrNeon
0 replies
20h2m

The tears of artists and copyright evangelists is so sweet.

ukuina
0 replies
21h36m

I apologize. I was trying to respond to inflammatory language ("poison") with similarly hyperbolic terms, and I should know better than to do that.

Let me rephrase: Would AI-powered upscaling/downscaling (not a simple deterministic mathematical scaling) not defeat this at a conceptual level?

spookie
0 replies
23h49m

Why would you train one?

jdiff
1 replies
23h13m

No, it's resistant to transformation. Rotation, cropping, scaling, the image remains poisonous. The only antidote known currently is active artist cooperation.

CaptainFever
0 replies
20h46m

Or Img2Img.

ang_cire
12 replies
23h19m

Setting aside the efficacy of this tool, I would be very interested in the legal implications of putting designs in your art that could corrupt ML models.

For instance, if I set traps in my home which hurt an intruder we are both guilty of crimes (traps are illegal and are never considered self defense, B&E is illegal).

Would I be responsible for corrupting the AI operator's data if I intentionally include adversarial artifacts to corrupt models, or is that just DRM to legally protect my art from infringement?

edit:

I replied to someone else, but this is probably good context:

DRM is legally allowed to disable or even corrupt the software or media that it is protecting, if it detects misuse.

If an adversarial-AI tool attacks the model, it then becomes a question of whether the model, having now incorporated my protected art, is now "mine" to disable/corrupt, or whether it is in fact out of bounds of DRM.

So for instance, a court could say that the adversarial-AI methods could only actively prevent the training software from incorporating the protected media into a model, but could not corrupt the model itself.

anigbrowl
4 replies
23h15m

None whatsoever. There is no right to good data for model training, nor does any contractual relationship exist between you and and a model builder who scrapes your website.

ang_cire
3 replies
22h30m

If you're assuming this is open-shut, you're wrong. I asked this specifically as someone who works in security. A court is going to have to decide where the line is between DRM and malware in adversarial-AI tools.

anigbrowl
1 replies
20h36m

I'm not. Malware is one thin, passive data poisoning is another. Mapmakers have long used such devices to detect/deter unwanted copying. In the US such 'trap streets' are not protected by copyright, but nor do they generate liability.

https://en.wikipedia.org/wiki/Trap_street

ang_cire
0 replies
6h10m

A trap street doesn't damage other data. Not even remotely useful as an analogy. That's to allow detection of copies, not to corrupt the copies from being useable.

ufocia
0 replies
22h7m

Worth trying but I doubt it unless we establish a right to train.

kortilla
2 replies
23h2m

That’s like asking if lying on a forum is illegal

ang_cire
1 replies
22h31m

No, it's much closer to (in fact, it is simply) asking if adversarial AI tools count as DRM or as malware. And a court is going to have to decide whether the model and or its output counts as separate software, which it is illegal for DRM to intentionally attack.

DRM can, for instance, disable its own parent tool (e.g. a video game) if it detects misuse, but it can't attack the host computer or other software on that computer.

So is the model or its output, having been trained on my art, a byproduct of my art, in which case I have a legal right to 'disable' it, or is it separate software that I don't have a right to corrupt?

danShumway
0 replies
16h55m

asking if adversarial AI tools count as DRM or as malware

Neither. Nightshade is not DRM or malware, it's "lying" about the contents of an image.

Arguably, Nightshade does not corrupt or disable the model at all. It feeds it bad data that leads the model to generate incorrect conclusions or patterns about how to generate images. This is assuming it works, which we'll have to wait and see, I'm not taking that as a given.

But the only "corruption" happening here is that the model is being fed data that it "trusts" without verifying that what the data is "telling" it is correct. It's not disabling the model or crashing it, the model is forming incorrect conclusions and patterns about how to generate the image. If Google translate asked you to rate its performance on a task, and you gave it an incorrect rating from what you actually thought its performance was, is that DRM? Malware? Have you disabled Google translate by giving it bad feedback?

I don't think the framing of this as either DRM or malware is correct. This is bad training data. Assuming it works, it works because it's bad training data -- that's why ingesting one or two images doesn't affect models but ingesting a lot of images does, because training a model on bad data leads the model to perform worse if and only if there is enough of that bad data. And so what we're really talking about here is not a question of DRM or malware, it's a question of whether or not artists have a legal obligation to make their data useful for training -- and of course they don't. The implications of saying that they did would be enormous, it would imply that any time you knowingly lied about a question that was being fed into an AI training set that doing so was illegal.

npteljes
0 replies
18h54m

I see it as no different than mapmakers inventing a nonexistent alley, to check who copies their maps verbatim ("trap street"). Even if this caused, for example, a car crash because of an autonomous driver, the onus I think would be on the one that made the car and used the stolen map for navigation, and not on the one that created the original map.

https://en.wikipedia.org/wiki/Trap_street

danShumway
0 replies
17h9m

The way Nightshade works (assuming it does work) is by confusing the features of different tags with each other. To argue that this is illegal would be to argue that mistagging a piece of artwork on a gallery is illegal.

If you upload a picture of a dog to DeviantArt and you label it as a cat, and a model ingests that image and starts to think that cats look like dogs, would anybody claim that you are breaking a law? If you upload bad code to Github that has bugs, and an AI model consumes that code and then reproduces the bugs, would anyone argue that uploading badly written code to Github is a crime?

What if you uploaded some bad code to Github and then wrote a comment at the top of the code explaining what the error was, because you knew that the model would ignore that comment and would still look at the bad code. Then would you be committing a crime by putting that code on Github?

Even if it could be proven that your intention was for that code or that mistagged image to be unhelpful to training, it would still be a huge leap to say that either of those activities were criminal -- I would hope that the majority of HN would see that as a dangerous legal road to travel down.

GaryNumanVevo
0 replies
23h1m

How would that situation be remotely related?

CaptainFever
0 replies
20h52m

Japan is considering it, I think? https://news.ycombinator.com/item?id=38615280

tbalsam
9 replies
16h11m

I put together a rather long Twitter post about this, but basically, (and I am not a lawyer), using Nightshade in most usecases (and especially as the authors are encouraging its use) is blatantly illegal under currently-very-well-established-law. Several lines of extremely well-established legal precedent are violated, including but not limited to Duty of Care.

Duty of Care is the legal line of reasoning that states it is illegal to poison your own sandwich if a coworker is stealing your lunch food, to give one example (bizzare as it may seem).

Poisoning your data carries significant risks.

All of this is my personal opinion, none of this is legal advice.

Source, or, for more info: https://en.m.wikipedia.org/wiki/Duty_of_care#:~:text=In%20to....

BadHumans
5 replies
16h1m

I'm not sold on your argument. I'm not an artist but I don't see how an artist using Nightshade is breaking the law. From an anti-AI point of view, you illegally took my artwork and used it. How is it my fault you didn't understand what you were stealing?

tbalsam
4 replies
15h51m

Currently it's not legally defined what is stealing and what is fair use with respect to this, which is of course why k feel it is such a strong issue, however, poisoning data and intentionally masking it to hide said poisoning is rather blatantly illegal.

Rather clearly I think most people support individual IP protection and that's not really a contested issue, however what is fair use and where things fall in that gray area is where things do get dicey.

thethimble
1 replies
15h28m

I wonder how your original claim balances with something like freedom of speech/expression? Ultimately it’s art and perhaps the Nightshade-modified version is a component of what makes the piece art?

tbalsam
0 replies
15h2m

Nightshade is intended to make minimal changes so I think that might be one argument though a difficult sell to some.

The original post is mainly about the current legality, I'm torn as to what the moral balance is, but poisoning data is similar to a lot of historical asymmetrical methods of topical influence through intentional property destruction, and that's an issue.

For example, the counterintuitive example of you poisoning a sandwich if your coworker is eating it makes you liable even if the original action is wrong on the coworker's part, if that makes sense.

It's not necessarily an intuitive legal construct but a necessary one at least.

That said, I think w.r.t. a final solution, grassroots vs big money legal will be the long-term playing field and I think efforts would be well spent on that. Though I worry about the perceived desperation of the field if people are this ready to poison open-field data in order to protect what they determine their own interests to be.

I think that's the big marker here that we should look at and be worried about, usually when ignored that tends to preclude different types of issues/problems.

BadHumans
1 replies
15h28m

The "blatantly illegal" part is the part I'm not understanding. I drew something, I did some processing on it, I put it online, someone grabbed it and used it without my permission. I'm not clear on the illegal part.

Rather clearly I think most people support individual IP protection and that's not really a contested issue

I think poking around HN or just this thread alone would tell you that at least people here don't support that.

tbalsam
0 replies
15h8m

I would recommend at least on the legal side of things looking into the Duty of Care history, it's not necessarily intuitive though I think it makes sense in the end. As another person said, oftentimes it comes down to intent (though negligence can play a role in the matter as I personally understand).

mattgreenrocks
1 replies
15h59m

Why is it my responsibility to ensure my data doesn’t inadvertently harm your model that I have zero knowledge of?

tbalsam
0 replies
15h53m

Duty of Care, additionally from an alternative perspective there are a number of laws covering the transmission of data that is intended to cause harm to another system.

tbalsam
0 replies
15h31m

(I understand that this is not a popular point, but I really want to emphasize that I am talking about what is _currently legal_ right now, not at all about the ethicality of large companies using people's data. The latter is a much harder topic.

This is mainly about what the law considers to be legal or not legal and is trying to avoid the more emotional side of the topic.)

enord
8 replies
23h37m

I’m completely flabbergasted by the number of comments implying copyright concepts such as “fair use” or “derivative work” apply to trained ML models. Copyright is for _people_, as are the entailing rights, responsibilities and exemptions. This has gone far beyond anthropomorphising and we need to like get it together, man!

ronsor
5 replies
22h51m

You act like computers and ML models aren't just tools used by people.

enord
4 replies
17h55m

What did I write to give you that impression?

Ukv
3 replies
15h34m

My initial interpretation was that you're saying fair use is irrelevant to the situation because machine learning models aren't themselves legal persons. But, fair use doesn't solely apply to manual creation - use of traditional algorithms (e.g: the snippets, caching, and thumbnailing done by search engines) is still covered by fair use. To my understanding, that's why ronsor pointed out that ML models are tools used by people (and those people can give a fair use defense).

Possibly you instead meant that fair use is relevant, but people are wording remarks in a way that suggests the model itself is giving a fair use defence to copyright infringement, rather than the persons training or using it?

enord
2 replies
5h49m

Well then I could have been much clearer because I meant something like the latter.

An ML model can neither have nor be in breach of copyright so any discussion about how it works, and how that relates to how people work or “learn” is besides the point.

What actually matters is firstly details about collation of source material, and later the particular legal details surrounding attribution. The last part involves breaking new ground legally speaking and IANAL so I will reserve judgement. The first part, collation of source material for training is emphatically not unexplored legal or moral territory. People are acting like none of the established processes apply in the case of LLMs and handwave about “learning” to defend it.

Ukv
1 replies
4h16m

and how that relates to how people work or “learn” is besides the point

It is important (for the training and generation stages) to distinguish between whether the model copies the original works or merely infers information from them - as copyright does not protect against the latter.

The first part, collation of source material for training is emphatically not unexplored legal or moral territory.

Similar to as in Authors Guild v. Google, Inc. where Google internally made entire copies of millions of in-copyright books:

> While Google makes an unauthorized digital copy of the entire book, it does not reveal that digital copy to the public. The copy is made to enable the search functions to reveal limited, important information about the books. With respect to the search function, Google satisfies the third factor test

Or in the ongoing Thomson Reuters v. Ross Intelligence case where the latter used the former's legal headnotes for training a language model:

> verbatim intermediate copying has consistently been upheld as fair use if the copy is "not reveal[ed] to the public."

That it's an internal transient copy is not inherently a free pass, but it is something the courts take into consideration, as mentioned more explicitly in Sega v. Accolade:

> Accolade, a commercial competitor of Sega, engaged in wholesale copying of Sega's copyrighted code as a preliminary step in the development of a competing product [yet] where the ultimate (as opposed to direct) use is as limited as it was here, the factor is of very little weight

And, given training a machine learning model is a considerably different purpose than what the images were originally intended for, it's likely to be considered transformative; as in Campbell v. Acuff-Rose Music:

> The more transformative the new work, the less will be the significance of other factors
enord
0 replies
9m

Listen, most website and book-authors want to be indexed by google. It brings potential audience their way, so most don’t make use of their _right_ to be de-listed. For these models, there is no plausible benefit to the original creators, and so one has to argue they have _no_ such right to be “de-listed” in order to get any training data currently under copyright.

CaptainFever
1 replies
20h47m

No one is saying a model is the legal entity. The legal entities are still people and corporations.

enord
0 replies
17h58m

Oh come on, you’re being insincere. Wether or not the model is learning from the work just like people is hotly debated as if it would make a difference. Fair use is even brought up. Fair use! Even if it applied, these training sets collate all of everything

I feel like I’m taking crazy pills TBQH

devmor
8 replies
1d3h

Baffling to see anyone argue against this technology when it is a non-issue to any model by simply acquiring only training data you have permission to use.

krapp
3 replies
1d2h

The reason people are arguing against this technology is that no one is using them in the way you describe. They actually wouldn't even be economically viable in that case.

devmor
2 replies
22h25m

If it is not economically viable for you to be ethical, then you do not deserve economic success.

Anyone arguing against this technology following the line of reasoning you present is operating in adverse to the good of society. Especially if their only motive is economic viability.

krapp
1 replies
21h59m

I feel like you read my comment and interpreted it in exactly the opposite way it was intended because I agree with you, and you're making the same point I was trying to make.

devmor
0 replies
18h23m

You are right, I read it as a defense rather than an explanation.

Ukv
2 replies
22h7m

I think people 100% have the right to use this on their images, but:

simply acquiring only training data you have permission to use

Currently it's generally infeasible to obtain licenses at the required scale.

When attempting to develop a model that can describe photos for visually impaired users, I had even tried to reach out to obtain a license from Getty. They repeatedly told me that they don't license images for machine learning[0].

I think it's easy to say "well too bad, it doesn't deserve to exist" if you're just thinking about DALL-E 3, but there's a huge number of positive and far less-controversial applications of machine learning that benefit from web-scale pretraining and foundation models - spam filtering, tumour segmentation, voice transcription, language translation, defect detection, etc.

[0]: https://i.imgur.com/iER0BE2.png

devmor
1 replies
18h20m

I don't believe it's a "doesn't deserve to exist" situation, because these things genuinely can be used for the public good.

However - and this is a big however - I don't believe it deserves the legal protection to be used for profit.

I am of the opinion that if you train your model on data that you do not hold the rights for, your usage should be handled similarly to most fair use laws. It's fine to use it for your personal projects, for research and education, etc. but it is not OK to use it for commercial endeavors.

Ukv
0 replies
17h58m

It's fine to use it for your personal projects, for research and education, etc. but it is not OK to use it for commercial endeavors.

Say I train a machine vision model that, after having pretrained on ImageNet or similar, detects deformities in a material for a small company that manufactures that material. Do you not think that would be fair use, despite being commercial?

To me it seems highly transformative (a defect detection model is entirely outside the original images' purposes) and does not at all impact the market of the works.

Moreover, you said it was "Baffling to see anyone argue against this technology" but it seems there are at least some models (like if my above detector was non-commercial) that you're ethically okay with and could be affected by this poisoning.

notfed
0 replies
17h26m

I don't know if asking permission of every copyright holder of every image on the Internet is as simple as you're implying.

alentred
6 replies
1d

With this "solution" it looks like the world of art enters the cat-and-mouse game the ad blockers were playing for the last decade or two.

isodev
3 replies
23h47m

I just tested it with Azure AI image classification and it worked - so this cat is yet to adapt to the mouse’s latest idea.

I still feel it is absolutely wrong to roam around the internet and scrape images (without consent) in order to power one’s cash cow AI. I hope more methods to protect artworks (including audio and other formats) become more accessible.

HKH2
2 replies
16h33m

Artists copy from each other all the time. Arguably, culture exists because of copying (folk stories by necessity); copyright makes culture top-down and stagnant, and you can't avoid it because they have the money to shove it right in your face. Who wants trickle-down culture?

blibble
1 replies
15h15m

it's not an artist, it's a piece of software

in the same way bittorrent or gzip is

HKH2
0 replies
5h35m

Sure. The person using it has intent. Now we have come to a point in which intent alone is art. Let there be light.

KTibow
1 replies
23h39m

I might be missing something because I don't know much about the architecture of either Nightshade or AI art generators, but I wonder if you could try to have a GAN-like architecture (an extra model trying to trick the model) for the part of the generator that labels images to build resistance to Nightshade-like filters.

the8472
0 replies
23h0m

It doesn't even have to be a full GAN, you only need to train the discriminator side to filter out the data. Clean reference images + Nightshade would be the generator side.

popohauer
5 replies
1d3h

I'm glad to see tools like Nightshade starting to pop up to protect the real life creativity of artists. I like AI art, but I do feel conflicted about its potential long term effects towards a society that no longer values authentic creativity.

Minor49er
4 replies
23h30m

Is the existence of the AI tool not itself a product of authentic creativity? Does eliminating barriers to image generation not facilitate authentic creativity?

23B1
3 replies
21h38m

No, it facilitates commoditization. Art – real art – is fundamentally a human-to-human transaction. Once everyone can fire perfectly-rendered perfectly-unique pieces of 'art' at each other, it'll just become like the internet is today: filled with extremely low-value noise.

Enjoy the short term novelty while you can.

BeFlatXIII
1 replies
4h51m

Art – real art – is fundamentally a human-to-human transaction.

Why is this hippie nonsense so popular?

23B1
0 replies
3h3m

Because some things are different than others, even though they might have the same word to describe them.

fulladder
0 replies
19h39m

This is the right prediction. Once machines can generate visual art, people will simply stop valuing it. We may see increased interest in other forms of art, e.g., live performance art like theater. It's hard to predict exactly how it'll play out, but once something becomes cheap to produce and widely available, it loses its luster for connoisseurs and then gradually loses its luster for everybody else too.

peter_d_sherman
5 replies
23h50m

To protect an individual's image property rights from image generating AI's -- wouldn't it be simpler for the IETF (or other standards-producing group) to simply create an

AI image exclusion standard

, similar to "robots.txt" -- which would tell an AI data-gathering web crawler that a given image or set of images -- was off-limits for use as data?

https://en.wikipedia.org/wiki/Robots.txt

https://www.ietf.org/

xg15
2 replies
23h44m

And then what? The scrapers themselves already happily ignore copyright, they won't be inclined to obey a no-ai.txt. So someone would have to enforce the standard. Currently I see no organisation who would be willing to do this or even just technologically able - as even just detecting such scrapers is an extremely hard task.

Nevertheless, I hope that at some not-so-far point in the future there will be more legal guidance about this kind of stuff, i.e. it will be made clear that scraping violates copyright. This still won't solve the problem of detectability but it would at least increase the risk of scrapers, should they be caught.

peter_d_sherman
1 replies
22h49m

The scrapers themselves already happily ignore copyright, they won't be inclined to obey a no-ai.txt.

Name two entities that were asked to stop using a given individuals' images that failed to stop using them after the stop request was issued.

Currently I see no organisation who would be willing to do this or even just technologically able - as even just detecting such scrapers is an extremely hard task.

// Part of Image Web Scraper For AI Image Generator ingestion psuedocode:

if fileExists("no-ai.txt") {

  // Abort image scraping for this site -- move on to the next site
} else {

  // Continue image scraping for this site
};

See? Nice and simple!

Also -- let me ask you this -- what happens to the intellectual property (or just plain property) rights of Images on the web after the author dies? Or say, 50 years (or whatever the legal copyright timeout is) after the author dies?

Legal grey area perhaps?

Also -- what about Images that exist in other legal jurisdictions -- i.e., other countries?

How do we know what set of laws are to apply to a given image?

?

Point is: If you're going to endorse and/or construct a legal framework (and have it be binding -- keep in mind you're going to have to traverse the legal jurisdictions of many countries, many countries!) -- you might as well consider such issues.

Also -- at least in the United States, we have Juries that can override any Law (Separation of Powers) -- that is, that which is considered "legally binding" -- may not be quite so "legally binding" if/when properly explained to a proper jury in light of extenuating (or just plain other) circumstances!

So kindly think of these issues prior to making all-encompasing proposals as to what you think should be "legally binding" or not.

I comprehend that you are just trying to solve a problem; I comprehend and empathize; but the problem might be a bit greater than you think, and there might be one if not serveral unexplored partial/better (since no one solution, legal or otherwise, will be all-encompassing) solutions -- because the problem is so large in scope -- but all of these issues must be considered in parallel -- or errors, present or future will occur...

xg15
0 replies
21h32m

Part of Image Web Scraper For AI Image Generator ingestion psuedocode:...

Yes, and who is supposed to run that code?

Name two entities that were asked to stop using a given individuals' images that failed to stop using them after the stop request was issued.

Github? OpenAI?[1] Stable Diffusion?[2] LAION?[3] What do you think why there are currently multiple high-profile lawsuits ongoing about exactly that topic?

Besides, that's not how things work. Training a foundation model takes months and currently costs a fortune in hardware and power - and once the model is trained, there is, as of now, no way to remove individual images from the model without restraining. So in practical terms it's impossible to remove an image if it has already been trained on.

So the better question would be, name two entities who have ignored an artist's request to not include their image when they encountered it the first time. It's still a trick question though because the point is that scraping happens in private - we can't know which images were scraped without access to the training data. The one indication that it was probably scraped is if a model manages to reproduce it verbatim - which is the basis for some of the above lawsuits.

[1] https://www.theverge.com/2022/11/8/23446821/microsoft-openai...

[2] https://www.theverge.com/2023/2/6/23587393/ai-art-copyright-...

[3] https://www.heise.de/hintergrund/Stock-photographer-sues-AI-...

potatolicious
1 replies
23h46m

Entities training models have no incentive to follow such metadata. If we accept the premise that "more input -> better models" then there's every reason to ignore non-legally-binding metadata requests.

Robots.txt survived because the use of it to gatekeep valuable goodies was never widespread. Most sites want to be indexed, most URLs excluded by the robots file are not of interest to the search engine anyway, and use of robots to prevent crawling actually interesting pages is marginal.

If there was ever genuine uptake in using robots to gatekeep the really good stuff search engines would've stopped respecting it pretty much immediately - it isn't legally binding after all.

peter_d_sherman
0 replies
23h1m

Entities training models have no incentive to follow such metadata. If we accept the premise that "more input -> better models" then there's every reason to ignore non-legally-binding metadata requests.

Name two entities that were asked to stop using a given individuals' images that failed to stop using them after the stop request was issued.

Robots.txt survived because the use of it to gatekeep valuable goodies was never widespread. Most sites want to be indexed, most URLs excluded by the robots file are not of interest to the search engine anyway, and use of robots to prevent crawling actually interesting pages is marginal.

Robots.txt survived because it was a "digital signpost" a "digital sign" -- sort of like the way you might put a "Private Property -- No Trespassing" sign in your yard.

Most moral/ethical/lawful people -- will obey that sign.

Some might not.

But the some that might not -- probably constitute about a 0.000001% minority of the population, whereas the majority that do -- probably constitute about 99.99999% of the population.

"Robots.txt" is a sign -- much like a road sign is.

People can obey them -- or they can ignore them -- but they can ignore them only at their own peril!

It's a sign which provides a hint for what the right thing to do in a certain set of circumstances -- which is what the Law is; which is what the majority of Laws are.

People can obey them -- or they can choose to ignore them -- but only at their own peril!

Most will choose to obey them. Most will choose to "take the hint", proverbially speaking!

A few might not -- but that doesn't mean the majority won't!

If there was ever genuine uptake in using robots to gatekeep the really good stuff search engines would've stopped respecting it pretty much immediately - it isn't legally binding after all.

Again, name two entities that were asked to stop using a given individuals' images that failed to stop using them after the stop request was issued.

jamesu
5 replies
1d2h

Long-term I think the real problem for artists will be corporations generating their own high quality targeted datasets from a cheap labor pool, completely outcompeting them by a landslide.

ufocia
3 replies
22h13m

It will democratize art.

23B1
1 replies
21h38m

then it won't be art anymore, it'll just be mountains of shit

sorta like what the laptop did for writing

jrflowers
0 replies
17h10m

This is a good point. There hasn’t been any writing since the release of the Gateway Solo in 1995

sussmannbaka
0 replies
9h55m

Art is already democratized. It has been for decades. Everyone can pick it up at zero cost. Even you!

The poorest people have historically produced great art. Training a model, however? Expensive. Running it locally? Expensive. Paying the sub? Expensive.

Nothing is being democratized, the only thing this does is devaluing the blood and sweat people have put into their work so FAANG can sell it to lazy suckers.

jdietrich
0 replies
21h30m

In the short-to-medium term, we're seeing huge improvements in the data efficiency of generative models. We haven't really started to see self-training in diffusion models, which could improve data efficiency by orders of magnitude. Current models are good at generalisation and are getting better at an incredible pace, so any efforts to limit the progress of AI by restricting access to training data is a speedbump rather than a roadblock.

eigenvalue
5 replies
19h56m

This seems like a pretty pointless "arms race" or "cat and mouse game". People who want to train generative image models and who don't care about what artists think about it at all can just do some basic post-processing on the images that is just enough to destroy the very carefully tuned changes this Nightshade algorithm makes. Something like resampling it to slightly lower resolution and then using another super-resolution model on it to upsample it again would probably be able to destroy these subtle tweaks without making a big difference to a human observer.

In the future, my guess is that courts will generally be on the side of artists because of societal pressures, and artists will be able to challenge any image they find and have it sent to yet another ML model that can quickly adjudicate whether the generated image is "too similar" to the artist's style (which would also need to be dissimilar enough from everyone else's style to give a reasonable legal claim in the first place).

Or maybe artists will just give up on trying to monetize the images themselves and focus only on creating physical artifacts, similar to how independent musicians make most of their money nowadays from touring and selling merchandise at shows (plus Patreon). Who knows? It's hard to predict the future when there are such huge fundamental changes that happen so quickly!

johnnyanmac
1 replies
18h45m

Or maybe artists will just give up on trying to monetize the images themselves and focus only on creating physical artifacts, similar to how independent musicians make most of their money nowadays from touring and selling merchandise at shows (plus Patreon).

As is, art already isn't a sustainable career for most people who can't get a job in industry. The most common monetization is either commissions or hiding extra content behind a pay wall.

To be honest I can see more proverbial "Furry artists" sprouting up in a cynical timeline. I imagine like every other big tech that the 18+ side of this will be clamped down hard by the various powers that be. Which means NSFW stuff will be shielded a bit by the advancement and you either need to find underground training models or go back to an artist. .

Gigachad
0 replies
15h6m

need to find underground training models

It's not particularly that hard. The furry nsfw models are already the most well developed and available models you can get right now. And they are spitting out stuff that is almost indistinguishable from regular art.

raincole
0 replies
13h40m

This seems like a pretty pointless "arms race" or "cat and mouse game".

If there is any "point" of this, it's that's going to push the AI models to become better at capturing how humans see things.

jMyles
0 replies
1h28m

musicians make most of their money nowadays from touring and selling merchandise at shows

Be reminded that this is - and has always been - the mainstream model of the lineages of what have come to be called "traditional" and "Americana" and "Appalachian" music.

The Grateful Dead implemented this model with great finesse, sometimes going out of their way to eschew intellectual property claims over their work, in the belief that such claims only hindered their success (and of course, they eventually formalized this advocacy and named it "The Electronic Frontier Foundation" - it's no coincidence that EFF sprung from deadhead culture).

hackernewds
0 replies
19h52m

the point is you could circumvent one nightshade, but as long as the cat and mouse game continues there can be more

dist-epoch
5 replies
1d2h

Remember when the music industry tried to use technology to stop music pirating?

This will work about as well...

Oh, I forget, fighting music pirating was considered an evil thing to do on HN. "pirating is not stealing, is copyright infringement", right? Unlike training neural nets on internet content which of course is "stealing".

snakeyjake
2 replies
23h34m

A more apt comparison is sampling.

AI is sampling other's works.

Musicians can and do sample. They also obtain clearance for commercial works, pay royalties if required, AND credit the samples if required.

AI "art" does none of that.

Minor49er
1 replies
23h23m

Musicians overwhelmingly do not even attempt to clear samples. This also isn't a great comparison since samples are taken directly out of the audio, not turned into a part of a pattern used to generate new sounds like what AI generators do with images

snakeyjake
0 replies
23h7m

Commercial musicians do not?

You sure about that?

Entire legal firm empires have been built on the licensing, negotiations, and fees that make up the industry.

I'm ain't talking about some dude on YouTube or Soundcloud. Few people care about some rando on Soundcloud. Those moles aren't big enough to whack. Vanilla Ice and MC Hammer were. OpenAI is as well.

There's even a company that specializes in sample clearance: https://sampleclearance.com

More info: https://www.soundonsound.com/sound-advice/sample-clearance

Also:

not turned into a part of a pattern used to generate new sounds like what AI generators do with images

This is demonstrably false. Multiple individuals have repeatedly been able to extract original images from AI generators.

Here's one-- Extracting Training Data from Diffusion Models https://arxiv.org/abs/2301.13188

Text, too: https://arxiv.org/abs/2311.17035

xigoi
0 replies
20h36m

The difference is that “pirating” is mostly done by individuals for private use, whereas training is mostly done by megacorporations looking to make more money.

kevingadd
0 replies
1d2h

FWIW, you're the only use of the word "steal" in this comment thread.

Many people would in fact argue that training AI on people's art without permission is copyright infringement, since the thing it (according to detractors) does is infringe copyright by generating knockoffs of people's work.

You will see some people use the term "stealing" but they're usually referring to how these AIs are sold/operated by for-profit companies that want to make money off artists' work without compensating them. I think it's not unreasonable to call that "stealing" even if the legal definition doesn't necessarily fit 100%.

The music industry is also not really a very good comparison point for independent artists... there is no Big Art equivalent that has a stranglehold on the legislature and judiciary like the RIAA/MPAA do.

brucethemoose2
5 replies
23h44m

What the article doesn't illustrate is that it destroys fine detail in the image, even in the thumbnails of the reference paper: https://arxiv.org/pdf/2310.13828.pdf

Also... Maybe I am naive, but it seems rather trivial to work around with a quick prefilter? I don't know if tradition denoising would be enough, but worst case you could run img2img diffusion.

reply

GaryNumanVevo
4 replies
23h1m

The poisoned images aren't intended to be viewed, rather scraped and pass a basic human screen. You wouldn't be able to denoise as you'd have to denoise the entire dataset, the entire point is that these are virtually undetectable from typical training set examples, but they can push prompt frequencies around at will with a small number of poisoned examples.

minimaxir
3 replies
22h30m

You wouldn't be able to denoise as you'd have to denoise the entire dataset

Doing that requires much less compute than training a large generative image model.

brucethemoose2
1 replies
22h2m

I guess the idea is that the model trainers are ignorant of this and wouldn't know to preprocess/wouldn't bother?

That's actually quite plausible.

BugsJustFindMe
0 replies
15h39m

I guess the idea is that the model trainers are ignorant of this

Maybe they're ignorant of it right up until you announce it, but then they're no longer ignorant of it.

GaryNumanVevo
0 replies
22h3m

the entire point is that these are virtually undetectable from typical training set examples

I'll repeat this point for clarity. After going over the paper again, denoising shouldn't affect this attack, it's the ability of plausible images to not be detected by human or AI discriminators (yet)

gmerc
4 replies
1d3h

Doing the work to increase OpenAIs moat

Drakim
3 replies
1d3h

Obviously AIs can just train on images that aren't poisoned.

jsheard
2 replies
1d3h

Is it possible to reliably detect whether an image is poisoned? If not then it achieves the goal of punishing entities which indiscriminately harvest data.

dist-epoch
0 replies
1d2h

AI's have learned much tougher things. You just need a small data set of poisoned images to learn it's features.

Kalium
0 replies
1d2h

You can use older images, collected from before the "poisoning" software was released. Then you don't have to.

This, of course, assumes that "poisoning" actually works. Glaze and Nightshade and similar are very much akin to the various documented attacks on facial recognition systems. The attack does not exploit some fundamental flaw in how the systems work, but specific characteristics in a given implementation and version.

This matters because it means that later versions and models will inevitably not have the same vulnerabilities. The result is that any given defensive transformation should be expected to be only narrowly effective.

squidbeak
2 replies
22h15m

I really don't understand the anxiety of artists towards AI - as if creatives haven't always borrowed and imitated. Every leading artist has had acolytes, and while it's true no artist ever had an acolyte as prodigiously productive as AI will be, I don't see anything different between a young artist looking to Picasso for cues and Stable Diffusion or DALL-E doing the same. Styles and methods haven't ever been subject to copyright - and art would die the moment that changed.

The only explanation I can find for this backlash is that artists are actually worried just like the rest of us that pretty soon AI will produce higher quality more inventive work faster and more imaginatively than they can - which is very natural, but not a reason to inhibit an AI's creative education.

jwells89
0 replies
21h39m

Imitation isn’t the problem so much as it is that ML generated images are composed of a mush of the images it was trained on. A human artist can abstract the concepts underpinning a style and mimic it by drawing all-new lineart, coloration, shading, composition, etc, while the ML model has to lean on blending training imagery together.

Furthermore there’s a sort of unavoidable “jitter” in human-produced art that varies between individuals that stems from vastly different ways of thinking, perception of the world, mental abstraction processes, life experiences, etc. This is why artists who start out imitating other artists almost always develop their imitations into a style all their own — the imitations were already appreciably different from the original due to the aforementioned biases and those distinctions only grow with time and experimentation.

There would be greatly reduced moral controversy surrounding ML models if they lacked that mincemeat/pink slime aspect.

beepbooptheory
0 replies
22h3m

This has been litigated over and over again, and there have been plenty of good points made and concerns raised over it by those who it actually affects. It seems a little bit disingenuous (especially in this forum) to say that that conclusion is the "only explanation" you can come up with. And just to avoid prompting you too much: trust me, we all know or can guess why you think AI art is a good thing regardless of any concerns one might bring up.

chris-orgmenta
2 replies
21h51m

I want progressive fees on copyright/IP/patent usage, and worldwide gov cooperation/legislation (and perhaps even worldwide ability to use works without obtaining initial permission, although let's not go into that outlandish stuff)

I want a scaling license fee to apply (e.g. % pegged to revenue. This still has an indirect problem with different industries having different profit margins, but still seems the fairest).

And I want the world (or EU, then others to follow suit) to slowly reduce copyright to 0 years* after artists death if owned by a person, and 20-30 years max if owned by a corporation.

And I want the penalties for not declaring usage** / not paying fees, to be incredibly high for corporations... 50% gross (harder) / net (easier) profit margin for the year? Something that isn't a slap on the wrist and can't be wriggled out of quite so easily, and is actually an incentive not to steal in the first place.)

[*]or whatever society deems appropriate.

[**]Until auto-detection (for better or worse) gets good enough.

IMO that would allow personal use, encourages new entrants to market, encourages innovation, incentivises better behaviour from OpenAI et al.

wraptile
0 replies
12h0m

Extremely naive to think that any of this could be enforced to any adequate level. Copyright is fundamentally broken and putting some plasters on it is not going to do much especially when these plasters are several decades too late.

Dylan16807
0 replies
16h24m

And I want the world (or EU, then others to follow suit) to slowly reduce copyright to 0 years* after artists death if owned by a person, and 20-30 years max if owned by a corporation.

Why death at all?

It's icky to trigger soon after death, it's bad to have copyright vary so much based on author age, and it's bad for many works to still have huge copyright lengths.

It's perfectly fine to let copyright expire during the author's life. 20-30 years for everything.

msp26
1 replies
1d2h

Like Glaze, Nightshade is computed as a multi-objective optimization that minimizes visible changes to the original image.

It's still noticeably visible.

kevingadd
0 replies
1d2h

Yeah, I've seen multiple artists complain about how glazing reduces image quality. It's very noticeable. That seems like an unavoidable problem given how AI is trained on images right now.

matt3210
1 replies
14h49m

Put a TOC on all your content that says “by using my content for AI you agree to pay X per image” and then send them a bill once you see it in an AI.

wruza
0 replies
14h24m

Once you see what exactly? “AI” isn’t some image filter from the early 2000s.

k__
1 replies
1d3h

How long will this work?

kevingadd
0 replies
1d2h

It's an arms race the bigger players will win, and it undermines the quality of the images. But it feels natural that artists would want to do something since they don't feel like anyone else is protecting them right now.

eddd-ddde
1 replies
1d3h

Isn't this just teaching the models how to better understand pictures as humans do? As long as you feed them content that looks good to a human, wouldn't they improve in creating such content?

lern_too_spel
0 replies
16h18m

You would think the economists at UChicago would have told these researchers that their tool would achieve the opposite effect of what they intended, but here we are.

In this case, the mechanism for how it would work is effectively useless. It doesn't affect OpenAI or other companies building foundation models. It only works on people fine-tuning these foundation models, and only if the image is glazed to affect the same foundation model.

ThinkBeat
1 replies
8h30m

In so far as anger goes against AIs being trained on particular intellectual properties.

A made up scenario¹ is that a person who is training an AI, goes to the local library and checks out 600 books on art. The person then lets the AI read all of them. After which they are returned to the library and another 600 books are borrowed

Then we can imagine the AI somehow visiting a lot of museums and galleries.

The AI will now have been trained on the style and looks of a lot of art from different artists

All the material has been obtained in a legal manner.

Is this an acceptable use?

Or can an artist still assert that the AI was trained with their IP without consent?

Clearly this is one of the ways a human would go about learning about styles, techniques etc..

¹ Yes you probably cannot borrow 600 books at a time. How does the AI read the books? I dont know. Simplicity would be that the researcher takes a photo of each page. This would be extremmly slow but for this hypothetical it is acceptable.

nanofus
0 replies
7h25m

I think the key difference here is that the most prominent image generation AIs are commercial and for-profit. The scenarios you describe are comparing a commercial AI to a private person. You cannot get a library card for a company, and you cannot bring a photography crew to a gallery without permission.

zirgs
0 replies
22h49m

Does it survive AI upscaling or img2img? If not - then it's useless. Nobody trains AI models without any preprocessing. This is basically a tool for 2022.

xg15
0 replies
23h53m

I wonder how this tool works if it's actually model independent. My understanding so far was that in principle each possible model has some set of pathological inputs for which the classification will be different than what a user sees - but that this set is basically different for each model. So did they actually manage to build an "universal" poison? If yes, how?

will5421
0 replies
22h13m

I think the artists need to agree to stop making art altogether. That ought to get people’s attention. Then the AI people might (be socially pressured or legally forced to) put their tools away.

whywhywhywhy
0 replies
21h36m

Why are there no examples?

ultimoo
0 replies
1d2h

would it have been that hard to include a sample photo and how it looks with the nightshade filter side by side in a 3 page document describing how it would look in great detail

tigrezno
0 replies
23h27m

Do not fight the AI, it's a lost cause, embrace it.

snerc
0 replies
10h11m

I wonder if we know enough about any of these systems to make such claims. This is all predicated on the fact that this tool will be in widespread use. If it is somehow widely used beyond the folks who have seen it at the top of HN, won't the big firms have countermeasures, ready to deploy?

rvba
0 replies
5h17m

The opening website is so poor - "what is nightshade" - then a whole paragraph that tells nothing, then another paragraph.. then no examples. This whole description should be reworked to be shorter and more to the point.

paulsutter
0 replies
18m

Cute. The effectiveness of any technique like this will be short-lived.

What we really need is clarification of the extent that copyright protection extends to similar works. Most likely from an AI analysis of case law.

paul7986
0 replies
14h48m

Any AI art/video/photography/music/etc generator company who generates revenue needs to add watermarks to let the public know its AI generator. This should be forced via legislation in all countries.

If they don't then whatever social network or other services where things can shared/viewed by large groups to millions & are posted publicly need to be labeled "We can not verify veracity of this content."

I want a real internet ..this AI stuff is just triple fold increasing fake crap on the Internet and in turn / time our trust in it!

nnevatie
0 replies
9h14m

The intention is good, from an AI-opponent's perspective. I don't think will work practically, though. The drawbacks for actual users of the image galleries, plus the level of complexity involved in poisoning the samples makes this unfeasible to implement at the scale required.

ngneer
0 replies
13h8m

I love it. This undermines the notion of ground truth. What separates correct information from incorrect information? Maybe nothing! I love how they acknowledge the never ending attack versus defense game. In stark contrast to "our AI will solve all your problems".

mmaunder
0 replies
14h27m

Trying to convince an AI it sees something and a human they don’t is probably a losing battle.

mjfl
0 replies
23h11m

Another way would be, for every 1 piece of art you make, post 10 AI generated arts, so that the SNR is really bad.

minimaxir
0 replies
22h23m

A few months ago I made a proof-of-concept on how finetuning Stable Diffusion XL on known bad/incoherent images can actually allow it to output "better" images if those images are used as a negative prompt, i.e. specifying a high-dimensional area of the latent space that model generation should stay away from: https://news.ycombinator.com/item?id=37211519

There's a nonzero chance that encouraging the creation of a large dataset of known tampered data can ironically improve generative AI art models by allowing the model to recognize tampered data and allow the training process to work around it.

mattszaszko
0 replies
10h1m

This timeline is getting quite similar to the second season of Pantheon.

matteoraso
0 replies
21h1m

Too little, too late. There's already very large high quality datasets to train AI art generators.

marcinzm
0 replies
22h52m

This feels like it'll actually help make AI models better versus worse once they train on these images. Artists are basically, for free, creating training data that conveys what types of noise does not change the intended meaning of the image to the artist themselves.

k__
0 replies
22h19m

What are LLMs that was trained with public domain content only?

I would believe there is enough content out there to get reasonably good results.

jdeaton
0 replies
18h54m

Sounds like free adversarial data augmentation.

ink404
0 replies
2d
iLoveOncall
0 replies
10h52m

I wonder if this is illegal in some countries. In France for example, there is the following law: "Obstructing or distorting the operation of an automated data processing system is punishable by five years' imprisonment and a fine of €150,000.".

If you ask me, this is 100% applicable in this case, so I wonder what a judge would rule.

gweinberg
0 replies
23h22m

For this to work, wouldn't you have to have an enormous number of artists collaborating on "poisoning" their images the same way (cow to handbag) while somehow keeping it secret form ai trainers that they were doing this? It seems to me that even if the technology works perfectly as intended, you're effectively just mislabeling a tiny fraction of the training data.

gumballindie
0 replies
1d3h

This is excellent. We need more tools like this, for text content as well. For software we need GPL 4 with ML restrictions (make your model open source or not at all). Potentially even DRM for text.

garg
0 replies
23h48m

Each time there is an update to training algorithms and in response poisoning algorithms, artists will have to re-glaze, re-mist, and re-nightshade all their images?

Eventually I assume the poisoning artifacts introduced in the images will be very visible to humans as well.

etchalon
0 replies
23h14m

My hope is these type of "poisoning tools" become ubiquitous for all content types on the web, forcing AI companies to, you know, license things.

drdrek
0 replies
5h33m

Only protection is adding giant gaping vaginas to your art, nothing less will deter scraping. If the Email spam community showed us something in the last 40 years is that no amount of defensive tech measures will work except financial disincentives.

aussieguy1234
0 replies
15h3m

The image generation models now are at the point where they can produce their own synthetic training images. So I'm not sure how big of an impact something like this would have.

arisAlexis
0 replies
21h24m

Wouldn't this be applicable to text too?

Zetobal
0 replies
21h52m

Well, at least for sdxl it's not working neither in LoRa nor dreambooth finetunes.

TenJack
0 replies
1d

Wonder if the AI companies are already so far ahead that they can use their AI to detect and avoid any poisoning?

Kuinox
0 replies
1d3h

More specifically, we assume the attacker:

• can inject a small number of poison data (image/text pairs) to the model’s training dataset

I think thoes are bad assumption, labelling is more and more done by some labelling AI.

GaggiX
0 replies
23h48m

These methods like Glaze usually works by taking the original image chaging the style or content and then apply LPIPS loss on an image encoder, the hope is that if they can deceive a CLIP image encoder it would confuse also other models with different architecture, size and dataset, while changing the original image as little as possible so it's not too noticeable to a human eye. To be honest I don't think it's a very robust technique, with this one they claim that a model instead of seeing for example a cow on grass the model will see a handbag, if someone has access to GPT4-V I want to see if it's able to deceive actually big image encoders (usually more aligned to the human vision).

EDIT: I have seen a few examples with GPT-4 V and how I imagine it wasn't deceived, I doubt this technique can have any impact on the quality of the models, the only impact that this could potentially have honestly is to make the training more robust.

Duanemclemore
0 replies
23h7m

For visual artists who don't want visible artifacting in the art they feature online, would it be possible to upload these alongside your un-poisoned art, but have them only hanging out in the background? So say having one proper copy and a hundred poisoned copies in the same server, but only showing the un-poisoned one?

Might this "flood the zone" approach also have -some- efficacy against human copycats?

Devasta
0 replies
17h19m

Delighted to see it. Fuck AI art.

Albert931
0 replies
22h52m

Artist are now fully dependent on Software Engineers for protecting the future of their career lol

24karrotts_
0 replies
11h46m

If you decrease quality of art, you give AI all the advantage in the market.