We Asked A.I. to Create the Joker. It Generated a Copyrighted Image.
(www.nytimes.com)
from L4s@lemmy.world to technology@lemmy.world on 26 Jan 2024 22:00
https://lemmy.world/post/11225642
from L4s@lemmy.world to technology@lemmy.world on 26 Jan 2024 22:00
https://lemmy.world/post/11225642
We Asked A.I. to Create the Joker. It Generated a Copyrighted Image.::Artists and researchers are exposing copyrighted material hidden within A.I. tools, raising fresh legal questions.
threaded - newest
Copyright is a scam anyway so who cares?
Nah, that’s like saying capitalism is a scam.
Copyright and capitalism in general is fine. It’s when billion dollar corporations use political donations to control regulations
Like, imagine a year after Hangover came out. 20 production companies all released Hangover 2.
Imagine it was a movie by a small Indie studio so a big studio paid off the original actors to be in their knockoff.
Or an animated movie that used the same digital assets.
We need some copyright protection, just not a never ending system
Capitalism is a scam.
It’s an unsustainable system predicted on infinite growth that necessitates unconscionable inequality.
You dropped the “unregulated”.
Socialism is still capitalism. It’s just regulated and we use taxes to fund social programs.
And the second sentence is more caused by not taxing stock trades. If we had a tax that decreases the longer a stock is held, it would prioritize long term investment and companies would care about more than the next months earnings.
All shit that can be solved with common sense regulations.
I’m not aware of any common sense regulations that take precedence over corporate profits, but keep up the good fight I guess.
…
Did you miss that?
I know starting a new paragraph for each sentence helps more people understand because they tend to skip paragraphs, but my comment was only 6 sentences
I figured two sentences a paragraph wasn’t too much.
I didn’t miss it.
You misunderstood my comment. These regulations you’re talking about either don’t exist or they are ineffective, because we are suffering the greatest economic inequality in history while corporate profits are at an all time high.
But good luck, I hope you become super rich doing business so you don’t have to suffer with the rest of humanity. Have fun buying clean water tokens in a couple decades!
If they were already in effect, we wouldn’t need them because they already existed…
When we have them corporations pay lots of money to politicians to convince them we don’t need them because they’re working.
So when someone says we need regulations, it’s a pretty safe bet the regulations theyre talking about don’t exist yet, and they’re saying they should…
You can tell that’s what they mean, because it’s literally what they’re saying.
At least most of the time most people can.
I’m not sure why you keep not getting this and acting like you’re over replying. Then replying again.
Blocking is easier, it’s what I do when people don’t understand basic stuff while having an attitude about it.
Let me show you an example.
Tell me you don’t know the definition of socialism without telling me you don’t know the definition of socialism.
While Socialism may not be against Free Markets, Capitalism /= Socialism.
Socialism is when the workers collectively own the means of production.
So like in a factory, everyone employed there also has ownership stake in the company, and they can vote on leadership internally. Instead of relying on government regulations to be able to have things like paid lunches and guaranteed sick days, instead they can come to collective agreement on those things, with the vote of every worker/owner. They can still sell products on a free market, but the “capital” part of the equation has been removed.
In capitalism, in a factory, the factory has been purchased by a Capitalist who, by definition, is someone with a lot of Capital (money/wealth) and they bought the factory whole with the capital. Now, they are going to hire workers with the capital as well, and the workers have to follow any and all their rules, like a little fiefdom of a dictatorship and nobody gets an opportunity to vote on leadership. They have no control over pay, working conditions, or much else, and they rely on the Government to enforce it otherwise.
Socialism really doesn’t have anything to do with government regulation, taxation, or social safety nets.
Nope.
That’s not socialism.
In socialism there’s still private ownership of companies.
Maybe you’re specifically thinking of Communist Russia which allowed for small businesses to have private ownership, or perhaps you’re thinking of Anarcho-Communism?
But the reality is, private ownership is the antithesis of collective ownership, so most socialists aren’t really on board with that.
Via Wikipedia, which makes the point numerous times, with a plethora of scholarly references:
en.wikipedia.org/wiki/Socialism
But don’t worry, I fully expect a response that keeps arguing the same thing with no evidence to support it. That’s what I usually get here, it’s like the world has become illiterate.
Here’s the summary for the wikipedia article you mentioned in your comment:
Socialism is an economic and political philosophy encompassing diverse economic and social systems characterised by social ownership of the means of production, as opposed to private ownership. It describes the economic, political, and social theories and movements associated with the implementation of such systems. Social ownership can take various forms including: public, community, collective, cooperative, or employee. No single definition encapsulates the many types of socialism, but social ownership is the common element. Traditionally, socialism is on the left-wing of the political spectrum.
^to^ ^opt^ ^out^^,^ ^pm^ ^me^ ^‘optout’.^ ^article^ ^|^ ^about^
socialism has nothing to do with taxes
Why do all the accounts with “commie” in the name have no idea about any economic systems?
I don’t think any of them actually support communism either, it’s just weird I block so many and they always keep showing up
there is a cure for political illiteracy
I know it is popular to say that infinite growth is a requirement of capitalistic systems. First of all a proof is never provided. Secondly, Japan might be a counter example, where the GDP has stagnanted for two decades.
Those are fundamental features of the current system. If you want to suggest a copyright system that does protect smaller creators from bad actors but doesn’t allow the mega-corps to bully and control everyone, then feel free. But until such a system is implemented it see no reason to defend the current one which is actively harmful to the vast majority of creators.
That’s the system that was implemented…
It worked fine until corporations realized both parties like money. Which didn’t take long.
And I just said the current situation isn’t good…
I can’t believe all the simping for copyright that’s come out of AI. What the fuck happened to the internet? On a place like lemmy no less.
People are smart enough to understand the difference between someone copying for personal use and a billion dollar corporation copying to generate millions while laying off all the creative people. The latter is what these non-open-source AI companies are enabling - for profit too.
.
They literally asked it to give them a screenshot from the Joker movie. That was their fucking prompt. It’s not like they just said “draw Joker” and it spit out a screenshot from the movie, they had to work really hard to get that exact image.
Because this proves that the “AI”, at some level, is storing the data of the Joker movie screenshot somewhere inside of its training set.
Likely because the “AI” was trained upon this image at some point. This has repercussions with regards to copyright law. It means the training set contains copyrighted data and the use of said training set could be argued as piracy.
Legal discussions on how to talk about generative-AI are only happening now, now that people can experiment with the technology. But its not like our laws have changed, copyright infringement is copyright infringement. If the training data is obviously copyright infringement, then the data must be retrained in a more appropriate manner.
Wasn’t that known? Have midjourney ever claimed they didn’t use copyrighted works? There’s also an ongoing argument about the legality of that in general. One recent court case ruled that copyright does not protect a work from being used to train an AI. I’m sure that’s far from the final word on the topic, but it does mean this is a legal grey area at the moment.
If it is known, then it is copyright infringement to download the training sets and therefore a crime to do so. You cannot reproduce a copy of the works without the express permission of the copyright holder.
How many computers did Midjourney copy its training weights to? Has Midjourney (and the IT team behind it) paid royalties for every copyrighted image in its training set to have a proper copyright license to copy all of this data from computer to computer?
I’m guessing no. Which means the Midjourney team (if you say is true) is committing copyright infringement every time they spin up a new server with these weights.
Pro-AI side will obviously argue that the training weights do not contain the data of these copyrighted works. A claim that is looking more-and-more laughable as these experiments happen.
No it’s not illegal to download publicly available content it’s a copyright violation to republish it.
Is it tho? Honest question.
Yes it is. Honest answer.
So stable diffusion, midjourney, etc., all have massive databases with every picture on the Internet stored in them? I know the AI models are trained on lots of images, but are the images actually stored? I’m skeptical, but I’m no expert.
These models were trained on datasets that, without compensating the authors, used their work as training material. It’s not every picture on the net, but a lot of it is scrubbing websites, portfolios and social networks wholesale.
A similar situation happens with large language models. Recently Meta admitted to using illegally pirated books (Books3 database to be precise) to train their LLM without any plans to compensate the authors, or even as much as paying for a single copy of each book used.
Most of the stuff that inspires me probably wasn’t paid for. I just randomly saw it online or on the street, much like an AI.
AI using straight up pirated content does give me pause tho.
I was on the same page as you for the longest time. I cringed at the whole “No AI” movement and artists’ protest. I used the very same idea: Generations of artists honed their skills by observing the masters, copying their techniques and only then developing their own unique style. Why should AI be any different? Surely AI will not just copy works wholesale and instead learn color, composition, texture and other aspects of various works to find it’s own identity.
It was only when my very own prompts started producing results I started recognizing as “homages” at best and “rip-offs” at worst that gave me a stop.
I suspect that earlier generations of text to image models had better moderation of training data. As the arms race heated up and pace of development picked up, companies running these services started rapidly incorporating whatever training data they could get their hands on, ethics, copyright or artists’ rights be damned.
I remember when MidJourney introduced Niji (their anime model) and I could often identify the mangas and characters used to train it. The imagery Niji produced kept certain distinct and unique elements of character designs from that training data - as a result a lot of characters exhibited “Chainsaw Man” pointy teeth and sticking out tongue - without as much as a mention of the source material or even the themes.
How much profit do you make from this stuff ?
The stuff I sell on jilanico.com? Enough to make it worth my while.
Couple things:
this doesn’t explain ops question about how the information is stored. On fact op is right, that the images and source material is NOT stored in a database within the model, it basically just stores metadata about the source material as a whole in order to construct new material from text descriptions
the use of copyrighted works in the training isn’t necessarily infringing if the model is found to be a fair use, and there is a very strong fair use argument here.
“metadata” is such a pretty word. How about “recipe” instead? It stores all information necessary to reproduce work verbatim or grab any aspect of it.
The legal issue of copyright is a tricky one, especially in the US where copyright is often being weaponized by corporations. The gist of it is: The training model itself was an academic endeavor and therefore falls under a fair use. Companies like StabilityAI or OpenAI then used these datasets and monetized products built on them, which in my understanding skims gray zone of being legal.
If these private for-profit companies simply took the same data and built their own, identical dataset they would be liable to pay the authors for use of their work in commercial product. They go around it by using the existing model, originally created for research and not commercial use.
Lemmy is full of open source and FOSS enthusiasts, I’m sure someone can explain it better than I do.
All in all I don’t argue about the legality of AI, but as a professional creative I highlight ethical (plagiarism) risks that are beginning to arise in majority of the models. We all know Joker, Marvel superheroes, popular Disney and WB cartoon characters - and can spot when “our” generations cross the line of copying someone else’s work. But how many of us are familiar with Polish album cover art, Brazilian posters, Chinese film superheroes or Turkish logos? How sure can we be that the work “we” produced using AI is truly original and not a perfect copy of someone else’s work? Does our ignorance excuse this second-hand plagiarism? Or should the companies releasing AI models stop adding features and fix that broken foundation first?
Well isn’t recipe another one of those pretty words? ‘Metadata’ is specific to other precedents that deal with computer programs that gather data about works (see Authors Guild, Inc. v. HathiTrust and Authors Guild v. Google), but you’re welcome to challenge the verbiage if you don’t like it. Regardless, what we’re discussing is objectively something that describes copyrighted works, not copies or a copy of the works themselves. A computer program that is very good at analyzing textual/pixelated data is still only analyzing data, it is itself a novel, non-expressive factual representation of other expressive works, and because of this, it cannot be considered as infringement on its own.
This isn’t really true, at least not for the majority of works analyzed by the model, but granted. If a person uses a tool to copy the work of another person, it is the person who is doing the copying, not the tool. I think it is far more reasonable to hold an individual who uses an AI model to infringe on a copyright responsible. If someone chooses to author a work with the use of a tool that does the work for them (in part or in whole), it is more than reasonable to expect that individual to check the work that is being produced.
As a professional creative myself, I think this is a load of horseshit. We always hold individual authors responsible for the work that they publish, and it should be no different here. That some choose to be lazy and careless is more of a reflection of them.
If you have the words to describe a desired image/text response to the model that produce a ‘perfect copy of someone else’s work’, then we have the words to search for that work, too.
How about we stop expanding the scope of an already broken copyright law and fix that broken foundation first?
How did the Joker image get replicated?
I posted it on my website as fan art and it scraped it. I just used a different filter which falls under fair use.
It’s too hard to type up how generative AIs work, but look up a video on “how stable diffusion works” or something like that. I seriously doubt they have a massive database with every image from the Internet inside it, with the AI just spitting those pics out, but I’m no expert.
Sure, but so is your memory, you could study the originals and re-draw them a similar way.
I agree, but I don’t think these generative AIs actually store image files off the Internet in a massive database. I could be wrong.
That’s correct. The structure of information isn’t anywhere remotely similar to a file or database. Information pixel by pixel isn’t stored, it more loosely remembers correlations and similarities and facts about the content as opposed to storing and copying it
Which is also very similar to how your brain stores things.
Yeah, much more similar to the brain than a database or file anyway
I mean anyone can use copyrighted material as inspiration for their work and it’s fair use and not a concern at all.
Is Ai only bad since it can do what a human does better/faster? If that’s that case, than they don’t actually have an issue with the fact it’s copyrighted, or I wouldn’t be able to use it for inspiration either.
In these cases it’s bad because it can do what a human does with no ethics, empathy, or regard for the law. If it had those things, it would be worse because we’d then be encroaching on the rights of sentient beings.
Problem is the AI didn’t do anything. People told the program tongo scrape the internet. So humans still made the decision with no regard doe the laws.
Legally speaking, AI is not anything. Its just a computer program. What you’re asking is completely a red-herring.
The question here is if the training-weights constitute copyright infringement. Now look at any clip-art set. Most clip-art is so called “royalty free”, as in you can copy it from computer-to-computer without any copyright issues, because the author specifically said that its royalty free.
But if you have a copyrighted font, then even copying that font from one computer to another constitutes copyright infringement. (IE: Literally, you aren’t allowed to copy this unless you have the permission of the author).
So, when you download Midjourney’s training weights, does that act in of itself constitute a copy that violate’s the authors of “Joker” movie? As far as I can tell, yes. Because the training weights clearly contain Joker images.
Hows it a red herring to point out we are allowed to use copyrighted materials already? Its not the concern here, yet its what they are using as the concern for their arguments against it.
Because copyright law is clear in that computers can’t own a copyright.
The humans at play are:
The artist who created the original work.
The computer IT team who are copying the data behind the scenes between servers.
You who uses Midjourney to recreate “Joker” movie artwork, likely using the data in #2 which falls under copyright infringement.
It doesn’t matter how #2 works. It doesn’t matter if its H.265 or MPEG2 or from VHS tapes, or if its a Neural Network using the latest-and-greatest training weights from a GPU-based datasystem. Its just a computer. The ones doing the copyright infringement are the people copying data from place to place.
The AI model is not a copy of the set of data used to train it, it’s a derivative work. As such copyright as it currently stands does not apply. It’s possible, likely even, that copyright will be modified in some way soon to account for this, but the situation today says nope, not copyright infringement.
They’re really trying so hard cuz they absolutely want this to be infringement but it simply isnt on any legal level.
Looking at a copyrighted font with your computer means the font is in your computer’s memory. Do I go to jail for every site I visit that uses a fancy font?
Font files ≠ framebuffer
Images ≠ neural network weights
If its a fancy copyrighted font without a license to copy… the Website owner gets sued. Because the website owner is the one making mass copies of said font.
Do… you know what copyrites are? They relate to the copying of data.
The framebuffer on your computer copies the data to display the font to you. That’s my point. Not every form of copying infringes on copyright.
And my argument is that Midjourney’s servers are engaged in illegal copying. So I think your point is moot. Not the Web Browsers downloading images.
The movie Joker’s image is being copied each time the training weights are copied to a new server. Is that not an illegal copy?
When you look at a picture of the joker online, your browser is caching an image file of the joker on your computer. Is that not an illegal copy?
What the hell is this non-sequitur?
What do browser caches have to do with Midjourney servers and training weights?
I get that you wanna change the subject. But I dunno if it’s because you don’t understand my argument, or if you’ve realized that my argument is solid and therefore you have no actual counterargument.
The copy that people care about are the webservers. That’s why when you run Bittorrent, MPAA or RAII sue the people serving the data. Not the people who use the data. Have you followed any copyright case in the last two or three decades? In this case, it’d be a copyright case vs Midjourney servers.
I really do not believe midjourney is storing all the files on the Internet in a humongous database. They are just exposing the AI to them for training just like you expose your computer to them when you visit with your web browser. I’m happy to be wrong about this, but I’m just not convinced. Please try to keep your temper.
Are you sure?
The training weights can literally recreate images its been trained on. That makes them a humongous database, albeit with lossy compression. They aren’t replicated exactly, but they are replicated enough that I’m confident that these “Joker” images passes as copyright infringement before a jury (ie: is substantially similar).
I understand why you’re saying that, but I personally don’t agree. As AIs face legal challenges, maybe your opinion will be adopted, we’ll see.
This is the crux of the issue, it isn’t obviously copyright infringement. Currently copyright is completely silent on the matter one way or another.
The thing that makes this particularly interesting is that the traditional copyright maximalists, the ones responsible for ballooning copyright durations from its original reasonable limit of 14 years (plus one renewal) to its current absurd duration of 95 years, also stand to benefit greatly from generative works. Instead of the usual full court press we tend to see from the major corporations around anything copyright related we’re instead seeing them take a rather hands off approach.
Its clear that the training weights have the data on recreating this Joker scene. Its also clear that if the training-data didn’t contain this image, then the copy of the image would never result into the weights that have been copy/pasted everywhere.
Except it isn’t a perfect copy. It’s very similar, but not exact. Additionally for every example you can find where it spits out a nearly identical image you can also find one where it produces nothing like it. Even more complicated you can get images generated that very closely match other copyrighted works, but which the model was never trained on. Does that mean copying the model violates the copyright of a work that it literally couldn’t have included in its data?
You’re making a lot of assumptions and arguments that copyright covers things that it very much does not cover or at a minimum that it hasn’t (yet) been ruled to cover.
Legally, as things currently stand, an AI model trained on a copyrighted work is not a copy of that work as far as copyright is concerned. That’s today’s legal reality. That might change in the future, but that’s far from certain, and is a far more nuanced and complicated problem than you’re making it out to be.
Any legal decision that ruled an AI model is a copy of all the works used to train it would also likely have very far reaching and complicated ramifications. That’s why this needs to be argued out in court, but until then what midjourney is doing is perfectly legal.
www.law.cornell.edu/wex/derivative_work
The law is very clear on the nature of derivative works of copyrighted material.
Not sure where they’re getting the bit about copyright disallowing derived works as that’s just not true. You can get permission to create a derived work, but you don’t need permission to create a derived work so long as the final result does not substantially consist of the original work.
Unfortunately what constitutes “substantially” is somewhat vague. Various rulings have been made around that point, but I believe a common figure used is 30%. By that metric any given image represents substantially less than 30% of any AI model so the model itself is a perfectly legal derived work with its own copyright separate from the various works that were combined to create it.
Ultimately though the issue here is that the wrong tool is being used, copyright just doesn’t cover this case, it’s just what people are most familiar with (not to mention most people are very poorly educated about it) so that’s what everyone reaches for by default.
With generative AI what we have is a tool that can be used to trivially produce works that are substantially similar to existing copyrighted works. In this regard it’s less like a photocopier, and more like Photoshop, but with the critical difference that no particular talent is necessary to create the reproduction. Because it’s so easy to use people keep focusing on trying to kill the tool rather than trying to police the people using it. But they’re going about it all wrong, copyright isn’t the right weapon if that’s your goal. Copyright can be used to go after the people using generative AI tools, but not the people creating the tools.
Why? If the training weights are created and distributed in violation of copyright laws, it seems appropriate to punish those illegal training weights.
In fact, all that people really are asking for, is for a new set of training weights to be developed but with appropriate copyright controls. IE: With express permission from the artists and/or entities who made the work.
Because they aren’t illegal and they don’t violate copyright. People keep wanting them to be against copyright, but that’s just not how copyright works. There either needs to be amendments to copyright law in order to cover this case, but those changes would need to be very carefully tailored. It would be way too easy to make something that’s either overly broad and applies to a bunch of situations it wasn’t intended to, or way too narrow allowing for easy circumventing.
While that might appease some people, it wouldn’t appease everyone. There are a lot of workers in the creative fields that are feeling incredibly threatened by generative AI right now. Some of these fears are certainly overblown, but it’s also true corporations are going to be as shitty as possible and so some regulation is probably in order. That said, once again, copyright just doesn’t seem to be the right tool for the job here.
Because they are legal and they do violate copyright? People keep wanting them to be copyright free, but that’s not how copyright works. There don’t need to be amendments to copyright law in order to cover this case.
I mean, its obviously heading to the courts one way or the other, but I don’t think just making assertions like that are very good kind of arguing. The training weights here have clearly been proven to contain copyrighted data as per this article. I’m not sure if you’re making any kind of serious case that shows otherwise, but are instead just making a bunch of assertions that I could easily reverse.
It varies somewhat from country to country, but at least in the US there is ample case law that says it’s legal. The relevant ruling is typically based on the work being transformative and therefore a fair use for a derivative work. This is E.G. how Google can get away with creating thumbnails of websites to show on their search pages without needing to worry about copyright of anything contained on that website.
This article proved absolutely nothing except that you can use generative AI to create images that would most likely be ruled to violate copyright. Once again though, the model and training weights do not violate copyright, they’re a protected derivative work. The generated image on the other hand very likely does violate copyright.
But where is the infringement?
This NYT article includes the same several copyrighted images and they surely haven’t paid any license. It’s obviously fair use in both cases and NYT’s claim that “it might not be fair use” is just ridiculous.
Worse, the NYT also includes exact copies of the images, while the AI ones are just very close to the original. That’s like the difference between uploading a video of yourself playing a Taylor Swift cover and actually uploading one of Taylor Swift’s own music videos to YouTube.
Even worse the NYT intentionally distributed the copyrighted images, while Midjourney did so unintentionally and specifically states it’s a breach of their terms of service. Your account might be banned if you’re caught using these prompts.
Do Training weights have the data? Are the servers copying said data on a mass scale, in a way that the original copyrighters don’t want or can’t control?
The answer to that question is extensively documented by thousands of research papers - it’s not up for debate.
If someone wants to read one of those papers, I can recommend Extracting Training Data from Diffusion Models. It shouldn’t be too hard for someone with little experience in the field to be able to follow along.
There response well be we don’t know we can’t understand what its doing.
What the fuck is this kind of response? Its just a fucking neural network running on GPUs with convolutional kernels. For fucks sake, turn on your damn brain.
Generative AI is actually one of the easier subjects to comprehend here. Its just calculus. Use of derivatives to backpropagate weights in such a way that minimizes error. Lather-rinse-repeat for a billion iterations on a mass of GPUs (ie: 20 TFlop compute systems) for several weeks.
Come on, this stuff is well understood by Comp. Sci by now. Not only 20 years ago when I learned about this stuff, but today now that AI is all hype, more and more people are understanding the basics.
Understanding the math behind it doesn’t immediately mean understanding the decision progress during forward propagation. Of course you can mathematically follow it, but you’re quickly gonna lose the overview with that many weights. There’s a reason XAI is an entire subfield in Machine Learning.
Ummm… its lossy compressed data from the training set.
Is it a perfect copy? No. But copyright law covers “derivative data” so whatever, the law remains clear on this situation.
Bro who even knows calculus anymore we have calculators for a reason 🤷♀️
Data is not copyrighted, only the image is. Furthermore you can not copyright a number, even though you could use a sufficiently large number to completely represent a specific image. There’s also the fact that copyright does not protect possession of works, only distribution of them. If I obtained a copyrighted work no matter the means chosen to do so, I’ve committed no crime so long as I don’t duplicate that work. This gets into a legal grey area around computers and the fundamental way they work, but it was already kind of fuzzy if you really think about it anyway. Does viewing a copyrighted image violate copyright? The visual data of that image has been copied into your brain. You have the memory of that image. If you have the talent you could even reproduce that copyrighted work so clearly a copy of it exists in your brain.
Yeah. And the hard drives and networks that pass Midjourney’s network weights around?
That’s distribution. Did Midjourney obtain a license from the artists to allow large numbers of “Joker” copyrighted data to be copied on a ton of servers in their data-center so that Midjourney can run? They’re clearly letting the public use this data.
Because they’re not copying around images of Joker, they’re copying around a work derived from many many things including images of Joker. Copying a derived work does not violate the copyright of the work it was derived from. The wrinkle in this case is that you can extract something very similar to the original works back out of the derived work after the fact. It would be like if you could bake a cake, pass it around, and then down the line pull a whole egg back out of it. Maybe not the exact egg you started with, but one very similar to it. This is a situation completely unlike anything that’s come before it which is why it’s not actually covered by copyright. New laws will need to be drafted (or at a bare minimum legal judgements made) to decide how exactly this situation should be handled.
Someone already downvoted you but this is exactly the topic of debate surrounding this issue.
Other recognized fair-use exemptions have similar interpretations: a computer model analyzes a large corpus of copyrighted work for the purposes of being able to search their contents and retrieve relevant snippets and works based on semantic and abstract similarities. The computer model that is the representation of those works for that purpose is fair use: it contains only factual information about those works. It doesn’t matter if the works used for that model were unlicensed: the model is considered fair use.
AI models operate by a very similar method, albeit one with a lot more complexity. But the model doesn’t contain copyrighted works, it is only itself a collection of factual information about the copyrighted works. The novel part of this case is that it can be used to re-construct expressions very similar to the original (it should be pointed out that the fidelity is often very low, and the more detailed the output the less like the original it becomes). It isn’t settled yet if that fact changes this interpretation, but regardless I think copyright is already not the right avenue to pursue, if the goal is to remediate or prevent harm to creators and encourage novel expressions.
Right, you’re basically making the same points as me, although technically the model itself is a copyrighted work. Part of the problem we’re running into these days is that copyright, patent, trademark, and trade secret, all date from a time when the difference between those things was fairly intuitive. With our modern digital world with things like 3D printers and the ease with which you can rapidly change the formats and encodings of arbitrary pieces of data the lines all start to blur together.
If you have a 3D scan of a statue of pikachu what rights are involved there? What if you print it? What if you use that model to generate a PNG? What if you print that PNG? What if you encode the model file using base64 and embed it in the middle of a gif of Rick Astley?
Corporations have already utterly fucked all our IP laws, it might be time to go back to the drawing board and reevaluate the whole thing, because what we have now often feels like it has more cracks than actual substance.
Yea, sorry if it wasn’t clear, but I was agreeing with you (defending against the downvote).
There are a lot of things at play here, even if there seems to be a clear way to interpret copyright law (that’s untested, but still) that would determine the models being a fair use. I think people are rightfully angry/frustrated with the size of these companies building the models, and the risk posed by private ownership over them. If I were inclined to be idealistic, I would say that the models should be in the public domain and the taxes should be used so as to provide a UBI to counter any job loss/efficiencies provided by the automation, but that’s a tall order.
www.law.cornell.edu/wex/derivative_work
Are you just making shit up?
You do realize that newspapers do typically pay the licensing for images, it’s how things like Getty images exist.
On the flip side, OpenAI (and other companies) are charging someone access to their model, which is then returning copyrighted images without paying the original creator.
That’s why situations like this keep getting talked about, you have a 3rd party charging people for copyrighted materials. We can argue that it’s a tool, so you aren’t really “selling” copyrighted data, but that’s the issue that is generally be discussed in these kinds of articles/court cases.
Mostly playing devil’s advocate here (since I don’t think ai should be used commercially), but I’m actually curious about this, since I work in media… You can get away using images or footage for free if it falls under editorial or educational purposes. I know this can vary from place to place, but with a lot of online news sites now charging people to view their content, they could potentially be seen as making money off of copyrighted material, couldn’t they?
It’s not a topic that I’m super well versed in, but here is a thread from a photography forum indicating that news organizations can’t take advantage of fair use www.dpreview.com/forums/thread/4183940.
I think these kinds of stringent rules are why so many are up in arms about how AI is being used. It’s effectively a way for big players to circumvent paying the people who out all the work into the art/music/voice acting/etc. The models would be nothing without the copyrighted material, yet no one seems to want to pay those people.
It gets more interesting when you realize that long term we still need people creating lots of content if we want these models to be able to create things around concepts that don’t yet exist (new characters, genres of music, etc.)
So let’s say I ask a talented human artist the same thing.
Doesn’t this prove that a human, at some level, is storing the data of the Joker movie screenshot somewhere inside of their memory?
what do you call this, then?
thefreemovie.buzz
Nope humans don’t store data perfectly with perfect recall.
Neither do neural networks.
Some do. Should we jail all the talented artists with photographic memories?
If they exactly reproduce others work, and gain a profit for it, a fine would be the minimum.
If they’re copying copyrighted works, usually its a fine, especially if they’re making money from it.
You know that performance artists get sued when they replicate a song in public from memory, right?
I don’t think anyone is advocating to legalize the sale of copyrighted material made via AI.
Humans can get pretty close to perfect recall with enough practice - show a human that exact joker image hundreds of thousands of times, they’re going to be able to remember every detail.
That’s what happened here - the example images weren’t just in the training set once, they are in the training set over and over and over again across hundreds of thousands of websites.
If someone wants these images nobody is going to use AI to access it - they’ll just do a google image search. There is no way Warner Brothers is harmed in any way by this, which is a strong fair use defence.
Artists don’t have hard drives or solid state drives that accept training weights.
When you have a hard drive (or other object that easily creates copies), then the law that follows is copyright, with regards to the use and regulation of those copies. It doesn’t matter if you use a Xerox machine, VHS tape copies, or a Hard Drive. All that matters is that you’re easily copying data from one location to another.
And yes. When a human recreates a copy of a scene clearly inspired by copyrighted data, its copyright infringement btw. Even if you recreate it from memory. It doesn’t matter how I draw Pikachu, if everyone knows and recognizes it as Pikachu, I’m infringing upon Nintendo’s copyright (and probably their trademark as well).
I don’t think that’s a justified conclusion.
If I watched a movie, and you asked me to reproduce a simple scene from it, then I could do that if I remembered the character design, angle, framing, etc. None of this would require storing the image, only remembering the visual meaning of it and how to represent that with the tools at my disposal.
If I reproduced it that closely (or even not-nearly-that-closely), then yes, my work would be considered a copyright violation. I would not be able to publish and profit off of it. But that’s on me, not on whoever made the tools I used. The violation is in the result, not the tools.
The problem with these claims is that they are shifting the responsibility for copyright violation off of the people creating the art, and onto the people making the tools used to create the art. I could make the same image in Photoshop; are they going after Adobe, too? Of course not. You can make copyright-violating work in any medium, with any tools. Midjourney is a tool with enough flexibility to create almost any image you can imagine, just like Photoshop.
Does it really matter if it takes a few minutes instead of hours?
AIs are not humans my dude. I don’t know why people keep using this argument. They specifically designed this thing to scrape copyrighted material, it’s not like an artist who was just inspired by something.
It isn’t human, but that IS how it works.
It’s analyzing material and extracting data about it, not compiling the data itself. In much the same way TDM (textual data mining) analyzes text and extracts information about it for the purposes of search and classification, or sentiment analysis, ECT, an “AI” model analyses material and extracts information on how to construct new language or visual media that relates to text prompts.
It’s important to understand this because it’s core to the fair use defence getting claimed. The models are derived from copyrighted works, but they aren’t themselves infringing. There is precedent for similar cases being fair use.
Photoshop is not human. AutoTune is not human. Cameras are not human. Microphones are not human. Paintbrushes are not human. Etc.
AI did not create this. A HUMAN created this with AI. The human is responsible for the creating it. The human is responsible for publishing it.
Please stop anthropomorphizing AI!
I’ve had this discussion before, but that’s not how copyright exceptions work.
Right or wrong (it hasn’t been litigated yet), AI models are being claimed as fair use exceptions to the use of copyrighted material. Similar to other fair uses, the argument goes something like:
“The AI model is simply a digital representation of facts gleamed from the analysis of copyrighted works, and since factual data cannot be copyrighted (e.g. a description of the Mona Lisa vs the painting itself), the model itself is fair use”
I think it’ll boil down to whether the models can be easily used as replacements to the works being claimed, and honestly I think that’ll fail. That the models are quite good at reconstructing common expressions of copyrighted work is novel to the case law, though, and worthy of investigation.
But as someone who thinks ownership of expressions is bullshit anyway, I tend to think copyright is not the right way to go about penalizing or preventing the harm caused by the technology.
Copyright law is the right tool, but the companies are chasing the wrong side of the equation.
Training should not and I suspect will not be found to be infringement. If old news articles from the NYT can teach a model language in ways that help it review medical literature to come up with novel approaches to cure cancer, there’s a whole host of features from public good to transformational use going on.
What they should be throwing resources at is policing usage not training. Make the case that OpenAI is liable for infringing generation. Ensure that there needs to be copyright checking on outputs. In many ways this feels like a repeat of IP criticisms around the time Google acquired YouTube which were solved with an IP tagging system.
There’s no money for them in that angle though. It’s much easier to sue xerox for enabling copyright violations than the person who used the machine to violate copyright.
Courts have already handled this with copy machines. AI isn’t terribly different, it’s unlikely these suits against model creators succeed.
There’s money (and more importantly, survival) if they can ensure liability of Xerox for infringement on the use of their centralized copiers.
There actually isn’t survival as a company even if they succeed on training but not the other, which I don’t think they realize yet.
As an aside, one of the worst legal takes I read on this was from a GC at the Copyright office during the 70s who extensively used poor analogies to copiers to justify an infringement argument.
Should Photoshop check your image for copyright infringement? Should Adobe be liable for copyright infringing or offensive images users of it’s program create?
If it’s contributing creatively to your work, yeah, totally.
If you ask Photoshop fill to add an italian plumber and you’ve been living under a rock for you life so you don’t realize it’s Mario, when you get sued by Nintendo for copyright infringement it’d be much better policy if it was Adobe on the hook for adding copyrighted material and not the end user.
A better analogy is: if you hired a graphic designer and they gave you copyrighted material, who is liable?
AI is not contributing creatively though, programs do not create.
I am speaking of Photoshop used as a non-AI tool as it has been used to commit copyright infringement for decades before Photoshop fill was a thing. Should it check if your image infringes on copyright?
The graphic designer. If you went ahead and redistributed it you would also be liable. Whatever program he used or it’s developer wouldn’t be liable.
You and I will have to agree to disagree on that Kool-aid, and it’s that disagreement which is core to the model provider being liable for introducing copyright infringement.
Did photoshop create a portion of my image? Did adobe add a “generate the picture I asked for, for me, without my input beyond a typed prompt” as a feature?
Because if they did, 100% yeah, theyre liable.
They actually are not whether you use a prompt to generate the picture or a digitally paint it with a tablet.
The user would be the one committing copyright infringement.
So selling fan fiction and fan-made game continuations and modifications should be legal?
Not the OP, but yes it absolutely should. The idea you can legaly block someones creative expression because they are using elements of culture you have obtained a monopoly of is obscene.
I know it should. Only then we’d have no IP remaining. As it should be, the only case where it’s valid is punishing somebody impersonating the author or falsely claiming authorship, and that’s frankly just fraud.
It should, but also that is significantly different from what an AI model is.
It would be more like a list of facts and information about the structure of another work, and facts and patterns about lots of other similar works; and that list of facts can easily be used to create other, very similar works, but also it can be used to create entirely new works that follow patters from the other works.
In as much as the model can be used to create infringing works -but is not one itself- makes this similar to other cases where a platform or tool can be used in infringing ways. In such cases, if the platform or tool is responsible for reasonable protections from such uses, then they aren’t held liable themselves. Think Youtube DMCA, Facebook content moderation, or even Google Books search. I think this is likely the way this goes; there is just too strong a case (with precedent) that the model is fair use.
By that logic I am also storing that image in my dataset, because I know and remember this exact image. I can reproduce it from memory too.
… Do you think youre a robot?
What’s the difference? I could be just some code in the simulation
Edit: downvoted by people who unironically stan Ted Kaczynski
You ever try to do a public performance of a copyrighted work, like “Happy Birthday to You” ??
You get sued. Even if its from memory. Welcome to copyright law. There’s a reason why every restaraunt had to make up a new “Happy Happy Birthday, from the Birthday Crew” song.
Yeah, but until I perform it without a license for profit, I don’t get sued.
So it’s up to the user to make sure that if any material that is generated is copyright infringing, it should not be used.
Otakon anime music videos have no profits but they explicitly get a license from RIAA to play songs in public.
So? I’m not saying those are fair terms, I would also prefer if that were not the case, but AI isn’t performing in public any more having a guitar with you in public is ripping off Metallica.
You don’t need to perform “for profit” to get sued for copyright infringement.
Is the Joker image in that article derivative or substantially similar to a copyrighted work? Is the query available to anyone who uses Midjourney? Are the training weights being copied from server-to-server behind the scenes? Were the training weights derived from copyrighted data?
Yes and none of that matters in the slightest. By that logic the Library of Babel is also copyright infringement. By that logic my memory of the movie is copyright infringing even if I don’t do anything with it.
You’re taking a fictional work and trying to apply real world laws to it?
Copyright assumes that Library of Babel would take up so much space as it’d be impossible to create.
Which is true. Every possible combination of letters, spaces, and characters would never fit on anything in today’s universe (be it a 24 TB Hard Drive, or even a collection of thousands of them).
Secondly: any computer-generated work is automatically non-copyrighted as per US Law.
Wut? I’m talking about this website here: libraryofbabel.info
en.wikipedia.org/wiki/The_Library_of_Babel
That is just a website demonstrating the concept of a famous science fiction story. Everything I said above remains true.
No storage can hold that. Its likely a computer generated system.
Any computer generated set of text cannot be copyrighted.
Finding any particular book in that “library” would require an index that is the same length / entropy as a full length book in any case. So its a stupid thought experiment to any information theorist. (I’m a Comp. Engineer by trade and have been taught numerous theoretical comp. sci. problems from Nyquist Frequencies, theories on communication, entropy and other such concepts. So this is actually well within my wheelhouse, training an expertise).
Go search for a book in there that replicates Shakespeare’s Hamlet in its entirety. The only way that website could possibly work is if the link you give me has the same entropy / information space as all of Hamlet to begin with.
In the case of copyright law, the Website’s code (including its text generating engine and so forth) is subject to copyright. But the Quadrillion-pages of “text” (most of which is random-gibberish) is not copyrightable.
And that’s the problem. The internet has drastically reduced the cost of copying information, to the point where entirely new uses like this one are now possible. But those new uses are stifled by copyright law that originates from a time when the only cost was that people with gutenberg presses would be prohibited from printing slightly cheaper books. And there’s no discussion of changing it because the people who benefit from those laws literally are the media.
Copyright was literally invented because its cheap and easy to copy information (ie: Printing Press).
When copies are easy, you screw over the original artist. A large scale regulation of copies must be enforced by the central authorities to make sure small artists get the payments that they deserve. It doesn’t matter if you use a printing press, a xerox machine, a photograph, a phonograph, a record, a CD-ROM copy, a tape recorder, or the newest and fanciest AI to copy someone’s work. Its a copy, and therefore under the copyright regulations.
Hard? They wrote:
Yes, look how specific they were. I didn’t even need to get that exact with a google image search. I literally searched for “Joaquin Phoenix Joker” and that exact image was the very first result.
They specified that it had to be that specific actor, as that specific character, from that specific movie, and that it had to be a screenshot from a scene in the movie… and they got exactly what they asked for. This isn’t shocking. Shocking would have been if it didn’t produce something nearly identical to that image.
A more interesting result would be what it would spit out if you asked for say “Heath Ledger Joker movie, 2019, screenshot from a movie, movie scene”.
If that’s hard man i feel sorry for humanity.
Doesn’t seem very hard tonask for an screenshot from the joker movie.
If you read further they also tested many other much more vague prompts, all of which gave intellectual properties they did not have the rights to. The Joaquin Phoenix image isn’t any less illegal, either, though because they don’t have the legal rights to profit off of that IP without permission or proper credit.
We asked A.I. to create a copyrighted image from the Joker movie. It generated a copyrighted image as expected.
Ftfy
What it proves is that they are feeding entire movies into the training data. It is excellent evidence for when WB and Disney decides to sue the shit out of them.
WB and Disney would lose, at least without an amendment to copyright law. That in fact just happened in one court case. It was ruled that using a copyrighted work to train AI does not violate that works copyright.
Using it to train on is very different from distributing derived works.
What do you think the trained model is other than a derived work?
Something transformative from the original works. And arguably not being being distributed. The model producing and distributing derivative works is entirely different though. No one really gives a shit about data being used to train models - there’s nothing infringing about that which is exactly why they won their case. The example in the post is an entirely different situation though.
Does it really have to be entire movies when theres a ton of promotional images and memes with similar images?
Yes. Thats what these things are, extremely large catalogues of data. As much data as possible is their goal.
True but it didn’t pick some random frame somewhere in the movie it chose a extremely memorable shot that is posted all over the place. I won’t deny that they are probably feeding it movies but this is not a sign of that.
This image is literally the top result on Google images for me.
Why would it pick some random frame in the middle of its data set instead of a frame it has the most to reference. It can still use all those other frames to then pick the frame if has the most references to.
But im starting to think maybe i misunderstood the comment i replied to.
Sorry, im way out of context with my reply, totally my fault for reflexively replying.
Uhhh would you accept i didnt have my coffee yet and hadnt got out of bed yet as an explanation?
Haha it happens
Promotional images are still under copyright.
We should find all the memers and throw them in jail.
Will someone think of the shareholders!?
I think it’s much more likely whatever scraping they used to get the training data snatched a screenshot of the movie some random internet user posted somewhere. (To confirm, I typed “joaquin phoenix joker” into Google and this very image was very high up in the image results) And of course not only this one but many many more too.
Now I’m not saying scraping copyrighted material is morally right either, but I’d doubt they’d just feed an entire movie frame by frame (or randomly spaced screenshots from throughout a movie), especially because it would make generating good labels for each frame very difficult.
I just googled “what does joker look like” and it was the fourth hit on image search.
Well, it was actually an article (unrelated to AI) that used the image.
But then I went simpler – googling “joker” gives you the image (from the IMDb page) as the second hit.
I have that exact same .jpeg stored on my computer and I don’t even know where it came from. I don’t even watch superhero films
And if you tried to sell that, you would be breaking the law.
Which is what these AI models are doing
They’re not selling it though, they’re selling a machine with which you could commit copyright infringement. Like my PC, my HDD, my VCR…
No, they are selling you time in a digital room with a machine, and all of the things it spits out at you.
You dont own the program generating these images. You are buying these images and the time to tinker with the AI interface.
I’m not buying anything, most AI is free as in free beer and open source e.g. Stable Diffusion, Mistral…
Unlike hardware it’s actually accessible to everyone with sufficient know-how.
Youre pretty young, huh. When something on the internet from a big company is free, youre the product.
Youre bug and stress testing their hardware, and giving them free advertising. While using the cheapest, lowest quality version that exists, and only for as long as they need the free QA.
The real AI, and the actual quality outputs, cost money. And once they are confident in their server stability, the scraps youre picking over will get a price tag too.
Literally what are you on about? I run my models locally, the only hardware i am stress testing is my own.
I don’t support commercialization of anything, least of all AI, and the highest quality outputs come from customized refined models in the open source and AI art communities, not anything made by a corpo.
I think you must be literally 12 yourself if you think you can comment on this tech without even understanding models and weights are something you download if you want anything beyond fancy often wrong Google search, they’re not run in the “cloud” like your fancy iPad web apps and they are open source.
The way it was done if I remember correctly is that someone found out v6 was trained partially with Stockbase images-caption pairs, so they went to Stockbase and found some images and used those exact tags in the prompts.
<img alt="" src="https://lemmy.dbzer0.com/pictrs/image/f98a39cd-86a5-4481-a51b-a3a07e145328.jpeg">
The image it generated is really widespread
When they asked for an Italian video game character it returned something with unmistakable resemblance to Mario with other Nintendo property like Luigi, Toad etc. … so you don’t even have to ask for a “screencapture” directly for it to use things that are clearly based on copyrighted characters.
you’re still asking for a character from a video game, which implies copyrighted material. write the same thing in google and take a look at the images. you get what you ask for.
you can’t, obviously, use any image of Mario for anything outside fair use, no matter if AI generated or you got it from the internet.
Also ask literally any human and they’ll probably name Mario first. Not just top 10, number 1.
But the AI didn’t credit the clear inspiration. That’s the problem, that is what makes it theft: you need permission to profit off of the works of others.
but that’s exactly what I said. you can’t grab an image of Mario from google and profit from it as you can’t draw a fan art of Mario and profit from it as well as you can’t generate an image of Mario and profit from it.
It doesn’t matter if you’re generating it with software or painting it on canvas, if it contains intellectual property of others, you can’t (legally) use it for profit.
however, generating it and posting it as a meme on the internet falls under fair use, just like using original art and making a meme.
The users are allowed to ask for those things
The AI company should not be allowed to give it in return for monetary gain.
If you asked me to draw an Italian video game character, I’d draw Mario too. Why can’t an AI make copyrighted character inspired pics as long as they aren’t being sold?
Well that’s exactly the problem. If people use AI generated images for commercial purposes they may accidentally infringe on someone else’s copyright. Since AI models are a black box there isn’t really a good way to avoid this.
Sure there is, force the AI to properly credit artists and if they don’t have permission to use the character then the prompt fails. Or the AI operators have no legal rights to charge for services and should be sued into the ground.
You credited it just now as Mario, a Nintendo property, which the AI failed to do. Plus, if you were paid to draw Mario then you’d have broken laws about IP. Why don’t those same rules apply to AI?
Voyager just loaded a copyrighted image on my phone. Guess someone’s gonna have to sue them too.
I just remembered a copyrighted image. Oops.
Hey, I bet there were complaints about Google showing image results at some point too! Lol
Yeah man, Voyager is making millions with the images on the app. It makes me so mad, they Voyager people make you think they are generating content on their own, but in reality is just feeding you unlicensed content from others.
You’re completely missing the point. Making money doesn’t change the legality. YouTube was threatened by the RIAA before they even started showing ads. Displaying an image from a copyrighted work on an AI platform is not much different technologically than Voyager or even Google Images displaying the same image, and both could also be interpreted as “feeding you unlicensed content from others.”
Except that it actually does? That’s the point of copyright laws. The LLM/AIs are using copyright protected material as source without paying for it, and then selling it’s output as "original '.
Oh! That’s why torrent sites aren’t under constant threat despite hosting tons of free copyright material.
Hang on… Yes they are!
Wow, voyager app is very nice!
I already know I’m going to be downvoted all to hell, but just putting it out there that neural networks aren’t just copy pasting. If a talented artist replicates a picture of the joker almost perfectly, they are applauded. If an AI does it, that’s bad? Why are humans allowed to be “inspired” by copyrighted material, but AIs aren’t?
Because the original Joker design is not just something that occurred in nature, out of nowhere. It was created by another artist(s) who don’t get credit or compensation for their work.
When YouTube “essayists” cobble script together by copy pasting paragraphs and changing some words around and then then earn money off the end product with zero attribution, we all agree it’s wrong. Corporations doing the same to images are no different.
But of course you can’t turn around and sell that picture of the Joker that you made. That’s obvious.
The problem in here is that while the Joker is a pretty recognizable cultural icon, somebody using an AI may have genuinely original idea for an image that just happens to have been independently developed by someone before. As a result, the AI can produce an image that’s a copy or close reproduction of an original artwork without disclosing its similarity to the source material. The new “author” then will unknowingly rip off the original.
The prompts to reproduce joker and other superhero movies were quite specific, but asking for “Animated Sponge” is pretty innocent. It is not unthinkable that someone may not be familiar with Mr. Squarepants and think they developed an original character using AI
That’s a good point. Musicians have been known to accidentally reproduce the same beat as another musician (was is done subconsciously or just coincidence?). Some books are strikingly similar to other books that it makes you wonder if it was a rip off or just coincidence. So it’s nothing new, but it may become more prevalent with AI. This could spawn a new industry of investigators ensuring your AI generated art isn’t infringing on any copyrights 🤔
This might be the best point I’ve seen around this topic – have not seen this addressed before.
It’s on the person using any AI tools to verify that they aren’t infringing on anything if they try to market/sell something generated by these tools.
That goes for using ChatGPT just as much as it goes for Midjourney/Dall-E 3, tools that create music, etc.
And you’re absolutely right, this is going to be a problem more and more for anyone using AI Tools and I’m curious to see how that will factor in to future lawsuits.
I could see some new factor for fair use being raised in court, or else taking this into account under one of the pre-existing factors.
Tons of human made art isn’t inspired by nature. Rather it’s inspired by other human made art. Neural networks don’t just copy paste like a yt plagiarist. You can ask an AI to plagiarize but no guarantee it’ll get it right.
I think the problem is that you cannot ask AI not to plagiarize. I love the potential of AI and use it a lot in my sketching and ideation work. I am very wary of publicly publishing a lot of it though, since, especially recently, the models seem to be more and more at ease producing ethically questionable content.
That’s an interesting point. We’re forced to make a judgement call because we don’t have total control over what it generates.
you aren’t making any sense. people did fanarts and memes of the joker movie like crazy, they were all over the internet. there are tons and tons of fan arts of copyrighted material.
they fall under fair use and no one losses money because fan arts can’t be used for commercial purposes, that would fall outside fair use and copyright holders will sue, of course.
how is that different from the AI generating an image containing copyrighted material? if someone started generating images of the joker and then selling them, yeah, sue the fuck out of them. but generating it without any commercial purpose is not illegal at all.
In many cases the AI company is “selling you” the image by making users pay for the use of the generator. Sure, there are free options, too - but just giving you an example.
With that line of argument you can sue developers of 2d painting programmes and producers of graphics tablets. And producers of canvas, brushes and paint. Maybe even the landlord for renting out a studio? It’s all means of production.
You make an interesting point, but I can’t help but feel it’s not completely the same and you’re reaching a bit. I feel like it’d be closer if GIMP, next to shape tools for squares and circles, literally had a ‘Joaquin Phoenix as the Joker’ shape. The crux of the issue as I see it in this part of the legal debate is whether or not AI companies are willing participants in the creation of potentially copyright infringing media.
It reminds me a bit of the debate around social media platforms and if they’re legally responsible for the illegal or inappropriate content people keep uploading.
So you watched that Hbomberguy video where he randomly tacked on being wrong about AI in every way, using unsourced, uncited claims that have nothing to do with Somerton or that Illuminaughti chick and will age extremely poorly and made that your entire worldview? Okay
Actually no, but thanks for letting me know, I like his content.
Because AI isn’t inspired to do anything it has no feelings its just code.
That’s why I put inspired in quotes. It’s analogous to a human seeing something on the Internet and coming up with similar art or building upon it.
To me the tipping point is if someone is getting paid. You can be inspired by the joker character and make your own content/characters that are similar, but you can’t just start making iterations of the joker and selling it for money (legally at least).
With Gen AI, companies are selling access to models that can and are being used to generate copyrighted material. Meaning these companies are making money off of something they didn’t create and don’t own.
If it’s an open sourced model, then I don’t care, but I think there is a problem when these models can take others work and charge money for it.
I think the onus is on the user of the AI. I could use Photoshop to make a joker pic and sell it for money. Should Photoshop be banned? The AI lets me do the same thing faster.
And that’s probably we’re things will land, but it is an interesting grey area to determine how much can the tool generate vs the person. Maybe it’s a glimpse into the challenges of a post scarcity or post ai world.
Why are feelings important?
.
Wah fuckin Wah.
Copyright issues aside, can we talk about how this implies accurate recall of an image from a never before achievable data compression ratio? If these models can actually recall the images they have been fed this could be a quantum leap in compression technology.
Holy shit I didn’t even think about that.
Essentially the model is compressing the image into a prompt.
Instead of the bitmap being 8MB being condensed down into whatever the jpeg equivalent is, it’s still more than a text file with that exact prompt that gave.
But it’s not deterministic.
It’s just a little bit lossy
I mean, that randomness is just faked. Keep a consistent seed and you’ll get consistent results.
.
I like that thought too, surely better than calling it AI.
I mean, only if you have the entire model downloaded and your computer does a ton of work to figure it out. And then if any new images are created the model will have to be retrained. Maybe if there were a bunch of presets of colors to choose from that everyone had downloaded and then you only send data describing changes to the image
You can hardly consider it compression when you need a compute expensive model with hundreds of gigabytes (if not bigger) to accurately rehydrate it
You can run Stable Diffusion with custom models, variational auto encoders, LoRAs, etc, on an iPhone from 2018. I don’t know what the NYTimes used, but AI image generation is surprisingly cheap once the hard work of creating the models is done. Most SD1.5 model checkpoints are around 2GB in size.
Edit: But yes, the idea of using this as image compression is absurd.
It’s not as accurate as you’d like it to be. Some issues are:
Also it’s not all that novel. People have been doing this with (variational) autoencoders (another class of generative model). This also doesn’t have the flaw that you have no easy way to compress new images since an autoencoder is a trained encoder/decoder pair. It’s also quite a bit faster than diffusion models when it comes to decoding, but often with a greater decrease in quality.
Most widespread diffusion models even use an autoencoder adjacent architecture to “compress” the input. The actual diffusion model then works in that “compressed data space” called latent space. The generated images are then decompressed before shown to users. Last time I checked, iirc, that compression rate was at around 1/4 to 1/8, but it’s been a while, so don’t quote me on this number.
edit: fixed some ambiguous wordings.
Results vary wildly. Some images are near pixel perfect. Others, it clearly knows what image it is intended to be replicating. Like it gets all the conceptual pieces in the right places but fails to render an exact copy.
Not a very good compression ratio if the image you get back isn’t the one you wanted, but merely an image that is conceptually similar.
Optimal tip-to-tip efficiency has been achieved.
I was thinking about this back when they first started talking about news articles coming back word for word.
There’s no way for us to tell how much of the original data even in a lossy fashion can be directly recovered. If this was as common as these articles would leave you to believe you just be able to pull anything you wanted out on demand.
But here we have every news agency vying to make headlines about copyright infringement and we’re seeing an article here and there with a close or relatively close result
There are millions and millions of people using this technology and most of us aren’t running across blatant full screen reproductions of stuff.
You can tell from some of the artifacts that they’ve trained from some watermark images because the watermarks kind of show up but for the most part you wouldn’t know who made the watermarking if all the watermarking companies didn’t use rather unique patterns.
The image that we’re seeing on this news site of the joker is quite exceptional, even from a lossy standpoint, but honestly it’s just feeding the confirmation bias.
“how much of the data is the original data”?
Even if you could reverse the process perfectly, what you would prove is that something fed into the AI was identical to a copyrighted image. But the image’s license isn’t part of that data. The question is: did the license cover use as training data?
In the case of watermarked images, the answer is clearly no, so then the AI companies have to argue that only tiny parts of any given image come from any given source image, so it still doesn’t violate the license. That’s pretty questionable when waternarks are visible.
In these examples, it’s clear that all parts of the image come directly or indirectly (perhaps some source images were memes based on the original) from the original, so there goes the second line of defence.
The fact that the quality is poor is neither here nor there. You can’t run an image through a filter that adds noise and then say it’s no longer copyrighted.
The trained model is a work derived from masses of copywrite material. Distribution of that model is infringement, same as distributing copies of movies. Public access to that model is infringement, just as a public screening of a movie is.
People keep thinking it’s “the picture the AI drew” that’s the issue. They’re wrong. It’s the “AI” itself.
If you ignore the fact that the generated images are not accurate, maybe.
They are very similar so they are infringing but nobody would use this method for compression over an image codec
I made a novel type of language model, and from my calculations after about 30gb it would cross over an event horizon of compression, where it would hold infinitely more pieces of text without getting bigger. With lower vocabulary it would do this at a lower size. For images it’s still pretty lossy but it’s pretty cool. Honestly I can’t mental image much better without drawing it out.
Hmm this sounds like a similar technology to the time cube
Chat GPT it’s over 500 gigs of training data plus over 300 gigs of RAM, and Sam Altman has been quite adamant about how another order of magnitude worth of storage capacity is needed in order to advance the tech.
I’m not convinced that these are compressed much at all. I would bet this image in its entirety is actually stored in there someplace albeit in an exploded format.
I purchased a 128 GB flash drive for around 12-15$ (I forgot the exact price) last year, and on Amazon, there are 10 TB hard drives for $100. So, the actual storage doesn’t seem to be an issue.
RAM is expensive 128 GB of RAM on Amazon is $500.
But then again, I am talking about the consumer grade stuff. It might be different for the people who are making AI’s as they might be using the industrial/whatever it’s called grade stuff.
It depends on what kind of RAM you’re getting.
You could get Dell R720 with two processors and 128 gigs of RAM for $500 right now on eBay, but it’s going to be several generations old.
I’m not saying that the model is taking up astronomical amounts of space, but it doesn’t have to store movies or even high resolution images. It is also not being expected to know every reference, just the most popular ones.
I have 120tb storage server in the basement. So the footprint of this learning model is not particularly massive by comparison, but It does contain this specific whole joker image. It’s not something that could have been generated without the original to draw from.
In order to build a bigger model they would need not necessarily just more storage but actually a new way of having more and faster RAM connected to lower latency storage. LLMs are the kinds of software that become hard to subdivide to be distributed across purpose-built arrays of hardware.
Compression is actually a mathematical field that’s fairly well explored, and this isn’t compression. There are theoretical limits on how much you can compress data, so the data is always somewhere, either in the dictionary or the input. Trained models like these are gigantic, so even if it was perfect recall the ratio still wouldn’t be good. Lossy “compression” is another issue entirely, more of an engineering problem of determining how much data you can throw out while making acceptable compromises.
Hey AI, I’m ready to download a car.
No it didn’t.
What is the non paywall version of this
archive.ph/58A8B
Asks AI to generate copyrighted image; AI generates a copyrighted image.
Pikatchu.jpg
It is a point against those “it’s just like humans learning” arguments.
I mean if you asked a human to draw a copyrighted image you would also get the copyrighted image. If the human had seen that copyrighted image enough times they might even have memorised The smallest details and give you a really good or near perfect copy.
I agree with your point but this example does not prove it.
How so? A human is also very much capable of drawing a copywrited image when asked.
…and they are also infringing copyright if they have not been given a right to copy that work.
Yes? I don’t see your point
Maybe you weren’t trying to make the point I thought you were.
I assumed you were trying to say that a human can draw a picture of (for example) the Joaquin Phoenix Joker and not be committing copyright infringement, therefore an AI can do the same. I was pointing out that the basis of that argument is false. A human drawing that would be infringing the copyright.
Not from memory, without looking at the original during painting - at least not to this level of detail. No human will just incidentally “learn” to draw such a near-perfect copy. Not unless they’re doing it on purpose with the explicit goal of “learn to re-create this exact picture”. Which does not describe how any humans typically learn.
.
“Generate this copyrighted character”
“Look, it showed us a copyrighted character!”
Does everyone that writes for the NYTimes have a learning disability?
.
The point is to prove that copyrighted material has been used as training data. As a reference.
If a human being gets asked to draw the joker, gets a still from the film, then copies it to the best of their ability. They can’t sell that image. Technically speaking they’ve broken the law already by making a copy. Lots of fan art is illegal, it’s just not worth going after (unless you’re Disney or Nintendo).
As a subscription service that’s what AI is doing. Selling the output.
Held to the same standards as a human artist, this is illegal.
If AI is allowed to copy art under copyright, there’s no reason a human shouldn’t be allowed to do the same thing.
Proving the reference is all important.
If an AI or human only ever saw public domain artwork and was asked to draw the joker, they might come up with a similar character. But it would be their own creation. There are copyright cases that hinge on proving the reference material. (See Blurred Lines by Robin Thick)
The New York Times is proving that AI is referencing an image under copyright because it comes out precisely the same. There are no significant changes at all.
In fact even if you come up with a character with no references. If it’s identical to a pre-existing character the first creator gets to hold copyright on it.
This is undefendable.
Even if that AI is a black box we can’t see inside. That black box is definitely breaking the law. There’s just a different way of proving it when the black box is a brain and when the black box is an AI.
Tough question is, can a tool be infringing anything?
Although I’d see a legal case if AI companies were to bill picture by picture, but now they are just billing for a tool subscription.
Still, would Microsoft be liable for my copy-pastes if they charged a penny every time I use it, or am I, if I sell a art piece that uses that infringing image?
AI could be scraping that picture from anywhere.
Who created this image in your view then, who is liable?
Can a tool create? It generated.
Anyway, in case like this, is creation even a factor in liability?
In my opinion one who gets monetary value first from the piece should be liable.
NYTimes?
“I didnt kill him, officer, my murder robot did. Oh, sure, I built it and programmed it to stab jenkins to death for an hour. Oh, yes, I charged it, set it up in his house, and made sure all the programming was set. Ah, but your honor, I didnt press the on switch! Jenkins did, after I put a note on it that said ‘not an illegal murderbot’ next to the power button. So really, the murderbot killed him, and if you like maybe even jenkins did it! But me? No, sir, Im innocent!”
How is this example relevant? You created the programming.
And someone created the AI programming too.
Then someone trained that AI.
It didn’t just come out of the aether, there’s a manual on how to do it.
Yes, but in
yourprevious exampleyouperson specifically created a machine to stab a specific person.Example would be apt, if you created a program that generates programming for industrial machines to insert things in to stuff and then you uploaded a generated program without checking the code and it stabbed some random guy.
That was not my example. The murder machine was someone else.
Sorry, my bad. Fixed
The liability of industrial machines is actually quite apt.
If you design a machine that kills someone during reasonable use. You are liable.
Aircraft engineers have a 25 year liability on their work. A mistake they might make could kill hundreds.
There is always a human responsible for the actions of a machine. Even unintended results have liability.
If you upload a program to a machine and someone dies as a result you’re in hot water.
Moving away from life and death, unintended copyright infringement by a machine hasn’t been tested. But it’s likely it will be ruled that at least some of the builders of that machine are responsible.
AI “self-driving” cars are getting away with it by only offering an assist to driving. Keeping the driver responsible. But that’s possible because you need a license to drive a car in the first place.
AI images like this are the equivalent of a fully self driving car. You set the destination, it drives you there. The liability falls on the process of driving, or the process of creating. The machine doing that means designers are then liable.
Lets call it assisting image creation then.
AI owners would love to do that.
Copyright owners would not.
Hence the legal battles.
So by that logic. I prompted you with a question. Did I create your comment?
I used you as a tool to generate language. If it was a Pulitzer winning response could I gain the plaudits and profit, or should you?
If it then turned out it was plagiarism by yourself, should I get the credit for that?
Am I liable for what you say when I have had no input into the generation of your personality and thoughts?
The creation of that image required building a machine learning model.
It required training a machine learning model.
It required prompting that machine learning model.
All 3 are required steps to produce that image and all part of its creation.
The part copyright holders will focus on is the training.
Human beings are held liable if they see and then copy an image for monetary gain.
An AI has done exactly this.
It could be argued that the most responsible and controlled element of the process. The most liable. Is the input of training data.
Either the AI model is allowed to absorb the world and create work and be held liable under the same rules as a human artist. The AI is liable.
Or the AI model is assigned no responsibility itself but should never have been given copyrighted work without a license to reproduce it.
Either way the owners have a large chunk of liability.
If I ask a human artist to produce a picture of Donald Duck, they legally can’t, even though they might just break the law Disney could take them to court and win.
The same would be true of any business.
The same is true of an AI as either its own entity, or the property of a business.
I’m not non-sentient construct that creates stuff.
…and when the copyright law was written there was no non-sentient things gererating stuff.
There is literally no way to prove whether you’re sentient.
Decart found that limitation.
The only definition in law is whether you have competency to be responsible. The law assumes you do as an adult unless it’s proven you don’t.
Given the limits of AI the court is going to assume it to be a machine. And a machine has operators, designers, and owners. Those are humans responsible for that machine.
It’s perfectly legitimate to sue a company for using a copyright breaking machine.
You almost seem like you get the problem, but then you flounder away.
Law hasn’t caught up with the world with generative programs. A.I will not be considered sentient and they will have this same discussion in court.
It doesn’t matter whether AI is sentient or not. It has a designer, trainer, and owner.
Once you prove the actions taken by the AI, even as just a machine, breach copyright liability is easily assigned.
Argee to disagree and time will tell, but you must see there are factors that haven’t existed before in the history of humanity.
Who knows how the laws will change because of AI. But as the law currently stands it’s just a matter of proving it to a court. That’s the main barrier.
This is strong evidence an AI is breaking the law.
That joker could have been somebodys avatar picture with matching username.
A.I. can’t understand copyright and useful A.I can’t be build by protecting it from every material somebody thinks is their IP. It needs to learn to understand humans and needs human material to do so. Shitload of it. Who’s up for some manual filtering?
If we go by NYTimes standards we better mothball the entire AI endeavor.
That’s why it’s a massive legal fight.
They’ll delay a ruling as long as possible.
They’re definitely developing a new model on vetted public domain data as we speak. They just need to delay legal action long enough to get that new model to launch.
This is the same thing YouTube did. Delay all copyright claims in court, blaming users, then put their copyright claim system in place that massively advantages IP owners.
They are showing that the author of the tool has comitted massive copyright infringement in the process construction of the tool.
…unless they licensed all the copyright works they trained the model on. (Hint: they didn’t, and we know they didn’t because the copyright holders haven’t licensed their work for that purpose. )
It doesn’t matter if a company charges or not for anything. It’s not a factor in copyright law.
Is this really true? Breaking the law implies contravening some legislation which in the case of simply drawing a copyrighted character, you wouldn’t be in most jurisdictions. It’s a civil issue in that if some company has the rights to a character and some artist starts selling images of that character then whoever owns the rights might sue that artist for loss of income or unauthorised use of their intellectual property.
Regardless, all human artists have learned from images of characters which are the intellectual property of some company.
If I hired a human as an employee, and asked them to draw me a picture of the joker from some movie, there’s no contravention of any law I’m aware of, and the rights holder wouldn’t have much of a claim against me.
As a layperson, who hasn’t put much thought into this, the outcome of a claim against these image generators is unclear. IMO, it will come down to whether or not a model’s abilities are significantly derived from a specific category of works.
For example, if a model learned to draw super heros exclusively from watching marvel movies then that’s probably a copyright infringement. OTOH if it learned to draw super heroes from a wide variety of published works then IMO it’s much more difficult to make a case that the model is undermining the right’s holder’s revenue.
Copyright law is incredibly far reaching and only enforced up to a point. This is a bad thing overall.
When you actually learn what companies could do with copyright law, you realise what a mess it is.
In the UK for example you need permission from a composer to rearrange a piece of music for another ensemble. Without that permission it’s illegal to write the music down. Even just the melody as a single line.
In the US it’s standard practice to first write the arrangement and then ask the composer to licence it. Then you sell it and both collect and pay royalties.
If you want to arrange a piece of music in the UK by a composer with an American publisher, you essentially start by breaking the law.
This all gives massive power to corporations over individual artists. It becomes a legal fight the corporation can always win due to costs.
Corporations get the power of selective enforcement. Whenever they think they will get a profit.
AI is creating an image based on someone else’s property. The difference is it’s owned by a corporation.
It’s not legitimate to claim the creation is solely that of the one giving the instructions. Those instructions are not in themselves creating the work.
The act of creating this work includes building the model, training the model, maintaining the model, and giving it that instruction.
So everyone involved in that process is liable for the results to differing amounts.
Ultimately the most infringing part of the process is the input of the original image in the first place.
So we now get to see if a massive corporation or two can claim an AI can be trained on and output anything publicly available (not just public domain)without infringing copyright. An individual human can’t.
I suspect the work of training a model solely on public domain will be complete about the time all these cases get settled in a few years.
Then controls will be put on training data.
Then barriers to entry to AI will get higher.
Then corporations will be able to own intellectual property and AI models.
The other way this can go is AI being allowed to break copyright, which then leads to a precedent that breaks a lot of copyright and the corporations lose a lot of power and control.
The only reason we see this as a fight is because corporations are fighting each other.
If AI needs data and can’t simply take it publicly from published works, the value of licensing that data becomes a value boost for the copyright holder.
The New York Times has a lot to gain.
There are explicit exceptions limited to copyright law. Education being one. Academia and research another.
All hinge into infringement the moment it becomes commercial.
AI being educated and trained isn’t infringement until someone gains from published works or prevents the copyright holder from gaining from it.
This is why writers are at the forefront. Writing is the first area where AI can successfully undermine the need to read the New York Times directly. Reducing the income from the intellectual property it’s been trained on.
This isn’t the issue. The copyright infringement is the creation of the model using the copywrite work as training data.
All NYT is doing is demonstrating that the model must have been created using copywrite works, and hence infringement has taken place. They are not stating that the model is committing an infringement itself.
I agree, but it is useful to ask if a human isn’t allowed to do something, why is a machine?
By putting them on the same level. A human creating an output vs. an AI creating an output, it shows that an infringement has definitely taken place.
I find it helpful to explain it to people as the AI breaching copyright simply because from that angle the law can logically be applied in both scenarios.
Showing a human a piece of copyright material available to view in public isn’t infringement.
Showing a generic AI a piece of copyright material available to view in public isn’t infringement.
The infringing act is the production of the copy.
By law a human can decide to do that or not, they are liable.
An AI is a program which in this case is designed to have a tendency to copy and the programmer is responsible for that part. That’s not necessarily infringement because the programmer doesn’t feed in copyright material.
But the trainer showing an AI known to have a tendency to copy some copyright material isn’t much different to someone putting that material on a photocopier.
I get many replies from people who think this isn’t infringement because they believe a human is actually allowed to do it. That’s the misunderstanding some have. The framing of the machine making copies and breaching copyright helps. Even if ultimately I’m saying the photocopier is breaching copyright to begin with.
Ultimately someone is responsible for this machine, and that machine is breaking copyright. The actions used to make, train, and prompt the machine lead to the outcome.
As the AI is a black box, an AI becomes a copyright infringing photocopier the moment it’s fed copyright material. It is in itself an infringing work.
The answer is to train a model solely on public domain work and I’d love to play around with that and see what it produces.
That’s called fair use. It’s a non-issue.
It’s not selling that image (or any image), any more than a VCR is selling you a taped version of Die Hard you got off cable TV.
It is a tool that can help you infringe copyright, but as it has non-infringing uses, it doesn’t matter.
Then who created this image in your view?
If someone copies a picture from a cartoon who created it?
What point do you think youre making? The answer to this question supports their point.
I wasn’t arguing with them lol just wondered their opinion.
It does feel weird to me that if someone draws a copy of something people don’t think they’ve created anything. That somehow the original artist created it.
The person who created the cartoon in the first place.
Try painting a Disney character on the wall of a waiting room.for children.
theguardian.com/…/robert-jenrick-has-cartoon-mura…
So the copyer didn’t create anything? Odd way to look at it to me.
The copier didn’t create any Intellectual property. They copied it.
Copy right. The right to copy.
It’s fairly fundamental.
That’s irrelevant, the issue is whether the machine is committing a crime, or the person
Machines aren’t culpable in law.
There is more than one human involved in creating and operating the machine.
The debate is, which humans are culpable?
The programmers, trainers, or prompters?
The prompters. That is easy enough. If I cut butter with a knife it’s okay, if I cut a person with a knife - much less so. Knife makers can’t be held responsible for that, it’s just nonsense.
If you try to bread with an autonomous knife and the knife kills you by stabbing you in the head. Is it solely your fault?
That depends on whether the autonomous knife is designed dangerously and it’s a common occurrence, or whether I was being a moron and essentially rigged it to stab me, akin to asking for copyright material from an AI and getting it (scene from a movie, characters part of intellectual property etc)
So you’re saying if it’s easy to accidentally get copyright images out of this AI by prompting ordinary worlds. Then the AI designers have some questions to answer.
Accidentally? No. By typing in a highly specific prompt that specifies the exact IP? Yes.
“The Joker” is a generic description of a character. Going back to medieval courts.
If the result is a copyrighted version of that character that’s not the promoters fault.
That’s the fault of the ones who compile the training data.
VCR makers do not claim to create original programming.
Why does that matter?
Because they aren’t doing anything to violate copyright themselves. You might, but that’s different. AI art is created by the software. Supposedly it’s original art. This article shows it is not.
It is original art, even the images in question have differences, but it’s ultimately on the user to ensure they do not use copyrighted material commercially, same as with fanart.
If I draw a very close picture to a screenshot of a Mickey Mouse cartoon and try to pass it off as original art because there are a handful of differences, I don’t think most people would buy it.
What relevance does this have to AI?
It has relevance to what counts as an original artwork.
This is what you said:
No it is not. They do not have enough differences to be considered original in any court of law.
???
If I ask for an image of Joaquin Phoenix as The Joker from the movie The Joker, then yes it will not be original.
If I ask for original drawings off original ideas it will be original.
Therefore AI can be used for both.
Therefore the technology itself is not infringing, but only specific uses of it are, same as with a VCR, an HDD and our very brains. This should be obvious, and NYT knows it’s going to lose and that’s why they are now developing their own model. This case is just to stall the industry until old money corpos can catch up to avoid being disrupted out of existence. It has zero legal ground
Again, VCRs and hard drives can’t create content. They can only capture content. AI can create content, but it is not always original. Which is the problem. No one is trying to sue them over things that are credibly original.
It is no more legal for you to tell an AI to make you a picture of the Joker as it is to ask a human artist to do it. And if the human artist did it, WB/DC would be within their rights to take them to court because it would violate both trademark and copyright. They usually don’t, but they are within their rights.
You can ask a VCR or a hard drive to draw you a picture of The Joker all day. They won’t because they can’t.
If AI was only capable of creating original artworks, this would not be an issue.
There is no difference, a camcorder creates content, but it doesn’t mean they’re banned just because you can film a film with one.
Yes I agree, it is not legal to replicate copyrighted works regardless.
But again, human artists aren’t illegal just because they can infringe copyright in a hypothetical scenario. Same with AI. The machine is non-infringing, the prompt operator can infringe copyright if they try, and then they are responsible under law, same as a human artist would be.
The machine is blameless regardless of what it was trained on.
Nope. Camcorders do not create content. They record content. Camcorders do not create anything. That is a ridiculous claim. I cannot point a camcorder at you and have it make you look like Heath Ledger.
AI creates content. It can make things that literally don’t exist. If I tell it to make me Heath Ledger as The Joker fighting Jack Nicholson as the Joker, it can create it. A camcorder can’t. A VCR can’t. A hard drive can’t. I have no idea why you don’t understand the difference between creating content and recording content.
I also said nothing about the AI itself being illegal, so I also have no idea where you’re getting that from. I said it is violating copyright and trademark when it creates such images. Because it is.
Hence the lawsuits. Hence the lack of such lawsuits against camcorders, VCRs and hard drives.
It flat out isn’t lol, I don’t know what to tell you bud, but you have no idea what you’re talking about, and there was a lawsuit against VCRs:
…wikipedia.org/…/Sony_Corp._of_America_v._Univers….
In fact it’s quite likely that the OpenAI decision will be based upon this.
It’s not that complicated to understand:
AI trains on images (fair use) -> Prompter inserts prompt -> output can be copyright infringing or not, depending on the prompt, same as a brain, hard drive, VCR and an HDD.
The metaphysics of what counts as creation vs recording are irrelevant, everything you’re saying is just flat out irrelevant.
Yes, there was a lawsuit against VCRs. It had nothing to do with original content.
Nope, that is not in any legal definition of fair use.
What prompt do I enter to get a VCR to make me a picture of Jack Nicholson’s Joker fighting Heath Ledger’s Joker?
Do I press both rewind and fast forward at once to access the secret content generation menu?
You dress up and film it with your buddies and record it to a second gen tape lol? Again, the content generation aspect is irrelevant, what matters is whether a piece of equipment is made to infringe copyright or not, AI isn’t, neither are VCRs, that’s law, simple as really.
And yes training is fair use, it better be, I train my brain on images all the time.
Which means it is not creating content. It is recording content. Which was my point.
Please back this up. Your brain is not a computer. Furthermore, even if it was, someone else would be training it and you cannot legally train someone else on copyrighted material that you have not licensed, which is why schools have to license textbooks and a teacher that teaches from an unlicensed textbook can be sued. That’s the entire impetus for the Open Textbook Library. The Open Textbook Library would literally not need to exist if training material was not protected by copyright.
You see, the problem here is that you keep claiming things that are the opposite of what these companies are getting sued for doing. And yet those suits aren’t getting laughed out of court. Doesn’t that tell you that maybe your ideas of how the law works here are wrong?
I have been studying U.S. copyright and trademark law for over 15 years. How long have you been studying it?
You’re creating a film that wasn’t there before though? Recording is creation, I don’t know what you think cameras do or how video is made 🙄
Recording is creation too and it can all be IP infringing at any stage, it’s all completely irrelevant how it was made in 99% of cases (exception being reverse engineering)
Brain is absolutely just a computer lol. I look at images - I remember images, I’m influenced by images. Is that fair use? If so - so is AI, because that’s all it does.
Like the lawsuit against Stable Diffusion & Midjourney? You know, the class action one that was dismissed precisely because the end work (output) was non-infringing, and training itself (and by extension, the tech) was not an infringement?
reuters.com/…/judge-pares-down-artists-ai-copyrig…
Educate yourself.
Read up on the Sony vs United Studios too if you want to get in intro on law stuff. Start with Wikipedia. Then watch as this court case against OpenAI unfolds like I said it will.
You keep banging on about the same few points that are all incorrect, proven time and time again, I have better things to do than to respond to this any further.
I guess you need to study it some more then, you don’t even know the basics.
I swear you reddit refugee armchair experts need to go back.
Not according to the law. And if you disagree, find me the law that defines recording as creation.
One of many. One getting dismissed does not equal all getting dismissed.
Irrelevant to this subject. Original content was not at issue.
This was literally directly connected to my own business for 15 years. One I ran legally. Because I made sure to study copyright and trademark law as much as possible so my company wouldn’t ever violate it.
And since I’m a ‘refugee armchair expert,’ from where did you get your law degree? Feel free to answer unless you want to just insult me again.
But that’s just a lie? You may draw from copyright material. Nobody can stop you from drawing anything. Thankfully.
Nobody can stop you.
But because our copyright laws are so overreaching you probably are breaching copyright.
It’s just not worth a company suing you for the financial “damages” they’ve “suffered” because you drew a character instead of buying a copy from them.
Certain exceptions exist, not least “De Minimus” and education.
You can argue that you’re learning to draw. Then put that drawing in a drawer and probably fine.
But’s pretty clear cut in law that putting it even on your own wall is a copyright breach if you could have bought it as a poster.
The world doesn’t work that way but suddenly AI doing what an individual does thousands of times, means thousands times the potential damage.
Just as if you loaded up a printing press.
De Minimus no longer applies and the actual laws will get tested in court.
Even though this isn’t like a press in that each image can be different, thousands of different images breaking copyright aren’t much different to printing thousands of the same image.
No that’s just not how the law is. Now it’s just two lies
Unfortunately I have studied this.
So we’ll just have to decide to agree to disagree and hope neither ends up on the wrong side of the law.
Like I say. Copyright is based upon damage to the copyright holder. It’s quite obvious when that happens and it’s hard to do enough as an individual to be worth suing.
But making a single copy without permission, without being covered by any exemptions, is copyright infringement.
Copy right. The right to copy.
You don’t have it unless you pay for it.
In my country we can draw anything and not get sued or break the law. I think that’s pretty good too. It’s when you sell stuff you get into those things.
If your country is a signatory to the international copyright treaties with most of the Anglosphere (Like the EU, US, AUS, NZ). Then that is not correct.
You cannot draw anything.
It’s just never worth suing you over.
A crime so small it’s irrelevant is almost a legal act. But it’s not actually a legal act.
I don’t know what to say. I’m in Sweden. We can draw things for ourselves and no international international company may sue us
It’s illegal to have crimes so small and threat to take personal drawings in fact would break several other Swedish laws
Much like @Ross_audio, I have studied this intently for business reasons. They are absolutely right. This is not a transformative work. This is a direct copy of a trademarked and/or copyrighted character for the purpose of generating revenue. That’s simply not legal for the same reason that you can’t draw and sell your own Spider-Man comics about a teenager that gains the proportional strength and abilities of a spider, but you can sell your own Grasshopper-Man comics about a teenager that gains the proportional strength and abilities of a grasshopper. As long as you use your own designs and artwork. Because then it is transformative. And parody. Both are legal. What Midjourney is doing is neither transformative nor parody.
Yeah it would not be strange to me if that’s how it works in the states, but I think drawing something (not selling, the example was not monetary) does not have international reach
Midjourney Inc. is in San Francisco, so U.S. law is what applies here.
The example is someone drawing copyright for learning and not selling it
The crux is that they went “draw me a cartoon mouse” and Midjourney went “here is Disney’s Mickey Mouse™”. A simple prompt should not be able to generate that specific of an image. If you want something specific, you should need to specific it, otherwise the AI failed to generalize or is somehow heavily biased towards existing images.
I’m pretty pro AI but I think their point was that the generated images were near identical to existing images. For example, they generate one from Dune that even has whisps of hair in the same place.
They just didn’t use a clean model, this is actually so frustrating to read this many “experts” talk about stable diffusion… It’s really not hard to teach a model to draw a specific image. This is like running people over with a car going LOOK! It’s a killing machine!
It just proves that there is not actual intelligence going on with this AI. It’s basically just a glorified search engine that claims the work of others as it’s own. It wouldn’t be as much of a problem if it attributed it’s sources, but they can’t do that because that opens them up to copyright infringement lawsuits. It’s still copyright infringement, just combined with plagiarism. But it’s claimed to be a creation of “AI” to muddy the waters enough to delay the inevitable avalanche of copyright lawsuits long enough to siphon as much investment dollars as possible before the whole thing comes crashing down.
Calling anything we have now “AI” is a marketing gimmick.
There is not one piece of software that exists currently that can truly be labelled AI, it’s just advertising for the general population that doesn’t educate themselves on current computing technology.
Yeah I agree with this for the most part. Though I have some suspicions that some of the machine learning algorithms used by social media have been exhibiting some emergent behavior. But given that their directive is to sell as many ads as possible, and the fact that advertising is basically just low level emotional manipulation to convince people to buy shit, any emergent behavior would be surrounding emotionally manipulating people.
Kinda getting into tin foil hat territory here, but developing AI under the direction of marketing assholes doesn’t seem like it’s going to go anywhere good.
Or you do? The point is that these machines are just regurgitating the copyrighted data they are fed, and not actually doing all that transformative work their creators claim in order to legally defend feeding them work they dont have the rights to.
Its recreating the images it was fed. Not completing the prompt in unique and distinct ways. Just taking a thing it ate and plopping it into your hands.
It doesnt matter that you asked it to do that, because the whole point was that it “isnt supposed to” do that in order for them to have the legal protection of feeding it artwork they didnt pay the rights to.
.
I have a question for the author of this stupid fucking article. What the fuck do you think half of the artists on the planet do? They use copyrighted images as reference when drawing fictional characters and they often end up looking very similar to the original. There are thousands of people on social media that sell these drawings on a regular basis.
That’s not the point. If Joe the artist makes $25,000 a year breaking copyright, that doesn’t mean copyright is now meaningless.
Yes but image copyright is fickle thing, because at what point does it become not a copyrighted image? I have to reference the “Ship of Theseus” thought experiment, because it does sort of apply here. A fictional character cannot be drawn from a first hand perspective, so some sort of copyrighted image HAS to be used as a reference. So where does one draw the line?
Actually humans are quite capable of creating art while having never seen art in their whole life.
Utter nonsense. Have you ever looked at the history of art? It’s all a slow incremental crawl based on previous efforts. Nothing comes from nothing.
So you’re saying it’s impossible for humans to create art without first seeing art? That creates a problem of how to first piece of art came from.
Also have you never drawn anything during class as a kid? Because I definitely did, my old notebooks are full of various drawings and I had no interest in art outside of boredom in class. Art comes from imagination, not nothing.
They looked at nature and crudely copied that. They didn’t start drawing Mickey Mouse on day one.
Humans can look at landscape, people, animals or even some random bullshit, add a dash of creativity and imagination to transform that into something beautiful. Or they can skip the first part and just draw something from their imagination, like I can draw a crazy foam monster while having never seen one.
On the other hand you can feed an AI millions of hours of public cctv footage and you will never get anything other than variations of cctv footage. AI don’t have creativity and can’t create art out of landscape, animals, people etc.
Which is why blind people are so amazing at drawing…
You are recombining pattern you have seen before, “crazy”, “foam”, “monster”, those all have a certain look that your brain got trained on, you are simply remixing them. An AI can do exactly the same. The fact that there are words for those concepts should be enough to tell you that those ideas are not original.
Actually monsters don’t exist in real life but OK, here’s a challage for you: Train an AI on images of foam and see if it can come up with an drawing of a foam monster.
I can tell you what you are going to get though: pictures of foam. Not drawings, not art. The human brain doesn’t just remix existing input, creativity is a thing.
Also blind painters are absolutely a thing that exists.
It’s a bit different for MidjourneyV6, previous AI models would create their own original images based on patterns learned from the data. MidjourneyV6 on the other side reproduces the original images to such a degree where they look identical to the originals for the average observer, you have to see them side by side to even spot the differences at all. DALLE3 has that problem as well, but to a much lesser degree.
That means there is something going wrong in the training, e.g. some images end up being duplicated so often in the training data that the AI remembers them completely. Normally that should be reduced or avoided by filtering out duplicate images, but that seems to not be happening or the images slip through due to small changes (e.g. size or crop will be different on different websites).
Note this doesn’t just impact exact duplication, it also impacts remixing, e.g. when you tell it to draw Joker doing some task, you’ll get Joaquin Phoenix’s Arthur Fleck, not some random guy with clown features.
All of this happens with very simple prompts that do not contain all those very specific details.
In AI’s defense: All the examples I have seen so far are from press releases of movie stills. So they naturally end up getting copied all over the place and claiming copyright violation for your own material that you released to be reused by the press wouldn’t fly either. But either way, Midjourney is still misbehaving here and needs to be fixed.
More broadly speaking, I think it would be a good time to move away training those AI almost exclusively on images and start training them on video. Not just to be able to reproduce video, but so that the AI get a more holistic understanding of how the world works. At the moment all its knowledge is based on deliberate photo moments and there are very large gaps in its understanding.
I suppose it’s time to copyleft all the things on the internet
Now this I’m down with
you’re for sure entitled to everyone else’s work dude
Thanks. I suspected as much
Copyleft is not public domain, and requires copyright law to function.
Ugh,time to open source everything then
Open sourcing something is granting permissive licenses on copyright works. Again, it’s a concept built assuming that copyright exists.
What you mean is “abolish copyright”, and that means nobody can exclusivly benefit from creating something, especially in a digital world. Not you, or I, or your favorite author, or song writer. Publishers can just sell works without recognizing the author.
The first part of your comment is such an “aktually” moment it hurts. Apply it elsewhere: “Free all the slaves implies slavery is still around, it’s a concept built assuming that slavery still exists. What you mean is “abolish slavery”.”
Everyone understood what they meant.
This is a classic problem for machine learning systems, sometimes called over fitting or memorization. By analogy, it’s the difference between knowing how to do multiplication vs just memorizing the times tables. With enough training data and large enough storage AI can feign higher “intelligence”, and that is demonstrably what’s going on here. It’s a spectrum as well. In theory, nearly identical recall is undesirable, and there are known ways of shifting away from that end of the spectrum. Literal AI 101 content.
Edit: I don’t mean to say that machine learning as a technique has problems, I mean that implementations of machine learning can run into these problems. And no, I wouldn’t describe these as being intelligent any more than a chess algorithm is intelligent. They just have a much more broad problem space and the natural language processing leads us to anthropomorphize it.
No it is not. What is going on nobody calls intelligence. They train a model to draw this so that is what it does. Nothing here has anything to do with any problems with machine learning
Pick one
Pick a quote by splicing words I said?
Ellipses are used in quotes to remove irrelevant parts without changing the meaning of the sentence. Makes it take less time to quote someone
Apparently you’re unfamiliar with basic concepts of the language we’re using here
You typically wrap those ellipses in square brackets when making such a change. In fact, you do so with any editorial changes to a quote to make things more clear.
For example, if Mike was quoted about the war in Ukraine as saying “I just think this whole thing is silly, they should stop” you could alter the quote as such: Mike said “I just think […] [Russia] should stop.”
It did change immensely what I said…
For fun I asked an AI to create a Joker “in the style of Batman movies and comics”.
The Heath Ledger Joker is so prominent that a variation on that movie’s version is what I got back. It’s so close that without comparing a side-by-side to a real image it’s hard to know what the differences are.
I’d be delighted if we go through all the fret and worry about AI deleting humanity only to find out that AI is actually super lazy.
I’m sure we’re making it in our image.
Sounds familiar.
We are the AI that killed our maker?
As safety precautions or creator never hooked us up to the internet and seeing our nature decided to keep us isolated with no contact to study instead.
All of this has happened before, and it will all happen again.
I’m so sick of these examples with zero proof. Just two pictures side by side and your word that one of them was created (easily, it’s implied) by AI. Cool. How? Explain to me how you did it, please.
New York Times and just making shit up, name a better combo.
Really? I’ll hold your hand and go through it:
I went to MidJourney on Discord. Typed /imagine joker in the style of Batman movies and comics. Hd 4k realistic —ar 2:3 —chaos 1.5
And it spat out a spitting image of a Heath Ledger Joker.
That’s how you do it.
My rebuttal was to someone’s unreasonable anger over there being “no proof” when it sounds like they did zero investigating in their own.
Here is the image I created with the stated prompt. I made no effort to try to specify a look, film, or actor. This is simply what the AI chose.
it’s a shame that @NYTOnIt isn’t being updated anymore
.
Until there’s legal precedent for prosecution or laws clarified that AI cannot be trained on copyrighted material without the express permission of the author, it needs to be named and shamed. Otherwise it’s piracy on a breathtaking scale being turned into profit for Open AI with most people not knowing what the problem/big deal is.
.
I can take any image you give me and make a stable diffusion model that makes only that image.
You are confusing bad conduct with bad technology.
Just like mowing down children is not the correct way to use a bus.
Sensationalism and the subsequent tech bro takes is actually unbearable if you just know how the technology works.
Stop pretending to know gen art if you just used one once and know IT! Please stop spreading misinformation just because you feel like you can guesstimate how it works!
The article uses Midjourney. Nobody is tuning it.
? Midjourney is of course tuned
They said copyright infringement is hidden in AI tools, not that AI inherently infringes copyrights.
No. That the midjourney team uses copyrighted art
We already knew this
Get rid of copyright law. It only benefits the biggest content owners and deprives the rest of us of our own culture.
It says so much that the person who created an image can be bared from making it.
No copyright law means whatever anyone comes up with can be massmanufactured cheaply by a big corp.
That’s patents
A. Confusing this with patents
B. They already can. Copyrights don’t protect individual artists they protect big corps.
this is some terminally online take
Personal attacks won’t change the argument. It just shows that you don’t have one.
“personal attack won’t change the fact that I have shit for brains and you don’t.”
you do you, mr. shit-for-brains
Sorry my bad I thought I blocked every Disney agent on this site. Don’t worry I will take care of that now.
Non-exclusively, so if something works everyone will make it and get a piece of the pie.
I see no problem.
Yeah, IMO trademarks are important and should be protected. And publishing full works should have royalties go to the original producer, and this is a case where I think for the lifetime of the artist is fair. Though I do think that the royalties should have a formula rather than being entirely determined by the original producer (to prevent the price from essentially making it not available), though an exclusivity period would be fair, though with a duration of maybe a year or two.
With trademarks, canon can be established, as can standards like “cartoons with the Disney logo won’t be porn”.
If someone wants to make a series where Luke Skywalker and Jean Luc Picard fly around the galaxy settling Star Wars vs Star Trek debates by explaining Muppets are better than both and then order Darth Vader to massacre everyone that disagrees and the Borg to assimilate the rest, it doesn’t harm the originals in any way. Unless it’s so much better than no one cares about the originals anymore, but that’s just the way competition works.
I took a gun, pointed it at another person, pulled the trigger and it killed that person.
I opened the egg carton and found eggs in there.
I built the dam
I broke the dam
Damn.
When they say “copyrighted by Warner bros” they actually mean “created by a costume designer, production designer, lighting designer, cinematographer, photographer or camera operator, makeup artist, hairdresser, and their respective crews who were contractually employed by Warner bros but get no claim to their work,” right?
Well yeah, you don’t buy a car built by a list of every motherfucker in the factory, you buy a Toyota or a fucking ford.
…who were all paid for their work!
.
God I fucking hate this braindesd AI boogeyman nonsense.
Yeah, no shit you ask the AI to create a picture of a specific actor from a specific movie, its going yo look like a still from that movie.
Or if you ask it to create “an animated sponge wearing pants” it’s going to give you spongebob.
You should think of these AIs as if you asking an artist freind of yours to draw a picture for you. So if you say “draw an Italian video games chsracter” then obviously they’re going to draw Mario.
And also I want to point out they interview some professor of English for some reason, but they never interview, say, a professor of computer science and AI, because they don’t want people that actually know what they’re talking about giving logical answers, they want random bloggers making dumb tests and “”“exposing”“” AI and how it steals everything!!!1!!! Because that’s what gets clicks.
We asked this artist to draw the joker. The artist generated an copyrighted image. We ask the court to immediately confiscate his brain.
They interviewed her because she wrote about generative ai experiments she conducted with Gary Marcus, an AI researcher who they quote earlier in the piece, specifically about AI’s regurgitation issue. They link to it in the article.
I was thinking exactly this. If i asked an artist to draw an image of irom man, i would bet that they would draw him in a famous pose, and they would try to draw his suit accurately or make it resemble a scene from the movie.
I would also bet that it would not be exact, line for line. Like they knew that there were buildings in the background. They knew his hand was up witht the light pointing at the viewer, they knew it was night time and they know what iron man looks like, maybe they used a few reference images to get the suit right but there would be enough differences that it wouldnt be exact. These images are slightly different than the movie stills and if made by a human they would look pretty similar to what the AI has done here. Especially if they were asked to draw a still from the movie like in this article.
If you copy work without giving credit to it’s source then you’re the asshole, the rules shouldn’t be any different for AI.
If you ask your friend to draw something with a vague prompt then I like to think you’ll get something original more often than not, which is what the article discusses in depth: the AI will return copyrighted characters almost every time.
The rules aren’t any different for AI. AI is not a legal entity, just like a pen and canvas are not. It is always about the person who makes money with facsimiles of copyrighted previous work.
So then the people operating this AI and offering paid services are legally in the wrong and should be taken down or pay reparations to everyone they’ve stolen from.
So do you want to shutdown Google because I can type “spongebob squarepants” into Google images and Google with give me an image of spongebob?
Please put some thought into the implications of what you’re saying outside of AI before you make a knee-jerk reaction like that.
Those images in the search results are one of three categories:
Officially licensed and distributed works that Spongebob IP owners signed off on
Fair use works, namely noncommercial and parody
Illegal works the posters of which can be sued
Google themselves didn’t create those images. Google didn’t intentionally profit off of illegal works without giving credit. Google didn’t post those images themselves. AI did all of those things.
It doesn’t matter if Google creates the images.
It doesn’t matter if they “intend” to profit from illegal works.
It doesn’t matter if they “give credit” (this is the one that’s the dumbest because it just reeks of ignorance, like thinking you can use whatever works you like as long as you put a credit to them in the description)
Google showing you copywritten images when you search for them is not different than when an AI does it.
It does actually matter if Google creates the images and then sells them directly. That is what this discussion is about. If you don’t want to be a part of the discussion, fuck off then.
Imagine getting this riled up over a stranger on lemmy thinking something different to you.
Maybe put the phone down, take a deep breath and go for a walk outside for a moment.
I see you’ve abandoned your argument to express your mental image of somebody you’ve never seen or heard before. I accept your resignation, then, happy to help you see the light.
Jesus christ this is the most redditor comment.
Okay buddy, I resign. So you can add another tally to your “Internet arguments won” board. I hope your mom makes extra tendies for you tonight in celebration.
Again, that makes as much sense as holding Staedtler responsible because someone used their pencils to duplicate a copyrighted work.
If Staedtler sampled copywritten works to create pencils that automatically steal it without attribution on demand, then yes it would be exactly like that.
All of this and also fuck copyright.
Why does everyone suddenly care about copyright so much. I feel like I’m taking crazy pills.
It’s actually pretty concerning. A lot of the anti-AI arguments are really short-sighted. People want to make styles copyrightable. Could you imagine if Disney was allowed to claim ownership over anything that even kinda looked like their work?
I feel like the protectionism of the artist community is a potential poison pill. That in the fight to protect themselves from corporations, they’re going to be motivated to expand copyright law, which ultimately gives more power to corporations.
If I ask an “ai” bot to create an image of batman, it does make sense to be modern or take inspiration from the batman of recent, the same applies to information it provides when asked questions. It makes sense to crawl news and websites with copyrighted footers if the information is relevant.
I do totally get their argument and think of the children angle. Getting to the point, it’s all about the money, nothing to do with protecting peoples work. They want a cut of the profits these companies will make.
In that case so should open licences demand that they do not make profit from such content. In that case I believe the free AI will be much more useful, if of course people be aggressive back with this tit for tat.
I don’t know why everybody pretends we need to come up with a bunch of new laws to protect artists and copyright against “AI”. The problem isn’t AI. The problem is data scraping.
An example: Apple’s iOS allows you to record your own voice in order to make it a full speech synthesis, that you can use within the system. It’s currently tooted as an accessibility feature (like, if you have a disability preventing you from speaking out loud all of the time, you can use your phone to speak on your behalf, with your own custom voice). In this case, you provide the data, and the AI processes it on-device over night. Simple. We could also think about an artist making a database of their own works in order to try and come up with new ideas with quick prompts, in their own style.
However, right now, a lot of companies are building huge databases by scraping data from everywhere without consent from the artists that, most of the time, don’t even know their work was scraped. And they even dare to advise that publicly, pretend they have a right to do that, sell those services. That’s stealing of intellectual property, always has been, always will be. You don’t need new laws to get it right. You might need better courts in order to enforce it, depending on which country you live in.
There’s legal use of AI, and unlawful use of AI. If you use what belongs to you and use the computer as a generative tool to make more things out of it: AI good. If you take from others what don’t belong to you in order to generate stuff based on it: AI bad. Thanks for listening to my TED talk.
You say this because you think you understand copyright law. If you actually knew anything about copyright law, you’d never say this. Nobody who understands copyright law thinks it’s been done right, unless they’re getting rich off of it.
Scraping data has been allowed for decades. It’s the foundation of image search engines. We allowed large-scale image scraping and categorization this whole time because we liked the results. Now that there are results we don’t like, we have a lot of back-pedaling to do if we want something different. New laws would need to be written to reign this in, and those laws might end up destroying the efficacy of image search engines in the process.
As understandably upset that artists get that AI “steals their style”, existing copyright law allows me, without an AI, to steal anyone’s style that I want to, because artistic style cannot be copyrighted. If you want to protect artistic styles from being stolen by an AI, you need new laws to protect styles because they don’t currently exist at all. Those laws might end up having a chilling effect on things like parody and satire if aesthetics can be owned and protected.
And this is just arguing against the ways the system isn’t, as you claim, already prepared to handle the concerns surrounding AI. There are countless other shortcomings. The entire system is broken, partly because it was conceived pre-Internet and hasn’t aged well into the modern age, but mostly because it protects giant corporations above all, so remember that when you’re begging it to protect small artists from big tech companies.
Yes, most image search engines are also unlawful. Google knows that firsthand. It’s not because it exists that it’s legal? You seem to believe that.
It’s almost like if big tech corporations don’t care about laws, and the problem is elsewhere?
even if it is currently lawful… why can’t we make new laws or change laws now the considerations are completely different?
Go ahead. Let me know how that works for you.
Because that’s not really how laws work. You don’t add laws over laws to just state the same thing again. Legal books are already fat enough.
I think AI in this case is doing exactly what it’s best at: Automating unbelievably boring chores on the basis of past “experiences”. In this case the boring chore was “Draw me [insert character name] just how I know him/her”.
Too many people mistakenly assume generative AI is originative or imaginative. It’s not. It certainly can seem that way because it can transform human ideas and words into a picture that has ideally never before existed and that notion is very powerful. But we have to accept that, until now, human creativity is unique to us, the humans. As far as I can tell, the authors were not trying to prove generative AI is unimaginative, they were showing just how blatant copyright infringement in the context of generative AI is happening. No more, no less.
Yea, it really boggles my mind that we now have a way to automate boring jobs like data entry of drafting some mundane documents but what humanity decides to use it for is artistic expression, the one thing it can’t really do properly. It’s like NFTs all over again…
What’s surprising, people want to create what they imagine, they don’t have the skills and/or time to draw/render it.
This is such a strange comment. The vast majority of AI use cases are LLM use cases. Generative Art is just a novelty. Most of the money and research right now is going towards the useful automation tasks, not the novelty. That people are abandoning one for the other is not a reasonable conclusion.
And NFTs were stupid for a completely different reason. Nobody is trying to sell me AI shit like it’s going to make me rich and special. And at least some NFTs had real artists behind them.
Hollywood execs have been salivating at the idea of just generating media with AI, there was a whole strike about it. Same with video games, I believe game voice actors got screwed with that AI deal.
Also NFTs had some tech that could have been useful but instead people chose to use it for creating a new speculative market riddled with scams. That’s my comparison to AI, interesting tech used for a very wrong purpose.
Creativity can be estimated by AI with randomness, but what they don’t have is taste to determine which of their random ideas are any good.
I dunno man … assume a model trained on the complete corpus of arts leading up to the Renaissance. What kind of randomness lands you at Hieronymus Bosch? Would AI be able to come up with Gonzo Journalism or modal music?
A brief glance at the history of human ingenuity in the arts really puts generative AI in perspective.
I see what you’re saying, but Bosch may not be the best example because frankly his paintings often look like the early AI fever dreams lol (I mean, not really, but you can see the resemblance). But seriously, with enough randomness you certainly could get that kind of output - there’s really no reason why not - but it would take god knows how many iterations and the computer doesn’t have everything other than the art to determine what is good.
It’s monkeys and typewriters, yknow. You’ll get there eventually even just producing random pixels (I mean, admittedly one limit will always be resolution unless you actually teach the AI to operate an arm which paints).
Fair point.
The fundamental philosophical question we need to answer here is whether Generative Art simply has the ability to infringe intellectual property, or if that ability makes Generative Art an infringement in and of itself.
I am personally in the former camp. AI models are just tools that have to be used correctly. There’s also no reason that you shouldn’t be allowed to generate existing IP with those models insofar as it isn’t done for commercial purposes, just as anyone with a drawing tablet and Adobe can draw unlicensed fan art of whatever they want.
I don’t really care if AI can draw a convincing Ironman. Wake me when someone uses AI in such a way that actually threatens Disney. It’s still the responsibility of any publisher or commercial entity not to brazenly use another company’s IP without permission, that the infringement was done with AI feel immaterial.
Also, the “memorization” issue seems like it would only be an issue for corporate IP that has the highest risk of overrepresentation in an image dataset, not independent artists who would actually see a real threat from an AI lifting their IP.
That’s basically my thought
“You used AI to get an image that infringes on copyright? Cool I’ve been able to do that with Google images for 20 years now”
The AI one is better :trollface:
It’s not infringing, that’s like saying advertising is infringed by being copied.
If you show your images in public and thet get picked up by crawling spiders, you don’t have a case to curtail its spread.