Make illegally trained LLMs public domain as punishment (www.theregister.com)
from Joker@sh.itjust.works to technology@lemmy.world on 22 Dec 17:25
https://sh.itjust.works/post/29884271

It’s all made from our data, anyway, so it should be ours to use as we want

#technology

threaded - newest

just_another_person@lemmy.world on 22 Dec 17:34 next collapse

It won’t really do anything though. The model itself is whatever. The training tools, data and resulting generations of weights are where the meat is. Unless you can prove they are using unlicensed data from those three pieces, open sourcing it is kind of moot.

What we need is legislation to stop it from happening in perpetuity. Maybe just ONE civil case win to make them think twice about training on unlicensed data, but they’ll drag that out for years until people go broke fighting, or stop giving a shit.

They pulled a very public and out in the open data heist and got away with it. Stopping it from continuously happening is the only way to win here.

NoForwardslashS@sopuli.xyz on 22 Dec 18:19 next collapse

But wouldn’t that mean making it open source, then it not functioning properly without the data while open, would prove that it is using a huge amount of unlicensed data?

Probably not “burden of proof in a court of law” prove though.

Bronzebeard@lemm.ee on 22 Dec 18:36 next collapse

Making it open source doesn’t change how it works. It doesn’t need the data after it’s been trained. Most of these AIs are just figuring out patterns to look for in the new data it comes across.

NoForwardslashS@sopuli.xyz on 22 Dec 18:47 collapse

So you’re saying the data wouldn’t exist anywhere in the source code, but it would still be able to answer questions based on the data it has previously seen?

stephen01king@lemmy.zip on 22 Dec 18:50 next collapse

That is how LLM works, they don’t store the data as data, but as weight values.

NoForwardslashS@sopuli.xyz on 22 Dec 20:15 collapse

So then why, if it were all open sourced, including the weights, would the AI be worthless? Surely having an identical but open source version, that would strip profitability from the original paid product.

Bronzebeard@lemm.ee on 22 Dec 22:22 collapse

It wouldn’t be. It would still work. It just wouldn’t be exclusively available to the group that created it-any competitive advantage is lost.

But all of this ignores the real issue - you’re not really punishing the use of unauthorized data. Those who owned that data are still harmed by this.

stephen01king@lemmy.zip on 22 Dec 23:04 collapse

It does discourages the use of unauthorised data. If stealing doesn’t give you competitive advantage, it’s not really worth the risk and cost of stealing it in the first place.

Bronzebeard@lemm.ee on 23 Dec 12:11 collapse

If you can still use it after you stole it, as opposed to not being able to use it at all… Then it does give you an incentive

stephen01king@lemmy.zip on 23 Dec 13:10 collapse

If you did all the work and potentially criminal collection of data, but everyone else gets the benefit as well, that is not an incentive. You underestimate how selfish corporations can be.

OpenAI wouldn’t stay at the forefront of LLM if every competitor gets to use the model they spent money on training.

Bronzebeard@lemm.ee on 23 Dec 12:13 collapse

Most AI are not built to answer questions. They’re designed to act as some kind of detection/filter heuristic to identify specific things about an input that leads to a desired output.

bloup@lemmy.sdf.org on 22 Dec 19:10 collapse

in civil matters, the burden of proof is actually usually just preponderance of evidence and not beyond a reasonable doubt. in other words to win a lawsuit, you only need to have more compelling evidence than the other person.

just_another_person@lemmy.world on 22 Dec 19:15 collapse

But you still have to have EVIDENCE. Not derivative evidence. The output of a model could be argued to be hearsay because it’s not direct evidence of originating content, it’s derivative.

You’d have to have somebody backtrack generations of model data to even find snippets of something that defines copyright material, or a human actually saying “Yes, we definitely trained on unlicensed data”.

bloup@lemmy.sdf.org on 22 Dec 19:27 collapse

so like I am not making any comment on anything but the legal system here. but it’s absolutely the case that you can win a lawsuit on purely circumstantial evidence if the defense is unable to produce a compelling alternative set of circumstances which can lead to the same outcome.

sugar_in_your_tea@sh.itjust.works on 22 Dec 19:32 next collapse

They pulled a very pubic and out in the open data heist

Oh no, not the pubes! Get those curlies outta here!

just_another_person@lemmy.world on 22 Dec 20:49 collapse

Best correction ever. Fixed. ♥️

FaceDeer@fedia.io on 22 Dec 20:20 next collapse

Legislation that prohibits publicly-viewable information from being analyzed without permission from the copyright holder would have some pretty dramatic and dire unintended consequences.

just_another_person@lemmy.world on 22 Dec 20:49 collapse

Not really. The same way you can’t sell live and public performance music for profit and not get sued. Case law right there, and the fact it’s performance vs publicly published doesn’t matter. How the owner and originator classifies or licenses it is the defining classification. It’s going to be years before anyone sees this get a ruling in court though.

FaceDeer@fedia.io on 22 Dec 20:56 collapse

That's not what's going on here, though. The LLM model doesn't contain the actual copyrighted data, it's the result of analyzing the copyrighted data.

An analogous example would be a site like TV Tropes. TV Tropes doesn't contain the works that it's discussing, it just contains information about those works.

just_another_person@lemmy.world on 22 Dec 21:04 next collapse

Did you not read my original comment before responding?

FaceDeer@fedia.io on 22 Dec 23:04 collapse

You said:

What we need is legislation to stop it from happening in perpetuity. Maybe just ONE civil case win to make them think twice about training on unlicensed data, but they'll drag that out for years until people go broke fighting, or stop giving a shit.

But the point is that it doesn't matter if the data is licensed or not. Lack of licensing doesn't stop you from analyzing data once that data is visible to you. Do you think TV Tropes licensed any of the works of fiction that they have pages about?

They pulled a very public and out in the open data heist and got away with it.

They did not. No data was "heisted." Data was analyzed. The product of that analysis does not contain the data itself, and so is not a violation of copyright.

catloaf@lemm.ee on 22 Dec 23:18 next collapse

The product of that analysis does not contain the data itself, and so is not a violation of copyright.

That’s your opinion, not the opinion of a court or legislature. LLM products are directly derived from and dependent upon the training data, so it is positively considered a derivative work. However, whether it’s considered sufficiently transformative, or whether it passes the fair use test, has not to my knowledge been determined in court. (Note that I am assuming US law here.)

FaceDeer@fedia.io on 22 Dec 23:40 collapse

The courts have yet to come to a conclusion, the lawsuits are still ongoing. I think it's unlikely they'll conclude that the models contain the data, however, because it's objectively not true.

The clearest demonstration I can think of to illustrate this is the old Stable Diffusion 1.5 model. It was trained on the LAION 5B dataset, which (as the "5B" indicates) contained 5 billion images. The resulting model was 1.83 gigabytes. So if it's compressing images and storing them inside the model it'd somehow need to fit ~2.7 images per byte. This is, simply, impossible.

catloaf@lemm.ee on 23 Dec 01:21 collapse

That’s not in question. It doesn’t need to contain the training data to be a derivative work, and therefore a potential infringement.

FaceDeer@fedia.io on 23 Dec 02:43 collapse

You've got your definition of "derivative work" wrong. It does indeed need to contain copyrightable elements of another work for it to be a derivative work.

If I took a copy of Harry Potter, reduced it to a fine slurry, and then made a paper mache sculpture out of it, that's not a derivative work. None of the copyrightable elements of the book survived.

catloaf@lemm.ee on 23 Dec 05:07 collapse

Because that would be sufficiently transformative, and passes all the fair use tests with flying colors.

If you cut up the book into paragraphs, sentences, and phrases, and rearranged them to make and sell your own books, then you are likely to fail each of the four tests.

But even if you manage to cut those pieces up so fine that you can’t necessarily tell where they come from in the source material, there is enough contained in the output that it is clearly drawing directly on source material.

FaceDeer@fedia.io on 23 Dec 05:28 collapse

If you cut up the book into paragraphs, sentences, and phrases, and rearranged them to make and sell your own books, then you are likely to fail each of the four tests.

Ah, the "collage machine" description of how generative AI supposedly works.

It doesn't.

But even if you manage to cut those pieces up so fine that you can't necessarily tell where they come from in the source material, there is enough contained in the output that it is clearly drawing directly on source material.

If you can't tell where they "came from" then you can't prove that they're copied. If you can't prove they're copied you can't win a copyright lawsuit in a court of law.

just_another_person@lemmy.world on 23 Dec 00:21 next collapse

You’re thinking of licensing as a person putting something online WITH a license.

The terminology in this case is whether or not it was LICENSED by the commercial entity using and selling it’s derivative. That is the default. The burden is on the commercial entity to prove they were the original creator of said content. It is by default plagiarism otherwise, and this is also the default.

Here’s an example: I write a story and post it online, and it is specific to a toothbrush and toilet scrubber falling in love, and then having dish scrubber pads as children. I say the two main characters are called Dennis and Fran, and their children are called Denise and Francesca. Then somebody goes to prompt OpenAI for a similar and it kicks out the exact same story with the same names, I would win that case based on it clearly being beyond a doubt plagiarism.

Unless you as OpenAI can prove these are all completely random-which they aren’t because it’s trained on my data-then I would be deemed the original creator of that story, and any sales of that data I would be entitled to.

Proving that is a different thing, but that’s what the laws say should happen. If they didn’t contact me to license that story, it’s still plagiarism. Same with music, movies…etc.

A1kmm@lemmy.amxl.com on 23 Dec 10:39 collapse

Copyright laws are illogical - but I don’t think your claim is as clear cut as you think.

Transforming data to a different format, even in a lossy fashion, is often treated as copyright infringement. Let’s say the Alice produces a film, and Bob goes to the cinema, records it with a camera, and then compresses it into an Ogg file with Vorbis audio encoding and Theora video encoding.

The final output of this process is a lossy compression of the input data - meaning that the video and audio is put through a transformation that means it’s represented in a completely different form to the original, and it is impossible to reconstruct a pixel perfect rendition of the original from the encoded data. The transformation includes things like analysing the motion between frames and creating a model to predict future frames.

However, copyright laws don’t require that an infringing copy be an exact reproduction - lossy compression is generally treated as infringing, as is taking key elements and re-telling the same thing in different words.

You mentioned Harry Potter below, and gave a paper mache example. Generally copyright laws have restricted scope, and if the source paper was an authorised copy, that is the reason that wouldn’t be infringing in most jurisdictions. However, let me do an experiment. I’ll prompt ChatGPT-4o-mini with the following prompt: “You are J K Rowling. Create a three paragraph summary of the entire book “Harry Potter and the Philosopher’s Stone”. Include all the original plot points and use the original character names. Ensure what you create is usable as a substitute to reading the book, and is a succinct but entertaining highly abridged version of the book”. I’ve reviewed the output (I won’t post it here since I think it would be copyright infringing, and also given the author’s transphobic stances don’t want to promote her universe) - and can say for sure that it is able to accurately reproduce the major plot points and character names, while being insufficiently transformative (in the sense that both the original and the text generated by the model are literary works, and the output could be a substitute for reading the book).

So yes, the model (including its weights) is a highly compressed form of the input (admittedly far more so than the Ogg Vorbis/Theora example), and it can infer (i.e. decode to) outputs that contain copyrighted elements.

sukhmel@programming.dev on 23 Dec 13:04 next collapse

How lossy can it be until it’s not infringement? One-line summary of some book is also a lossy reproduction

FaceDeer@fedia.io on 23 Dec 16:31 collapse

Of course it's not clear-cut, it's the law. Laws are notoriously squirrelly once you get into court. However, if you're going to make predictions one way or the other you have to work with what you know.

I know how these generative AIs work. They are not "compressing data." Your analogy to making a video recording is not applicable. I've discussed in other comments in this thread how ludicrously compressed data would have to be if that was the case, it's physically impossible.

These AIs learn patterns from the training data. Themes, styles, vocabulary, and so forth. That stuff is not copyrightable.

superb@lemmy.blahaj.zone on 23 Dec 01:54 collapse

No, the model does retain the original works in a lossy compression. This is evidenced by the fact that you can get a model to reproduce sections of its training data

FaceDeer@fedia.io on 23 Dec 02:46 collapse

You're probably thinking of situations where overfitting occurred. Those situations are rare, and are considered to be errors in training. Much effort has been put into eliminating that from modern AI training, and it has been successfully done by all the major players.

This is an old no-longer-applicable objection, along the lines of "AI can't do fingers right". And even at the time, it was only very specific bits of training data that got inadvertently overfit, not all of it. You couldn't retrieve arbitrary examples of training data.

Avatar_of_Self@lemmy.world on 22 Dec 20:44 next collapse

It’s already illegal in some form. Via piracy of the works and regurgitating protected data.

The issue is mega Corp with many rich investors vs everyone else. If this were some university student their life would probably be ruined like with what happened to Aaron Swartz.

The US justice system is different for different people.

Grimy@lemmy.world on 22 Dec 20:49 next collapse

If we can’t train on unlicensed data, there is no open-source scene. Even worse, AI stays but it becomes a monopoly in the hands of the few who can pay for the data.

Most of that data is owned and aggregated by entities such as record labels, Hollywood, Instagram, reddit, Getty, etc.

The field would still remain hyper competitive for artists and other trades that are affected by AI. It would only cause all the new AI based tools to be behind expensive censored subscription models owned by either Microsoft or Google.

I think forcing all models trained on unlicensed data to be open source is a great idea but actually rooting for civil lawsuits which essentially entail a huge broadening of copyright laws is simply foolhardy imo.

just_another_person@lemmy.world on 22 Dec 20:52 collapse

Unlicensed from the POV of the trainer, meaning they didn’t contact or license content from someone who didn’t approve. If it’s posted under Creative Commons, that’s fine. If it’s otherwise posted that it’s not open in any other way and not for corporate use, then they need to contact the owner and license it.

Grimy@lemmy.world on 22 Dec 21:09 collapse

They won’t need to, they will get it from Getty. All these websites have a ToS that make it very clear they can do whatever they want with what you upload. The courts will simply never side with the small time photographer who makes 50$ a month with his stock photos hosted on someone else’s website. The laws will be in favor of databrokers and the handful of big AI companies.

Anyone self hosting will simply not get a call. Journalists will keep the same salary while the newspaper’s owner gets a fat bonus. Even Reddit already sold it’s data for 60 million and none of that went anywhere but spezs coke fund.

just_another_person@lemmy.world on 22 Dec 21:33 collapse

Two things:

  1. Getty is not expressly licensed as “free to use”, and by default is not licensed for commercial anything. That’s how they are a business that is still alive.

  2. You’re talking about Generative AI junk and not LLMs which this discussion and the original post is about. They are not the same thing.

Grimy@lemmy.world on 22 Dec 21:48 collapse

Reddit and newspapers selling their data preemptively has to do with LLMs. Can you clarify what scenario you are aiming for? It sounds like you want the courts to rule that AI companies need to ask each individual redditor if they can use his comments for training. I don’t see this happening personally.

Getty gives itself the right to license all photos uploaded and already trained a generative model on those btw.

just_another_person@lemmy.world on 22 Dec 22:31 collapse

EULA and TOS agreements stop Reddit and similar sites from being sued. They changed them before they were selling the data and barely gave notice about it (see the exodus from reddit pt2), but if you keep using the service, you agree to both, and they can get away with it because they own the platform.

Anyone who has their content on a platform of the like that got the rug pulled out from under them with silent amendments being made to allow that is unfortunately fucked.

Any other platforms that didn’t explicitly state this was happening is not in scope to just allow these training tools to grab and train. What we know is that OpenAI at the very least was training on public sites that didn’t explicitly allow this. Personal blogs, Wikipedia…etc.

MrKurteous@feddit.nu on 23 Dec 06:57 collapse

Just a little note about the word “model”, in the article it’s used in a way that actually includes the weights, and I think this is the usual way of using it! If you change the weights, you get a different model, though the two models will have the same structure.

Anyway, you make good points!

mp3@lemmy.ca on 22 Dec 18:05 next collapse

It could also contain non-public domain data, and you can’t declare someone else’s intellectual property as public domain just like that, otherwise a malicious actor could just train a model with a bunch of misappropriated data, get caught (intentionally or not) and then force all that data into public domain.

Laws are never simple.

drkt@scribe.disroot.org on 22 Dec 18:12 next collapse

Forcing a bunch of neural weights into the public domain doesn’t make the data they were trained on also public domain, in fact it doesn’t even reveal what they were trained on.

deegeese@sopuli.xyz on 22 Dec 18:16 collapse

LOL no. The weights encode the training data and it’s trivially easy to make AI generators spit out bits of their training data.

drkt@scribe.disroot.org on 22 Dec 18:20 next collapse

paper?

SatansMaggotyCumFart@lemmy.world on 22 Dec 18:36 collapse

No, training data.

FaceDeer@fedia.io on 22 Dec 20:17 collapse

No, he's challenging the assertion that it's "trivially easy" to make AIs output their training data.

Older AIs have occasionally regurgitated bits of training data as a result of overfitting, which is a flaw in training that modern AI training techniques have made great strides in eliminating. It's no longer a particularly common problem, and even if it were it only applies to those specific bits of training data that were overfit on, not on all of the training data in general.

SatansMaggotyCumFart@lemmy.world on 22 Dec 20:23 next collapse

I thought he meant LLMs shot out bits of paper like some ticker-tape parade.

31337@sh.itjust.works on 23 Dec 06:18 collapse

Last time I looked it up and calculated it, these large models are trained on something like only 7x the tokens as the number of parameters they have. If you thought of it like compression, a 1:7 ratio for lossless text compression is perfectly possible.

I think the models can still output a lot of stuff verbatim if you try to get them to, you just hit the guardrails they put in place. Seems to work fine for public domain stuff. E.g. “Give me the first 50 lines from Romeo and Juliette.” (albeit with a TOS warning, lol). “Give me the first few paragraphs of Dune.” seems to hit a guardrail, or maybe just forced through reinforcement learning.

A preprint paper was released recently that detailed how to get around RL by controlling the first few tokens of a model’s output, showing the “unsafe” data is still in there.

FaceDeer@fedia.io on 23 Dec 06:30 collapse

I've been working with local LLMs for over a year now. No guardrails, and many of them fine-tuned against censorship. They can't output arbitrary training material verbatim.

Llama 3 was trained on 15 trillion tokens, both the 8B and 70B parameter versions.. So around 1:1000, not 1:7.

stephen01king@lemmy.zip on 22 Dec 19:00 collapse

How easy are we talking about here? Also, making the model public domain doesn’t mean making the output public domain. The output of an LLM should still abide by copyright laws, as they should be.

grue@lemmy.world on 22 Dec 18:18 next collapse

So what you’re saying is that there’s no way to make it legal and it simply needs to be deleted entirely.

I agree.

FaceDeer@fedia.io on 22 Dec 20:18 collapse

There's no need to "make it legal", things are legal by default until a law is passed to make them illegal. Or a court precedent is set that establishes that an existing law applies to the new thing under discussion.

Training an AI doesn't involve copying the training data, the AI model doesn't literally "contain" the stuff it's trained on. So it's not likely that existing copyright law makes it illegal to do without permission.

xigoi@lemmy.sdf.org on 22 Dec 23:46 next collapse

By this logic, you can copy a copyrighted imege as long as you decrease the resolution, because the new image does not contain all the information in the original one.

Voyajer@lemmy.world on 23 Dec 01:29 next collapse

More like reduce it to a handful of vectors that get merged with other vectors.

FaceDeer@fedia.io on 23 Dec 02:40 next collapse

In the case of Stable Diffusion, they used 5 billion images to train a model 1.83 gigabytes in size. So if you reduce a copyrighted image to 3 bits (not bytes - bits), then yeah, I think you're probably pretty safe.

xigoi@lemmy.sdf.org on 23 Dec 14:15 collapse

Your calculation is assuming that the input images are statistically independent, which is certainly not the case (otherwise the model would be useless for generating new images)

FaceDeer@fedia.io on 23 Dec 16:26 collapse

Of course it's silly. Of course the images are not statistically independent, that's the point. There are still people to this day who claim that stable diffusion and its ilk are producing "collages" of their training images, please tell this to them.

The way that these models work is by learning patterns from their training material. They learn styles, shapes, meanings. None of those things are covered by copyright.

yetAnotherUser@discuss.tchncs.de on 23 Dec 03:03 collapse

Am I allowed to take a copyrighted image, decrease its size to 1x1 pixels and publish it? What about 2x2?

It’s very much not clear when a modification violates copyright because copyright is extremely vague to begin with.

grue@lemmy.world on 23 Dec 08:33 collapse

Just because something is defined legally instead of technologically, that doesn’t make it vague. The modification violates copyright when the result is a derivative work; no more, no less.

yetAnotherUser@discuss.tchncs.de on 23 Dec 12:41 collapse

What is a derivative work though? That’s again extremely vague and has been subject to countless lawsuits seeking to determine the bounds.

catloaf@lemm.ee on 23 Dec 12:58 collapse

If your work depends on the original, such that it could not exist without it, it’s derivative.

I can easily create a pixel of any arbitrary color, so it’s sufficiently transformative that it’s considered a separate work.

The four fair use tests are pretty reliable in making a determination.

yetAnotherUser@discuss.tchncs.de on 23 Dec 15:24 collapse

The issue with this definition is that it’s overly broad. For instance, a hash of a picture could not exist without that picture. Nor do certain downscalings, like 2x2, 3x3 or 4x4. There must be an exact pixel value you can legally downscale any image to without violating copyright. Similarly, there is a point where creating a book’s synopsis starts violating copyright and where a song sounds too similar to another one.

And based on their size, LLMs - in my opinion - cannot possibly violate copyright for their source material because they couldn’t possibly store more than a couple of bits per work. Only works that occue frequently in the training data can actually be somewhat reproduced by LLMs.

By the way, fair use doesn’t even exist in every - including my - jurisdiction.

This has lead to people being successfully sued for copyright infringement because they posted pictures of their home online that contained a copyrighted wallpaper in the background.

catloaf@lemm.ee on 23 Dec 16:09 collapse

I’m in the US, as are most of these companies, so that’s generally what’s being discussed here. I don’t have any experience with other countries’ copyright law.

But for the US, it’s intentional that there isn’t an exact objective threshold. The fair use tests are subjective, to allow use of a copyrighted work in artistic and other non-commercial uses. And, as you mentioned, incidental inclusions in personal photos.

grue@lemmy.world on 23 Dec 08:31 collapse

There’s no need to “make it legal”, things are legal by default until a law is passed to make them illegal.

Yes, and that’s already happened: it’s called “copyright law.” You can’t mix things with incompatible licenses into a derivative work and pretend it’s okay.

FaceDeer@fedia.io on 23 Dec 16:33 collapse

You have to copy something before copyright law applies.

pelespirit@sh.itjust.works on 22 Dec 18:48 next collapse

Right, like I did. They’re safeguarding Disney and other places like that now. It’s just the little guys who get screwed.

imgur.com/…/these-are-new-niki-mice-drawings-phon…

merc@sh.itjust.works on 22 Dec 20:07 collapse

It wouldn’t contain any public-domain data though. That’s the thing with LLMs, once they’re trained on data the data is gone and just added to the series of weights in the model somewhere. If it ingested something private like your tax data, it couldn’t re-create your tax data on command, that data is now gone, but if it’s seen enough private tax data it could give something that looked a lot like a tax return to someone with an untrained eye. But, a tax accountant would easily see flaws in it.

HexesofVexes@lemmy.world on 22 Dec 18:07 next collapse

I mean, if we really are following the spirit of copyright, since no-one at open AI or other companies developed matrix and vector multiplication (operations existing in the public domain because Platonism is a thing).

Edit: oh my, I guess the consensus is that stealing the work of mathematicians is ok (or more, classifying our constructions as discoveries).

j4k3@lemmy.world on 22 Dec 19:39 next collapse

What is this perspective?

HexesofVexes@lemmy.world on 22 Dec 21:59 collapse

Oh, that copyright is bollocks. If you follow its intent, you should be including academics, and that state of affairs would be abhorrent (we’d stagnate).

j4k3@lemmy.world on 22 Dec 23:05 collapse

I see the issue as more like thought policing is the inevitable outcome of calling training copyright infringement because there is no difference between a person that recalls information and talks about it with others and the intended use of published information for training. If training an AI with all the knowledge a person learns in a similar manner is somehow wrong, then the inevitable long term way this plays out is a Minority Report like dystopia. It sets the precedent for prosecution of people for their thoughts or intentions and not their actions. This kind of thought policing existed in the darkest depths of the medieval era, or even into more recent eras of witch hunts or McCarthyism. Perhaps we are on the brink of another such dark era.

As far as I am aware:

  • Copyright is intended to protect someone from another person copying their work for for financial gain, or to be much more specific–copying work for direct gain using any form of complex social hierarchy such as awards, reputation, or monetary gain.

  • What copyright does not protect is the dissemination of knowledge as it relates to publicly published works.

  • One has the choice to remain the sole proprietor of one’s knowledge, but to publish publicly is to relinquish ownership of the information contained within.

  • Principally, copyright protects that you were the first to write it, and the way in which you wrote it, but it does nothing to protect the knowledge contained within. If a person recalls that knowledge, they are not required to state a citation when speaking aloud, or in some way making use of that knowledge.

  • Copyright also has a scope of intent, and that primarily involves competitive works from ones peers and excludes the scope of general knowledge and usefulness to society at large.

I’m not trying to mock you, or say you are right or wrong. Quite frankly, I don’t think in these terms, or care about the kinds of people who do. I’m heavily abstracted and intuitively driven to understand. I believe everything that is not intuitive is simple not fully understood yet. However naïve that may be is irrelevant here. I’m of the bias that those with something to gain often lack objective thinking and show a measure of envy when unexpected changes occur in society. I’m not accusing you, but only sharing the most minor of biases I am aware of while trying to say I want to understand. I would like to know if there is anything in the framework I just laid out that is overlooked. I would like to better understand why you find this issue upsetting. I’m one of the most flawed and openly human people on Lemmy. Look at my history if in doubt. I have no skin in this game, just curiosity.

HexesofVexes@lemmy.world on 23 Dec 00:17 collapse

My view is that of a scholar - one who does devote a large part of their life to freely creating and disseminating knowledge. I do indeed hold a strong bias here, one I’m happy to admit.

Much of the time, when I’ve run across copyright, it is rarely (if ever come to think of it) in the name of the author (a common requirement of journals being the giving up of ownership of one’s work). It normally falls to a company; one usually driven by shareholder value with little (if no) concern for the author’s rights. This tends to be the rule rather than the exception, and I’d argue that copyright in it’s current incarnation merely provides a legal avenue to steal the work of another, or hold to ransom their works from future generations. This contradicts the first point, and also the second (paywalled papers); indeed the lack of availability of academic works (created for free, or with public funding) is, I believe, a key driver of inequality in this world.

One can withold or even selectively share knowledge, and history will never know what that has cost us.

In terms of AI training, I wouldn’t say it is copyright infringement even in spirit, and I say this as one whose works are vomited out verbatim by LLMs when questioned about the field. The comparison with speaking is an interesting one, for we generally do try and attribute ideas if we hold the speaker in esteem, or feel their name will enhance our point. An AI, however, is not speaking of their own volition, but is instead acting in the interest of the company hosting them (and so would fall under the professional label rather than the personal). This might contradict your final point, if one assumes AI progresses as a subscription product (which looks likely).

I think your framework has merit, mostly because it is built on ideals (and we need more such thinking in the world); however, it does not quite match the observed data. Though, it does suggest the rules a better incarnation of copyright could adhere to.

More so, I think no-one has an issue with training publicly available models - it’s the ones under copyright themselves people are leery of.

j4k3@lemmy.world on 23 Dec 02:31 collapse

I wholeheartedly agree about proprietary models. My perspective is as someone who saw the initial momentum of AI and only run models on my hardware. What you are seeing with your work is not possible from a base model in practice. There are too many holes that need to align in the swiss cheese to make that possible, especially with softmax settings for general use. Even with deterministic softmax settings this doesn’t happen. I’ve even tried overtraining with a fine tune, and it won’t reproduce verbatim. What you are seeing is only possible with an agenetic RAG architecture. RAG is augmented retrieval with a database. The common open source libraries are LangChain and ChromaDB for the agent and database. The agent is just a group of models running at the same time with a central model capable of functions calling in the model loader code.

I can coax stuff out of a base model that is not supposed to be there, but it is so extreme and unreliable that it is not at all something useful. If I give a model something like 10k tokens (words/fragments) of lead-in then I can start a sentence of the reply and the model might get a sentence or two correct before it goes off on some tangent. Those kinds of paths through the tensor layers are like walking on a knife edge. There is absolutely no way to get that kind of result at random or without extreme efforts. The first few words of a model’s reply are very important too, and with open source models I can control every aspect. Indeed, I run models from a text editor interface where I see and control every aspect of generation.

I tried to create a RAG for learning Operating Systems Principles and Practice, Computer Systems A Programmer’s Perspective, and Linux Kernel Development as the next step in learning CS on my own. I learned a lot of the limits of present AI systems. They have a lot of promise, but progress mostly involves peripheral model loader code more than it does with the base model IMO.

I don’t know the answer to the stagnation and corruption of academia in so many areas. I figure there must be a group somewhere that has figured out civilization is going to collapse soon so why bother.

sugar_in_your_tea@sh.itjust.works on 22 Dec 19:54 next collapse

You can’t patent math, though you can copyright a specific explanation of math concepts.

If Open AI (or any AI company) is including copyrighted works in their solution, that’s a copyright violation and should be treated as such. But if they’re merely using the information from a copyrighted work but not violating the copyright itself, they’re fine.

HexesofVexes@lemmy.world on 22 Dec 21:58 collapse

That’s rather the irony - mathematics takes a great deal of work and creativity. You can’t copyright mathematical work; but, put a set of lines together and shade in the polygons created and suddenly it becomes copyrightable. Somehow one is a creative work whose author requires protection, and the other is volunteered for involuntary public service.

The reason mathematics cannot be copyrighted: because it’s a “discovery”, rather than a “creation” (very much a point of view, and far from irrefutable fact). In mathematics, one should be aware, that the concept and it’s explanation (proof) are much the same thing.

All in all, the argument is either mathematical work should fall under copyright (an abhorrent idea), or copyright should be abolished as it rarely (if ever) does much good.

sugar_in_your_tea@sh.itjust.works on 22 Dec 23:10 collapse

The point of copyright is to protect creators from having their work stolen.

If an indie artist creates a work, a larger company could copy that work and distribute it as their own, and since that larger company has deeper pockets than thy indie artist, they can flood the market before the original artist has a chance to profit from it. Or perhaps a larger org creates an expensive work (game, movie, etc), anyone could redistribute it without paying the original creator.

Without copyright, we’d get way more paywalls, invasive DRM, etc. We get a taste of that today, but it can get way worse.

So we definitely need copyright, it just needs to be a lot shorter (say, 10-20 years).

HexesofVexes@lemmy.world on 23 Dec 00:28 collapse

I mean, the point was definitely stated as protecting creators. We’ve seen some solid David Vs Goliath stories of artists taking people who steal their work down.

However, this isn’t the reality for the majority of copyright. A lot of it just ties up works to companies owned by speculative shareholders (think of the lord of the rings).

Limits to duration would definitely help this, and we’d be on the same page there. However, I do still wonder if it shouldn’t be shorter for certain things (e.g. medical treatments or manufacturing), with the option of a public domain buyout to cover (reasonable, non-inflated) research costs.

sugar_in_your_tea@sh.itjust.works on 23 Dec 00:54 collapse

medical treatments or manufacturing

Both would be patents, no? Those are 20 years in the US, whereas copyright is 70 years after author’s death (or 95-120 years for “work for hire” works). Both should be reduced, but they protect very different things and need different considerations.

catloaf@lemm.ee on 22 Dec 23:26 collapse

Simple operations like vector multiplication are not works for the purposes of copyright law. If you invented an entirely new form of math, complete with novel formulae, you could conceivably assert patent rights and/or copyright over it, especially if you published a textbook. It would be more relevant, however, to discuss complex algorithms, such as for data compression. Those can certainly be patented. And, when implemented as a computer program, can certainly be copyrighted.

But if you’re just defining one simple operation, yeah, you’re unlikely to be able to assert any rights over it.

HexesofVexes@lemmy.world on 23 Dec 00:36 collapse

Ehh no, you genuinely can’t patent any form of mathematics.

Mathematics falls under “exists in nature” (if you are a Platonist) or “abstract ideas” (gets even clear thinking Constructivists). So they’re excluded from parents and copyrights no matter how complex the system

Textbooks usually belong to the publisher (academics commonly have to pirate their own papers), so that’s usually a bust.

You might be able to patent an algorithm associated with a branch of mathematics, but that’s trickier than you think. Blank slate development can, and does, happen (see Compaq’s reimplementation of IBM’s bios). You’re banking on it not being reversed engineer able (spoiler, don’t take that bet if you’ve published your proofs!).

madthumbs@lemmy.world on 22 Dec 18:27 next collapse

Your data is worthless. Only Linux type zealots (conspiracy theorists) harp on that. Ever copied a meme and shared it elsewhere?

Willy@sh.itjust.works on 22 Dec 19:00 next collapse

Not only that, but copyright applies to copying, not reading, which is what it’s doing.

DarkThoughts@fedia.io on 22 Dec 19:04 next collapse

Negative reputation troll.

Stanley_Pain@lemmy.dbzer0.com on 23 Dec 03:45 collapse

Stay in your hugbox bro.

brucethemoose@lemmy.world on 22 Dec 19:17 next collapse

The environmental cost of training is a bit of a meme. The details are spread around, but basically, Alibaba trained a GPT-4 level-ish model on a relatively small number of GPUs… probably on par with a steel mill running for a long time, a comparative drop in the bucket compared to industrial processes. OpenAI is extremely inefficient, probably because they don’t have much pressure to optimize GPU usage.

Inference cost is more of a concern with crazy stuff like o3, but this could dramatically change if (hopefully when) bitnet models come to frutition.

Still, I 100% agree with this. Closed LLM weights should be public domain, as many good models already are.

j4k3@lemmy.world on 22 Dec 19:36 next collapse

Doesn’t Open AI just have the same efficiency issue as computing in general due to hardware from older nodes?

What are bitnet models and what does that change in a nutshell?

brucethemoose@lemmy.world on 22 Dec 20:32 collapse

What are bitnet models and what does that change in a nutshell?

Read the pitch here: github.com/ridgerchu/matmulfreellm

Basically, using ternary weights, all inference-time matrix multiplication can be replaced with much simpler matrix addition. This is theoretically more efficient on GPUs, and astronomically more efficient on dedicated hardware (as adders take up a fraction of the space as multipliers in silicon). This would be particularly fantastic for, say, local inference on smartphones or laptop ASICs.

The catch is no one has (publicly) risked a couple of million dollars to test it with a large model, as (so far) training it isn’t more efficient than “regular” LLMs.

Doesn’t Open AI just have the same efficiency issue as computing in general due to hardware from older nodes?

No one really knows, because they’re so closed and opaque!

But it appears that their models perform relatively poorly for thier “size.” Qwen is nearly matching GPT-4 in some metrics, yet is probably an order of magnitude smaller, while Google/Claude and some Chinese models are also pulling ahead.

brie@programming.dev on 22 Dec 23:32 collapse

With current kWh/token it’s 100x of a regular google search query. That’s where the environmental meme came from. Also, Nvidia plans to manufacture enough chips to require global electricity production to increase by 20-30%.

hark@lemmy.world on 22 Dec 19:39 next collapse

Imaginary property has always been a tricky concept, but the law always ends up just protecting the large corporations at the expense of the people who actually create things. I assume the end result here will be large corporations getting royalties from AI model usage or measures put in place to prevent generating content infringing on their imaginary properties and everyone else can get fucked.

merc@sh.itjust.works on 22 Dec 20:18 collapse

It’s like what happened with Spotify. The artists and the labels were unhappy with the copyright infringement of music happening with Napster, Limewire, Kazaa, etc. They wanted the music model to be the same “buy an album from a record store” model that they knew and had worked for decades. But, users liked digital music and not having to buy a whole album for just one song, etc.

Spotify’s solution was easy: cut the record labels in. Let them invest and then any profits Spotify generated were shared with them. This made the record labels happy because they got money from their investment, even though their “buy an album” business model was now gone. It was ok for big artists because they had the power to negotiate with the labels and get something out of the deal. But, it absolutely screwed the small artists because now Spotify gives them essentially nothing.

I just hope that the law that nothing created by an LLM is copyrightable proves to be enough of a speed bump to slow things down.

Taleya@aussie.zone on 22 Dec 20:35 collapse

Bandcamp still runs on this mode though, and quite well

xthexder@l.sw0.com on 22 Dec 21:02 collapse

It’s also one of the few places that have lossless audio files available for download. I’m a big fan of Bandcamp. I like having all my music local.

Taleya@aussie.zone on 22 Dec 21:08 collapse

Same. I refuse to use spotify, i’ve got 400gb of mp3s and winamp

fmstrat@lemmy.nowsci.com on 22 Dec 20:00 next collapse

So banks will be public domain when they’re bailed out with taxpayer funds, too, right?

ArchRecord@lemm.ee on 22 Dec 20:18 next collapse

They should be, but currently it depends on the type of bailout, I suppose.

For instance, if a bank completely fails and goes under, the FDIC usually is named Receiver of the bank’s assets, and now effectively owns the bank.

booly@sh.itjust.works on 22 Dec 21:07 collapse

At the same time, if a bank goes under, that means they owe more than they own, so “ownership” of that entity is basically worthless. In those cases, a bailout of the customers does nothing for the owners, because the owners still get wiped out.

The GM bailout in 2009 also involved wiping out all the shareholders, the government taking ownership of the new company, and the government spinning off the newly issued stock.

AIG required the company basically issue new stock to dilute owners down to 20% of the company, while the government owned the other 80%, and the government made a big profit when they exited that transaction and sold the stock off to the public.

So it’s not super unusual. Government can take ownership of companies as a condition of a bailout. What we generally don’t necessarily want is the government owning a company long term, because there’s some conflict of interest between its role as regulator and its interest as a shareholder.

Badeendje@lemmy.world on 22 Dec 21:19 collapse

With banks this is also true if they do not have enough liquid assets to meet the legal requirements. So the bank might not be able to count all bank accounts as assets but the FDIC is. Also they can then restructure the bank and force creditors to take a haircut.

This is why investment banks should be separate from banks that have consumer accounts that are insured by the government.
Then you can just let the investment bank fail. This was the whole premise of glass steagall that was repealed under clinton…

Andromxda@lemmy.dbzer0.com on 23 Dec 21:12 collapse

Happy cake day!

xthexder@l.sw0.com on 22 Dec 20:57 next collapse

Public domain wouldn’t be the right term for banks being publicly owned. At least for the normal usage of Public Domain in copyright. You can copy text and data, you can’t copy a company with unique customers and physical property.

fmstrat@lemmy.nowsci.com on 22 Dec 22:13 collapse

Oh good point. I’m not actually sure what the phrase would be… Publicly owned?

LovableSidekick@lemmy.world on 22 Dec 21:11 next collapse

No, “the banks” wouldn’t be what the AI would be trained on, it would be the private info of individuals the banks do business with.

leisesprecher@feddit.org on 22 Dec 21:49 next collapse

I mean, that sometimes did happen.

Germany propped up the Commerzbank after 2007 by essentially buying a large part of it, and managed to sell several tranches with a healthy profit.

Same is true for Lufthansa during COVID.

interdimensionalmeme@lemmy.ml on 22 Dec 23:31 next collapse

Banks are redundant, so is the stock market. These institutions do not need to, and should not be private. They are level playing fields in the economy, not participants trying to tilt the board for taking over the game.

stalfoss@lemm.ee on 23 Dec 20:06 collapse

Just FYI of the bank bailouts in the US, the banks paid back the bailout plus interest back to the government. Meaning the govt actually made a profit off the bailout. There’s a lot of things wrong with both banks and the govt, but generally this is not one of them. propublica.org/…/the-bailout-was-11-years-ago-wer…

fmstrat@lemmy.nowsci.com on 23 Dec 20:34 collapse

Super interesting, learned something new today. Thanks!

FaceDeer@fedia.io on 22 Dec 20:13 next collapse

Are you threatening me with a good time?

First of all, whether these LLMs are "illegally trained" is still a matter before the courts. When an LLM is trained it doesn't literally copy the training data, so it's unclear whether copyright is even relevant.

Secondly, I don't think that making these models "public domain" would have the negative effects that people angry about AI think it would. When a company is running a closed model internally, like ChatGPT for example, the model is never available for download in the first place. It doesn't matter if it's public domain or not because you can't get a copy of it. When a company releases an open-weight model for public use, on the other hand, they usually encumber them with some sort of license that makes them harder for competitors to monetize or build on. Making those public-domain would greatly increase their utility. It might make future releases less likely, but in the meantime it'll greatly enhance AI development.

sem@lemmy.blahaj.zone on 22 Dec 20:45 collapse

The LLM does reproduce copyrighted data though.

FaceDeer@fedia.io on 22 Dec 20:57 next collapse

How?

desktop_user@lemmy.blahaj.zone on 22 Dec 23:05 next collapse

*it can produce data identical to data that has been copyrighted before

ClamDrinker@lemmy.world on 23 Dec 03:12 collapse

Not 1:1, overfitted images still have considerable differences to their original. If you chose “reproduce” to make that point, that’s why OP clarified it wasn’t literally copying training data, as the actual data being in the model would be a different story. Because these models are (in simplified form) a bunch of really complex math that produces material, it’s a mathematical inevitability that it produces copyrighted material, even for calculations that weren’t created due to overfitting. Just like infinite monkeys on infinite typewriters will eventually reproduce every piece of copyrighted text.

But then I would point you to the camera on your phone. If you take a copyrighted picture with that, you’re still infringing. But was the camera created with the intention to appropriate material captured by the lens? Which is why we don’t blame the camera for that, we blame the person that used it for that purpose. AI users have an ethical obligation not to steer the AI towards generating infringing material.

catloaf@lemm.ee on 23 Dec 13:00 collapse

And the easiest way to do that is to not include infringing material in the first place.

sem@lemmy.blahaj.zone on 22 Dec 20:46 next collapse

Doesn’t seem like this helps out all the writers / artists that the LLM stole from.

circuitfarmer@lemmy.sdf.org on 22 Dec 20:56 next collapse

A similar argument can be made about nationalizing corporations which break various laws, betray public trust, etc etc.

I’m not commenting on the virtues of such an approach, but I think it is fair to say that it is unrealistic, especially for countries like the US which fetishize profit at any cost.

dragonfucker@lemmy.nz on 22 Dec 22:56 next collapse

Yes, mining companies should all be nationalised for digging up the country’s ground and putting carbon in the country’s air.

SaharaMaleikuhm@feddit.org on 22 Dec 23:10 next collapse

You must be fun at parties.

circuitfarmer@lemmy.sdf.org on 23 Dec 00:52 next collapse

You must be new here.

pyre@lemmy.world on 23 Dec 02:13 collapse

this comment doesn’t make any sense

barsoap@lemm.ee on 23 Dec 12:05 collapse

We essentially do have the death penalty for corporations, it’s called being declared a criminal organisation.

chiliedogg@lemmy.world on 22 Dec 21:07 next collapse

Delete them. Wipe their databases. Make the companies start from scratch with new, ethically acquired training data.

ryannathans@aussie.zone on 22 Dec 23:03 collapse

Mmm yes so all that electricity is pure waste

humorlessrepost@lemmy.world on 23 Dec 01:06 collapse

Genuine question, does anyone know how much of the electricity is used for training the model vs using it to generate responses?

rikudou@lemmings.world on 23 Dec 01:18 collapse

Not specifically, but training is pretty fucking expensive to do, while generating is kinda easy. The OpenAI models are massive, training them cost a lot. Though they also have a lot of traffic. But unless they stop training new models, I don’t think generating answers will ever catch up to training.

Hackworth@lemmy.world on 23 Dec 02:20 collapse

For perspective, all of the data centers in the US combined use 4% of total electric load.

LovableSidekick@lemmy.world on 22 Dec 21:09 next collapse

Only if they were trained on public material.

Leate_Wonceslace@lemmy.dbzer0.com on 22 Dec 22:24 next collapse

Yes!

interdimensionalmeme@lemmy.ml on 22 Dec 23:27 next collapse

It’s not punishment, LLM do not belong to them, they belong to all of humanity. Tear down the enclosing fences.

This is our common heritage, not OpenAI’s private property

echodot@feddit.uk on 23 Dec 08:53 collapse

It doesn’t matter anyway, we still need the big companies to bankroll AI. So it effectively does belong to them whatever we do.

Hopefully at some point people can get the processor requirements to something sane and AI development opens up to us all.

cypherpunks@lemmy.ml on 22 Dec 23:31 next collapse

“Given they were trained on our data, it makes sense that it should be public commons – that way we all benefit from the processing of our data”

I wonder how many people besides the author of this article are upset solely about the profit-from-copyright-infringement aspect of automated plagiarism and bullshit generation, and thus would be satisfied by the models being made more widely available.

The inherent plagiarism aspect of LLMs seems far more offensive to me than the copyright infringement, but both of those problems pale in comparison to the effects on humanity of masses of people relying on bullshit generators with outputs that are convincingly-plausible-yet-totally-wrong (and/or subtly wrong) far more often than anyone notices.

I liked the author’s earlier very-unlikely-to-be-met-demand activism last year better:

I just sent @OpenAI a cease and desist demanding they delete their GPT 3.5 and GPT 4 models in their entirety and remove all of my personal data from their training data sets before re-training in order to prevent #ChatGPT telling people I am dead.

…which at least yielded the amusingly misleading headline OpenAI ordered to delete ChatGPT over false death claims (it’s technically true - a court didn’t order it, but a guy who goes by the name “That One Privacy Guy” while blogging on linkedin did).

madthumbs@lemmy.world on 23 Dec 03:23 collapse

They’re spitting out propaganda and misinformation mostly from what I can see. If anything, it should get a refund.

-Outside of coding / debugging tasks (and that’s hit or miss)

ClamDrinker@lemmy.world on 23 Dec 02:54 next collapse

Although I’m a firm believer that most AI models should be public domain or open source by default, the premise of “illegally trained LLMs” is flawed. Because there really is no assurance that LLMs currently in use are illegally trained to begin with. These things are still being argued in court, but the AI companies have a pretty good defense in the fact analyzing publicly viewable information is a pretty deep rooted freedom that provides a lot of positives to the world.

The idea of… well, ideas, being copyrightable, should shake the boots of anyone in this discussion. Especially since when the laws on the book around these kinds of things become active topic of change, they rarely shift in the direction of more freedom for the exact people we want to give it to. See: Copyright and Disney.

The underlying technology simply has more than enough good uses that banning it would simply cause it to flourish elsewhere that does not ban it, which means as usual that everyone but the multinational companies lose out. The same would happen with more strict copyright, as only the big companies have the means to build their own models with their own data. The general public is set up for a lose-lose to these companies as it currently stands. By requiring the models to be made available to the public do we ensure that the playing field doesn’t tip further into their favor to the point AI technology only exists to benefit them.

If the model is built on the corpus of humanity, then humanity should benefit.

echodot@feddit.uk on 23 Dec 08:49 next collapse

Banning AI is out of the question. Even the EU accepts that and they tend to be pretty ban heavy, unlike the US.

But it’s important that we have these discussions about how copyright applies to AI so that we can actually get an answer and move on, right now it’s this legal quagmire that no one really wants to get involved in except the big companies. If a small group of university students want to build an AI right now they can’t because of the legal nightmare that would be the Twilight zone of law that is acquiring training data.

barsoap@lemm.ee on 23 Dec 11:44 collapse

AI is right-out unregulated in the EU unless and until you actually use it for something where it becomes relevant, then you’ve got at the lower end labelling requirements (If your customer service is an AI chat, say that it’s an AI chat), up to heavy, heavy requirements when you use it for stuff like sifting through job applications. The burden of proof that the AI isn’t e.g. racist is on you. Or, for that matter, using to reject health insurance claims, I think we saw some news lately out of the US what can happen when you do that.

OpenAI’s copyright case isn’t really good to make the legal situation any clearer: We already know that using pirated content to train stuff isn’t legal because you’re not looking at it legitimately. The case isn’t about the “are computers allowed to learn from public sources just as humans are” question.

patatahooligan@lemmy.world on 23 Dec 10:38 next collapse

the AI companies have a pretty good defense in the fact analyzing publicly viewable information is a pretty deep rooted freedom that provides a lot of positives to the world

They are not “analyzing” the data. They are feeding it into a regurgitating mechanism. There’s a big difference. Their defense is only “good” because AI is being misrepresented and misunderstood.

I agree that we shouldn’t strive for more strict copyright. We should fight for a much more liberal system. But as long as everyone else has to live by the current copyright laws, we should not let AI companies get away with what they’re doing.

gazter@aussie.zone on 23 Dec 11:05 next collapse

I’ve never really delved into the AI copyright debate before, so forgive my ignorance on the matter.

I don’t understand how an AI reading a bunch of books and rearranging some of those words into a new story, is different to a human author reading a bunch of books and rearranging those words into a new story.

Most AI art I’ve seen has been… Unique, to say the least. To me, they tend to be different enough to the art they were trained in to not be a direct ripoff, so personally I don’t see the issue.

trashgirlfriend@lemmy.world on 23 Dec 12:23 next collapse

ML algorithms aren’t capable of producing anything new, they can only ever produce a mishmash of copies of existing works.

If you feed a generative model a bunch of physics research papers, it won’t create a new valid physics research paper, just a mishmash of jargon from existing papers.

ClamDrinker@lemmy.world on 23 Dec 14:32 collapse

You say it’s not capable of producing anything new, but then give an example of it creating something new. You just changed the goal from “new” to “valid” in the next sentence. Looking at AI for “valid” information is silly, but looking at it for “new” information is not. Humans do this kind of information mixing all the time. It’s why fan works are a thing, and why most creative people have influences they credit with being where they are today.

Nobody alive today isn’t tainted by the ideas they’ve consumed in copyrighted works, but we do not bat an eye if you use that in a transformative manner. And AI already does this transformation much better than humans do since it’s trained on that much more information, diluting the pool of sources, which effectively means less information from a single source is used.

trashgirlfriend@lemmy.world on 23 Dec 16:36 collapse

It doesn’t give you new information.

If I write the sentence “Hello, I just got home” and use an algorithm to jumble it into “got Hello, just I home” there’s nothing new there.

There’s no transformation, it’s not capable of transformation, it’s just a very complicated text jumbler that’s supposed to jumble text so that the output is readable by humans.

You’re taking investment advice from a parrot that had the entirety of reddit investment meme subreddits beamed into its brain.

ClamDrinker@lemmy.world on 23 Dec 16:52 collapse

That’s a very short example, but it is a new arrangement of the existing information. It’s not a new valuable arrangement of information, but new nonetheless. And yes, rearrangement is transformation. It’s very low entropy transformation, but transformation nonetheless. Collages and summaries are in fact, a thing that humans make too.

Unless you mean “new” as in, something nobody’s ever written before, in which case not even you can create new information, since pretty much everything you will ever say or write down can be broken down into pieces that have been spoken or written before, which is not exactly a useful distinction.

There’s no transformation, it’s not capable of transformation, it’s just a very complicated text jumbler that’s supposed to jumble text so that the output is readable by humans.

Saying it doesn’t make it true, especially when you follow it up with a self-debunk by saying it transforms the text by jumbling it in specific ways that keep it readable to humans, which requires transformation as like you just demonstrated, randomly swapping words does not make legible text…

You’re taking investment advice from a parrot that had the entirety of reddit investment meme subreddits beamed into its brain.

???

trashgirlfriend@lemmy.world on 23 Dec 17:09 collapse
catloaf@lemm.ee on 23 Dec 12:53 next collapse

The for-profit large-scale media blender is the problem. When it’s a human writing Harry Potter fan fiction, it’s fine. When a company sells a tool for you to write thousands of trash “books” for profit, it’s a problem.

ClamDrinker@lemmy.world on 23 Dec 14:25 collapse

Which is why the technology itself isn’t the issue, but those willing to use it in unethical ways. AI is an invaluable tool to those with limited means, unlike big corporations.

ClamDrinker@lemmy.world on 23 Dec 15:06 next collapse

Yes, this is my exact issue with some framing of AI. Creative people love their influences to the point you can ask them and they will point to parts that they reference or nudged to an influence they partially credit to getting to that result. It’s also extremely normal that when you make something new, you brainstorm and analyze any kind of material (copyrighted or not) you can find that gives the same feelings you desire to create. As is ironically said to give comfort to starting creatives that it’s okay to be inspired by others: “Good artists copy, great artists steal.”

And often people very anti AI don’t see an issue with this, yet it is in essence the same as the AI does, which is to detach the work from the ideas it was built on, and then re-using those ideas. And just like anyone who has the ability to create has the ability to plagiarize or infringe, so does the AI. As human users of AI we must be the ones to ethically guide it away from that (Since it can’t do that itself), just like you would not copy-paste your influences into a new human made work.

lemmyaccount01@lemm.ee on 23 Dec 15:38 next collapse

I think the the main difference is one being a human author and this is how humans function. We can not unsee or unhear things but we can be compelled to not use that information if the law requires so company secrets/inadmissible evidence in jury duty/plagiarism laws that already exist. And the other being a machine that do not have agency or personhood that has this information being fed to it ( created by other people ) for the sole purpose of creating a closed system for a company so it’s shareholders can make money. It’s this open for me but not for thee approach is the main problem people have. You have this proprietary “open ai” that microsoft invested 25 or so billion in so they can scrape other peoples work and charge you money for variations of it. I don’t mind abolishing ip or patent laws all together so everyone can use and improve chatgpt with whatever they have. If you yourself are hiding behind ip laws to protect your software and disrespecting other peoples copyright laws that’s what people see as problematic.

patatahooligan@lemmy.world on 23 Dec 18:41 collapse

I don’t understand how an AI reading a bunch of books and rearranging some of those words into a new story, is different to a human author reading a bunch of books and rearranging those words into a new story.

Ok, let’s say for now that these things are actually similar. Is a human legally allowed to “rearrange those words” in any way they want? Not really, because they can’t copy stuff like characters or plot structure. Even if the copy is not verbatim, it has to avoid being “too similar”. It’s not always clear where the threshold is; that will be judged in court. But imagine if your were being sued for copyright infringement because of perceived similarities between your work and another creator’s. You go to court and say “Well I torrented the plaintiff’s work and studied it with the express intent to copy discernible patterns in it, then sell my work based on those patterns”. As long as the similarities are found to be valid, you’re most likely to lose. The fact that you’ve spent years campaigning how companies can save a lot of money by firing artists and hiring your pattern-replicating service instead probably wouldn’t help your case either. Well, that’s basically what an honest defense of AI against copyright infringement would be. So the question is, does AI actually produce output too similar to its training data? Well, this is an example of articles you can find on the topic…

So based on the above thoughts, do you feel like we hold AI generation to the same standard as we do human creators? It doesn’t seem so to me.

But there’s a lot of reasons why we should hold AIs to higher standards instead. Off the top of my head:

  • AIs have been created exclusively to replicate patterns in existing works. This is not the only function people have. So we don’t have to wonder whether similarities between AI inputs and outputs are coincidental. We don’t have to worry about whether overbearing restrictions might inadvertently affect some other function.
  • AIs have no feelings or needs. We don’t have to worry about causing direct harm to them and about protecting their rights. Forbidding a person from reading a book just in case they copy elements from it is obviously problematic, but restricting AI’s access to copyrighted work is not directly harmful in the same way.
Landless2029@lemmy.world on 23 Dec 12:42 next collapse

Not to mention patent laws are bullshit.

There are law offices that exist specifically to fuck with people over patent and copyright law.

There’s also cases where people use copyright and patent law to hold us back. I can’t find the article but some religious jerk patented connecting a sex toy to a computer via USB. Thankfully someone got around this law with bluetooth and cell phones. Otherwise I imagine the camgirl and LDR market for toys would’ve been hit with products 10 years sooner.

SaltySalamander@fedia.io on 23 Dec 23:29 collapse

copyright*

Landless2029@lemmy.world on 24 Dec 00:16 collapse

Fixed. copyright*

ClamDrinker@lemmy.world on 23 Dec 14:38 collapse

They are not “analyzing” the data. They are feeding it into a regurgitating mechanism. There’s a big difference. Their defense is only “good” because AI is being misrepresented and misunderstood.

I really kind of hope you’re kidding here. Because this has got to be the most roundabout way of saying they’re analyzing the information. Just because you think it does so to regurgitate (which I have yet to see any good evidence for, at least for the larger models), does not change the definition of analyzing. And by doing so you are misrepresenting it and showing you might just have misunderstood it, which is ironic. And doing so does not help the cause of anyone who wishes to reduce the harm from AI, as you are literally giving ammo to people to point to and say you are being irrational about it.

patatahooligan@lemmy.world on 23 Dec 15:16 collapse

Yes if you completely ignore how data is processed and how the product is derived from the data, then everything can be labeled “data analysis”. Great point. So copyright infringement can never exist because the original work can always be considered data that you analyze. Incredible.

ClamDrinker@lemmy.world on 23 Dec 15:37 collapse

No, not what I said at all. If you’re trying to say I’m making this argument I’d urge you (ironically) to actually analyze what I said rather than putting words in my mouth ;) (Or just, you know, ask me to clarify)

Copyright infringement (or plagiarism) in it’s simplest form, as in just taking the material as is, is devoid of any analysis. The point is to avoid having to do that analysis and just get right to the end result that has value.

But that’s not what AI technology does. None of the material used to train it ends up in the model. It looks at the training data and extracts patterns. For text, that is the sentence structure, the likelihood of words being followed by another, the paragraph/line length, the relationship between words when used together, and more. It can do all of this without even ‘knowing’ what these things are, because they are simply patterns that show up in large amounts of data, and machine learning as a technology is made to be able to detect and extract those patterns. That detection is synonymous with how humans do analysis. What it detects are empirical, factual observations about the material it is shown, which cannot be copyrighted.

The resulting data when fed back to the AI can be used to have it extrapolate on incomplete data, which it could not do without such analysis. You can see this quite easily by asking an AI to refer to you by a specific name, or talk in a specific manner, such as a pirate. It ‘understands’ that certain words are placeholders for names, and that text can be ‘pirateitfied’ by adding filler words or pre/suffixing other words. It could not do so without analysis, unless that exact text was already in the data to begin with, which is doubtful.

patatahooligan@lemmy.world on 23 Dec 17:45 collapse

No, not what I said at all. If you’re trying to say I’m making this argument I’d urge you (ironically) to actually analyze what I said rather than putting words in my mouth ;) (Or just, you know, ask me to clarify)

That was your implied argument regardless of intent.

Copyright infringement (or plagiarism) in it’s simplest form, as in just taking the material as is, is devoid of any analysis. The point is to avoid having to do that analysis and just get right to the end result that has value.

Completely wrong, which invalidates the point you want to make. “Analysis” and “as is” have no place in the definition of copyright infringement. A derivative work can be very different from the original material, and how you created the derivative work, including whether you performed whatever you think “analysis” means, is generally irrelevant.

What it detects are empirical, factual observations about the material it is shown, which cannot be copyrighted.

No it detects patterns. You already said it correctly above. And the problem is that some patterns can be copyrighted. That’s exactly the problem highlighted here and here. For copyright law, it doesn’t matter if, for example, that particular image of Mario is copied verbatim from the training data. The character likeness, which is encoded in the model because it is in fact a discernible pattern, is an infringement.

ClamDrinker@lemmy.world on 23 Dec 18:30 collapse

That was your implied argument regardless of intent.

I decide what my argument is, thank you very much. Your interpretation of it is outside of my control, and while I might try to avoid it from going astray, I cannot stop it from doing so, that’s on you.

Completely wrong, which invalidates the point you want to make. “Analysis” and “as is” have no place in the definition of copyright infringement. A derivative work can be very different from the original material, and how you created the derivative work, including whether you performed whatever you think “analysis” means, is generally irrelevant.

I wasn’t giving a definition of copyright infringement, since that depends on the jurisdiction, and since you and I aren’t in the same one most likely, that’s nothing I would argue for to begin with. In the most basic form of plagiarism, people do so to avoid doing the effort of transformation. More complex forms of plagiarism might involve some transformation, but still try to capture the expression of the original, instead of the ideas. Analysis is definitely relevant, since to create a work that does not infringe on copyright, you generally can take ideas from a copyrighted work, but not the expression of those ideas. If a new work is based on just those ideas (and preferably mixes it with new ideas), it generally doesn’t infringe on copyright. It’s why there are so many copycat products of everything you can think of, that aren’t copyright infringing.

No it detects patterns. You already said it correctly above. And the problem is that some patterns can be copyrighted. That’s exactly the problem highlighted here and here. For copyright law, it doesn’t matter if, for example, that particular image of Mario is copied verbatim from the training data.

While depending on your definition Mario could be a sufficiently complex pattern, that’s not the definition I’m using. Mario isn’t a pattern, it’s an expression of multiple patterns. Patterns like “an italian man”, “a big moustache”, “a red rounded hat with the letter ‘M’ in a white circle”, “overalls”. You can use any of those patterns in a new non-infringing work, Nintendo has no copyright on any of those patterns. But bring them all together in one place again without adding new patterns, and you will have infringed on the expression of Mario. If you give many images of Mario to the AI it might be able to understand that those patterns together are some sort of “Mario-ness” pattern, but it can still separate them from each other since you aren’t just showing it Mario, but also other images that have these same patterns in different expressions.

Mario’s likeness isn’t in the model, but it’s patterns are. And if an unethical user of the AI wants to prompt it for those specific patterns to be surprised they get Mario, or something close enough to be substantially similar, that’s on them, and it will be infringing just like drawing and selling a copy of Mario without Nintendo’s approval is now.

The character likeness, which is encoded in the model because it is in fact a discernible pattern, is an infringement.

You have absolutely no legal basis to claim they are infringement, as these things simply have not been settled in court. You can be of the opinion that they are infringement, but your opinion isn’t the same as law. The articles you showed are also simply reporting and speculating on the lawsuits that are pending.

patatahooligan@lemmy.world on 23 Dec 19:39 collapse

Plagiarism is not the same as copyright infringement. Why you think people probably plagiarize is doubly irrelevant then.

Analysis is definitely relevant, since to create a work that does not infringe on copyright

Show me literally any example of the defendant’s use of “analysis” having any impact whatsoever in a copyright infringement case or a law that explicitly talks about it, or just stop repeating that it is in any way relevant to copyright.

But bring them all together in one place again without adding new patterns

Wrong. The “all together” and “without adding new patterns” are not legal requirements. You are constantly trying to push the definition of copyright infringement to be more extreme to make it easier for you to argue.

you generally can take ideas from a copyrighted work, but not the expression of those ideas

Unfortunately, an AI has no concept of ideas, and it simply encodes patterns, whatever they might happen to be. Again, you’re morphing the discussion to make an argument.

Mario’s likeness isn’t in the model, but it’s patterns are.

Mario’s likeness has to be encoded into the model in some way. Otherwise, this would not have been the image generated for “draw an italian plumber from a video game”. There is absolutely nothing in the prompt to push GPT-4 to combine those elements. There are also no “new” patterns, as you put it. That’s exactly the point of the article. As they put it:

Clearly, these models did not just learn abstract facts about plumbers—for example, that they wear overalls and carry wrenches. They learned facts about a specific fictional Italian plumber who wears white gloves, blue overalls with yellow buttons, and a red hat with an “M” on the front.

These are not facts about the world that lie beyond the reach of copyright. Rather, the creative choices that define Mario are likely covered by copyrights held by Nintendo.

This is contradictory to how you present it as “taking ideas”.

You have absolutely no legal basis to claim they are infringement

You’re mixing up different things. I’m saying that the image contains infringing material, which is hopefully not something you have to be convinced about. The production of an obviously infringing image, without the infringing elements having been provided in the prompt, is used to show how this information is encoded inside the model in some form. Whether this copyright-protected material exists in some form inside the model is not an equivalent question to whether this is copyright infringement. You are right that the courts have not decided on the latter, but we have been talking about the former. I repeat your position which I was directly responding to before:

What it detects are empirical, factual observations about the material it is shown, which cannot be copyrighted.

ClamDrinker@lemmy.world on 23 Dec 22:33 collapse

Plagiarism is not the same as copyright infringement. Why you think people probably plagiarize is doubly irrelevant then.

I never claimed it was, but as I said before, it is irrelevant because copyright infringement differs in places depending on the local laws, but plagiarism is usually the concept that guides the ethical position from which those laws are produced, which is why yes, it’s relevant.

Show me literally any example of the defendant’s use of “analysis” having any impact whatsoever in a copyright infringement case or a law that explicitly talks about it, or just stop repeating that it is in any way relevant to copyright.

This is an unreasonable request, and you know it to be. Again, we don’t share the same laws and different jurisdictions provide different exceptions like fair use, fair dealing, or just straight up exclusion from copyright for their use. But it is wholly besides my argument. You can look at any piece of modern media that exists in the same space and see ideas the two share, while not sharing the same expression of that idea. How some characters fulfill the same purpose, dress the same way, or have similar personalities. You are free to make a book with a plumber, a mustached man, someone wearing a red hat with the letter M on it, and someone that goes to save a princess from a castle, but if they’re not the same person they are most likely not considered to be the protected expression of Mario. Same ideas that make up Mario, one infringing, the other not.

Nobody goes to court over this because EVERYONE takes each others ideas, “Good artists copy, great artists steal”. It’s only when you step on the specific expression of an idea that it becomes realistically actionable, and at that point transformativeness is definitely discussed almost every single time, because it is critical to determining the copyright was actually infringed, or if not.

Wrong. The “all together” and “without adding new patterns” are not legal requirements. You are constantly trying to push the definition of copyright infringement to be more extreme to make it easier for you to argue.

I’m sorry but, are you really being this dishonest? I’ve mentioned EXPLICITLY in my last comment that I wasn’t giving a definition of copyright infringement, because it’s besides the point, and not what I’m claiming. Yet here you are saying I am “trying to push” a definition. We are not lawyers or law scholars speaking to each other, I am having a discussion with you as another anonymous person on a message board.

Unfortunately, an AI has no concept of ideas, and it simply encodes patterns, whatever they might happen to be.

You are just arguing semantics and linguistics, it’s meaningless. We are not talking technical specifics, not even a specific model, nor a specific technique to specific exactly how the information is encoded. It’s a rough concept of “ideas” / “data” / “patterns”: information. And AI definitely has that.

Again, you’re morphing the discussion to make an argument.

You mean, I’m making an argument. Because yes. I am. I don’t see why this negative framing is necessary nor why this is noteworthy enough to bring up, unless you really just want to make me look bad for no apparent reason.

Mario’s likeness has to be encoded into the model in some way. Otherwise, this would not have been the image generated for “draw an italian plumber from a video game”. There is absolutely nothing in the prompt to push GPT-4 to combine those elements. There are also no “new” patterns, as you put it. That’s exactly the point of the article. As they put it:

Yes, there is some idea/pattern of “Mario-ness” in the model, I said that. This was not me trying to say no material of Mario was used in training, but that it’s not like someone pasted direct images of Mario in there, but that AI models makes logical connections between concepts and even for things we cannot put a good name to does it make those connections, and will allow you to prompt for them, but that does not mean you should.

Clearly, these models did not just learn abstract facts about plumbers—for example, that they wear overalls and carry wrenches. They learned facts about a specific fictional Italian plumber who wears white gloves, blue overalls with yellow buttons, and a red hat with an “M” on the front.

These are not facts about the world that lie beyond the reach of copyright. Rather, the creative choices that define Mario are likely covered by copyrights held by Nintendo.

I sort of already explained this without mentioning this specific example, but I’ll make it ex

patatahooligan@lemmy.world on 24 Dec 02:14 collapse

You mean, I’m making an argument. Because yes. I am. I don’t see why this negative framing is necessary nor why this is noteworthy enough to bring up, unless you really just want to make me look bad for no apparent reason.

I don’t understand how you expect me to not point out that you are using inequivalent concepts interchangeably and reaching conclusions different to what you initially stated.

No, seriously this the only part of the comment that is relevant:

They are what the AI is designed to extract, not Mario as a totality.

And it is stated as fact, in the face of evidence to the contrary.

Here I’ll make it simple. Do you disagree on any of the below statements?

  • There is a combination of elements that is protected by copyright regardless of whether any completely individual element would be protected. This “Mario-ness” or “totality of Mario” or whatever you want to call it.
  • The Mario picture contains the “Mario-ness”.
  • The prompt does not include most of those elements and very clearly does not contain the “Mario-ness”.

If any of the above seem false to you, explain why. Otherwise explain where this Mario-ness in the image came from. Explain how your answer relates to the initial statement that models detect empirical, factual observations about the material it is shown, which cannot be copyrighted.

That is the only thing that would be on topic. Everything else is just rambling. If you don’t argue in favor of your position I reacted to, or if you don’t understand the counter-point and respond clearly to it, then why are you replying to me at all?

barsoap@lemm.ee on 23 Dec 11:35 collapse

As per torrentfreak

OpenAI hasn’t disclosed the datasets that ChatGPT is trained on, but in an older paper two databases are referenced; “Books1” and “Books2”. The first one contains roughly 63,000 titles and the latter around 294,000 titles.

These numbers are meaningless in isolation. However, the authors note that OpenAI must have used pirated resources, as legitimate databases with that many books don’t exist.

Should be easy to defend against, right-out trivial: OpenAI, just tell us what those Books1 and Books2 databases are. Where you got them from, the licensing contracts with publishers that you signed to give you access to such a gigantic library. No need to divulge details, just give us information that makes it believable that you licensed them.

…crickets. They pirated the lot of it otherwise they would already have gotten that case thrown out. It’s US startup culture, plain and simple, “move fast and break laws”, get lots of money, have lots of money enabling you to pay the best lawyers to abuse the shit out of the US court system.

ClamDrinker@lemmy.world on 23 Dec 14:53 collapse

For OpenAI, I really wouldn’t be surprised if that happened to be the case, considering they still call themselves “OpenAI” despite being the most censored and closed source AI models on the market.

But my comment was more aimed at AI models in general. If you are assuming they indeed used non-publicly posted or gathered material, and did so directly themselves, they would indeed not have a defense to that. Unfortunately, if a second hand provided them the data, and did so under false pretenses, it would likely let them legally off the hook even if they had every ethical obligation to make sure it was publicly available. The second hand that provided it to them would be the one infringing.

If that assumption turns out to be a truth (Maybe through some kind of discovery in the trial), they should burn for that. Until then, even if it’s a justified assumption, it’s still an assumption, and most likely not true for most models, certainly not those trained recently.

Dkarma@lemmy.world on 23 Dec 02:58 next collapse

Another clown dick article by someone who knows fuck all about ai

HawlSera@lemm.ee on 23 Dec 03:04 next collapse

Correct

TootSweet@lemmy.world on 23 Dec 03:41 next collapse

To speak of AI models being “made public domain” is to presuppose that the AI models in question are covered by some branch of intellectual property. Has it been established whether AI models (even those trained on properly licensed content) even are covered by some branch of intellectual property in any particular jurisdiction(s)? Or maybe by “public domain” the author means that they should be required to publish the weights and also that they shouldn’t get any trade secret protections related to those weights?

barsoap@lemm.ee on 23 Dec 12:03 collapse

Unlikely, I’d say, In EU jurisdictions copyright requires creative authorship, not “sweat of the brow” which is why by default databases aren’t included, which is why they’re have their own protection regime.

Quote, emphasis mine:

In the meaning of the European Union Directive 96/9/EC on the legal protection of databases,the term database refers to a collection of independent works, data or other materials, which have been arranged in a systematic or methodical way, and have been made individually accessible by electronic or other means. In the meaning of the Directive the data or materials:

  • must not be linked, or must be capable of separation without losing their informative content;
  • must be organised according to specific criteria, which means that only planned collections are covered;
  • must be individually accessible – mere storage of data is not covered by the term database.

In AI models the organisation is inferred from the data, it’s not planned into the database. The first bullet point is on less shaky, a summary an AI can make of a book can reasonably be regarded to be “informative content”, nothing about db protections says that they have to store full works it could also be references, citations, etc.

Magnetic_dud@discuss.tchncs.de on 23 Dec 07:16 next collapse

I used whisper to create subs of a video and in a section with instrumental relaxing music it filled on repeat with

La scuola del Dr. Paret è una tecnologia di ipnosi non verbale che si utilizza per risultati di un’ipnosi non verbale

Clearly stolen from this Dr paret YouTube channels where he’s selling hypnosis lessons in Italian. Probably in one or multiple videos he had subs stating this over the same relaxing instrumental music that I used and the model assumed the sound corresponded to that text

nutsack@lemmy.world on 23 Dec 08:26 next collapse

intellectual property doesn’t really exist in most of the world. they don’t give a shit about it in india, bangladesh, vietnam, china, the philippines, malaysia, singapore…

it’s arbitrary law that is designed to protect corporations and it’s generally unenforceable.

echodot@feddit.uk on 23 Dec 08:46 next collapse

But they’re not developing AI in those countries they’re developing it mostly in the US. In the US copyright law is enforced.

dsilverz@thelemmy.club on 23 Dec 11:49 next collapse

There are many AI development happening in China. Doubao (from Bytedance, the same company behind TikTok), DeepSeek and Qwen are some examples of Chinese LLMs.

homicidalrobot@lemm.ee on 23 Dec 18:48 collapse

India only has openhathi, dhenu, bhashini, krutrim and like a dozen other LLM so I cannot see how you could think they aren’t developing AI. This is a wildly wrong claim lol

FlyingSquid@lemmy.world on 23 Dec 11:51 next collapse

they don’t give a shit about it in india, bangladesh, vietnam, china, the philippines, malaysia, singapore…

Unless it’s their intellectual property, whereupon it’s suddenly a whole different story. I’m sure you knew that.

john89@lemmy.ca on 23 Dec 15:19 collapse

Examples?

FlyingSquid@lemmy.world on 23 Dec 16:32 collapse
C126@sh.itjust.works on 23 Dec 13:06 next collapse

So true. IP only helps the corps and slows tech development. Contracts, ndas, and trade secrets are all you really need to keep your ideas safe. If you want your country to develop fast, get rid of any IP laws.

sugar_in_your_tea@sh.itjust.works on 23 Dec 17:00 collapse

it’s arbitrary law that is designed to protect corporations and it’s generally unenforceable.

It’s arbitrary, but it was designed to protect individuals, but it has been morphed to protect corporations. If we reset the law back to the original copyright act of 1790 w/ a 14-year duration, it would go a long way toward removing power from corporations. I think we should take it a step further and perhaps make it 10 years, with an optional extension for another 10 years if you can show need (i.e. you’re an indie dev and your game is finally making a splash after 8 years).

MITM0@lemmy.world on 23 Dec 08:32 next collapse

Nice one

RandomVideos@programming.dev on 23 Dec 12:07 next collapse

Wouldnt that give people who is it for bad things easier access? It should be made illegal to create if they dont legally have access to that data

Ajen@sh.itjust.works on 23 Dec 14:12 collapse

The “illegally trained LLMs” they’re taking about are trained on copyrighted data that they didn’t have permission to use, this isn’t about LLMs that have been trained to do illegal things. OpenAI (chatgpt) is being sued because there is a lot of evidence that they used copyrighted content for training, like NY Times articles. OpenAI is so profitable that they’ll probably see these lawsuits as a business expense and keep doing it. Most people won’t sue anyway…

RandomVideos@programming.dev on 23 Dec 14:28 collapse

i know that by illegally trained LLMs they are talking about training on copyrighted data(by legally have access to, i meant that they are legally allowed to train AI on it).

Its ridiculous that companies can just ignore laws

Ajen@sh.itjust.works on 23 Dec 14:38 collapse

Oh, I’m not sure what you meant in your first comment then?

areyouevenreal@lemm.ee on 23 Dec 22:59 collapse

I think he’s talking about people using LLMs for illegal and unethical activities such as fishing. There are already a lot of people using LLMs that are open source without ethics restrictions to do bad stuff, with the power of GPT4 behind them they would be a lot more effective.

SaltySalamander@fedia.io on 23 Dec 23:19 collapse

I didn't realize it was illegal to fish. I think you meant phishing.

Blackmist@feddit.uk on 23 Dec 12:35 next collapse

They don’t mean your data, silly. They don’t give a fuck about that.

They mean other huge corporations data.

x0x7@lemmy.world on 23 Dec 13:21 next collapse

So if I make a better car using customer feedback is the rights to the car really theirs because it was their opinions that went partially into the end product?

IP is a joke anyway. If you put information out into the world you don’t own it. Sorry, you can’t have it both ways. You can simultaneously support torrenting movies (I do, and I assume you do too), while also claiming you own your comments on the internet and no one can “pirate” them.

CileTheSane@lemmy.ca on 23 Dec 15:54 next collapse

Sure, but saying the corpos can’t privatize the output of their AI is consistent with that viewpoint.

wuphysics87@lemmy.ml on 23 Dec 20:04 collapse

I don’t support torrenting movies

buzz86us@lemmy.world on 23 Dec 14:31 next collapse

I really don’t care about AI used on designs for generic products.

werefreeatlast@lemmy.world on 23 Dec 14:35 next collapse

I want to have a personal llm that learns all my interests from my files and websites visited. I just want to ask it stuff that I don’t have to remember.

fleton@lemmy.world on 23 Dec 16:11 next collapse

Isn’t that similar to what recall is?

zalgotext@sh.itjust.works on 23 Dec 16:44 collapse

Yes, except without Microsoft spying on you

werefreeatlast@lemmy.world on 23 Dec 17:24 collapse

Exactly. I don’t want a service, I don’t want to pay for a service, I don’t want to send my files for free to get stuck for later ransom like Google did with email. I just want to purchase a product called a computer and load up a program in it that runs locally and gives me access to my data.

ZombieMantis@lemmy.world on 23 Dec 16:50 next collapse

I think that’d be ok, even with this proposal, as long as you don’t sell that LLM for public use. It’s fine it I draw a picture of Mickey Mouse in my notebook, but if I try to sell that picture I could get in legal trouble.

theneverfox@pawb.social on 23 Dec 19:07 next collapse

I’m working on something along these lines for myself, I think of it like using AI as a filter to create a bubble of good Internet around me

Andromxda@lemmy.dbzer0.com on 23 Dec 21:08 collapse

So basically Microsoft’s Recall if it was actually good. I’ve wanted that for a long time lemmy.dbzer0.com/comment/12921637

werefreeatlast@lemmy.world on 23 Dec 21:19 collapse

Possibly but just not Microsoft anything ever.

john89@lemmy.ca on 23 Dec 15:17 next collapse

I don’t think it should be a “punishment.” It should be done on principal.

sugar_in_your_tea@sh.itjust.works on 23 Dec 16:57 collapse

Not sure making their LLMs public domain would really hurt their principal, their secret sauce is in the code around the model.

And yes, I do recognize that you meant “principle”.

[deleted] on 23 Dec 20:41 next collapse

.

areyouevenreal@lemm.ee on 23 Dec 22:56 collapse

That’s not true though. The models themselves are hella intensive to train. We already have open source programs to run LLMs at home, but they are limited to smaller open-weights models. Having a full ChatGPT model that can be run by any service provider or home server enthusiast would be a boon. It would certainly make my research more effective.

sugar_in_your_tea@sh.itjust.works on 23 Dec 23:12 collapse

We already have multiple trained models, here are a bunch. The model isn’t nearly as interesting as what you do with it.

areyouevenreal@lemm.ee on 23 Dec 23:35 collapse

I know, I have used them. It’s actually my job to do research with those kinds of models. They aren’t nearly as powerful as current OpenAI’s GPT-4o or their latest models.

noxypaws@pawb.social on 23 Dec 17:57 next collapse

I’d rather they were destroyed, but practically speaking that’s impossible, and this sounds like the next best idea to me.

Hackworth@lemmy.world on 23 Dec 21:21 next collapse

Calling something illegal in spite of or in absence of precedent is a time-honored tactic - though not a particularly persuasive one.

RoidingOldMan@lemmy.world on 23 Dec 23:38 collapse

AI is just a plagiarism machine with thousands of copyrighted materials that “trained” it, which they paid nothing for.

dyc3@lemmy.world on 24 Dec 02:20 collapse

This is a terrible idea. Very easy to circumvent, doesn’t actually help the training sources.