445

Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

top 50 comments
sorted by: hot top controversial new old
[-] Gibdos@feddit.de 94 points 11 months ago

I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.

[-] adriaan@sh.itjust.works 58 points 11 months ago

That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.

[-] Shurimal@kbin.social 35 points 11 months ago

just reinforcement learning models

...like the naturally occuring neural networks are.

[-] Khalic@kbin.social 45 points 11 months ago

The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.

[-] canihasaccount@lemmy.world 14 points 11 months ago

Yeah, accurately simulating a single pyramidal neuron requires an eight-layer deep neural network:

https://www.cell.com/neuron/pdf/S0896-6273(21)00501-8.pdf

load more comments (1 replies)
[-] FaceDeer@kbin.social 8 points 11 months ago

It’s a LOT more complicated and frankly, barely understood.

Yet you confidently state that the brain doesn't work the way LLMs do?

Obviously it doesn't work exactly the same way that LLMs do, if only because of the completely different substrates. But when you get to more nebulous concepts like "creativity" and "inspiration" it's not so clear.

load more comments (2 replies)
load more comments (1 replies)
[-] lemmyvore@feddit.nl 26 points 11 months ago

Tell you what, you get a landmark legal decision classifying LLM as people and then we'll talk.

Until then it's software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.

[-] newthrowaway20@lemmy.world 42 points 11 months ago* (last edited 11 months ago)

That's an interesting take, I didn't know software could be inspired by other people's works. And here I thought software just did exactly as it's instructed to do. These are language models. They were given data to train those models. Did they pay for the data that they used to train for it, or did they scrub the internet and steal all these books along with everything everyone else has said?

load more comments (14 replies)
[-] Wander@kbin.social 31 points 11 months ago

Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?

[-] Grimy@lemmy.world 19 points 11 months ago

Clearly not. He's saying that other authors have done the same as the software does. The software creators implemented the same principle into their llm. You are being daft on purpose.

[-] newthrowaway20@lemmy.world 26 points 11 months ago* (last edited 11 months ago)

It's not the same principle. Large language models aren't 'inspired' to write new works. Software can't be inspired. It follows instructions. Even though large language models might feel like somebody is talking back to you and giving you new information, it's just code following instructions designed to predict output based on the input provided and the data supplied. There's no inspiration to be had, and to attribute inspiration to language models is a huge mischaracterization of what's happening under the hood. Can a language model, without being told what to do, actually use any of the data it was fed to create something? No. Every single large language model requires some sort of input from a user to act as a seed before any sort of response can begin.

This is why it's so stupid to call this shit AI, because people start thinking it's actual intelligence. Really, It's just a fancy illusion.

load more comments (1 replies)
[-] mojo@lemm.ee 12 points 11 months ago

They purchased their books to get inspiration from, the original author gets paid, and the author consented to selling it. That's the difference.

Also the LLM can post entire snippets or chapters of books, which of course you'll take at face value even if it hallucinates and makes the author look like a worse author then they are.

[-] elbarto777@lemmy.world 28 points 11 months ago

These are machines, though, not human beings.

I guess I'd have to be an author to find out how I'd feel about it, to be fair.

[-] Shurimal@kbin.social 14 points 11 months ago

These are machines, though, not human beings.

What's the difference? On the most fundamental level it's all the same.

[-] AnonStoleMyPants@sopuli.xyz 16 points 11 months ago

The same thing as with tooooooons of things: scale.

Nobody cares if one dude steals office supplies at work. Now, if everyone stats doing it, or if the single guy steals everything, then action is taken.

Nobody cares if a random person draws in the same style and with same characters as you, but if they start to sell them, or god forbid, out-sell you, then there is a problem.

Nobody cares (except police I guess) if a random driver drives double the speed limit and annoys people living next to the road on the weekends, but when tons of people do it, you get speed bumps.

Nobody cares if few people pirate movies, but when it gets to mainstream and companies notice that there might be money being lost. Then you get whatever we have now.

Nobody cares if the mudhill behind your house erodes a bit and you get mud on your shoes. Have a bunch of that erode and you realise the danger...

You have been fine-tuning your own writing style for a decade and random schmuck starts to write similarly, you probably don't care. No harm done. Now, get an AI to write 10 000 books in a weekend and someone starts to sell them... well now you have a completely different problem.

On a fundamental level the exact same thing is happening, yet action is only taken after a certain threshold is step over.

load more comments (1 replies)
[-] Wander@kbin.social 14 points 11 months ago

Unless you think theres no difference between killing a person and closing a program, I think we can agree they should be treated differently in the eyes of the law.

And so theres a difference between a person reading a book and being inspired by it, and someone writing a program that automatically transforms the book in data that can create new books.

[-] brygphilomena@lemmy.world 13 points 11 months ago

A human, regardless of how many books they read, will have personal experiences that are undeniably unique to themselves. They will interpret the works they read differently from each other based on their worldly experiences. Their writing, no matter how many books they read and get inspired on, will always be influenced by their own personal lives. They can experience love, hate, heartbreak, empathy, sadness, and happiness.

This is something a LLM does not have, and in my opinion, is a massive distinguishing factor. So on a "fundamental" level, it is not the same. It is no where near the same.

load more comments (2 replies)
[-] elbarto777@lemmy.world 9 points 11 months ago

Wait. Are human beings machines?

load more comments (2 replies)
load more comments (7 replies)
load more comments (4 replies)
[-] mojo@lemm.ee 58 points 11 months ago* (last edited 11 months ago)

Here's an idea, legally force companies like OpenAI to rely on opt-in data, rather then build their entire company on stealing massive amounts of data. That includes requiring to retrain from scratch. Sam Altman was crying for regulations for scary AI, right?

[-] FaceDeer@kbin.social 24 points 11 months ago

Would search engines only be allowed to show search results for sources that had opted in? They "train" their search engine on public data too, after all.

[-] mojo@lemm.ee 21 points 11 months ago

They aren't reselling their information, they're linking you to the source which then the website decides what to do with your traffic. Which they usually want your traffic, that's the point of a public site.

That's like trying to say it's bad to point to where a book store is so someone can buy from it. Whereas the LLM is stealing from that bookstore and selling it to you in a back alley.

[-] PsychedSy@sh.itjust.works 14 points 11 months ago

AI isn't either. It's selling statistical data about the books.

[-] mojo@lemm.ee 22 points 11 months ago

It literally shares passages verbatim

[-] BetaDoggo_@lemmy.world 8 points 11 months ago

So does any site that quotes the book. Just being trained on a work doesn't give the model the ability to cite it word for word. For most of the books in this set you wouldn't even be able to get a single accurate quote out of most models. The models gain the ability to cite passages from training on other sources citing these same passages.

load more comments (5 replies)
load more comments (1 replies)
[-] 0ddysseus@lemmy.world 55 points 11 months ago

This is no different than every other capitalist enterprise. The whole system works on taking a public resource, claiming private ownership of it, and then selling it back to the public for profit.

First it was farmland, then coal and minerals, oil, seafood, and now ideas. Its how the system works and is the whole reason people have been trying to stop it for the past 150 years.

The people making the laws are there because they and/or their parents and/or grandparents did the exact same thing. As despicable and corrupt as it is you won't change it by complaining and no-one is going to make a law to stop it.

[-] Franzia 14 points 11 months ago

God damned right. Every "new" thing tends to be stolen. In more event history, its stolen from other capital, or from innovation with a free license, rather than artwork. Publishers might actually be able to make a problem out of this.

[-] pavnilschanda@lemmy.world 33 points 11 months ago

I hope they can at least get compensated.

load more comments (6 replies)
[-] JackGreenEarth@lemm.ee 18 points 11 months ago

I would be proud, but you do you.

load more comments (7 replies)
[-] Smoogs@lemmy.world 13 points 11 months ago

Ok so it’s been stealing art now it’s coming for authors. At what point do we hold the coalition who started this shit culpable for numerous accounts of plagiarism?

load more comments (1 replies)
[-] originalucifer@moist.catsweat.com 13 points 11 months ago

do they also complain when their books are used to train wet networks in public schools? those networks are also later exploited by corporations who dont give back the writers. hmmmmmmm

[-] cynar@lemmy.world 18 points 11 months ago

They do get paid for that, however. They get a share of the value of each book sold. Those schools are paying for the books.

There is also the catch that those wet networks are of finite lifespan and are output throttled. This limits the losses caused. A lot of authors also consider improving those networks a big part of why they write.

It's the difference between someone hand drawing a Micky mouse birthday card for their sibling, and hallmark mass producing them for sale. The former is considered acceptable, the latter is grounds for a law suit.

[-] Gutless2615@ttrpg.network 13 points 11 months ago

Everyone’s a fan of fair use until it’s their work that is transformed.

load more comments (2 replies)
[-] leaky_shower_thought@feddit.nl 8 points 11 months ago

There's an idea by Barath Raghavan about an AI dividend that companies pay each netizen a share for the data they use to train these models.

I am into this idea if companies can't even do a simple opt-in mechanism.

[-] RalphWolf@lemmy.ca 8 points 11 months ago

Does this fall under fair-use part of copyright?

[-] FaceDeer@kbin.social 10 points 11 months ago

It hasn't been tested in court yet but I don't see why it shouldn't.

load more comments (6 replies)
load more comments (2 replies)
[-] Pyr_Pressure@lemmy.ca 8 points 11 months ago

Curious if the AI company actually bought those books or if they just came across them by pirating.

load more comments (1 replies)
load more comments
view more: next ›
this post was submitted on 29 Sep 2023
445 points (100.0% liked)

Technology

58125 readers
3937 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS