this post was submitted on 29 Sep 2023
398 points (93.8% liked)

Technology

58157 readers
3886 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

you are viewing a single comment's thread
view the rest of the comments
[–] Gibdos@feddit.de 24 points 11 months ago (4 children)

I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.

[–] adriaan@sh.itjust.works 33 points 11 months ago (2 children)

That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.

[–] Hackerman_uwu@lemmy.world 9 points 11 months ago (1 children)

More to the point: they replicate patterns of words.

[–] lloram239@feddit.de 7 points 11 months ago (1 children)
[–] FontMasterFlex@lemmy.world 1 points 11 months ago

That's a Bingo!

[–] Shurimal@kbin.social 8 points 11 months ago (2 children)

just reinforcement learning models

...like the naturally occuring neural networks are.

[–] Khalic@kbin.social 26 points 11 months ago (2 children)

The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.

[–] canihasaccount@lemmy.world 11 points 11 months ago (1 children)

Yeah, accurately simulating a single pyramidal neuron requires an eight-layer deep neural network:

https://www.cell.com/neuron/pdf/S0896-6273(21)00501-8.pdf

[–] demonsword@lemmy.world 3 points 11 months ago

that was an interesting read, thank you

[–] originalucifer@moist.catsweat.com -2 points 11 months ago

so its barely understood, but this definitely is not it. got it.

[–] lemmyvore@feddit.nl 10 points 11 months ago (1 children)

Tell you what, you get a landmark legal decision classifying LLM as people and then we'll talk.

Until then it's software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.

[–] Touching_Grass@lemmy.world 6 points 11 months ago* (last edited 11 months ago) (1 children)

What exactly was not permitted by the license? Reading?

[–] sab@lemmy.world 14 points 11 months ago (3 children)

Using it to (create a tool to) create derivatives of the work on a massive scale.

[–] SirGolan@lemmy.sdf.org 8 points 11 months ago

Wikipedia: In copyright law, a derivative work is an expressive creation that includes major copyrightable elements of a first, previously created original work.

I think you may be off a bit on what a derivative work is. I don't see LLMs spouting out major copyrightable elements of books. They can give a summary sure, but Cliff Notes would like to have a word if you think that's copyright infringement.

[–] lloram239@feddit.de 5 points 11 months ago (1 children)

Better tell that Google and their search index, book scanning project and knowledge graph.

[–] sab@lemmy.world -1 points 11 months ago* (last edited 11 months ago)

I didn't know those were LLMs, TIL.

[–] Touching_Grass@lemmy.world -1 points 11 months ago (1 children)

Well when that happens we have laws. So no problems

[–] sab@lemmy.world 2 points 11 months ago (1 children)

Would you be okay with applying that argument for any crime?

[–] Touching_Grass@lemmy.world -1 points 11 months ago (1 children)

Ever heard of the early 21st century classic Minority Report

[–] sab@lemmy.world 3 points 11 months ago

You're missing the point. I'll make your example more specific.

Well when fraud/rape/murder happens we have laws. So no problems.

Those things happen. Creating a LLM based on copyrighted material without permission happens - it's not a hypothetical. But even then, giving a punishment after the fact does not make the initial crime "no problem", as you put it.

[–] newthrowaway20@lemmy.world 29 points 11 months ago* (last edited 11 months ago) (1 children)

That's an interesting take, I didn't know software could be inspired by other people's works. And here I thought software just did exactly as it's instructed to do. These are language models. They were given data to train those models. Did they pay for the data that they used to train for it, or did they scrub the internet and steal all these books along with everything everyone else has said?

[–] lloram239@feddit.de -4 points 11 months ago (1 children)

And here I thought software just did exactly as it’s instructed to do.

AI isn't software. Everything the AI knows is from the books. There is no human instructing the AI what to do. All the human does is build the scaffolding to let the AI learn, everything else is in the data.

[–] elbarto777@lemmy.world 18 points 11 months ago (4 children)

These are machines, though, not human beings.

I guess I'd have to be an author to find out how I'd feel about it, to be fair.

[–] Touching_Grass@lemmy.world 8 points 11 months ago

Machines that aren't reproducing or distributing works

[–] sab@lemmy.world 3 points 11 months ago

I don't think anyone is faulting the machines for this, just the people who instruct the machines to do it.

[–] dutchkimble@lemy.lol 2 points 11 months ago (1 children)

But terminator said neural networks

[–] Shurimal@kbin.social -3 points 11 months ago (4 children)

These are machines, though, not human beings.

What's the difference? On the most fundamental level it's all the same.

[–] brygphilomena@lemmy.world 12 points 11 months ago (2 children)

A human, regardless of how many books they read, will have personal experiences that are undeniably unique to themselves. They will interpret the works they read differently from each other based on their worldly experiences. Their writing, no matter how many books they read and get inspired on, will always be influenced by their own personal lives. They can experience love, hate, heartbreak, empathy, sadness, and happiness.

This is something a LLM does not have, and in my opinion, is a massive distinguishing factor. So on a "fundamental" level, it is not the same. It is no where near the same.

[–] lloram239@feddit.de 0 points 11 months ago

A human, regardless of how many books they read, will have personal experiences that are undeniably unique to themselves.

So will every AI. ChatGPT will give you different answers than Bard or WizardLM, since they are all trained on different books. And every StableDiffusion model creates different images, different styles, different topics, etc. It's all in the data they "experienced".

[–] originalucifer@moist.catsweat.com -1 points 11 months ago

do you really think we are that far off... from giving a foundational memory and motivation layers to these LLMs, that could mimic.. or even.. generate the generic thoughts youre indicating?

i dont think so. you seem to imply its impossibility, i expect its inevitability. the human brain will not be a black box forever... it still exists in a world of physics we can emulate, even if rudimentary.

[–] Wander@kbin.social 10 points 11 months ago

Unless you think theres no difference between killing a person and closing a program, I think we can agree they should be treated differently in the eyes of the law.

And so theres a difference between a person reading a book and being inspired by it, and someone writing a program that automatically transforms the book in data that can create new books.

[–] AnonStoleMyPants@sopuli.xyz 10 points 11 months ago (1 children)

The same thing as with tooooooons of things: scale.

Nobody cares if one dude steals office supplies at work. Now, if everyone stats doing it, or if the single guy steals everything, then action is taken.

Nobody cares if a random person draws in the same style and with same characters as you, but if they start to sell them, or god forbid, out-sell you, then there is a problem.

Nobody cares (except police I guess) if a random driver drives double the speed limit and annoys people living next to the road on the weekends, but when tons of people do it, you get speed bumps.

Nobody cares if few people pirate movies, but when it gets to mainstream and companies notice that there might be money being lost. Then you get whatever we have now.

Nobody cares if the mudhill behind your house erodes a bit and you get mud on your shoes. Have a bunch of that erode and you realise the danger...

You have been fine-tuning your own writing style for a decade and random schmuck starts to write similarly, you probably don't care. No harm done. Now, get an AI to write 10 000 books in a weekend and someone starts to sell them... well now you have a completely different problem.

On a fundamental level the exact same thing is happening, yet action is only taken after a certain threshold is step over.

[–] sab@lemmy.world 1 points 11 months ago
[–] elbarto777@lemmy.world 3 points 11 months ago (2 children)

Wait. Are human beings machines?

[–] jennraeross@lemmy.world 4 points 11 months ago

Please do not take this as support of ai use of copyrighted works (I don’t), but as far as I can tell, yes we are machines. This rant is just me being aspie atm, so feel free to ignore it.

We are thinking machines programmed by our genetics, predispositions, experiences, and circumstances. A 2 part explanation of how humans are merely products of their circumstances was once put forward to me. The first part is that humans can do anything, but only the thing we want to do most.

For instance, a common rebuttal is that people can choose go to the gym even when they find the experience of exercise undesirable. However, when that happens, it’s merely a case of other wants out balancing the want to not go to the gym, typically they want to be fit.

We want to not spend money, but we want to not rush going to jail for stealing more, usually. We want to not work overtime, but sometimes we want the extra cash more than that.

The second part of the argument is that we can’t choose what we want. When someone talks themselves out of the slice of cheesecake, they aren’t changing what they want, they’re resolving said want against the larger want they have to lose weight.

And if we make decisions by our wants, while said wants are not decided by us, then despite appearances we are little more than complex automata.

[–] lloram239@feddit.de 3 points 11 months ago

Biological machines, yes.

[–] Wander@kbin.social 15 points 11 months ago (1 children)

Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?

[–] Grimy@lemmy.world -2 points 11 months ago (2 children)

Clearly not. He's saying that other authors have done the same as the software does. The software creators implemented the same principle into their llm. You are being daft on purpose.

[–] newthrowaway20@lemmy.world 16 points 11 months ago* (last edited 11 months ago) (1 children)

It's not the same principle. Large language models aren't 'inspired' to write new works. Software can't be inspired. It follows instructions. Even though large language models might feel like somebody is talking back to you and giving you new information, it's just code following instructions designed to predict output based on the input provided and the data supplied. There's no inspiration to be had, and to attribute inspiration to language models is a huge mischaracterization of what's happening under the hood. Can a language model, without being told what to do, actually use any of the data it was fed to create something? No. Every single large language model requires some sort of input from a user to act as a seed before any sort of response can begin.

This is why it's so stupid to call this shit AI, because people start thinking it's actual intelligence. Really, It's just a fancy illusion.

[–] lloram239@feddit.de -2 points 11 months ago

This is why it’s so stupid to call this shit AI

It is using the term as defined. Maybe stop being a stupid parrot just repeating crap you heard else where and use your brain for a moment. I am losing hope that humans are capable of thought reading all this junk.

[–] mojo@lemm.ee 8 points 11 months ago

They purchased their books to get inspiration from, the original author gets paid, and the author consented to selling it. That's the difference.

Also the LLM can post entire snippets or chapters of books, which of course you'll take at face value even if it hallucinates and makes the author look like a worse author then they are.