Arthur Besse to [email protected]English • edit-22 years agoSarah Silverman and other authors are suing OpenAI and Meta for copyright infringement, alleging that they're training their LLMs on books via Library Genesis and Z-Librarywww.thedailybeast.comexternal-linkmessage-square129fedilinkarrow-up1217cross-posted to: [email protected]
arrow-up1217external-linkSarah Silverman and other authors are suing OpenAI and Meta for copyright infringement, alleging that they're training their LLMs on books via Library Genesis and Z-Librarywww.thedailybeast.comArthur Besse to [email protected]English • edit-22 years agomessage-square129fedilinkcross-posted to: [email protected]
minus-square@[email protected]linkfedilinkEnglish62•2 years agoThis is a strawman. You cannot act as though feeding LLMs data is remotely comparable to reading.
minus-squarepipslinkfedilinkEnglish5•2 years agoBecause the LLM is also outputting the copyrighted material.
minus-square@[email protected]linkfedilinkEnglish3•2 years agoSo could any human that got inspired by something…
minus-square@[email protected]linkfedilinkEnglish18•2 years agoBecause reading is an inherently human activity. An LLM consuming data from a training model is not.
minus-square@[email protected]linkfedilinkEnglish7•2 years agoLLMs forcing us to take a look at ourselves and see if we’re really that special. I don’t think we are.
minus-square@[email protected]linkfedilinkEnglish6•2 years agoFor now, we’re special. LLMs are far more training data-intensive, hardware-intensive, and energy-intensive than a human brain. They’re still very much a brute-force method of getting computers to work with language.
This is a strawman.
You cannot act as though feeding LLMs data is remotely comparable to reading.
Why not?
Because the LLM is also outputting the copyrighted material.
So could any human that got inspired by something…
Because reading is an inherently human activity.
An LLM consuming data from a training model is not.
LLMs forcing us to take a look at ourselves and see if we’re really that special.
I don’t think we are.
For now, we’re special.
LLMs are far more training data-intensive, hardware-intensive, and energy-intensive than a human brain. They’re still very much a brute-force method of getting computers to work with language.