Arthur Besse@lemmy.ml to Technology@beehaw.orgEnglish · edit-22 年前Sarah Silverman and other authors are suing OpenAI and Meta for copyright infringement, alleging that they're training their LLMs on books via Library Genesis and Z-Librarywww.thedailybeast.comexternal-linkmessage-square129fedilinkarrow-up1219arrow-down10
arrow-up1219arrow-down1external-linkSarah Silverman and other authors are suing OpenAI and Meta for copyright infringement, alleging that they're training their LLMs on books via Library Genesis and Z-Librarywww.thedailybeast.comArthur Besse@lemmy.ml to Technology@beehaw.orgEnglish · edit-22 年前message-square129fedilink
minus-squareag_roberston_author@beehaw.orglinkfedilinkEnglisharrow-up62·2 年前This is a strawman. You cannot act as though feeding LLMs data is remotely comparable to reading.
minus-squareag_roberston_author@beehaw.orglinkfedilinkEnglisharrow-up18·2 年前Because reading is an inherently human activity. An LLM consuming data from a training model is not.
minus-squareTheBurlapBandit@beehaw.orglinkfedilinkEnglisharrow-up7·2 年前LLMs forcing us to take a look at ourselves and see if we’re really that special. I don’t think we are.
minus-squareDominic@beehaw.orglinkfedilinkEnglisharrow-up6·2 年前For now, we’re special. LLMs are far more training data-intensive, hardware-intensive, and energy-intensive than a human brain. They’re still very much a brute-force method of getting computers to work with language.
minus-squarepips@lemmy.filmlinkfedilinkEnglisharrow-up5·2 年前Because the LLM is also outputting the copyrighted material.
minus-squareDouble_A@discuss.tchncs.delinkfedilinkEnglisharrow-up3·2 年前So could any human that got inspired by something…
This is a strawman.
You cannot act as though feeding LLMs data is remotely comparable to reading.
Why not?
Because reading is an inherently human activity.
An LLM consuming data from a training model is not.
LLMs forcing us to take a look at ourselves and see if we’re really that special.
I don’t think we are.
For now, we’re special.
LLMs are far more training data-intensive, hardware-intensive, and energy-intensive than a human brain. They’re still very much a brute-force method of getting computers to work with language.
Because the LLM is also outputting the copyrighted material.
So could any human that got inspired by something…