this post was submitted on 24 Oct 2024
133 points (94.0% liked)

Technology

59422 readers
2819 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Grimy@lemmy.world 22 points 3 weeks ago (3 children)

Using public facing data to build machine learning model is not against copyright laws. There is a transformative clause for a reason.

Strengthening copyright laws will only hurt the open source scene and give companies like openai and google a soft monopoly.

Not only that but the money is going to go to data brokers and platforms like reddit and getty. Individuals aren't getting a dime.

[–] pennomi@lemmy.world 10 points 3 weeks ago

I agree with you in principle. Legally, the courts are still deciding.

[–] unrelatedkeg@lemmy.sdf.org 2 points 3 weeks ago

Yes. Just as entering through an unlocked door isn't trespassing.

Most of the sources also have copyright notices the model gobbles up, effectively making it more like trespassing and taking the no trespassing sign home as a souvenir.

[–] kibiz0r@midwest.social -1 points 3 weeks ago

This is where we need something other than copyright law. The problem with generative AI companies isn't that somebody looked at something without permission, or remixed some bits and bytes.

It's that their products are potentially incredibly harmful to society. They would be harmful even if they worked perfectly. But as they stand, they're a wide-open spigot of nonsense, spewing viscous sludge into every single channel of human communication.

I think we can bring out antitrust law against them, and labor unions are also a great tool. Privacy, and a right to your own identity factor in, too. But I think we're also going to need to develop some equivalent of ecological protections when it comes to information.

For a long time, our capacity to dump toxic waste into the environment was minuscule compared to the scale of the natural world. But as we automated more and more, it became clear that the natural world has limits. I think we're headed towards discovering the same thing for the world of information.