this post was submitted on 30 Sep 2023
555 points (97.8% liked)

World News

39041 readers
2660 users here now

A community for discussing events around the World

Rules:

Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.

We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.

All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.


Lemmy World Partners

News !news@lemmy.world

Politics !politics@lemmy.world

World Politics !globalpolitics@lemmy.world


Recommendations

For Firefox users, there is media bias / propaganda / fact check plugin.

https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] diffuselight@lemmy.world 32 points 1 year ago* (last edited 1 year ago) (3 children)

That’s a fundamental misunderstanding of how diffusion models work. These models extract concepts and can effortlessly combine them to new images.

If it learns woman + crown = queen

and queen - woman + man = king

it is able to combine any such concept together

As Stability has noted. any model that has the concept of naked and the concept of child in it can be used like this. They tried to remove naked for Stable Diffusion 2 and nobody used it.

Nobody trained these models on CSAM and the problem is a dilemma in the same way a knife is a dilemma. We all know a malicious person can use a knife for murder, including of children Yet society has decided that knives sufficient other uses that we still allow their sale pretty much everywhere.

[–] grepe@lemmy.world 5 points 1 year ago (1 children)

This can be used by pedophiles is used as an argument to ban cryptography... I wonder if someone will apply that to the generative AI.

[–] piecat@lemmy.world 1 points 1 year ago

Depends how profitable it is.

If it can replace workers no, if it threatens the big players like Disney yes.

[–] 0ddysseus@lemmy.world 0 points 10 months ago

Here you go bud, no misunderstanding at all. The image generators are trained on CSAM, as I said.

https://www.independent.co.uk/news/ap-study-developers-thorn-canada-b2467386.html

[–] 0ddysseus@lemmy.world -2 points 1 year ago* (last edited 10 months ago) (1 children)

Editing this reply to say that I was in fact right and I did not have any fundamental misunderstanding of anything. And the database in question here is called LAIOn and contains 6 billions images scraped from the web, including CSAM images.

Thanks for that. As I said, I'm not big into how AI works, so not surprised I got that wrong. The databases of everything that has come across the clear web are still there though and are available for use by people with access.

[–] BreakDecks@lemmy.ml 5 points 1 year ago (2 children)

What are you referring to by "the database of everything that has come across the clear web"?

[–] 0ddysseus@lemmy.world 1 points 10 months ago

See this new article. The image database they looked into is called LAIOn. There are others though of course. I don't mean google crawlers, I mean image databases for training image generators.

https://www.independent.co.uk/news/ap-study-developers-thorn-canada-b2467386.html

[–] NightAuthor@lemmy.world 1 points 1 year ago

NSA servers? jkjk, kinda

I think they mean Google's web-crawler index, but I don't think that the index works that way.... well, on the other hand, they do cache some stuff.