this post was submitted on 22 Aug 2023
340 points (81.4% liked)

Technology

60129 readers
2701 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
 

Each of these reads like an extremely horny and angry man yelling their basest desires at Pornhub’s search function.

you are viewing a single comment's thread
view the rest of the comments
[–] GBU_28@lemm.ee 4 points 1 year ago (3 children)

You do know how LLM are trained right?

[–] diffuselight@lemmy.world 13 points 1 year ago (2 children)

I just retained an LLM on your comment you put on the public internet. You feel violated enough to equate it to physical violation?

[–] urist@lemmy.blahaj.zone 1 points 1 year ago (4 children)

How would you respond to photo realistic porn that looks like your mother, daughter, [insert person you care about here] especially if they found it distressing?

How would you feel if it was posted on facebook? How would you feel if they had to deal with it at work? From coworkers? From clients?

We are entering uncharted waters. You know why this is different than training a model on text, and your reply to @GBU_28@lemm.ee is hostile and doesn't acknowledge why people would be upset about AI porn featuring their likeness.

[–] diffuselight@lemmy.world 11 points 1 year ago* (last edited 1 year ago) (1 children)

you are answering a question with a different question. LLMs don’t make pictures of your mom. And this particular question?. One that has roughly existed since Photoshop existed.

It just gets easier every year. It was already easy. You could already pay someone 15 bucks on Fiver to do all of that, for years now.

Nothing really new here.

The technology is also easy. Matrix math. About as easy to ban as mp3 downloads. Never stopped anyone. It’s progress. You are a medieval knight asking to put gunpowder back into the box, but it’s clear it cannot be put back - it is already illegal to make non consensual imagery just as it is illegal to copy books. And yet printers exist and photocopiers exist.

Let me be very clear - accepting the reality that the technology is out there, it’s basic, easy to replicate and on a million computers now is not disrespectful to victims of no consensual imagery.

You may not want to hear it, but just like with encryption, the only other choice society has is full surveillance of every computer to prevent people from doing “bad things”. everything you complain about is already illegal and has already been possible - it just gets cheaper every year. What you want to have protection from is technological progress because society sucks at dealing with the consequences of it.

To be perfectly blunt, you don’t need to train any generative AI model for powerful deepfakes. You can use technology like Roop and Controlnet to synthesize any face on any image from a singe photograph. Training not necessary.

When you look at it that way, what point is there to try to legislate training with these arguments? None.

[–] urist@lemmy.blahaj.zone -1 points 1 year ago (1 children)

I’m not making an argument to ban it. I’m just pointing out you’re pretending a model from text someone wrote is similar to a model that makes nonconcentual porn.

I don’t think it can be banned, it’s just something they will need to encorperate into revenge porn laws, if it isn’t already covered.

I’m just pointing out your comment sucked.

[–] diffuselight@lemmy.world 5 points 1 year ago* (last edited 1 year ago) (1 children)

It’s already covered under those laws. So what are you doing that’s different from ChatGPT hallucinating here ?

Those laws don’t spell out the tools (photoshop); they hinge on reproducing likeness.

[–] urist@lemmy.blahaj.zone -5 points 1 year ago

Oh good, someone who has read every revenge porn law, ever. I’m glad they work exactly the same, in every nation and state.

Anyway, I must be hallucinating, true, because it seems you keep attacking what I’m saying, instead of defending the comment you made earlier that I took issue with, the one that points out you’re being needlessly hostile.

[–] stevedidWHAT@lemmy.world 7 points 1 year ago* (last edited 1 year ago) (1 children)

I can do this right now with photoshop dude what are you talking about. This just points to the need for more revenge porn laws.

We don’t have to sit in the fire when we can crawl out. Are we still on fire? Yeah. Can we do something about that? Yeah!

It seems like so many people these days want perfect solutions but the reality is that sometimes we have to make incremental solutions to erase the problem as much as we can.

[–] polymer@lemmy.world 5 points 1 year ago* (last edited 1 year ago) (1 children)

And incidentally, this need for revenge porn laws is also a symptomatic issue with a separate cause. Technology always moved forward and with no relation to social advancement, where there is also no realistic “Genie being forced back in the bottle” scenario either.

That being said, easier access to more powerful technology with lackluster recognition of personal responsibility doesn’t exactly bring happy prospects. lol…

[–] stevedidWHAT@lemmy.world 1 points 1 year ago

Agreed, personal responsibility went out the window a long time ago. Apathy reigns supreme.

[–] Donjuanme@lemmy.world 4 points 1 year ago

Revenge porn/blackmail/exploitation will hopefully become much less obscene, not to the "let's not prosecute this" levels, but maybe people can stop living in fear of their lives being ruined by malicious actors (unless that's your kink, you do you).

It will take/drive/demand a massive cultural shift, but if you asked me which world I would rather live in, and the options are one where people are abused and exploited, or one where people can visualize their perversions more easily (but content creators have a harder time making a living) I'll take the second. Though I may have straw-manned a bit, it's not something I've thought of outside of this forum thread.

[–] afraid_of_zombies@lemmy.world 2 points 1 year ago

I wouldn't be happy about it but me not being happy about something doesn't mean I just get an override.

I think the boat has sailed a bit on this one. You can't really copyright your own image and even if you were some famous person who is willing to do this and fight the legal battles you still have to go up against the fact that no one is making money off of it. You might be able to get a news source to take down that picture of you but it is another thing to make it so the camera company can't even record you.

But hey I was saying for years that we need to change the laws forbidding photography of people and property without consent and everyone yelled at me that they have the right to use a telescoping lense to view whomever they wanted blocks away.

The creeps have inherited the earth.

[–] GBU_28@lemm.ee 1 points 1 year ago

Why would I? Folks who have had real nudes of them posted on the Internet haven't felt "physical violation" but they've certainly been violated.

If you had photos of me and trained a porn generating LLM on my photos and shared porn of me, in an identifiable way, I would consider that violation.

But simply taking my words in that simple sentence isn't identifiable, unique, or revealing. So no.

Further, the original point was about the ethics of AI porn. You can't get something from nothing.