67
all 45 comments
sorted by: hot top controversial new old
[-] boomer478@lemmy.ml 54 points 9 months ago

People have been creating and posting realistic looking fake celebrity nudes for quite literally decades now, but now they're using AI and its suddenly a problem?

[-] kromem@lemmy.world 41 points 9 months ago

I'm not sure if you noticed, but people who write for a living have suddenly started writing quite a lot about how technology that can write and generate media are bad.

[-] hazeebabee@slrpnk.net 9 points 9 months ago

Which is so silly, because AI writing still needs a human editor. I write for a living and there tons of work that involves using AI as a tool to increase productivity rather than to replace writers completely... like photoshop didnt put photographers out of business it just changed the work flow.

[-] H_Interlinked@lemmy.world 11 points 9 months ago* (last edited 9 months ago)

I work in a clinical setting where some Doctors are trying an AI program for generating their clinical notes out of the casual conversation between them and the patient. It's way off its mark for what we demand in quality. It requires significant editing from the healthcare provider, and if the note is very robust it quickly becomes more of a chore than modern voice transcription. Our review is not great so far.

[-] kromem@lemmy.world 6 points 9 months ago

That's a terrible way to be using a LLM for generating clinical notes.

Sounds more like trying to use a screwdriver to hammer in screws than an issue with the screwdriver itself.

[-] Eggyhead@kbin.social 8 points 9 months ago* (last edited 9 months ago)

Sounds more like trying to use a screwdriver to hammer in screws

This is what I think about AI being forced into many things these days. Feels more like an attempt to justify subscription plans than anything actually productive.

[-] kromem@lemmy.world -2 points 9 months ago

In part this is because the SotA model is by far GPT-4, but OpenAI has pigeon holed it into 'chatbot.'

The earliest versions of it pre-release when it was being incorporated into Bing were amazing. Probably the most impressive thing I've seen in tech.

But it was too human-like and freaking users out, so rather than wait for the market to adjust they did extensive fine tuning to make the large language model trained to predict human ouput be less likely to produce human-like output.

The problem is that they don't have a scalpel for this sort of thing and ended up with a model that's very good as a chatbot within a certain scope, but significantly impaired at some of the outside the box mechanics visible early on.

And because it's the SotA, everyone is now using it to fine tune their own models.

So the entire industry is being set back in practical applications outside of "kind of boring chatbot."

[-] H_Interlinked@lemmy.world 5 points 9 months ago

Right. It seemed like a reach when I first heard of it, but that's how it's advertised and the Hospital was sold on at least trying it out.

[-] Dukeofdummies@kbin.social 35 points 9 months ago* (last edited 9 months ago)

It's because a person can crank out a deep fake in 3 hours, and a crappy one in one. It never cropped up because... well lets be real it was a couple of weirdos that were doing it, unless it bubbles up from the dark corners of the internet you risk the Streisand effect by bringing attention to it.

AI can crank out 40 in a minute. 7200 in three hours. That's an entirely different beast. The sheer mass and volume ramps up the odds of any image bubbling up from the dark corners of the web falling into the limelight and now this problem that wasn't big enough to merit thought is rearing up it's ugly head right in front of us.

You can generate unique pictures of Taylor Swift faster than even Taylor swift can generate pictures of Taylor Swift. Within one hour of Taylor swift being seen with a man (and you have enough images of the man) you can create a dozen images of her on a date with that man and attempt to sell them to paparazzi.

The problem is volume. Just like how email made everyone connected and allowed the Nigerian Prince scandal to occur.

[-] echo64@lemmy.world 22 points 9 months ago

It's also not just limited to Taylor swift.

People don't care now because people don't thin famous people should have any right to dignity, but their minds are gonna change really quickly when it's their sister, mother or daughter.

[-] macattack@lemmy.world 7 points 9 months ago

I don't think many have gone viral on social media before or took less than 5 minutes to create. My uneducated guess is that previously this stuff would be in some niche forum in the recesses of the internet

[-] gian@lemmy.grys.it 3 points 9 months ago

My uneducated guess is that previously this stuff would be in some niche forum in the recesses of the internet

Not really. Back at the time there were public usenet groups specifically dedicated to the [hot actress of the day] fake porn.

[-] finthechat@kbin.social 4 points 9 months ago

I can see why people are upset, I can agree that distribution of these images can be an issue, but this has the same energy of "I am mad that a certain picture of me is on the internet, I demand that they take it down." Sorry, that's not going to happen any time soon.

[-] echo64@lemmy.world 2 points 9 months ago

It's now massively accessible and realistic. Yes, it's a problem.

[-] Brunbrun6766@lemmy.world 13 points 9 months ago

Google.com/images

Damn look at that, been accessible for decades

[-] echo64@lemmy.world 2 points 9 months ago

The tools are accessible. I wish this place wasn't full of weirdo ai tech bros sometimes.

[-] BruceTwarzen@kbin.social 4 points 9 months ago

Because downloading gimp is really hard.

[-] echo64@lemmy.world -3 points 9 months ago* (last edited 9 months ago)

I'm sure you do not need me to explain to you the cavernous difference between gimp and deepfake. I trust you are capable of that.

[-] magnusrufus@lemmy.world 2 points 9 months ago

You could be honest and acknowledge that there is a massive difference in time investment and skill required between the old way of creating fake porn of unconsenting people and the new way.

[-] Stalinwolf@lemmy.ca 20 points 9 months ago

I was a part of a Blursed AI group on Facebook that had been a lot of fun, but suddenly this week it first shifted to Taylor Swift porn, and then to alt-right MAGA shit very quickly. The comments on the Trump and MAGA related images were very on-board with it too. The change was so abrupt that I got the fuck out of there. Seemed orchestrated.

[-] jballs@sh.itjust.works 19 points 9 months ago
[-] M137@lemmy.world 2 points 9 months ago

Hahaha, that's glorious!

[-] kamenlady@lemmy.world 13 points 9 months ago

Platforms, including X, where the images were first shared

They also now have a dedicated platform at their disposal. Expect more trash to be posted first on X in the future.

[-] gian@lemmy.grys.it 1 points 9 months ago

Like back at the time they had Usenet groups, rotten.com and a long list of other sites to post fake porn and other discutible content.

Not that I am saying that is something right, just that is nothing new.

[-] dditty@lemm.ee 10 points 9 months ago

!https://i.kym-cdn.com/entries/icons/facebook/000/038/932/tco_-_2021-11-24T113041.108.jpg

[-] Treczoks@lemmy.world 8 points 9 months ago

She probably has the money to sue each and every poster into oblivion.

[-] aniki@lemm.ee -5 points 9 months ago* (last edited 9 months ago)

For what? She's a public figure who does photo shoots. That's fair game buddy.

[-] magnusrufus@lemmy.world 4 points 9 months ago

The fuck is wrong with you?

[-] aniki@lemm.ee -3 points 9 months ago* (last edited 9 months ago)

I didn't say it's moral or good or just -- but it's not illegal. My usage of fair game is in reference to fair use which is allowed under copyright law.

Try not eating the crayons.

[-] magnusrufus@lemmy.world 3 points 9 months ago

If you meant a different term with a different meaning then you should use that term and even in that case I doubt that the fair use clause would stand as a defense.

[-] aniki@lemm.ee -3 points 9 months ago* (last edited 9 months ago)

I mean -- no one has taken it to court so you can't say one way or another. Is it fair use if I put a cover of a magazine into stable diffusion and have it change the background? If I put michelle obama on mars did i just break the law? [no] This is a legal rats nest.

[-] white_shotgun@aussie.zone 6 points 9 months ago

The grover one was hilarious... Jus sayin

[-] GuyDudeman@lemmy.world -2 points 9 months ago

Is it a problem though? I mean, it just makes Rule34 pics that much easier to create. And you wouldn't want to kink-shame anyone, now, would you? Why is it always heterosexual men who are kink-shamed? Why is liking naked women a bad thing?

[-] denshirenji@lemmy.world 43 points 9 months ago

Alright, I'll bite. To start, this is a real person that we are talking about. A real person who did not consent. Does that mean anything to you? The fact that there is a very real person that exists in very real life that has had this happen to them?

Otherwise, I agree. Nothing wrong with the male libido.

[-] andrai@feddit.de 15 points 9 months ago

I don't need someone's consent to draw lewd fanart of them.

[-] admin@lemmy.my-box.dev 11 points 9 months ago* (last edited 9 months ago)

I wouldn't be so sure. Depending on where you are in the world, there dozens of laws that might interfere. Ranging from publicity rights to slander, especially if the images are photorealistic (enough).

[-] denshirenji@lemmy.world 10 points 9 months ago

Do you need their permission to distribute that art?

[-] Grimy@lemmy.world 9 points 9 months ago

100%, which is why the fault lies with the bad actors and the platforms that let this proliferate and not the tool itself.

Can you imagine this headline but with Photoshop instead of AI? It would be utterly silly.

This is orchestrated to create anger against AI. There's a lot of money involved in it and that money triples if consumers aren't allowed to run and distribute models on their own PC.

[-] denshirenji@lemmy.world 4 points 9 months ago

I can agree with you on principal. Just as long as you aren't distributing things like this I don't really see an issue. Not the tool, the distributor / platform. I also agree that these articles are meant to ensure that those technology can be held behind locked doors. I fully support the idea of making AI something that is self-hostable.

That being said there are people in this thread that see nothing wrong with distributing lewd pictures of real people (drawn, ai generated, or even hacked and stolen). That is the only thing that I was addressing.

[-] Grimy@lemmy.world 1 points 9 months ago

Ya I agree, there's a very ugly side of things and It's a shame certain people are rolling themselves in it.

There's also the fact that lewds of celebs and similar material will grow exponentially but the same is also true for all other media so hopefully it balances itself out.

I think that's the main problem with a lot of these articles, they are missing the forest for the trees so to speak. We are looking at an explosion of culture, the bad stuff is just along for the ride. I'm personally excited for it.

[-] BarbecueCowboy@kbin.social 0 points 9 months ago

This is actually super super tricky.

So, there's an exemption for 'Transformative' art, and while this is obviously pretty shady, it feels like there's a good chance this would qualify as transformative. Basically, you can't copy an existing photograph you don't own, but you can take an existing person and paint a new original picture of them.

We had a big lawsuit just last year where the Supreme Court clarified the line a bit. In that case, the art was found to be not Transformative, but they did a lot to explain why, and based on that, this would be super likely to fall on the side of 'Legally Allowed'.

[-] NoIWontPickaName@kbin.social -1 points 9 months ago
[-] denshirenji@lemmy.world 2 points 9 months ago

I'm not a lawyer and can't even begin to answer that question. I was merely trying to get the conversation starting down that logical track, because I, personally, think that it is at the heart of the matter.

Looking this up, it seems that, at least in California, it probably would be considered illegal, at least according to this site.

[-] Candelestine@lemmy.world 6 points 9 months ago

So, this line of thought is not going to get AI fakes permitted, it's going to get rule 34 banned.

We've been censoring sex shit for a long ass time, in case you haven't noticed. The recent trend of information freedom is not going to defeat that old religious puritanical bs, and people's wishes for privacy on top of it.

Trump would shut that shit down fast. Conservatives want people reproducing, not masturbating, and he has christian supporters to keep in line, who do not like porn.

this post was submitted on 25 Jan 2024
67 points (79.6% liked)

Technology

59197 readers
848 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS