295
submitted 1 year ago by L4s@lemmy.world to c/technology@lemmy.world

WhatsApp’s AI shows gun-wielding children when prompted with ‘Palestine’::By contrast, prompts for ‘Israeli’ do not generate images of people wielding guns, even in response to a prompt for ‘Israel army’

you are viewing a single comment's thread
view the rest of the comments
[-] Valmond@lemmy.mindoki.com 11 points 1 year ago

It's not about "adding code" or any other bullshit.

AI today is trained on datasets (that's about it), the choice of datasets can be complicated, but that's where you moderate and select. There is nothing "AI learns of its own" sci-fi dream going on.

Sigh.

[-] Serdan@lemm.ee 6 points 1 year ago

It's reasonable to refer to unsupervised learning as "learning on its own".

[-] Torvum@lemmy.world 2 points 1 year ago

Really wish the term virtual intelligence was used (literally what it is)

[-] GiveMemes@jlai.lu 8 points 1 year ago

We should honestly just take the word intelligence out of the mix for rn bc these machines aren't "intelligent". They can't do things like critically think, form its own opinions, etc. They're just super efficient data aggregation at the end of the day, whether or not they're based on the human brain.

We're so far off from 'intelligent' machine learning that I think it really throws off how people think about it to call it intelligence of any sort.

[-] Torvum@lemmy.world 1 points 1 year ago

Techbros just needed to use the search engine optimization buzzword tbh.

[-] Serdan@lemm.ee 1 points 1 year ago

LLMs can reason about information. It's fine to call them intelligent systems.

[-] ichbinjasokreativ@lemmy.world -1 points 1 year ago

One of the many great things about the mass effect franchise is its separation of AI and VI, the latter being non-conscious and simple and the former being actually 'awake'

[-] theyoyomaster@lemmy.world 1 points 1 year ago

It is about adding code. No dataset will be 100% free of undesirable results. No matter what marketing departments wish, AI isn't anything close to human "intelligence," it is just a function of learned correlations. When it comes to complex and sensitive topics, the difference between correlation and causation is huge and AI doesn't distinguish. As a result, they absolutely hard code AI models to avoid certain correlations. Look at the "[character] doing 9/11" meme trend. At the fundamental level it is impossible to restrict undesirable outcomes by avoiding them in training models because there are an infinite combinations of innocent things that become sensitive when linked in nuanced ways. The only way to combat this is to manually delink certain concepts; they merely failed to predict it correctly for this specific instance.

this post was submitted on 04 Nov 2023
295 points (93.0% liked)

Technology

59598 readers
3933 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS