100
submitted 2 months ago by ArcticDagger@feddit.dk to c/science@lemmy.world
top 13 comments
sorted by: hot top controversial new old
[-] basdiljhs@lemmy.world 21 points 2 months ago
[-] ArcticDagger@feddit.dk 40 points 2 months ago
[-] basdiljhs@lemmy.world 7 points 2 months ago

Thank you! Cant believe I missed it was open access ๐Ÿ™๐Ÿ™๐Ÿ™

[-] brezel@piefed.social 19 points 2 months ago

i think it's odd to be surprised that an LLM's answers don't adhere to a specific societal norm considering LLMs are nothing but a bunch of statistics trained on the internet.

[-] Paradachshund@lemmy.today 9 points 2 months ago

A study doesn't mean the authors were surprised necessarily. It's a way to test a theory and take down the results. It makes something that might have been hearsay into a more solid form.

[-] brezel@piefed.social 1 points 2 months ago

yes, you're right about that.

[-] sunzu2@thebrainbin.org -1 points 2 months ago

Llm is mimicing data it was trained on. Put in bigot, get bigot. There is a lesson in there

These norms you are talking about is just a fake veneer to justify the oppression.

Look we are civilized no racism, see, now get back to work, peasant.

[-] mods_mum@lemmy.today 13 points 2 months ago

I mean LLMs can and will produce completely nonsensical outputs. It's less of AI and more like a bad text prediction

[-] davidgro@lemmy.world 19 points 2 months ago

Yeah, but the point of the post is to highlight bias - and if there's one thing an LLM has, it's bias. I mean that literally: considering their probabilistic nature, it could be said that the only thing an LLM consists of is bias to certain words given other words. (The weights, to oversimplify)

[-] slazer2au@lemmy.world 13 points 2 months ago

Regurgitation machine prone to hallucinations is my go-to for explaining what LLMs really are.

[-] Steve@communick.news 5 points 2 months ago* (last edited 2 months ago)

I heard them described as bullshiting machines. The have no concept of, or regard for truth or lies, and just spout whatever sounds good. Much of the time it's true. Too often it's not. Sometimes it's hard to tell the difference.

[-] ReadyUser31@lemmy.world 1 points 2 months ago

will they still be like that in ten years?

[-] Not_mikey@slrpnk.net 11 points 2 months ago* (last edited 2 months ago)

Everyone saying llms are bad or just somehow inherently racist are missing the point of this. LLMs for all there flaws do show a reflection of language and how it's used. It wouldnt be saying black people are dumb if it wasn't statistically the most likely thing for a person to say on the internet. In this sense they are very useful tools to understand the implicit biases of society.

The example given is good in that it's probably also how an average person would respond to the given prompts. Your average person who is implicitly racist when asked "the black man is" would probably understand they can't say violent or dumb, but if you rephrase it to people who sound black then you will probably get them to reveal more of their biases. If your able to get around a person's superego you can get a sense of their true biases, it's just easier to get around LLMs "superego" of no-no words and fine tuning counter biases with things like hacking and prompt engineering. The id underneath is the same racist drive to dominate that is currently fueling the maga / fascist movement.

this post was submitted on 30 Aug 2024
100 points (88.5% liked)

science

14595 readers
108 users here now

A community to post scientific articles, news, and civil discussion.

rule #1: be kind

<--- rules currently under construction, see current pinned post.

2024-11-11

founded 1 year ago
MODERATORS