404
submitted 10 months ago by db0@lemmy.dbzer0.com to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] Prandom_returns@lemm.ee 51 points 10 months ago

Can we fucking stop anthropomorphising software?

[-] db0@lemmy.dbzer0.com 76 points 10 months ago

"Hallucinate" is the standard term used to explain the GenAI models coming up with untrue statements

[-] Draegur@lemm.ee 24 points 10 months ago* (last edited 10 months ago)

in terms of communication utility, it's also a very accurate term.

when WE hallucinate, it's because our internal predictive models are flying off the rails filling in the blanks based on assumptions rather than referencing concrete sensory information and generating results that conflict with reality.

when AIs hallucinate, it's due to its predictive model generating results that do not align with reality because it instead flew off the rails presuming what was calculated to be likely to exist rather than referencing positively certain information.

it's the same song, but played on a different instrument.

[-] arken@lemmy.world 5 points 10 months ago

when WE hallucinate, it's because our internal predictive models are flying off the rails filling in the blanks based on assumptions rather than referencing concrete sensory information and generating results that conflict with reality.

Is it really? You make it sound like this is a proven fact.

[-] CosmicCleric@lemmy.world 4 points 10 months ago* (last edited 10 months ago)

Is it really? You make it sound like this is a proven fact.

I believe that's where the scientific community is moving towards, based on watching this Kyle Hill video.

[-] PipedLinkBot@feddit.rocks 0 points 10 months ago

Here is an alternative Piped link(s):

this Kyke Hill video

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

[-] Dasus@lemmy.world 2 points 10 months ago

I know I'm responding to a bot, but... how does a PipedLinkBot get "Kyle Hill" wrong to "Kyke Hill"? More AI hallucinations?

[-] FarceOfWill@infosec.pub 4 points 10 months ago

Op has a pencil in the top right, looks like it was edited

[-] Dasus@lemmy.world 2 points 10 months ago

True, I missed that

[-] KillingTimeItself@lemmy.dbzer0.com 2 points 10 months ago

i mean, idk about the assumptions part of it, but if you asked a psych or a philosopher, im sure they would agree.

Or they would disagree and have about 3 pages worth of thoughts to immediately exclaim otherwise they would feel uneasy about their statement.

[-] UmeU@lemmy.world 1 points 10 months ago

Better than one of those pesky unproven facts

[-] assassinatedbyCIA@lemmy.world 2 points 10 months ago

I think a more accurate term would be confabulate based on your explanation.

[-] Draegur@lemm.ee 1 points 10 months ago

you know what, i like that! I like that a lot!

load more comments (19 replies)
[-] planish@sh.itjust.works 12 points 10 months ago

No?

An anthropomorphic model of the software, wherein you can articulate things like "the software is making up packages", or "the software mistakenly thinks these packages ought to exist", is the right level of abstraction for usefully reasoning about software like this. Using that model, you can make predictions about what will happen when you run the software, and you can take actions that will lead to the outcomes you want occurring more often when you run the software.

If you try to explain what is going on without these concepts, you're left saying something like "the wrong token is being sampled because the probability of the right one is too low because of several thousand neural network weights being slightly off of where they would have to be to make the right one come out consistently". Which is true, but not useful.

The anthropomorphic approach suggests stuff like "yell at the software in all caps to only use python packages that really exist", and that sort of approach has been found to be effective in practice.

[-] Nom@lemm.ee 2 points 10 months ago
[-] PipedLinkBot@feddit.rocks 0 points 10 months ago

Here is an alternative Piped link(s):

too late

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

this post was submitted on 28 Mar 2024
404 points (97.2% liked)

Technology

61833 readers
1017 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS