147
you are viewing a single comment's thread
view the rest of the comments
[-] istanbullu@lemmy.ml 14 points 7 months ago

I don't buy into this "AI is dangerous" hype. Humans are dangerous.

[-] funkless_eck@sh.itjust.works 11 points 7 months ago

"ooh it's more advanced but don't worry- it's not conscious"

is as much a marketing tactic as "how it feels to chew 5 gum" or buzzfeedesque "top 10 celebrity mistakes - number 3 will blow your mind"

it's a tech product that runs a series of complicated loops against a large series of texts and returns the closest comparison, as it stands it's never going to be dangerous in and of itself.

[-] kromem@lemmy.world 2 points 7 months ago

it's a tech product that runs a series of complicated loops against a large series of texts and returns the closest comparison, as it stands it's never going to be dangerous in and of itself.

That's not how it works. I really don't get what's with people these days being so willing to be confidently incorrect. It's like after the pandemic people just decided that if everyone else was spewing BS from their "gut feelings," well gosh darnit they could too!

It uses gradient descent on a large series of texts to build a neural network capable of predicting those texts as accurately as possible.

How that network actually operates ends up a black box, especially for larger models.

But research over the past year and a half in simpler toy models has found that there's a rather extensive degree of abstraction. For example, a small GPT trained only on legal Othello or Chess moves ends up building a virtual representation of the board and tracks "my pieces" and "opponent pieces" on it, despite never being fed anything that directly describes the board or the concept of 'mine' vs 'other'. In fact, in the Chess model, the research found there was even a single vector in the neural network that could be flipped to have the model play well or play like shit regardless of the surrounding moves fed in.

It's fairly different from what you seem to think it is. Though I suspect that's not going to matter to you in the least, as I've come to find that explaining transformers to people spouting misinformation about them online has about the same result as a few years ago explaining vaccine research to people spouting misinformation about that.

[-] funkless_eck@sh.itjust.works 2 points 7 months ago* (last edited 7 months ago)

I dont know if saying "it's not a loop! it's an iterative process using a series of steps!" is that much of a burn.

my dude, that's a loop.

[-] Chakravanti@sh.itjust.works 2 points 7 months ago

Well He That Remains came by just to show that everything we experience is always part of a bigger loop. You can fucking kill him and even slam the break; crash to his design of the the highest number of alternate dimensions and then some and it won't stop the loop. 99.99% of the time he'll be back. We only need to consciously accept the concept of no more than the notion to summon his return. Even if we were to successfully crack the time management mech and undo his manipulation, he'll be back when we track him down to build another one.

The Loop is more nature than matter to energy combined. When everything in all of reality would expand infinitely far apart, the whole shebang goes lateral mirror again with a whole new dimension. There is no end to any aspect of reality. Anywhere it would be, turns out it's "just" "another" Loop Mirror.

[-] Thorny_Insight@lemm.ee 2 points 7 months ago

Generative AI and LLMs is not what people mean when they're talking about the dangers of AI. What we worry about doesn't exist yet.

[-] funkless_eck@sh.itjust.works 2 points 7 months ago

I dont think AI sentience as danger is going to be an issue in our lifetimes - we're 123 years in January since the first well known story featuring this trope (Karel Čapek's Rossumovi Univerzáiní Robotī)

We are a long way off from being able to copy virtual perception, action and unified agency of even basic organisms right now.

Therefore all claims about the "dangers" of AI are only dangers of humans using the tool (akin to the dangers of driving a car vs the dangers of cars attacking their owners without human interaction) and thus are just marketing hyperbole

in my opinion of course

[-] Thorny_Insight@lemm.ee 1 points 7 months ago

Well yeah perhaps, but isn't that kind of like knowing that an asteroid is heading towards earth and feeling no urgency about it? There's non-zero chance that we'll create AGI withing the next couple years. The chances may be low but consequences have the potential to literally end humanity - or worse.

[-] funkless_eck@sh.itjust.works 1 points 7 months ago

"non zero" isnt exactly convincing, to me. there is also a non-zero chance God exists.

[-] hikaru755@feddit.de 2 points 7 months ago

I mean... It might be. Just depends on how much potential there still is to get models up to higher reasoning capabilities, and I don't think anyone really knows that yet

[-] Thorny_Insight@lemm.ee 3 points 7 months ago

Yeah maybe. I just personally don't think LLMs are actually intelligent. They're just capable of faking intelligence but at the same time making errors that perfectly indicate that it's basically just bluffing. I'd be more worried about an AI that knows less things but demonstrates higer capability for logic and reasoning.

[-] Thorny_Insight@lemm.ee 7 points 7 months ago* (last edited 7 months ago)

AI can be dangerous. The point is not that it's likely but that in the very unlikely event of it going rogue it can at worst have civilication ending consequences.

Imagine how easy it is to trick a child as an adult. The difference in intelligence between a human and superintelligent AGI would be orders of magnitude greater that that.

[-] conciselyverbose@sh.itjust.works 2 points 7 months ago

An actual AI (that modern tools don't even vaguely resemble) could maybe theoretically be dangerous.

An LLM cannot be dangerous. There's no path to anything resembling intelligence or agency.

[-] kromem@lemmy.world 1 points 7 months ago

Exactly. People try to scare into regulatory capture talking about paperclip maximizers when meanwhile it's humans and our corporations that are literally making excess shit to the point of human extinction.

To say nothing for how often theorizing around 'superintelligence' imagines the stupidest tendencies of humanity being passed on to it while denying our smartest tendencies as "uniquely human" despite existing models largely already rejecting the projected features and modeling the 'unique' ones like empathy.

this post was submitted on 19 Apr 2024
147 points (88.5% liked)

Technology

59598 readers
1925 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS