362
submitted 11 months ago* (last edited 11 months ago) by SnotFlickerman@lemmy.blahaj.zone to c/asklemmy@lemmy.ml

Money wins, every time. They're not concerned with accidentally destroying humanity with an out-of-control and dangerous AI who has decided "humans are the problem." (I mean, that's a little sci-fi anyway, an AGI couldn't "infect" the entire internet as it currently exists.)

However, it's very clear that the OpenAI board was correct about Sam Altman, with how quickly him and many employees bailed to join Microsoft directly. If he was so concerned with safeguarding AGI, why not spin up a new non-profit.

Oh, right, because that was just Public Relations horseshit to get his company a head-start in the AI space while fear-mongering about what is an unlikely doomsday scenario.


So, let's review:

  1. The fear-mongering about AGI was always just that. How could an intelligence that requires massive amounts of CPU, RAM, and database storage even concievably able to leave the confines of its own computing environment? It's not like it can "hop" onto a consumer computer with a fraction of the same CPU power and somehow still be able to compute at the same level. AI doesn't have a "body" and even if it did, it could only affect the world as much as a single body could. All these fears about rogue AGI are total misunderstandings of how computing works.

  2. Sam Altman went for fear mongering to temper expectations and to make others fear pursuing AGI themselves. He always knew his end-goal was profit, but like all good modern CEOs, they have to position themselves as somehow caring about humanity when it is clear they could give a living flying fuck about anyone but themselves and how much money they make.

  3. Sam Altman talks shit about Elon Musk and how he "wants to save the world, but only if he's the one who can save it." I mean, he's not wrong, but he's also projecting a lot here. He's exactly the fucking same, he claimed only he and his non-profit could "safeguard" AGI and here he's going to work for a private company because hot damn he never actually gave a shit about safeguarding AGI to begin with. He's a fucking shit slinging hypocrite of the highest order.

  4. Last, but certainly not least. Annie Altman, Sam Altman's younger, lesser-known sister, has held for a long time that she was sexually abused by her brother. All of these rich people are all Jeffrey Epstein levels of fucked up, which is probably part of why the Epstein investigation got shoved under the rug. You'd think a company like Microsoft would already know this or vet this. They do know, they don't care, and they'll only give a shit if the news ends up making a stink about it. That's how corporations work.

So do other Lemmings agree, or have other thoughts on this?


And one final point for the right-wing cranks: Not being able to make an LLM say fucked up racist things isn't the kind of safeguarding they were ever talking about with AGI, so please stop conflating "safeguarding AGI" with "preventing abusive racist assholes from abusing our service." They aren't safeguarding AGI when they prevent you from making GPT-4 spit out racial slurs or other horrible nonsense. They're safeguarding their service from loser ass chucklefucks like you.

you are viewing a single comment's thread
view the rest of the comments
[-] flashgnash@lemm.ee 1 points 11 months ago* (last edited 11 months ago)

The billions of those in squishy APUs don't have the ability and resources to make exact copies of themselves, or scale up said APUs to improve the speed they can think at and the number of things they can think about at once

Whether it would decide to use that to take over the world is a different question entirely but left to its own devices, if it wanted to it could bring that about in a way that human beings can't do nearly as efficiently

If a human tries to do horrible stuff they eventually die and can't do said horrible stuff anymore. They also can't gain more than one lifetime's worth of knowledge

AGI in the form it is generally considered would be like a human that lives forever, can clone itsself perfectly, requires no sleep, food, etc, can teleport anywhere in the world provided there's a computer there for it to use and could modify its own brain or even create an entirely new one

Take over the world doesn't mean anything specific. It means a theoretical AGI connected to the internet could do pretty much anything it felt like doing whether humans liked it or not (provided it has enough time to gain enough of a foothold)

[-] MudMan@kbin.social 2 points 11 months ago* (last edited 11 months ago)

They literally do and have done for tens of thousands of years. One may say that's how they got to AGI in the first place, the squishies. And then they learned to write for that whole "one lifetime of knowledge thing" and you wouldn't believe the kind of stuff they got into after that. Scary stuff.

Also, they have hands. Big advantage, the hands. Great for grabbing things. Remarkably hard to stay plugged in if your rival has hands and you don't. Big competitive disadvantage.

Alright, I think this conversation has derailed enough. We can maybe pick it back up when we have a firm standard for world takeovers. If you guys boil it down to a set of steps I may even give it a go. I don't have anything better to do this week.

this post was submitted on 20 Nov 2023
362 points (88.3% liked)

Asklemmy

43728 readers
1228 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy ๐Ÿ”

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~

founded 5 years ago
MODERATORS