362
submitted 11 months ago* (last edited 11 months ago) by SnotFlickerman@lemmy.blahaj.zone to c/asklemmy@lemmy.ml

Money wins, every time. They're not concerned with accidentally destroying humanity with an out-of-control and dangerous AI who has decided "humans are the problem." (I mean, that's a little sci-fi anyway, an AGI couldn't "infect" the entire internet as it currently exists.)

However, it's very clear that the OpenAI board was correct about Sam Altman, with how quickly him and many employees bailed to join Microsoft directly. If he was so concerned with safeguarding AGI, why not spin up a new non-profit.

Oh, right, because that was just Public Relations horseshit to get his company a head-start in the AI space while fear-mongering about what is an unlikely doomsday scenario.


So, let's review:

  1. The fear-mongering about AGI was always just that. How could an intelligence that requires massive amounts of CPU, RAM, and database storage even concievably able to leave the confines of its own computing environment? It's not like it can "hop" onto a consumer computer with a fraction of the same CPU power and somehow still be able to compute at the same level. AI doesn't have a "body" and even if it did, it could only affect the world as much as a single body could. All these fears about rogue AGI are total misunderstandings of how computing works.

  2. Sam Altman went for fear mongering to temper expectations and to make others fear pursuing AGI themselves. He always knew his end-goal was profit, but like all good modern CEOs, they have to position themselves as somehow caring about humanity when it is clear they could give a living flying fuck about anyone but themselves and how much money they make.

  3. Sam Altman talks shit about Elon Musk and how he "wants to save the world, but only if he's the one who can save it." I mean, he's not wrong, but he's also projecting a lot here. He's exactly the fucking same, he claimed only he and his non-profit could "safeguard" AGI and here he's going to work for a private company because hot damn he never actually gave a shit about safeguarding AGI to begin with. He's a fucking shit slinging hypocrite of the highest order.

  4. Last, but certainly not least. Annie Altman, Sam Altman's younger, lesser-known sister, has held for a long time that she was sexually abused by her brother. All of these rich people are all Jeffrey Epstein levels of fucked up, which is probably part of why the Epstein investigation got shoved under the rug. You'd think a company like Microsoft would already know this or vet this. They do know, they don't care, and they'll only give a shit if the news ends up making a stink about it. That's how corporations work.

So do other Lemmings agree, or have other thoughts on this?


And one final point for the right-wing cranks: Not being able to make an LLM say fucked up racist things isn't the kind of safeguarding they were ever talking about with AGI, so please stop conflating "safeguarding AGI" with "preventing abusive racist assholes from abusing our service." They aren't safeguarding AGI when they prevent you from making GPT-4 spit out racial slurs or other horrible nonsense. They're safeguarding their service from loser ass chucklefucks like you.

you are viewing a single comment's thread
view the rest of the comments
[-] SnotFlickerman@lemmy.blahaj.zone -5 points 11 months ago
  1. LLM /= AGI

  2. Models made for specific purpose instead of general purpose are of course going to need less CPU cycles because you aren't creating an AGI, you are creating a specialized tool.

  3. It still takes far longer to produce a result on smaller hardware. An AGI that takes days to do anything isn't exactly that dangerous.

[-] gladflag@lemmy.ml 3 points 11 months ago
  1. Why are you even talking about AGI needing a certain amount of compute then? You’re using LLM numbers.

Go have a read of some doomsday scenarios. I’m not saying they’re right, but it feels plausible to me.

[-] SnotFlickerman@lemmy.blahaj.zone -5 points 11 months ago* (last edited 11 months ago)

Go have a read of some doomsday scenarios. I’m not saying they’re right, but it feels plausible to me.

I have, and I have been working with computer hardware and networking my whole life. I have a degree in network administration. I think the fears are absolutely overblown by people who don't understand hardware.

Most people don't own fancy new computers. Most people are still running shit from 10 years ago and don't want to have to upgrade. The idea that the world could be taken over by an AGI seems literally fancifully absurd to me.

[-] 520@kbin.social 1 points 11 months ago* (last edited 11 months ago)

Bro, I run AI projects on a consumer fucking laptop. Sure, they aren't exactly LLM levels of complexity, but anybody with a need for serious hardware for a bit will just rent it off AWS or so.

[-] flashgnash@lemm.ee 3 points 11 months ago

I understand that LLMs are not agi, but as agis don't currently exist I think it's fair to assume the same concept that applies to literally all software of over time people discover more efficient ways to do things will also apply to it

Also we don't know how slow or fast it will end up being, some deep learning models are incredibly fast, some are slow

[-] SnotFlickerman@lemmy.blahaj.zone -3 points 11 months ago* (last edited 11 months ago)

Once you get down to individual bits, you can't make code any smaller. You have a finite number of bits to work with. In networking, especially.

There is literally an upper (lower?) limit on how small you can make code.

Like others in the thread, I think you're confusing the great pace at which we have increased the hardware speed of computers and the miniaturization of computer components with "code" somehow getting "smaller" which... isn't really a thing when you're dealing with something as complex as this. You can't run an LLM on the same number of lines you can print up "Hello World!"

It's way more that we have more CPU speed, more RAM, and faster storage with more space for data to live.

[-] flashgnash@lemm.ee 2 points 11 months ago

print(1) print(2) print(3) print(4) print(5)

for I=1,5: print(I)

There you go I made code smaller

I also never said anything about making code smaller I said making it more efficient. It's not about compressing it it's about finding better, less CPU expensive ways to do things, which we absolutely do

Another AI based example, video chats currently work streaming video, but there's a technology in development that takes one screenshot, sends that, then sends expression data to be reconstructed on the other side

Far more efficient network wise

Hardware speed has increased, sure but that applies to both consumer hardware and servers, all a theoretical AGI would have to do is improve on its own training/code enough that it will run at all on consumer level hardware (which language models currently will do

(For reference, llama 40B runs just fine on my ThinkPad from 2016, pre-trained models are not that difficult to run, training is the expensive part)

[-] SnotFlickerman@lemmy.blahaj.zone -1 points 11 months ago* (last edited 11 months ago)

You're misunderstanding what I mean by "making code smaller." Because.... that's not that much smaller. Each Unicode character is 2 bytes, with some being as many as 4 bytes. This code snippet is 64 bytes. Can you magically make Unicode characters smaller than 2 bytes? You can't. There's a literal physical limit on how small you can make code.

Sure, you can come up with clever ways to use less code. But my point is there is a limit on how much less code you can use, and that always is based on physical hardware limitations. Just because modern hardware makes it feel limitless doesn't mean it is.

EDIT: Got my data sizes mixed up.

[-] flashgnash@lemm.ee 1 points 11 months ago

Ok, but who says we're anywhere near that limit with AI? It's still very new technology

Used to be you'd need a massive disk for megabytes of slow storage, now we've got 4tb nvmes smaller than a credit card

As I said, this theoretical AGI does not have to make itself tiny, all it would have to do is be able to run at a reasonable speed on the average gaming PC for example which is feasible considering the heaviest pre-trained AI models will run on hardware from 2016 albeit somewhat slowly

I don't think the argument there's a physical limit works here, as it's entirely unknown how efficient existing models can be made currently, let alone a purely hypothetical AGI which can and will be used to improve on itself

this post was submitted on 20 Nov 2023
362 points (88.3% liked)

Asklemmy

43728 readers
1228 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~

founded 5 years ago
MODERATORS