4
you are viewing a single comment's thread
view the rest of the comments
[-] saucerwizard@awful.systems 2 points 1 year ago
[-] TinyTimmyTokyo@awful.systems 4 points 1 year ago

That reminds me. If the world is about to FOOM into a kill-all-humans doomscape, why is he wasting time worrying about seed oils?

A lot of rationalism is just an intense fear of death. Simulation hypothesis? Means that maybe you can live forever if you're lucky. Superintelligence? Means that your robot god might grant you immortality someday. Cryogenics? Means that there's some microscopic chance that even if you pass away you could be revived in the future at some point. Long terminism? Nothing besides maybe someday possibly making me immortal could possibly matter.

I mean don't get me wrong I'd give a lot for immortality, but I try to uhh... stay grounded in reality.

[-] bobthomson70@mastodon.social 3 points 1 year ago

@sailor_sega_saturn @TinyTimmyTokyo Been thinking and saying this for a while. These powerful billionaire types are terrified of death as it’s so egalitarian - nobody escapes it; No matter how much money and power they accumulate, they can’t get control over this one thing and it drives them up the wall.

load more comments (38 replies)
load more comments (38 replies)
load more comments (38 replies)
this post was submitted on 21 Aug 2023
4 points (100.0% liked)

SneerClub

968 readers
3 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 1 year ago
MODERATORS