14
submitted 1 year ago* (last edited 1 year ago) by BrickedKeyboard@awful.systems to c/sneerclub@awful.systems

First, let me say that what broke me from the herd at lesswrong was specifically the calls for AI pauses. That somehow 'rationalists' are so certain advanced AI will kill everyone in the future (pDoom = 100%!) that they need to commit any violent act needed to stop AI from being developed.

The flaw here is that there's 8 billion people alive right now, and we don't actually know what the future is. There are ways better AI could help the people living now, possibly saving their lives, and essentially eliezer yudkowsky is saying "fuck em". This could only be worth it if you actually somehow knew trillions of people were going to exist, had a low future discount rate, and so on. This seems deeply flawed, and seems to be one of the points here.

But I do think advanced AI is possible. And while it may not be a mainstream take yet, it seems like the problems current AI can't solve, like robotics, continuous learning, module reuse - the things needed to reach a general level of capabilities and for AI to do many but not all human jobs - are near future. I can link deepmind papers with all of these, published in 2022 or 2023.

And if AI can be general and control robots, and since making robots is a task human technicians and other workers can do, this does mean a form of Singularity is possible. Maybe not the breathless utopia by Ray Kurzweil but a fuckton of robots.

So I was wondering what the people here generally think. There are "boomer" forums I know of where they also generally deny AI is possible anytime soon, claim GPT-n is a stochastic parrot, and make fun of tech bros as being hypesters who collect 300k to edit javascript and drive Teslas*.

I also have noticed that the whole rationalist schtick of "what is your probability" seems like asking for "joint probabilities", aka smoke a joint and give a probability.

Here's my questions:

  1. Before 2030, do you consider it more likely than not that current AI techniques will scale to human level in at least 25% of the domains that humans can do, to average human level.

  2. Do you consider it likely, before 2040, those domains will include robotics

  3. If AI systems can control robotics, do you believe a form of Singularity will happen. This means hard exponential growth of the number of robots, scaling past all industry on earth today by at least 1 order of magnitude, and off planet mining soon to follow. It does not necessarily mean anything else.

  4. Do you think that mass transition where most human jobs we have now will become replaced by AI systems before 2040 will happen

  5. Is AI system design an issue. I hate to say "alignment", because I think that's hopeless wankery by non software engineers, but given these will be robotic controlling advanced decision-making systems, will it require lots of methodical engineering by skilled engineers, with serious negative consequences when the work is sloppy?

*"epistemic status": I uh do work for a tech company, my job title is machine learning engineer, my girlfriend is much younger than me and sometimes fucks other dudes, and we have 2 Teslas..

you are viewing a single comment's thread
view the rest of the comments
[-] self@awful.systems 15 points 1 year ago

look I don’t want to shock you but that’s basically what they get paid to do. and (perverse) incentives apply - of course goog isn’t just going to spend a couple decabillion then go “oh shit, hmm, we’ve reached the limits of what this can do. okay everyone, pack it in, we’re done with this one!”, they’re gonna keep trying to milk it to make some of those decabillions back. and there’s plenty of useful suckers out there

a lot of corporations involved with AI are doing their damndest to damage our relationship with the scientific process by releasing as much fluff disguised as research as they can manage, and I really feel like it’s a trick they learned from watching cryptocurrency projects release an interminable amount of whitepapers (which, itself, damaged our relationship with and expectations from the engineering process)

[-] fasterandworse@awful.systems 8 points 1 year ago* (last edited 1 year ago)

As someone who went from high school directly into a publishing company as a “web designer” in 1998 I spent the next 20 years assuming that academic work was completely uninfluenced by commercial interests. HCI was academic, UX was commercial. Wasn’t till around 2019 that I started reading ACM papers about HCI from the 70s up. Fuck me was I surprised with how mixed up it all is. ACM interactions magazine published monthly case studies for Apple or did profiles on Jef Raskin talking about HCI for brand loyalty.

Anyway. Point is a published paper doesn’t mean shit if you just read a few because an article pointed you to them. I don’t know. This thread sucks

[-] TerribleMachines@awful.systems 7 points 1 year ago

Preach, as someone inside academia, the bullcrap is real. I very rarely read a paper that hasn't got a major stats issue—an academic paper is only worth something if you understand it enough to know how wrong it is or there's plenty of replication/related work building on it, ideally both. (And it's a technical field with an objective measure of truth but don't let my colleagues in humanities hear me say that—its not that their work is worthless, its just its not reliable.)

[-] froztbyte@awful.systems 7 points 1 year ago

"shitcoiners or oil companies.. who wore it best?"

but the rest of your reply reminds me that someone (I think steve or blake?) mentioned a thing here recently about a book on blaming guthenberg for this state of fucking everything up. I want to go read that, and I really need to get around to writing my rantpost about the "the problem of information transfer at scale is that scale is lossy, and this is why ..... [handwaves at many problems, examples continue]" thing that at least 8 friends of mine have had to put up with in DM over the last few years

[-] BrickedKeyboard@awful.systems -4 points 1 year ago

They also hyped autonomous cars and the Internet itself including streaming video for years before it was practical. Your filter of "it's all hype" only works 99 percent of the time.

[-] dgerard@awful.systems 15 points 1 year ago* (last edited 1 year ago)

autonomous cars aren't

look, there is no way on earth you didn't lose a fortune in crypto last year

this post was submitted on 01 Sep 2023
14 points (79.2% liked)

SneerClub

983 readers
1 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 1 year ago
MODERATORS