188
submitted 2 weeks ago by misk@sopuli.xyz to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] Mikina@programming.dev 81 points 2 weeks ago

Lol. We're as far away from getting to AGI as we were before the whole LLM craze. It's just glorified statistical text prediction, no matter how much data you throw at it, it will still just guess what's the next most likely letter/token based on what's before it, that can't even get it's facts straith without bullshitting.

If we ever get it, it won't be through LLMs.

I hope someone will finally mathematically prove that it's impossible with current algorithms, so we can finally be done with this bullshiting.

[-] SlopppyEngineer@lemmy.world 18 points 2 weeks ago

There are already a few papers about diminishing returns in LLM.

[-] feedum_sneedson@lemmy.world 9 points 2 weeks ago

I just tried Google Gemini and it would not stop making shit up, it was really disappointing.

[-] bitjunkie@lemmy.world 3 points 2 weeks ago

I'm not sure that not bullshitting should be a strict criterion of AGI if whether or not it's been achieved is gauged by its capacity to mimic human thought

[-] finitebanjo@lemmy.world 2 points 2 weeks ago

The LLM aren't bullshitting. They can't lie, because they have no concepts at all. To the machine, the words are all just numerical values with no meaning at all.

[-] 11111one11111@lemmy.world 0 points 2 weeks ago* (last edited 2 weeks ago)

Just for the sake of playing a stoner epiphany style of devils advocate: how does thst differ from how actual logical arguments are proven? Hell, why stop there. I mean there isn't a single thing in the universe that can't be broken down to a mathematical equation for physics or chemistry? I'm curious as to how different the processes are between a more advanced LLM or AGI model processing data is compares to a severe case savant memorizing libraries of books using their home made mathematical algorithms. I know it's a leap and I could be wrong but I thought I've heard that some of the rainmaker tier of savants actually process every experiences in a mathematical language.

Like I said in the beginning this is straight up bong rips philosophy and haven't looked up any of the shit I brought up.

I will say tho, I genuinely think the whole LLM shit is without a doubt one of the most amazing advances in technology since the internet. With that being said, I also agree that it has a niche where it will be isolated to being useful under. The problem is that everyone and their slutty mother investing in LLMs are using them for everything they are not useful for and we won't see any effective use of an AI services until all the current idiots realize they poured hundreds of millions of dollars into something that can't out perform any more independently than a 3 year old.

[-] finitebanjo@lemmy.world 0 points 2 weeks ago* (last edited 2 weeks ago)

First of all, I'm about to give the extreme dumbed down explanation, but there are actual academics covering this topic right now usually using keywords like AI "emergent behavior" and "overfitting". More specifically about how emergent behavior doesn't really exist in certain model archetypes and that overfitting increases accuracy but effectively makes it more robotic and useless. There are also studies of how humans think.

Anyways, human's don't assign numerical values to words and phrases for the purpose of making a statistical model of a response to a statistical model input.

Humans suck at math.

Humans store data in a much messier unorganized way, and retrieve it by tracing stacks of related concepts back to the root, or fail to memorize data altogether. The values are incredibly diverse and have many attributes to them. Humans do not hallucinate entire documentations up or describe company policies that don't exist to customers, because we understand the branching complexity and nuance to each individual word and phrase. For a human to describe procedures or creatures that do not exist we would have to be lying for some perceived benefit such as entertainment, unlike an LLM which meant that shit it said but just doesn't know any better. Just doesn't know, period.

Maybe an LLM could approach that at some scale if each word had it's own model with massive amounts more data, but given their diminishing returns displayed so far as we feed in more and more processing power that would take more money and electricity than has ever existed on earth. In fact, that aligns pretty well with OpenAI's statement that it could make an AGI if it had Trillions of Dollars to spend and years to spend it. (They're probably underestimating the costs by magnitudes).

[-] naught101@lemmy.world 0 points 1 week ago

emergent behavior doesn’t really exist in certain model archetypes

Hey, would you have a reference for this? I'd love to read it. Does it apply to deep neural nets? And/or recurrent NNs?

[-] finitebanjo@lemmy.world 1 points 1 week ago* (last edited 1 week ago)

There is this 2023 study from Stanford which states AI likely do not have emergent abilities LINK

And there is this 2020 study by.... OpenAI... which states the error rate is predictable based on 3 factors, that AI cannot cross below the line or approach 0 error rate without exponentially increasing costs several iterations beyond current models, lending to the idea that they're predictable to a fault LINK

There is another paper by DeepMind in 2022 that comes to the conclusion that even at infinite scales it can never approach below 1.69 irreducable error LINK

This all lends to the idea that AI lacks the same Emergent behavior in Human Language.

[-] naught101@lemmy.world 2 points 1 week ago
[-] GamingChairModel@lemmy.world 1 points 2 weeks ago

I hope someone will finally mathematically prove that it's impossible with current algorithms, so we can finally be done with this bullshiting.

They did! Here's a paper that proves basically that:

van Rooij, I., Guest, O., Adolfi, F. et al. Reclaiming AI as a Theoretical Tool for Cognitive Science. Comput Brain Behav 7, 616–636 (2024). https://doi.org/10.1007/s42113-024-00217-5

Basically it formalizes the proof that any black box algorithm that is trained on a finite universe of human outputs to prompts, and capable of taking in any finite input and puts out an output that seems plausibly human-like, is an NP-hard problem. And NP-hard problems of that scale are intractable, and can't be solved using the resources available in the universe, even with perfect/idealized algorithms that haven't yet been invented.

This isn't a proof that AI is impossible, just that the method to develop an AI will need more than just inferential learning from training data.

[-] Mikina@programming.dev 2 points 1 week ago

Thank you, it was an interesting read.

Unfortunately, as I was looking more into it, I've stumbled upon a paper that points out some key problems with the proof. I haven't looked into it more and tbh my expertise in formal math ends at vague memories from CS degree almost 10 years ago, but the points do seem to make sense.

https://arxiv.org/html/2411.06498v1

[-] naught101@lemmy.world 0 points 1 week ago

Doesn't that just say that AI will never be cheap? You can still brute force it, which is more or less how back propagation works.

I don't think "intelligence" needs to have a perfect "solution", it just needs to do things well enough to be useful. Which is how human intelligence developed, evolutionarily - it's absolutely not optimal.

[-] GamingChairModel@lemmy.world 2 points 1 week ago

You can still brute force it, which is more or less how back propagation works.

Intractable problems of that scale can't be brute forced because the brute force solution can't be run within the time scale of the universe, using the resources of the universe. If we're talking about maintaining all the computing power of humanity towards a solution and hoping to solve it before the sun expands to cover the earth in about 7.5 billion years, then it's not a real solution.

[-] naught101@lemmy.world 1 points 1 week ago

Yeah, maybe you're right. I don't known where the threshold is.

I wonder if the current computational feasibility will cap out improvment of current generation LLMs soon?

[-] TheFriar@lemm.ee 1 points 2 weeks ago

The only text predictor I want in my life is T9

[-] 7rokhym@lemmy.ca 0 points 2 weeks ago

Roger Penrose wrote a whole book on the topic in 1989. https://www.goodreads.com/book/show/179744.The_Emperor_s_New_Mind

His points are well thought out and argued, but my essential takeaway is that a series of switches is not ever going to create a sentient being. The idea is absurd to me, but for the people that disagree? They have no proof, just a religious furver, a fanaticism. Simply stated, they want to believe.

All this AI of today is the AI of the 1980s, just with more transistors than we could fathom back then, but the ideas are the same. After the massive surge from our technology finally catching up with 40-60 year old concepts and algorithms, most everything has been just adding much more data, generalizing models, and other tweaks.

What is a problem is the complete lack of scalability and massive energy consumption. Are we supposed to be drying our clothes at a specific our of the night, and join smart grids to reduce peak air conditioning, to scorn bitcoin because it uses too much electricity, but for an AI that generates images of people with 6 fingers and other mangled appendages, that bullshit anything it doesn't know, for that we need to build nuclear power plants everywhere. It's sickening really.

So no AGI anytime soon, but I am sure Altman has defined it as anything that can make his net worth 1 billion or more, no matter what he has to say or do.

[-] HawlSera@lemm.ee 1 points 1 week ago

Until you can see the human soul under a microscope, we can't make rocks into people.

[-] RoidingOldMan@lemmy.world 0 points 1 week ago

a series of switches is not ever going to create a sentient being

Is the goal to create a sentient being, or to create something that seems sentient? How would you even tell the difference (assuming it could pass any test a normal human could)?

[-] 7rokhym@lemmy.ca 1 points 1 week ago

Powering off a pile of switches is turning it off. Powering off a sentient being is killing it. Not to mention a million other issues it raises.

this post was submitted on 27 Dec 2024
188 points (94.8% liked)

Technology

60346 readers
941 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS