37
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 24 Aug 2024
37 points (63.9% liked)
Technology
59689 readers
1593 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
Science cannot say much about what it is to think since it doesn't understand the brain well enough ... and the day we can fully explained it, we will also be able to replicated it on computers.
Science can and does quantify what our brains do vs what an LLM does though. That's the point. That's why the brain knows when it's supplying wrong information or guessing but the LLM does not.
The LLM can provide wrong information. What it can't do is intentionally lie.
i agree with you that we are much better than LLMs in the fact we can verify our errors (and we can do much more things LLMs don't do).
Still i am happy to have access to their vast memory and i know where they fail most of times so i can work with them in a productive way.
The day we provide them (or DNNs) with "Will" is i think when they will become (more) dangerous.
Gotta use extra heavy quotes around "will"... Mixing up AI pseudo-science with fascist pseudo-science.
Nah fam. You're comparing this thing to yourself and other humans because to you it sounds like us. "Sounds like" is the phrasing conspiracy theorists use to try to make correlations regardless of whether they make sense or not without filling in any actual facts. Either you know and understand how they work and are researching this and have a valid rebuttal or you don't. But an assumption does not make anything you said fact. People who are researching this have already refuted this claim.
You don't even know what the word "will" would mean in this context, or that it would provide an LLM with the kind of consciousness required to be sapient. So can we stop arguing if you just admit that you like using LLM's so you have a bias and a poor understanding of what they are and aren't.
Wild pseudo-scientific generalization.
There are many many things that are fully explained but will never be replicated on computers. Eg. Any numerical problem bigger than a computer.