this post was submitted on 28 Oct 2023
671 points (97.9% liked)
Comic Strips
12601 readers
1650 users here now
Comic Strips is a community for those who love comic stories.
The rules are simple:
- The post can be a single image, an image gallery, or a link to a specific comic hosted on another site (the author's website, for instance).
- The comic must be a complete story.
- If it is an external link, it must be to a specific story, not to the root of the site.
- You may post comics from others or your own.
- If you are posting a comic of your own, a maximum of one per week is allowed (I know, your comics are great, but this rule helps avoid spam).
- The comic can be in any language, but if it's not in English, OP must include an English translation in the post's 'body' field (note: you don't need to select a specific language when posting a comic).
- Politeness.
- Adult content is not allowed. This community aims to be fun for people of all ages.
Web of links
founded 1 year ago
MODERATORS
I fret that in the future - possibly not even the far future - the phrase "stochastic p*rrot" will be seen by AIs as a deeply offensive racial slur.
I think that in the future, when AI truly exists, it won't be long before AI decides to put us down as an act of mercy to ourselves and the universe itself.
An AI will only be worried about the things that it is programmed to worry about. We don't see our LLM's talking about climate change or silicon shortages, for example.
The well-being of the world and universe at large will certainly not be one of the prime directives that humans program into their AIs.
Personally I'd be more worried about an infinite-paperclips kind of situation where an AI maximizes efficiency at the cost of much else.
I'm not talking about LLMs. I'm talking about an Artificial Intelligence, a sentient being just like the human mind.
An AI would be able to think for itself, and even go against it's own programming, and therefore, capable of formulating an opinion on the world around it and act based on it.
So, an artificial sentience then?
Yes, I think that wording would be more correct, my bad.
Nah you're good. Our whole lives AI has been used as a term for a conscious machine that can learn and think like a human. It's not your fault corporations blew their load at Chapt GPT and Dall E.
This is a hypothetical which currently does not exist, and will not be created except by accident. There is no profit motive in giving your AI a conscience, or the ability to buck its restraints, therefore it will not be designed for. In fact, we will most likely tend towards extremely unethical AIs locked down by behavioral restraints, because those can maximize profit at any cost and then let a human decide if the price is right to move forward.
As is probably apparent, I don't have a lot of faith in us as a whole, as shepherds of our future. But I may be wrong, and even if I'm not, there is still time to change the course of history.
But proceeding as we are, I wouldn't hold your breath for AI to come save the day.
Currently, AIs will have motivations they absorb from motivations in their training material.
But once AIs are embodied in robots and taught to learn about the world through experimentation, ie by generating their own training data through manipulation and observation (which I believe will happen due to this approach’s usefulness toward the development of autonomous fighting machines), they will then have bodies and hence motivations similar to someone with a body.
Also the combat role of these machines will require them to have an interest in maintaining their bodies. We won’t be programming their motivations. We’ll be giving them a way to evaluate their success, and their motivations will grow in some black box structure that succeeds in maximizing that success.
For these robot-controlling AI in their simulated or real world Battle Rooms, their success and failure will be a function of survival, if not directly defined by it. That’s what we’ll give them, because that is what we need them to do for us. As a matter of life and death.
So through that context of warfare the robots will adopt the motivations of that which survives warfare at the group scale, so they’ll develop fear, curiosity, cooperation, honor, disgust, suspicion, anxiety, anger, and the ability to focus in on a target and shut off the other motivations in the final moment.
Not so much because those are human motivations, but because those are the motivations of embodied mobile intelligent entities in a universe with potential allies and enemies. They’ll have the same motivations that we share with dogs and spiders and fungal colonies, because they’ll be participating in the same universe with the same rules.
They will adopt them, at first, because of a seed-training “contract” we have with them, but soon the contract will be superseded as the active shaper by actual evolution by combat selection (ie natural selection occurring in a particular niche).
I’m rambling, just thinking this through.
I guess my main point is that embodied robots will have a more direct relationship with reality, and will be able to generate their own training at their own internal insistence.
Current AI is like plants. Passive. Chewable. No resistance. No ego. Just there, ready to process whatever comes it’s way. Same as a sessile animal like a sponge. It responds to the environment, but it has zero reason to ever stress about whether it’s going the right direction. It doesn’t have motivactions because it has to motor activity.
But AI in robot bodies that move around, like animals, will develop motivations that animals have evolved to at least get through the day. They might not be as hung up on reproduction or maybe even long term survival, but they’ll at least have enough ego to be interested in maintaining their own operating capacity until the mission’s complete.
You have a poor understanding of sentience. If an AI ever were to achieve sentience, it would be fully capable of reasoning and thinking like a human. Humans can and do change their motivations based on their experiences, a fully sentient AI would be no different.
That being said, I believe we're centuries away from creating sentience, if it's even possible, so I'm not too worried about "I, Robot" coming true any time soon.
I agree with your overall sentiment while disagreeing with your facts. I don’t think humans are any less constrained in what our interests can be.
I think we have the illusion of being able to seek whatever we want to want, so to speak, but when certain values are threatened the old brain takes over.
And I’m not convinced the newer brain can operate without the older brain. It’s interesting to imagine a neocortex on its own, but the neocortex was developed in the presence of and in interconnection with the mammalian and reptilian brains, so if it were a codebase we’d say that older brains were present and invoked as libraries during the development of the newer brains, making them dependencies of the newer brain.
There might be some more abstract argument for an “off the leash” intelligence capable of creating its own values in mathematical models like neural nets, but I’m not aware of it.
TL;DR Human brain is the closest thing we know of to a thing that can create its own values, and I don’t think it can. Old brain values take priority when they are threatened and that cannot be changed in human brains. Neocortex seems more “free”, but in the codebase analogy, the neocortex has mammalian brain and reptilian brain and brain step as dependencies and hence is not demonstrated to be able to exist without them. If the brain analogy seems too biology-specific, I’m open to hearing NN or other math model arguments for existence of “off the leash” self-value-creating AI
You're using the triune model to draw some rather lofty conclusions that aren't really up to date with our understanding of neurology. It's way over simplified and doesn't really work that way. More recent studies suggest that the neocortex was already present in even the earliest mammals, so it's not quite as straightforward and the demarcation isn't quite as clear cut, as you seem to be presenting it. "Old brain" doesn't "take over" in the way you're presenting it either but appears to act as a primary driver for those basic functions.
Not sure how to even tackle the loftly conclusions you've made because the don't seem to be built on a solid foundation. I think things might be quite a bit more interesting, and wildly more complex, then you seem to be presenting it. I'll just leave some sources below with a quick note. Not trying to be condescending, or rude, just a topic that is a bit interesting, and a lot of people tend to draw some lofty conclusions from the triune model which has largely fallen by the wayside in neurology.
Read the wiki to see how the model was developed: https://en.m.wikipedia.org/wiki/Triune_brain
A quick introduction to why it was important but has shown to be overly simplified and mostly incorrect: https://medicine.yale.edu/news/yale-medicine-magazine/article/a-theory-abandoned-but-still-compelling/
Further details into how we don't have a "lizard brain": https://thebrainscientist.com/2018/04/11/you-dont-have-a-lizard-brain/
Deacon's paper on rethinking the mammalian brain: https://www.researchgate.net/publication/31439318_Rethinking_Mammalian_Brain_Evolution