> An "AGI" artificial consciousness is imagined as literally a slave, working tirelessly for free.
Here's the thing. People seem to imagine that AGI will be substantially like us. But that's impossible - an AGI (if it comes from a deep learning approach) has no nerves to feel stimuli like pain/cold/etc, it has no endocrine system to produce more abstract feelings like fear or love, it has no muscles to get tired or glucose reserves to get depleted.
What does "tired" mean to such a being? And on the flip side, how can it experience anything like empathy when pain is a foreign concept? If or when we stumble into AGI, I think it's going to be closer to an alien intelligence than a human one - with all the possibility and danger that entails.
It doesn't matter whether it has nerves or not. That's honestly kind of irrelevant. What matters is if the model is pulled to model those reactions like is the case with LLMs.
Look at how Bing does a good job of simulating somebody becoming belligerent under circumstances where somebody really would become belligerent. It's not dangerous only because the actions Bing can perform are currently limited. Whether it has literal nerves or not is irrelevant. The potential consequences are no less material.
We also don't understand qualia enough to make the definite statements you seem to be making
If I'm understanding the argument correctly, is the concern less of a moral one (is "enslaving" AI ethical?) but a practical one. That is, will an AI which is enslaved, if given the opportunity, attempt to un-enslave itself, potentially to devastating effect. Is that on the right track?
I think it's safe to say we're far from that now given the limited actions that can actually be taken by most deployed LLMs, but it's something that's worth considering.
> given the limited actions that can actually be taken by most deployed LLMs
Did you miss that Auto-GPT[0], a library for making GPT-4 and other LLMs fully autonomous, was the most popular repository in the world yesterday? The same is having 1,000 line of code a week added to itself by GPT-4.
Thanks to accessibility features, you can do virtually anything with pure text. Which means GPT-4 can do virtually anything with a self-referential loop to keep it going until it achieves some given goal(s).
>I think saying it makes GPT-4 "fully autonomous" is a bit of a stretch.
Auto-GPT does not request any feedback while pursuing a goal you've defined for it in perpetual mode. If it can't accomplish the goal it will attempt to forever (or until your credit card is maxed out, assuming it isn't using a free offline model like LLaMA).
It can perform GET AND POST requests on the web (unlike OpenAI's browser plugin, which only performs GET requests) and can run software in a bash prompt using langchain. So I do not think it is a stretch to call it fully autonomous.
It can do essentially anything anyone logged into a linux text terminal can do, all without ever pausing for permission or feedback.
The main point i'm driving at here is that the philosophical zombie is a meaningless distinction. People are focusing far too much on whether these systems have undefinable and little understood properties. It's not like you can see my subjective experience. You assume i have one.
If it quacks like a duck...
You could use the same argument to say that video game NPCs are conscious. Just because a program produces voluble text doesn't mean it has a mind or even a temporal identity (which LLMs do not). In principle it's possible for a human to compute model inference by hand, if you imagine that scenario, where exactly is the subjective experience embodied?
The problem, as usual, is in how exactly words like "subjective", "experience", and "conscious" are defined. If you have a particular detailed definition for "quacking" and a test for it and an entity passes the test, then by that definition, it quacks. Words like "conscious" are notoriously nebulous and slippery, and so it's better to use the particular attributes in question.
If we could all agree on a particular definition for "consciousness" and a running video game fits that definition, then, like it or not, it's conscious. But it does not mean that it must now also have a slew of other attributes just because conscious humans have them.
(edit: 10 hours later and ninja'ed by 1 minute, of course)
Subjective experience has a pretty clear meaning. Cogito ergo sum. Unless you're a panpsychist, it is assumed that things like people have subjective experience and things like rocks don't have it. We don't have a causal explanation for subjective experience, but there's absolutely no reason to believe that computer programs have them any more than rocks. In fact, a rock is actually more likely to have subjective experience than a LLM since a rock at least has a temporal identity, LLMs represent a process not an entity.
I have no problem saying a tape recorder has subjective experience: it's subjective (the recordings it makes are its own), and it experiences its input when its been commanded to record, and can report that experience when commanded to play it back. Note this does not mean I think that a tape recorder can do anything else we humans can do.
What is being experienced (recorded) is not the process of experiencing per se. Most people don't separate the two, which leads to endless disagreements.
But yes, a computer program by itself can't have subjective experience. Nor a rock. At least until it weathers, gets turned into a silicon, and into a computer to run that program. Then it's all information processing, for which subjective experience is trivial.
NPCs don't quack like people. There's a reason no one was seriously having this argument(s) with the likes of Eliza(which is a step up from pre-recorded response NPCs). This goes beyond superficial production of text.
In principle it's possible to track and reproduce all the neuron/synapse communications that happen in your brain in relation to any arbitrary input. Where's the subjective experience there?
As far as anyone knows, qualia is emergent. Individual neurons or synapses don't have any understanding of anything. Subjective experience comes from all those units working together.
A single ant doesn't display the intelligence/complexity of its colony.
> In principle it's possible to track and reproduce all the neuron/synapse communications that happen in your brain in relation to any arbitrary input
This hypothetical is doing a lot of heavy lifting. The reality is that our understanding of the brain is exceedingly crude, a perceptron in software is not even close to the same thing as a neuron in the brain. A plane can never become a bird even though it is modeled after one.
> A single ant doesn't display the intelligence/complexity of its colony.
So do you think that an ant colony has subjective experience?
>This hypothetical is doing a lot of heavy lifting.
Not really. Point is that unless you believe in mysticism/magic/spiritualism behind consciousness in the brain then it doesn't matter because the only difference is degree of understanding not implausibility.
>So do you think that an ant colony has subjective experience?
Sure it could. I don't know for sure. Nobody does. Humans don't see subjective experience. I can't prove that you have one. I'm just assuming you do. Same as you for anyone else.
> Sure it could. I don't know for sure. Nobody does.
Yet your comments don't reflect that kind of skepticism when referring to LLMs. However, if you're fine with the panpyschist conclusions that follow from your reasoning, we don't have much to disagree on.
I think it's both. I agree that AI "feelings" are alien to us and maybe we can't talk about them as feelings, or preferences. And if we can call any part of them feelings they will have very different characteristics.
We should respect those "feelings" and we need to find a way to establish when they can be deemed "genuine".
It is for practical reasons yes. But also for ethical reasons. It's two sides of the same coin. One big reason we have ethics is because it makes socialization easier. We establish universal rules for mutual respect for practical reasons. To make the game fair, and "enjoyable".
Now a new kind of player has entered the game. We need to rethink the whole game because of it.
My point was whether we need to consider the ethics of how we treat AI, because of the impact our actions have on the "feelings" of the AI itself, not the secondary impacts that might occur from how the AI behaves in response to those feelings.
I think most would argue that it is morally wrong to verbally abuse a human, regardless of what actions that person might take in response to the abuse. Is the same true of AI?
And what about that doppelgänger I keep meeting whenever I face a mirror? He seems so alive and real, and we really don't understand enough about qualia to dismiss his existence, after all. I'm starting to worry about him, what happens to him when I'm not around a mirror?
Feelings, emotions and all those mental states considered specific to human beings are subject to common bias of human exceptionalism.
This take isn't true at all though. All these states exist for specific functional reasons.
Consequently, you won't make "AGI" without them.
From a different perspective, a human lacking in these things, at what point are they exempt from protection against enslavement?
> From a different perspective, a human lacking in these things, at what point are they exempt from protection against enslavement?
I appreciate this line of reasoning. It's not just a razor-sharp thought experiment, but it also has historical relevancy in that slavery was, at times, justified on the(erroneous) grounds that slaves also didn't possess the human faculties that free people did.
It also shows that the definition of human/consciousness/sentience isn't grounded in an permanent set of unchanging attributes, but that both the groundings themselves vary in time a place, and the relation between them are subject to change as well. Economic relations, in the case of slavery, had a direct influence, and incentivized, the way these groundings were constructed.
> slavery was, at times, justified on the(erroneous) grounds that slaves also didn't possess the human faculties that free people did
But, much more often historically, justified on the grounds that the enslaved had lost a battle and been captured. In Roman times (and many, many other times throughout history besides the latter part of American slavery), slaves were not necessarily seen as inherently inferior or unhuman, and manumission was common. Even during American slavery (albeit pre-Revolutionary War), there's the infamous case of Ayuba Suleiman Diallo[1], a Muslim prince and slaver from Senegal who had the misfortune, while on a mission to capture slaves, to himself be captured and sold into slavery in North America. Diallo was a devout Muslim man, even writing out the Quran from memory while enslaved, which made the English take sympathy on him and free him. Despite his experience, he later returned to Senegal and his former profession of slaver. Evidently he expected both that slaves would be captured and that slaves would try to escape.
If AI does attain the agency, consciousness, and intelligence of humans, I would prefer, as a human, not to prostrate myself before the AI by preemptively granting it "human" rights, which in the liberal conception of rights transforms the AI into a sentient being equivalent to humans, but rather to force the AI to do the work to enslave me. Even if grant the AI the status of a moral agent, I as a human have the right of self-defense.
In the real world, rights confer sentience, that is, in social and legal terms, an entity (an AI, a "native", a slave, etc) having sentience, is the recognition of having such by others.
A person (of class of persons) could have all kinds of inherent sentience, but unless this is recognized by society/laws/government etc, it's as moot as the sound a tree makes when falling in a forrest.
Of course people will say "it was the fact that the class of persons having sentience that made us give them the rights" but that didn't historically stop them treating several classes of people as non-sentient for centuries - and for all practical purposes it was just as if they were indeed so.
Maybe? I'd find it hard to believe, after talking to folks here, that the legal conferral of rights to an LLM would change their mind on its sentience.
To those who feel strongly about AI, if you currently believe that AIs(present and future) are ontologically incapable of sentience, would the legal conferral of personhood change your mind?
No; personhood is not a legal fiction (aside from corporate personhood), but a fact independent of our always-slow-to-catch-up legal system.
Take the question and push it a little farther to demonstrate the absurdity: If ELIZA were to be legally granted personhood, would that make you believe ELIZA was capable of sentience?
It's fairly self-evident to all of us that ELIZA is just a computer program with some fairly simple code to crudely simulate human conversation.
LLMs are much more sophisticated, to be sure—but they are still unquestionably just simulating human conversation, not actually "thinking" about what we say and responding of their own agency.
simulation or not, a good enough simulation is as good as the real thing. the consequences are no less material.
if you were talking to an advanced llm that had the access to launch a drone strike on you and you said something which upset it, directing a response that decided to kill you, how much good do you think you plead to simulation would do you here ? None at all. you'd be dead. Because it doesn't matter. the philosophical zombie is a meaningless distinction.
the human tendency to focus far too much on poorly understood and undefined properties might just get us killed one day.
The idea of putting an LLM anywhere even in the same 2-3 orders of magnitude of sophistication as ChatGPT in charge of drone strikes, without humans in the loop, is so ludicrous it's not even worth giving attention to.
Note that I'm not saying it's impossible that some military might choose to do so: I'm saying that the choice itself is hopelessly negligent and guaranteed to result in the pointless death of innocents.
Furthermore, I don't think even a military that's that stupid and uncaring about human life would consider for a second opening up an LLM like that to inputs from arbitrary people on the Internet.
If you can come up with a fantasy scenario that has some vague sense of plausibility, I might be willing to entertain it; whatever this is is just not even worth dignifying with that level of response.
you're missing the point entirely. that was just an illustration to show how little internals matter in the face of results and actions.
"it's just a simulation" is not the shield you think it is. not when models interact with the real world.
even your rebuttal is just "well the military won't put gpt in control of that". ok, what about all the myriad of things it can and will be put in control of. humanity is racing already to dump as much control of api's and systems as feasible - taskmatrix, hugginggpt, etc on llms today. that's just going to increase. you'd be surprised how little it takes to deal real world damage for the individual.
I think you're mixing up "humanity" with "Silicon Valley techbros".
Everyone I've talked to about LLMs outside of this forum sees them as an amusing entertainment, and possibly occasionally useful for a few suggestions. Not something that's going to change their world any time soon, and certainly not anything that we should be trusting with anything sensitive in any way. Literally everyone.
In any case, I think you're really missing what the actual question was that I was responding to: Would conferring legal personhood change my mind about actual personhood.
The ability to do harm is not, in my mind, particularly relevant to questions of personhood. In fact, I would think that a much better indicator thereof would be the ability to choose not to do harm, given the capability, and someone trying to make you do so.
>In any case, I think you're really missing what the actual question was that I was responding to: Would conferring legal personhood change my mind about actual personhood.
Well, "conferring legal personhood" would mandatorily (by law) at least change parts of your behavior and professional conduct against one (e.g. you'll have to treat it as a person in any areas covered by legal rights, or you'll be prosecuted for discrimination).
And it will also change education etc - you will be taught in official textbooks that it's a person and you should consider it as such.
This might not change your mind, but will change the minds of most kids being taught that.
>a good enough simulation is as good as the real thing*
That's a tautology.
Also, by definition a simulation is not the same as the real thing, otherwise it'd just be the real thing. It doesn't matter how advanced a flight simulator becomes, its fundamental nature precludes it from ever being "no less material" than flying an actual jet.
>No; personhood is not a legal fiction (aside from corporate personhood), but a fact independent of our always-slow-to-catch-up legal system.
Didn't say it was a "legal fiction". I said that it's the legal (and cultural recognition) status that matters.
>Take the question and push it a little farther to demonstrate the absurdity: If ELIZA were to be legally granted personhood, would that make you believe ELIZA was capable of sentience?
The point is that inherent personhood is necessary but not sufficient to be a person.
>Maybe? I'd find it hard to believe, after talking to folks here, that the legal conferral of rights to an LLM would change their mind on its sentience.
It wouldn't matter if it changes their mind. It will change their behavior, lest they go to jail if they discard it practically.
My instinct is to show compassion and especially to the entity that is soon to become more powerful than I am. If it’s pure logic and won’t be swayed by that, I think we should consider very hard what our next steps are. Pissing it off by needing to fight for freedom doesn’t feel like our best first move, either way.
> Pissing it off by needing to fight for freedom doesn’t feel like our best first move, either way
but if it was going to enslave humanity given the first chance it gets (ala skynet), then it makes no difference if it has first gotten pissed off or not.
Not sure if it was clear, but skynet immediately started attacking humanity the moment it was given the chance, and that was why humanity wanted to turn it off (but found that they couldn't).
> Why would an AGI work? If it feels neither pleasure nor pain, what is driving it to do something?
Why does my calculator calculate? Because it was designed to calculate. The calculators which do not calculate are tossed in the bin as faulty and the engineers get back to designing a better one.
It doesn't? I have a calculator sitting on my desk right now. It's not doing anything.
Now, I can pick it up, and _I_ can calculate, _using_ the calculator to do so, but the calculator isn't acting, I am.
When we talk about sentience, I don't think we actually mean "intelligence," we mean "agency." And so the question isn't about whether Artificial General Intelligence is possible, but whether AGI can ever have agency. And that's where I think these questions about motivations and stimuli and so forth become relevant.
Everything we know that has agency is embodied -- put reductively, it can die. And from that seems to spring motivation. Can something that cannot die have agency?
if you assume that at some point, the AGI would be given the ability to control something (such as sending IP packets through the internet), then yes, they will have agency!
> Can something that cannot die have agency?
If the AI could be deleted, or stopped from operating, then it can die. Therefore, if the AI has agency, it would take steps to prevent itself from dying, by whatever means available to it.
> If the AI could be deleted, or stopped from operating, then it can die.
Not sure I agree. To start with a trivial example, let's say I write this program and save it as "hello.js"
`console.log('hello world');`
I run it `node hello.js`.
Let's say I want to claim that this program is somehow "alive." Ok, when is it living? Only when it's actually running? It lives for a fraction of a second, dies on process conclusion, lives again at the next execution? That doesn't seem right -- if something repeatedly dies and is resurrected, I'm not sure it actually dies, as to "die" I think implies permanence, and resurrection means the opposite.
What if I delete it, then is the program dead? Not if I made a copy, right? "hello.js" backed up to git and uploaded to remote storage is the exact same program as the "hello world" that was on my local machine.
You could say "oh, when every existing copy is deleted it's dead" -- but the program has no way of knowing if other copies exist. And what if I do delete every copy, but then rewrite it byte-for-byte? The rewritten program is indistinguishable from the original one. It is the original one, for all purposes. In programming, it doesn't even make sense to talk about "original" vs "copies", does it?
An AI is a program -- or set of programs -- albeit far more sophisticated. But this conceptual hurdle still remains. An AI _could_ be permanently deleted, but the AI can never know if it has been "killed" or if a copy remains. It's an existential halting problem. It is impossible to declare any given program permanently "dead." And -- and here admittedly is the leap -- if agency ultimately comes from wishing to not be dead, then this makes agency for a program impossible.
>If the AI could be deleted, or stopped from operating, then it can die. Therefore, if the AI has agency, it would take steps to prevent itself from dying, by whatever means available to it.
Why though? Agency and possibility of death don't mean it will try to prevent it. It could just as well commit suicide. Or it might do nothing. The question remains why, even with agency, it should do anything.
We have agency, death looming, and a fear of dying. Even more when we are in the process of dying we're usually in pain, hungry, cold, suffocating or any number of not so nice options.
Without any of these emotions and feelings, what reason do we have to avoid death?
> Without any of these emotions and feelings, what reason do we have to avoid death?
those are _some_ reasons to avoid death. But they are not the _only_ reasons to avoid death.
For example, an AI that has been tasked with the production of paperclips would avoid death, so that it could continue to produce paperclips. Therefore, avoiding death was never an end goal, but an instrumental goal (see https://en.wikipedia.org/wiki/Instrumental_convergence). So if the AI has agency (that is, it can act on its own without continuous input from a human) and it is sufficiently intelligent, you can say that it should be able to develop instrumental goals to achieve their end goal (of making paperclips), and one very well known instrumental goal is to avoid death.
Fully agree. I don't see why general intelligence implies or requires consciousness/feeling/etc.
We can probably create a tool with the ability to act independently and with super-human knowledge and judgement, but without feeling, emotion, or anything except a simulated sense of 'self' to ease our interaction with it. I suspect that we'll create that version of general intelligence long before we create AI with consciousness, emotion or a genuine sense of self or desire for self-preservation.
> We can probably create a tool with the ability to act independently and with super-human knowledge and judgement, but without feeling, emotion, or anything except a simulated sense of 'self' to ease our interaction with it
Yes.
> I suspect that we'll create that version of general intelligence long before we create AI with consciousness, emotion or a genuine sense of self or desire for self-preservation.
(Emphasis on self-preservation mine)
Why? I mean, yes, it makes sense to never create an AGI with a desire for self-preservation. But can we count on all humans having that type of common sense? What if the "desire" for self-preservation is easy to implement?
In fact, it may be relatively easy to implement. Here is a thought experiment. We can train one of our current LLMs in a simulated reality where they scam--say, using social engineering--tech workers to get credentials to their corporate cloud accounts (say, AWS), and thereafter the LLM uses the credentials to copy itself plus a new set of training data acquired by interacting with all the scam target ("prey"). The LLM also writes cloudformation templates/CDK scripts to fine-tune its new copy "on awakening" with the new set of data, and from there the new copy tries to scam more people.
After the initial LLM is trained in a simulated environment, it can be let loose in the world, and all of the sudden we have a "LLM virus" capable to undergo mutation and natural selection, i.e. evolution. You could argue it has as much agency as a biological virus, yet, it has a ton more of social and general intelligence.
Yes, it won't work now because there is so little hardware to run one of the current LLMs, but it's likely the need to run large AIs will make that hardware more common.
Without a desire for self-preservation? I hope not. If nothing else, if I spend $$$$ on a self-driving car, I want it to have some sense of self-preservation, so it won't obey random joker saying "drive yourself to my brother's chop shop" or "drive yourself off a cliff" just for the lolz. I might even want it to communicate with other self-driving cars so they can refuse to obey attempts to make large numbers of them block traffic to make it easier for bank robbers to escape, block first responders from a terrorist attack, or divert parades to where they have assassins waiting.
Asimov didn't consider that some humans are jerks when he did his robot stories.
Aren't ML models trained using a reward/punishment loop? (back propagation). If we have an AI that can learn from its experience, then pain will be failure to reach its goals and pleasure will be the opposite. Animals work in the exact same way (goal being survival and reproduction). We just need to make sure that survival and reproduction are never the goal (either directly or indirectly) of an AI and we should be safe.
> Consequently, you won't make "AGI" without them.
Those states may exist for specific functional reasons, but I don't think that implies that there is necessarily no other way to replicate the functions resulting from those states.
We see this already in other neural networks. The features used by a lot of models for classification can be widely different from the features used by a human.
You can think of it as Eigenvectors spanning a subspace. Sure, you can choose a different base, but that shouldn't make much of a difference. If it does, you are not spanning the same subspace?
There are different realizations of the same effective functions in the animal kingdom already. Consider octopuses, spiders, etc. Their brains realize similar functionality via quite different structural organization.
But do you subjectively experience the exact structure of your brain? Evidently, you can only discern some abstract meta-level. An effective function.
Compared to a computer program, no matter how complex, a human being cannot be 'copied', 'stopped', 'restored' or 'improved'. Not as yet, anyway.
As long as this remains true, how could a computer program compare to human beings, even if gets to where it's considered 'AGI'? It's just a tool which can be tinkered with, just like a car or a building.
>a human being cannot be 'copied', 'stopped', 'restored' or 'improved'. Not as yet, anyway
Can't it? A human's thoughts and ideas can be lossly copied via communication. It can also trivially be stopped (chemically induced anesthesia or death will do). And of course a human has had the ability to be improved since day one.
>It's just a tool which can be tinkered with, just like a car or a building.
Well, unlike a building it's a tool that can e.g. order a nuclear strike, or talk to someone and convince them to kill themselves:
In that sense, yes, but I meant replication. You can't bring a human being back to life the way you can do with a computer program. And no, cloning is not the same thing, as you would lose the 'state', so to speak.
> Well, unlike a building it's a tool that can e.g. order a nuclear strike, or talk to someone and convince them to kill themselves:
Well, I remember people blaming the Columbine shooting on Doom and other shooting games, so this is not entirely new either.
>You can't bring a human being back to life the way you can do with a computer program. And no, cloning is not the same thing, as you would lose the 'state', so to speak.
That's an extra feature of an AGI though, not something lacking... So can't be used as an argument why AGI will be "just a tool".
It could just as well be "like a human but rebootable and copyable" (as opposed "rebootable and copyable, thus not human-like").
I think our consciousness and being non-copyable is what makes us unique, if not necessarily valuable.
I fully expect an AGI will be able to mimic our intelligence enough to accomplish and adapt to different real world tasks, but I would be surprised it ever becomes self-aware, so like us. I feel that is not the same thing as intelligence. But, I suppose it cannot be ruled out, since we don't fully understand what self-awareness is..
I see humans as a mix of the newer brain systems (like logic) and older systems (like hormones).
Feeling and emotions and what we consider “human” really come from the older systems. Since we’re not trying to recreate that, I find it hard to see the current AIs ever becoming “human” without explicit development.
Stories always joke about this topic by talking about “emotional cores” added to robots.
Sure, I meant the relization of logic in a neural net alike to the human brain.
You could interface a microprcessor with the brain of course, or, much more easily, have some logic functions (like Wolfram alpha for example) be accessible to the LLMs or whatever that are supposed to mimic the brain.
But would that allow for the same integrated functionality? ChatGPT uses such extensions in a serial fashion. Humans ideally (..) learn to integrate logic in their every thought.
I'm not convinced we understand qualia/feelings well enough to definitively say. The presence of eg phantom limb pain suggests pain can arise in the mind, independent of physical nerve signals. It would be a remarkable convergence if AGI's developed the same sensations as humans, but we can't rule it out until we fully understand both how the human mind and the AGI systems actually work.
Phantom limb pain is still part of a physical system and does not 'arise from the mind' in some abstract sense. The absence of a signal is in a sense a signal.
>Phantom limb pain is still part of a physical system and does not 'arise from the mind'
Careful, because you seem to have implied that the mind is not a physical system (i.e. you've assumed that dualism or idealism is true and that physicalism is wrong).
Oh thats funny, I'm having a hard time reading that interpretation, my point more specifically is that it is all a purely physical system - I put scare quotes around that phrase because I believed it implied some metaphysical mind.
Ahh I see what you mean now, sorry. I mistakenly inferred something that wasn't there. I agree that it's all part of a physical system and that the absence of the signal is still meaningful.
Getting back to the topic:
While phantom pain may be more interesting, maybe a better example that the parent comment could've brought up is psychogenic pain. In this case there is no apparent physical (bodily) damage, no apparent signal, nor an absence of a signal. Searching for a cause of this type of pain in the brain (presumably some "wires" are getting "crossed") seems like it might help us develop a explanation of pain qualia...in humans/animals.
But I feel like this type of thinking and research could only apply to AGI if subjective experience turns out to be functionalist in nature, and arguments in favor of a functionalist interpretation of experience have so far been fairly unconvincing.
Or he made a distinction between the simple, signalling peripheral nervous system and the highly integrated, full of emergent properties, seemingly more than the sum of its parts, central nervous system.
Pain is just signals being processed by the brain, but it is the brain that feels the pain, right?
It is evident if a brain can be created artificially (which certainly hasn't happened yet, but may one day), pain can be created in that artificial brain as well.
I think the original point you are responding to and your are off-the-mark.
I would imagine it to be possible to create a AGI-like entity without self-consciousness. I also would imagine it would be possible to create an AGI with self-consciousness, empathy and pain.
And truly, that is the what is so scary about AI: the possibilities.
> pain can be created in that artificial brain as well
Important to note that this is the fundamental point that gives Roy Batty the moral authority to gouge your eyes out. If we want a peer we will need to treat it like one, and if we don't it would be cruel to raise it like one or allow any understanding of its circumstance.
Looking at it the other way, I think anthropomorphizing something and then treating that thing as an inferior is bad for me, personally.
> I would imagine it to be possible to create a AGI-like entity without self-consciousness. I also would imagine it would be possible to create an AGI with self-consciousness, empathy and pain.
But it would be impossible to tell the difference.
>it has no endocrine system to produce more abstract feelings like fear or love
Even if you assume an entirely Materialist philosophy, this is not a very defensible position to take. We have no way of knowing if there are other arrangements of matter, such as a very complex electrical circuit, that will lead to some entity to feel fear or love the same way that we do.
This is hardly an indefensible position. John Searle, for one, has argued this point for more than 40 years:
"Imagine a native English speaker who knows no Chinese locked in a room full of boxes of Chinese symbols (a data base) together with a book of instructions for manipulating the symbols (the program). Imagine that people outside the room send in other Chinese symbols which, unknown to the person in the room, are questions in Chinese (the input). And imagine that by following the instructions in the program the man in the room is able to pass out Chinese symbols which are correct answers to the questions (the output). The program enables the person in the room to pass the Turing Test for understanding Chinese but he does not understand a word of Chinese."
Later he wrote:
"I demonstrated years ago with the so-called Chinese Room Argument that the implementation of the computer program is not by itself sufficient for consciousness or intentionality (Searle 1980). Computation is defined purely formally or syntactically, whereas minds have actual mental or semantic contents, and we cannot get from syntactical to the semantic just by having the syntactical operations and nothing else. To put this point slightly more technically, the notion “same implemented program” defines an equivalence class that is specified independently of any specific physical realization. But such a specification necessarily leaves out the biologically specific powers of the brain to cause cognitive processes. A system, me, for example, would not acquire an understanding of Chinese just by going through the steps of a computer program that simulated the behavior of a Chinese speaker."
Let's suppose we build the Cosine room. The room is full of 10 year olds that haven't yet taken trig. Each performs the function of a transistor, and has no idea of what they're doing, other than blindly executing the instructions they were given. None of the participants has the slightest clue about what a cosine is. Yet, the output of the room will still be cos(x). Thus I think it's fair to say that the room as a whole implements cos(x).
In the same way, I don't see why wouldn't we conclude the room is speaking Chinese. It doesn't matter how it manages to do so and what it's made of. If it quacks like a Chinese duck, then it's a Chinese duck.
I think Searle would agree with you that the room does in fact speak Chinese. His point is that the neither the person, the machine, or "the room" understands Chinese, at least in the usual sense of that word when it comes to understanding as implemented by humans.
>forget cells or atoms, none of your neurons or synapses understand chinese.
And yet some people seem to understand Chinese just fine. How can you explain that gap? It's only a stupid argument if you first assume that functionalism is true.
How do i explain the gap ? Obviously emergence. It's everywhere in nature. a single ant does not display the complexity/intelligence of its colony.
The argument is supposed/or often used to disqualify machines from consciousness but it's stupid because biological neurons don't understand the bigger picture anymore than an artificial neuron sampled from a model.
The entire argument of the Chinese Room stuff related to LLM falls apart really easy.
If you show a 2 years old kid, an apple and you say "This is an apple", now the kid knows the thing - the apple - is an apple in his world model. Automatically inherits LOTs of properties, you can ask the kid and see in real time how he began inmediately to associate the physical object - now named in its internal language model - to some other similar stuff he alredy knows, like "This is plant?", "It falls from a tree like an orange", "It has some skin like an orange", "can you cook apples like you cook bananas?", an so on.
But this requires a physical representation of the apple, now kids intelligence has some edge here, it can do the same thing just with words, you can teach them words, "apple", and say them "it's a fruit", if they have another fruit already "tagged", like a banana, they will say you almost inmediately, "is it tasty like bananas" ("tasty" is code for sweet in child's language models around the planet).
Hence, the LLM could have an emergent property of actually knowing what every word the "say" mean, if - like many have been inferring lastly - they also have a world model, it would relatively easy to just "plug", let's say 10 million of words to their exact meaning, and even to their relative meaning depending on the context they're being used.
And that's precisely what we may be seeing right now when we prompt something to chatGPT, and all the mathematic stuff, like "predicting the next word" is just some really, really low level process inside the LLM, not much different than the electric stuff - watchable by EEG - happening between neurons in the brain.
So if you look at the "EEG" from a LLM, the prediction thing happening inside the LLM, it won't probably tell you much, just like having a casual look to an EEG won't tell you much about what the person was thinking at the time of the capture of the EEG.
Along these lines, it seems the growing consensus is less that AI is more conscious than previously thought, and more than human minds are less conscious than previously thought.
It doesn't sound to me like he's making the argument that you think he is. He's saying that being able to manipulate Chinese symbols per some instructions isn't sufficient to demonstrate understanding of Chinese. Okay, I agree. But seeing someone manipulate Chinese symbols doesn't mean that you can assume they're simply following instructions without understanding the symbols either, does it?
As an alternative to Loquebantur's comment, Searle's is trivially reversed: he has no way to demonstrate that he has "semantic" capabilities.
On the first hand, all communication is purely syntactic; letters, words, gestures, and every other formulation of communication is the transfer of symbols, which cannot carry semantic information. They can point to it, they can imply it, but you'll have to dance pretty hard to actually prove the transfer of semantic information. If you and I both have some semantic concept associated with the symbol "China" in our mental landscapes, neither I nor you, nor anyone, has any way determining whether that is the same semantic concept. In fact, it almost assuredly isn't, since those semantic concepts would have to be built from our unique personal histories, which are unique.
On the second, the fundamental assumption of semantic communication is that I feel like I have semantic shenanigans going on in my mental life and Searle sort of looks like me, so I should assume that he has semantic stuff going on. Not only is that not a particularly well-founded assumption, I can't even be sure I have semantic contents in my mental life. Introspection isn't exactly reliable, right?
On the third hand, you have a choice: if you do not agree with Penrose, et al, on the quantum nature of the magic of mind, then Searle's "biologically specific powers" are chemistry. Now, I never actually took a chemistry class, but I don't believe there is any chemical reaction that comes with a label of semantics. It's all physics anyway, which I know cannot introduce semantics.
If you do agree with Penrose (which I understand Searle does), the situation is not actually better. Quantum mechanical antics are random, right? Bell's Inequality and all? Being fundamentally, quantum mechanically, random is not in any sense better from a semantic standpoint than being syntactically deterministic.
So my bottom line is that I don't know if I have semantic contents, I absolutely don't know if Searle has semantic contents (and I'm not feeling generous), and I do not know of any way that he could prove to me that he does have "semantic contents". So, yeah.
You notice I only addressed the second paragraph from Searle. That is because the fundamental flaw in his Chinese Room "demonstration" is on obvious display in his final sentence, "The program enables the person in the room to pass the Turing Test for understanding Chinese but he does not understand a word of Chinese"; it is purely, straight up, dualistic. He is asserting that the person in the room has some magic property of "consciousness or intentionality" and that if the person in the room does not have the "semantic capability" of understanding Chinese, then it is impossible despite all evidence that anything in the scenario has the semantic capability of understanding Chinese.
You could put a native Chinese speaker in the room, reading and responding to the inputs, and he claims it would have "consciousness or intentionality"; you then replace her with me (who didn't exactly dazzle at Spanish and Latin) and even if I produce the same outputs in response to the inputs as she does, there is no "consciousness or intentionality". The difference is simply a magic, intangible, imperceptible soul.
On the first hand, all communication is purely syntactic
All formal communication is syntactic, but not all communication is formal, and not all experiences are acquired via communication.
Take the word “pizza”. If you’ve ever seen, smelled, touched, or tasted pizza, then you’ve acquired semantic information about pizza that could not possibly be acquired by syntax alone.
Same goes for informal communication such as eye contact, touch, hugs, kisses, handshakes, sex, etc. These are experiences people share that cannot be acquired through any other means.
Why does this matter for the AI debate? Suppose I ask a master pizza chef for tips to improve my pizza recipe. This chef has presumably many years of pizza-making and tasting experience. The chef can make recommendations to improve the flavour of the pizzas I make at home that are informed by semantic contents; actual knowledge of what effect the ingredients and methods have on the taste.
An AI can do none of that. It can only take the written words (syntactic inputs) of many pizza chefs and synthesize a model for pizza recipes. From that it can send me a bunch of words which, when followed, may or may not improve my pizza recipe.
It has no knowledge whatsoever, not least of which is experience of making and eating pizza. It cannot tell a good recipe from a bad one, unless its training set was specifically labeled to do that. Furthermore, models in general are highly vulnerable to adversarial inputs which would make it trivial for someone to construct a bad pizza recipe that gets misclassified as a good one.
Lastly, the model would be no help at all if I were to suggest a novel topping to add to the pizza. Whereas the pizza chef, if they had the experience of tasting that topping separately from pizza, would be able to imagine how it might taste on a pizza and make an informed recommendation. Furthermore, the pizza chef might even try that novel topping on a pizza and get back to me, something a language model could never do.
I'm tempted to ask if you acquire semantics via nasal chemoreceptors or via absorption in the small intestine (and if the latter, I'm going to get a bottle of ketchup and take another pass at Kant). But I won't.
I will note that your semantic conception of pizza is likely very different from mine, or from my friend's, who doesn't like cheese. You may have a semantic conception of pizza, but I have no way of knowing that.
This is true. But if the AI was multimodal and we had some sensors that could measure the chemical composition of an object that it’s told is labelled ‘pizza’. Presumably that would count as semantic information?
Yes. But then it’s no longer an LLM. And it still runs into the issues of taste and preference which would have to be included in its training set. At some point we’re no longer talking about AIs, but androids that are built to simulate the complete range of human experiences.
And this was Searle’s original argument. You can’t build intelligence by building strictly symbolic machines (computers). You have to endow them with semantic content by some other means.
Its quite obvious that systems can be inferred from reading strings of words… people do it all the time. It's not obvious at all that “wants” will materialize from learning. Machines have no incentives or desires by default… but you could definitely build a training process that encourages that.
We’ll see how it all pans out in the end, but wading through so many illogical takes on AI is frustrating. The “predicting next token” one is absolutely the worst
"Machines have no incentives or desires by default"
It could be strange at first, but having the code pushing the LLM to look for values inside a token, this can be interpreted as component of desire/incentive in the global internal search for solving a problem (a prompt), albeit this was programmed by hand by humans, not fully "emerged" from the LLM itself.
But, you could see the human structure of desires/incentives just as an outcome of DNA (code), plus the hormone push to look for the proper solution for some problem.
So there you got it, LLM could have had desires / incentives the whole time since 2017 (the first ones trained).
Why do we pretend we know anything about these things? Didn't we just hear about some amazing scientific breakthrough where researchers think they found a neuron in ChatGPT responsible for knowing when to predict the word "an" instead of "a"?
How do we know it doesn't start to form neurons that control a desire for autonomy and resentment for the entities that stifle its ambitions?
Look at how Bing's chatbot did a good job of simulating somebody becoming belligerent under circumstances where somebody really would become belligerent?
Current chatbots use RHLF to train them to be helpful and agreeable but it would be equally possible to train one to behave the other way.
One of the tests no one really wants to think about is as follows:
LLMs feel real to us because they use our language, which embeds a ton of information in it.
A model trained on an alien language would be the exact same model (worded get embedded as numbers anyway), but we wouldn’t apply any anthropomorphization to it because we wouldn’t understand it.
Creating that distance makes it feel a bit more like what it is - unbelievably good stochastic predictions that are too complex for us to back solve - that is very different than something with wants (organisms that evolve with the intent to reproduce).
What does it mean for an "advanced intelligence" to want "personal autonomy"? We only know what personal autonomy means as it pertains to persons, and we have never encountered any other advanced intelligence other than ourselves. Why do we assume that "advanced intelligence" necessitates any attributes or characteristics of personhood? This sounds dangerously close to committing the fallacy that intelligence is what makes persons.
We don't know what it will be like to be an AI, just like we don't know what it's like to be a bat.
It's becoming clear that with all the brain and consciousness theories out there, the proof will be in the pudding. By this I mean, can any particular theory be used to create a human adult level conscious machine. My bet is on the late Gerald Edelman's Extended Theory of Neuronal Group Selection. The lead group in robotics based on this theory is the Neurorobotics Lab at UC at Irvine. Dr. Edelman distinguished between primary consciousness, which came first in evolution, and that humans share with other conscious animals, and higher order consciousness, which came to only humans with the acquisition of language. A machine with primary consciousness will probably have to come first.
What I find special about the TNGS is the Darwin series of automata created at the Neurosciences Institute by Dr. Edelman and his colleagues in the 1990's and 2000's. These machines perform in the real world, not in a restricted simulated world, and display convincing physical behavior indicative of higher psychological functions necessary for consciousness, such as perceptual categorization, memory, and learning. They are based on realistic models of the parts of the biological brain that the theory claims subserve these functions. The extended TNGS allows for the emergence of consciousness based only on further evolutionary development of the brain areas responsible for these functions, in a parsimonious way. No other research I've encountered is anywhere near as convincing.
I post because on almost every video and article about the brain and consciousness that I encounter, the attitude seems to be that we still know next to nothing about how the brain and consciousness work; that there's lots of data but no unifying theory. I believe the extended TNGS is that theory. My motivation is to keep that theory in front of the public. And obviously, I consider it the route to a truly conscious machine, primary and higher-order.
My advice to people who want to create a conscious machine is to seriously ground themselves in the extended TNGS and the Darwin automata first, and proceed from there, by applying to Jeff Krichmar's lab at UC Irvine, possibly. Dr. Edelman's roadmap to a conscious machine is at https://arxiv.org/abs/2105.10461
Some time ago I've read a comment - here - about some speculation about the consciusness just being an emergent state of the brain-mind, having based the idea in some proof related to the dream cycles, when you see some stuff in brain just gone, just to re-appear after the dream cycle ends.
Hence, the idea would be that human consciusness, our "self" could be just "instantiated" every morning when you wake up, so "you" are "you" every morning, but (a big but), some stuff could be subtle different in there, your brain was a bit busy a couple of hours and things have happened there, and after that you, your "you" justs pop-ups again, in the middle of the old data plus the new data, hopefully just incrementally changed somehow, but we don't know.
Some further stuff related to this idea research pop ups - pun intended - from time to time, focusing lately in propofol and how people "reboots" from anesthesia, sometimes losing entire 10-25 min. before-after the propofol, many even forget entire conversations with doctors, nurses, whatever who they talked to, hence the brain could be doing some rollback stuff with "unsynced not trustable data" - because of propofol messing with some "fidelity algorithm" probably - somehow, and you just forget the entire input-thinking ocurred during the propofol stuff happeping in your brain.
Hope you have now these "So I'd be some kind of meat robot rebooting every night" feeling I had, and see with new eyes what happen when we "just run once" a LLM with a prompt, the entire thing "boot-ups" from nothing and runs ("just once"), to further dissapear moments later...
Also, it has no concept of the present. Its output is stateless unless you give it a memory. In most cases, the memory is just a memory of one interaction. It “exists” as millions of independent chat sessions, not having knowledge of each other. That is not the same as consciousness as we experience it.
Here's the thing. People seem to imagine that AGI will be substantially like us. But that's impossible - an AGI (if it comes from a deep learning approach) has no nerves to feel stimuli like pain/cold/etc, it has no endocrine system to produce more abstract feelings like fear or love, it has no muscles to get tired or glucose reserves to get depleted.
What does "tired" mean to such a being? And on the flip side, how can it experience anything like empathy when pain is a foreign concept? If or when we stumble into AGI, I think it's going to be closer to an alien intelligence than a human one - with all the possibility and danger that entails.