Well no one can prove they have a mind to anyone other than themselves.
And to extend that, there’s obviously a way for electrical information processing to give rise to consciousness. And no one knows how that could be possible.
Meaning something like a true, alien AI would probably conclude that we are not conscious and instead are just very intelligent meat computers.
So, while there’s no reason to believe that current AI models could result in consciousness, no one can prove the opposite either.
I think the argument currently boils down to, “we understand how AI models work, but we don’t understand how our minds work. Therefore, ???, and so no consciousness for AI”
Consciousness seems to arise from a need for sense of time and space. Navigation basically. Finding home, finding food, finding mates. I say this after decades of exploring mind altering substances and going on a decade of nearly daily meditation practice.
A friend of mine suggested it’s just this simple and that even worms are conscious. They’re aware of themselves to some degree and the when and the where. I’m sure they experience things way different than us, having different senses for assessing when and where and a different neuro structure for processing information from their bodies and the environment.
So, no point beyond consciousness being more common than I think people assume and actually not that difficult to define.
Consciousness is the sense of time and space. And most animals seem to have it. Do machines? I don’t know enough about the technology to have an educated opinion.
Mammalian intelligence is based on repurposing spatial mapping circuitry but that’s not consciousness itself, that is, the Miller Number: 7±2 things we can keep in conscious at the same time. That sense of time of space has a very specific quality to its qualias, they’re all, well, spacious. That thing as “just the room, no map in it” is also part of the Buddhist Jhanas (“boundless empty space”), but there’s plenty of stuff going on in the mind that isn’t part of that – say, the pure impression of “bright” when your SO dares open the window blinds does not have a navigational “bright from the window which is in that direction” to it, that’s an additional layer, a where, on top of the primitive what.
My best inference is that the function of consciousness is to flexibly make connections between different parts of the whole, and that on the level of learning / writing memory instead of automatic response: It is, in fact, possible to avoid running into a lamppost without being conscious of it, been there, done that, the let’s call it motor cortex first acting and then making me conscious, as if to say “have I been a good boy?”. That is it’s actually a quite passive process, being thrown left and right by systems wanting to do some connection, and shouldn’t be equated with will at all.
If you’re familiar with Buddhism then you’re familiar with the six and eight consciousness models?
Like in your lamppost example, I would argue part of you (body and eye consciousness) were quite conscious of the lamppost even if the consciousness mind was paying more attention to something else. Keeping as much of the senses (including sense of mind) in mind as you’re able to based on the depth of your practice and guarding against distractions away from what is happening now, is mindfulness.
In the eighth consciousness model, again in your lamp post example, we could say the seventh consciousness was occupied chasing after the past or future and mindfulness was barely present. Thankfully your other consciousnesses reacted and kept you safe. Manas becomes aware of this after the fact because its nature is ignorance.
The eighth consciousness is the base. The root. It’s more fundamental than I making. Which is probably what you were doing when you nearly walked into something. Thinking about what you’re doing later. I should do some laundry when I get home, maybe?
People mistake sense of agency (I making, manas, ego) with the base of consciousness. But consciousness is effortless and grasping at me and mine takes effort, its just more subtle effort than most people are aware of. When this grasping stops, awareness continues. In my personal experience.
So, I think it’s possible machines are conscious. If they have a sense of agency maybe the question Western science and the media keep asking. Maybe they just don’t have the models or personal experience to delineate between ego and consciousness. The people asking I mean. Hence the we don’t even know what consciousness is bit I keep hearing. Maybe not Western science. But human beings have been exploring these questions with the tools of Buddhist practice for 2500 years. I trust their definitions and they passed my own smell test.
I would argue part of you (body and eye consciousness) were quite conscious of the lamppost even if the consciousness mind was paying more attention to something else.
That’s semantics. My major objection to that kind of definition is that it knows no bound and distinction: Where do you stop assuming consciousness? Electrons are reacting to, influencing, and interacting with other electrons, is that also a form of consciousness? One could say so, but then everything is conscious which is the same thing as saying as nothing is conscious because without anything to delineate, terms are meaningless. I prefer language such as that what you call “body and eye consciousness” has agentive properties, that it can learn, that it generally wants to cooperate and be of service to the whole, such things. Lumping it up with consciousness risks confusing interpretations of messages of the thing (which is all we’re ever conscious of) for the thing itself.
guarding against distractions away from what is happening now, is mindfulness.
What was happening then is that I was using the way from home to the supermarket to think about code, with ample trust in mind so that I did not fear the lamppost. What good would have keeping my consciousness on the external world have done? The body/eye did not need integrative oversight, while my modelling mind very much could use a helping hand. Imposing it on the former and denying it to the latter would’ve been inflicting violence on myself.
Be careful to not moralise around “distraction”. Bluntly said when your teacher chided you for day-dreaming you probably weren’t distracted you were thinking about something more pertinent to your immediate development than calculus. Where discipline in directing consciousness comes into play is keeping your mind free from neurosis, within parameters in which you use your faculties according to their nature, as well as self-conditioning, e.g. if you’re addicted to potato chips, make sure that a) you don’t deny yourself potato chips and b) eat. every. potato. chip. with. full. consciousness. That’s to connect the act of eating up those chips to all the negative opinions you have about your behaviour, instead of it being only connected to something maladaptive. Scientifically proven and neurologically explained that and how that works, btw. In that sense “distraction” is “false, incomplete, sense of comfort”.
Also my last post was purely in regards to the first part of yours. I appreciate the insight into moralizing distraction and will retead it when I’m not distracted by the meat of our interesting conversation.
So mind consciousness trusted body/eye consciousness. I know what you mean, I dance and do this to enter flow state.
In the early Buddhist model consciousness would be the aggregate of the six sense consciousness. In the eight consciousness model the seventh consciousness might identify more strongly with one of these six, generally mind.
The store consciousness is the aggregate of all eight and that’s what I’m arguing is fundamentally what all experience arises from. The perception of emptiness, i.e. no self (consciousness itself is an aggregate and can’t be separated from its objects) and impermanence (change or time). Sense of time and space. To be conscious is to be aware of something. Movement through electrical synapses stimulated by sense impressions, even just the impression of sound from our own thoughts or the impression of limitless space in the fifth jhana.
I understand your objections to assumptions matter could be conscious based on this model. I think it would be inaccurat because not all matter has six sense bases and the storehouse is itself an aggregate.
But we are matter, and we’re conscious, so the fundamental conditions are there in some simple form. The movement of electrons as you stated.
But fire is fire when it’s fire, and ash when it’s ash. Even if the potential is there we don’t say fire is already ash when it’s not.
I think it would be inaccurate because not all matter has six sense bases and the storehouse is itself an aggregate.
The first five are basically one, in the sense that a blind or deaf person is not fundamentally less of a human than the rest of us. The model also misses some stuff, e.g. mere touch doesn’t include proprioception or sense of balance and if you read it as if it did (“body sense”) then why distinguish touch from e.g. sense of taste. Seventh I’d say is a subsystem (and so pervasive that the Stoics allow for both preferred and unpreferred indifferents – yes you can prefer pudding over gruel or the other way round just don’t think it’s a virtue), eighth is a stage of development, what you get when everything aligns well. The impression of a well-lubed machine.
I understand your objections to assumptions matter could be conscious based on this model. I think it would be inaccurate because not all matter has six sense bases and the storehouse is itself an aggregate.
I generally have no real idea of where to put the line. This stuff here might help, anything less than a T3 system can’t have experience of mind (they can’t learn to learn, which requires feeding back information about the changes in mind (for lack of better term) into the mind), OTOH that doesn’t mean that all T3 systems are actually integrating different sources, or balancing them: If you only ever were conscious of one aspect, there could be no conflict or interaction with another aspect, and thus consciousness would serve no role (and not evolve in the first place). It’s a matter of a required number of subsystems needing coordinating, and that coordinating itself having a necessary level of adaptiveness, be T3. Also I can authoritatively say that the human mind is not made to think about that kind of stuff. It’s all maps and models, direct knowledge fails I’m not sure the territory can even understand the question. Look, a squirrel!
In the case of the deaf blind person, and this is an aside, I don’t believe we’re born a blank slate in this regard even if the physical eyes don’t see. Helen Keller described actively in what I would call eye consciousness. Closed eye visual space/dream space. I’m sure there’s variance here depending on the nature of a person’s disability. I.e. neurology vs physiology.
I’m interested in the link and will read it. I’m only an amateur when it comes to coding and a layperson with AI.
And yeah, I don’t think it’s something we can put in a box. The map isn’t the territory and at some point describing consciencness from within using concepts (thinking, with language or otherwise) turns into a dog chasing its own tail. Thanks for the reminder not to bite myself in the chasing (your comment earlier about subtle forms of self aggression).
Okay, I’ve read over the article and a few links to help my understanding.
I think in the yogacara/eight consciousness model, we’d say store consciousness is the t4 evolving consciousness that stores the collective and individual seeds. T3 would be discriminative awareness or volitional awareness, i.e. what I want (manas or the wisdom of equality in its enlightened state). T2 would be mind consciousness, which through skillful application we can find that well lubed machine you mentioned (aligning itself and manas with store consciousness through practice and deep looking, which resolves fear of death as we’re able to look beyond our individual lives). And T1 the sense gates with only reflexive awareness.
I appreciate this conversation btw and hear what you’re saying about the maps only being maps and how they miss some things like sense of balance.
So yeah, I don’t think machines are T3 systems. No sense of agency. A working space for learning (like mind consciousness) but not ‘self aware’.
In principle subsystems that aren’t awareness can also be T3 systems, I suspect that at least from the motor cortex, mine does seem to have gotten more effective at learning from moment to moment, meaning it learned how to learn better and that’s T3. At least I think it’s not just me learning to not micro-manage it as much, it’s very hard to be sure about any of this, too many intersecting possibilities.
From the cybernetic/information theory side we don’t really know how these kinds of systems work in the first place, we’re barely getting started understanding T2 systems. All the AI tech we have is basically ways to breed fruit flies to fly left or right when seeing certain patterns, with enough computing power thrown at it to look impressive. We already had that kind of tech in the 50s (first implementations 54 for genetic algorithms, 57 for the perceptron), of course less impressive.
Well no one can prove they have a mind to anyone other than themselves.
And to extend that, there’s obviously a way for electrical information processing to give rise to consciousness. And no one knows how that could be possible.
Meaning something like a true, alien AI would probably conclude that we are not conscious and instead are just very intelligent meat computers.
So, while there’s no reason to believe that current AI models could result in consciousness, no one can prove the opposite either.
I think the argument currently boils down to, “we understand how AI models work, but we don’t understand how our minds work. Therefore, ???, and so no consciousness for AI”
“No brain?”
“Oh, there’s a brain all right. It’s just that the brain is made out of meat! That’s what I’ve been trying to tell you.”
“So … what does the thinking?”
“You’re not understanding, are you? You’re refusing to deal with what I’m telling you. The brain does the thinking. The meat.”
“Thinking meat! You’re asking me to believe in thinking meat!”*
Consciousness seems to arise from a need for sense of time and space. Navigation basically. Finding home, finding food, finding mates. I say this after decades of exploring mind altering substances and going on a decade of nearly daily meditation practice.
A friend of mine suggested it’s just this simple and that even worms are conscious. They’re aware of themselves to some degree and the when and the where. I’m sure they experience things way different than us, having different senses for assessing when and where and a different neuro structure for processing information from their bodies and the environment.
So, no point beyond consciousness being more common than I think people assume and actually not that difficult to define.
Consciousness is the sense of time and space. And most animals seem to have it. Do machines? I don’t know enough about the technology to have an educated opinion.
Mammalian intelligence is based on repurposing spatial mapping circuitry but that’s not consciousness itself, that is, the Miller Number: 7±2 things we can keep in conscious at the same time. That sense of time of space has a very specific quality to its qualias, they’re all, well, spacious. That thing as “just the room, no map in it” is also part of the Buddhist Jhanas (“boundless empty space”), but there’s plenty of stuff going on in the mind that isn’t part of that – say, the pure impression of “bright” when your SO dares open the window blinds does not have a navigational “bright from the window which is in that direction” to it, that’s an additional layer, a where, on top of the primitive what.
My best inference is that the function of consciousness is to flexibly make connections between different parts of the whole, and that on the level of learning / writing memory instead of automatic response: It is, in fact, possible to avoid running into a lamppost without being conscious of it, been there, done that, the let’s call it motor cortex first acting and then making me conscious, as if to say “have I been a good boy?”. That is it’s actually a quite passive process, being thrown left and right by systems wanting to do some connection, and shouldn’t be equated with will at all.
If you’re familiar with Buddhism then you’re familiar with the six and eight consciousness models?
Like in your lamppost example, I would argue part of you (body and eye consciousness) were quite conscious of the lamppost even if the consciousness mind was paying more attention to something else. Keeping as much of the senses (including sense of mind) in mind as you’re able to based on the depth of your practice and guarding against distractions away from what is happening now, is mindfulness.
In the eighth consciousness model, again in your lamp post example, we could say the seventh consciousness was occupied chasing after the past or future and mindfulness was barely present. Thankfully your other consciousnesses reacted and kept you safe. Manas becomes aware of this after the fact because its nature is ignorance.
The eighth consciousness is the base. The root. It’s more fundamental than I making. Which is probably what you were doing when you nearly walked into something. Thinking about what you’re doing later. I should do some laundry when I get home, maybe?
People mistake sense of agency (I making, manas, ego) with the base of consciousness. But consciousness is effortless and grasping at me and mine takes effort, its just more subtle effort than most people are aware of. When this grasping stops, awareness continues. In my personal experience.
So, I think it’s possible machines are conscious. If they have a sense of agency maybe the question Western science and the media keep asking. Maybe they just don’t have the models or personal experience to delineate between ego and consciousness. The people asking I mean. Hence the we don’t even know what consciousness is bit I keep hearing. Maybe not Western science. But human beings have been exploring these questions with the tools of Buddhist practice for 2500 years. I trust their definitions and they passed my own smell test.
That’s semantics. My major objection to that kind of definition is that it knows no bound and distinction: Where do you stop assuming consciousness? Electrons are reacting to, influencing, and interacting with other electrons, is that also a form of consciousness? One could say so, but then everything is conscious which is the same thing as saying as nothing is conscious because without anything to delineate, terms are meaningless. I prefer language such as that what you call “body and eye consciousness” has agentive properties, that it can learn, that it generally wants to cooperate and be of service to the whole, such things. Lumping it up with consciousness risks confusing interpretations of messages of the thing (which is all we’re ever conscious of) for the thing itself.
What was happening then is that I was using the way from home to the supermarket to think about code, with ample trust in mind so that I did not fear the lamppost. What good would have keeping my consciousness on the external world have done? The body/eye did not need integrative oversight, while my modelling mind very much could use a helping hand. Imposing it on the former and denying it to the latter would’ve been inflicting violence on myself.
Be careful to not moralise around “distraction”. Bluntly said when your teacher chided you for day-dreaming you probably weren’t distracted you were thinking about something more pertinent to your immediate development than calculus. Where discipline in directing consciousness comes into play is keeping your mind free from neurosis, within parameters in which you use your faculties according to their nature, as well as self-conditioning, e.g. if you’re addicted to potato chips, make sure that a) you don’t deny yourself potato chips and b) eat. every. potato. chip. with. full. consciousness. That’s to connect the act of eating up those chips to all the negative opinions you have about your behaviour, instead of it being only connected to something maladaptive. Scientifically proven and neurologically explained that and how that works, btw. In that sense “distraction” is “false, incomplete, sense of comfort”.
Also my last post was purely in regards to the first part of yours. I appreciate the insight into moralizing distraction and will retead it when I’m not distracted by the meat of our interesting conversation.
Okay.
So mind consciousness trusted body/eye consciousness. I know what you mean, I dance and do this to enter flow state.
In the early Buddhist model consciousness would be the aggregate of the six sense consciousness. In the eight consciousness model the seventh consciousness might identify more strongly with one of these six, generally mind.
The store consciousness is the aggregate of all eight and that’s what I’m arguing is fundamentally what all experience arises from. The perception of emptiness, i.e. no self (consciousness itself is an aggregate and can’t be separated from its objects) and impermanence (change or time). Sense of time and space. To be conscious is to be aware of something. Movement through electrical synapses stimulated by sense impressions, even just the impression of sound from our own thoughts or the impression of limitless space in the fifth jhana.
I understand your objections to assumptions matter could be conscious based on this model. I think it would be inaccurat because not all matter has six sense bases and the storehouse is itself an aggregate.
But we are matter, and we’re conscious, so the fundamental conditions are there in some simple form. The movement of electrons as you stated.
But fire is fire when it’s fire, and ash when it’s ash. Even if the potential is there we don’t say fire is already ash when it’s not.
The first five are basically one, in the sense that a blind or deaf person is not fundamentally less of a human than the rest of us. The model also misses some stuff, e.g. mere touch doesn’t include proprioception or sense of balance and if you read it as if it did (“body sense”) then why distinguish touch from e.g. sense of taste. Seventh I’d say is a subsystem (and so pervasive that the Stoics allow for both preferred and unpreferred indifferents – yes you can prefer pudding over gruel or the other way round just don’t think it’s a virtue), eighth is a stage of development, what you get when everything aligns well. The impression of a well-lubed machine.
I generally have no real idea of where to put the line. This stuff here might help, anything less than a T3 system can’t have experience of mind (they can’t learn to learn, which requires feeding back information about the changes in mind (for lack of better term) into the mind), OTOH that doesn’t mean that all T3 systems are actually integrating different sources, or balancing them: If you only ever were conscious of one aspect, there could be no conflict or interaction with another aspect, and thus consciousness would serve no role (and not evolve in the first place). It’s a matter of a required number of subsystems needing coordinating, and that coordinating itself having a necessary level of adaptiveness, be T3. Also I can authoritatively say that the human mind is not made to think about that kind of stuff. It’s all maps and models, direct knowledge fails I’m not sure the territory can even understand the question. Look, a squirrel!
In the case of the deaf blind person, and this is an aside, I don’t believe we’re born a blank slate in this regard even if the physical eyes don’t see. Helen Keller described actively in what I would call eye consciousness. Closed eye visual space/dream space. I’m sure there’s variance here depending on the nature of a person’s disability. I.e. neurology vs physiology.
I’m interested in the link and will read it. I’m only an amateur when it comes to coding and a layperson with AI.
And yeah, I don’t think it’s something we can put in a box. The map isn’t the territory and at some point describing consciencness from within using concepts (thinking, with language or otherwise) turns into a dog chasing its own tail. Thanks for the reminder not to bite myself in the chasing (your comment earlier about subtle forms of self aggression).
Okay, I’ve read over the article and a few links to help my understanding.
I think in the yogacara/eight consciousness model, we’d say store consciousness is the t4 evolving consciousness that stores the collective and individual seeds. T3 would be discriminative awareness or volitional awareness, i.e. what I want (manas or the wisdom of equality in its enlightened state). T2 would be mind consciousness, which through skillful application we can find that well lubed machine you mentioned (aligning itself and manas with store consciousness through practice and deep looking, which resolves fear of death as we’re able to look beyond our individual lives). And T1 the sense gates with only reflexive awareness.
I appreciate this conversation btw and hear what you’re saying about the maps only being maps and how they miss some things like sense of balance.
So yeah, I don’t think machines are T3 systems. No sense of agency. A working space for learning (like mind consciousness) but not ‘self aware’.
In principle subsystems that aren’t awareness can also be T3 systems, I suspect that at least from the motor cortex, mine does seem to have gotten more effective at learning from moment to moment, meaning it learned how to learn better and that’s T3. At least I think it’s not just me learning to not micro-manage it as much, it’s very hard to be sure about any of this, too many intersecting possibilities.
From the cybernetic/information theory side we don’t really know how these kinds of systems work in the first place, we’re barely getting started understanding T2 systems. All the AI tech we have is basically ways to breed fruit flies to fly left or right when seeing certain patterns, with enough computing power thrown at it to look impressive. We already had that kind of tech in the 50s (first implementations 54 for genetic algorithms, 57 for the perceptron), of course less impressive.
deleted by creator