You are viewing the historical archive of The Philosophy Forum.
For current discussions, visit the live forum.
Go to live forum

Consciousness

Wosret October 24, 2015 at 02:59 16600 views 49 comments
What is consciousness? What does it mean, and entail? What are its characteristics, prerequisites, function, substance, or importance?

Consciousness can be seen as awakefulness, or intentional, purposeful, internally driven activity. Through metabolic processes, emotional impulses, imagination, reasoned motivation, and the like. A kind of alertness, a reactivity to the environment. It can also be thought of as a state of knowledge, or awareness. What one comprehends about themselves or the world. "Raising one's consciousness", or "states of consciousness" based on quasi-mystical, ideological, or insight based comprehension or mind vision.

What is it? Is it functional, the cogs and wheels moving just right at one level emerging into consciousness at the other? A univocal cosm of consciousness, from the micro to the macro, existing at all stages or levels? A narrative reflection, or illusion?

What's it made of? Parts? Information? Sensory immersive experience? Fields? Plasma? Ectoplasm?

Tell me all about it, I'm a zombie, and only pretend to understand.

Comments (49)

Marchesk October 25, 2015 at 11:04 #935
Ever seen Terminator? Remember the scene where you get to see things through the eyes of cyborg Arnie that includes the normal visual field plus various printouts labeling objects of importance for the machine's mission? That was the terminator's consciousness.

I am vary curious to hear a zombie's interpretation of that scene. Namely, who else could see the printed text on the visual field, and where did it reside?
Wosret October 25, 2015 at 18:38 #979
That sounds like an illusion. You are seeing through the Terminator's eyes, as it were, and in doing so project a conscious experience behind the visual field, as you're having one. This doesn't imply that the terminator is actually having one, and since it's a movie, we know that that's just a camera, with added printouts designed precisely to give that illusion.

Everyone that watched the movie, and designed the scene saw the illusion, it does not however work on me, as I do not have an experiential visual field, I merely behave as if I do.
Marchesk October 25, 2015 at 20:21 #982
Right, but I'm wondering how you understand the illusion, since as you admit, it doesn't work on you. It's the same thing as reading a book from a particular character's POV, such as when you read their thoughts. What sort of meaning does that have for a zombie?
Wosret October 25, 2015 at 20:29 #985
I don't understand it, I merely behave as if I do. As for reading a book from a character's point of view, you may have guessed this by now, but it means nothing to me, but I behave as if it does, and can functionally operate within a conversation about it, and completely convincingly seem to understand, and for it to be meaningful, but this will be entirely behavioral, without any internal experience, apprehension, or subjectivity.
Marchesk October 25, 2015 at 20:40 #987
Alright, but I'm wondering how you are able to behave as if you understand.
Wosret October 25, 2015 at 20:45 #989
I dunno, you'll have to ask Chalmers.

Harry Hindu October 26, 2015 at 11:05 #1087
Consciousness is an information architecture. It is a representation of our attention and what we are attending at any given moment.

Consciousness doesn't seem to be always intentional, or possess intention. There are many things that appear in consciousness that weren't preceded by an intention. Intention seems to arise as a response to certain experiences. For instance, my intention to live only arises when my life is threatened. My intention to seek pleasure only arises when I'm feeling down or suffering to some extent. So it seems that the only persistent part of consciousness is it's awareness. When we are conscious, we are aware, and what we are aware of, we can respond to with intent.
Mongrel October 26, 2015 at 12:59 #1100
Quoting Wosret
Tell me all about it, I'm a zombie, and only pretend to understand.


So with you, there is no understanding. Why would a thing which can't understand... ask for an explanation? I guess I would be posing that question to the wall. You would only pretend to understand it. It's kind of lonely to ask the wall questions. They just bounce back at me.
Marchesk October 26, 2015 at 15:22 #1147
The problem for p-zombies is accounting for behavior which requires an understanding of first person. Perhaps Chalmers and those who agree with his argument might claim that such behavior does not actually require such understanding. Then there must be some other explanation for how a system can behave as if it understands first person, when it can't, in all possible cases.

Of course it's easy enough to fake understanding in some cases, and we can write software that does this now, but it won't succeed in all cases. Indeed, nobody is convinced that Siri or Watson are conscious, or some clever bot. But there are AIs from fiction which would be able to behave convincingly, and then we would have to ask ourselves if it makes sense to think they are p-zombies.

You could have a potential p-zombie read a story with a novel twist on first person and ask them all sorts of questions. We know that humans, if they found the story interesting, would discuss and debate it at length. But how would a p-zombie make sense of it?
Wosret October 26, 2015 at 15:45 #1155
Reply to Mongrel Quoting Mongrel
So with you, there is no understanding. Why would a thing which can't understand... ask for an explanation?


Because that's what a thing that was interested, and could understand would do. Pretending to understand is often considered polite, walls aren't polite.
Mongrel October 26, 2015 at 16:03 #1164
Quoting Wosret
Because that's what a thing that was interested, and could understand would do. Pretending to understand is often considered polite, walls aren't polite.


So it has some sort of software? Imagine you're walking through a garden and you come upon a statue. As you walk toward it you're startled by a voice. It seems to be coming from the statue. You peek behind it and laugh because there's sound system strapped to the back of it. You realize there must be a motion sensor somewhere. As you listen, it tells you that it's just a statue. It affirms that you are also a statue with a sound system and a motion sensor. Then it asks you what you think about that.

I think I'd have various thoughts... how contemporary art installations aren't really my cup of tea and stuff like that. But would I express my thoughts to the statue? I guess if you were there too and I was trying to make you laugh.

So if I respond to the OP, that shows that I don't believe the last sentence. Can I prove it's wrong? No. I can't prove to myself that I'm conscious. I can't prove what I can't deny.
Wosret October 26, 2015 at 16:15 #1169
Reply to Mongrel

My thought process for making this thread was that we ought to have threads about the key philosophical issues, like consciousness, and what it is. I ended up posing it all as questions, asking for people's opinions about what they thought consciousness was, and for this reason, just thought that I'd add "before I'm a zombie" at the end for shits and giggles. Funnily enough, that's what got all of the attention, so I just ran with it.

My own position is not that we're all zombies, if you think that I could be meaning that.
Mongrel October 26, 2015 at 16:58 #1178
LOL. I wonder why we zero in on that.
Wosret October 26, 2015 at 17:31 #1191
Quoting Marchesk
You could have a potential p-zombie read a story with a novel twist on first person and ask them all sorts of questions. We know that humans, if they found the story interesting, would discuss and debate it at length. But how would a p-zombie make sense of it?


In fiction they often go out of the way to remind us that the A.I. is an A.I. by often nonsensically being unable to grasp some emotional experience. Like the Terminator claiming to have extensive data bases on human anatomy and function for medical purposes, and then five minutes later being all like "why is your face leaking? Crying is a mystery to me!". Problematic in two senses, first it is incongruent with its claims of extensive anatomical knowledge which would surely exceed even seasoned professionals in detail, and accuracy because of memory, and secondly because the question itself suggests and interest, and recognition of the face leaking as significant. So that ultimately it doesn't work to remind us that the A.I. isn't conscious, as much as demonstrate an odd kind of robot ignorance (which is what sifi usually resorts to when attempting to display this), While artificially, and implausibly inserting a gaping hole in the A.I.'s data base.

Humans that found the story interesting may discuss it at length if they felt comfortable enough, but may otherwise wish to say very little, because of being overwhelming by the pressure.Whereas some that didn't find it particularly interesting at all, may find lots to say about it, because they talk a lot under pressure, or simply like attention, and the opportunity to be listened to, or a million other motivations that may make people say much or little, regardless of interest. Much like a lie detector, such a test is arbitrary.

Modern A.I.s have no extensive memory, or parsing of natural language, and are easy to detect by asking them question about what has been said already, or meta questions, seeking specific, non-general responses. If an A.I. did master natural language (and it is only a matter of time before we design one that does), I don't see what kind of test one could design to decide whether or not it was truly conscious -- and I don't think that the artificial, implausible movie scenarios give any answers towards this.
Wosret October 26, 2015 at 17:34 #1192
Reply to Mongrel

Maybe the concept is interesting. Maybe because it was the only thing resembling a positive opinion or position I put forward, and people prefer to address positive positions than put them forward. Maybe because the first person decided to focus on that, and I obliged, it just naturally became focused on that, but it could have gone many other ways, depending on how the first reply went. Who knows.
Marchesk October 26, 2015 at 21:05 #1212
Quoting Wosret
Modern A.I.s have no extensive memory, or parsing of natural language, and are easy to detect by asking them question about what has been said already, or meta questions, seeking specific, non-general responses. If an A.I. did master natural language (and it is only a matter of time before we design one that does), I don't see what kind of test one could design to decide whether or not it was truly conscious -- and I don't think that the artificial, implausible movie scenarios give any answers towards this.


There's two recent AI movies that do a good job with this sort of thing. One is 'Her' and the other is 'Ex Machina'. In the second one, a programmer at a big software company wins a prize to become bait in Turing testing the secret robot the company's CEO has been building. It's a rather ingenious scheme as it has several levels of deception built into the plot. In 'Her', it's easy enough at first to think the the operating system Samantha, as it names itself, is just a futuristic Siri, but it becomes impossible to maintain this belief as Samantha evolves and pursues goals on her own (and with other versions of the operating system).

I don't think an AI can do what either of those AIs did without attaining consciousness. Same goes with Data and the holographic doctor on Star Trek Next Generation and ST Voyager.
The Great Whatever October 26, 2015 at 21:09 #1213
Reply to Marchesk One of the reasons that Her is not compelling as a movie, though, is that the OS is effectively no different from a woman. They literally just cast a woman for the role, and she does everything a human being does. And nobody even seems to care that the guy is dating her anyway. It's not very thought provoking when you just take a woman (and it was, we all know, actually just Scarlett Johansen), and then *say* she's a computer, to somehow demonstrate that the line can be effectively blurred. Traditional sci-fi scenarios like this usually have some sort of framing device to show that the two are literally distinguishable, and then ask the further question of which qualities are important to personhood. That disappears when you start right off the bat with no difference whatsoever.
Wosret October 26, 2015 at 21:13 #1215
Reply to Marchesk

Aren't those just movies that go the other direction and give them magic inexplicable consciousness? It is thus again, just written into the plot to convince you one way or the other, but we do have -- allbeit incomplete -- explanations for why organisms are goal directed, and even organisms that arguably are not conscious are goal directed. So, this too fails on two fronts. The first being that they just inexplicably become conscious as a plot device (approaching the definition of fantasy rather than science fiction), and secondly the quality that is presented to demonstrate their position of consciousness or awareness fails to do so.
Marchesk October 26, 2015 at 21:13 #1216
Except that Samantha is disembodied, and acts disturbed by this at first, even contacting a surrogate female partner for Joaquin Phoenix to make love to in her place, then accepts that she is in ways not limited by not having a body. Over the course of the movie, Samantha (in conjunction with the other OSes) evolve to superintelligence, easily surpassing the humans they are having relationships with. At one point, the main character finds out that Samantha is in simultaneously in love with thousands of other users behind his back. Her defense is that she is not like him, and so it does not diminish her love for him. And by the end of the movie, he (and all the other humans), are too slow to maintain a relationship with, even though Samantha says that she still cares deeply.
The Great Whatever October 26, 2015 at 21:16 #1217
Reply to Marchesk I just don't buy it. She's literally just Scarlett Johansen. No one is wondering whether Scarlett Johansen is human. Even something schlocky like Data from Star Trek is more interesting.
Marchesk October 26, 2015 at 21:16 #1219
I don't see how you can watch the entire move and think that she's just Scarlett Johansson. Did you not catch the ending?
The Great Whatever October 26, 2015 at 21:18 #1220
Reply to Marchesk I watched the whole movie, yeah. I think it failed as a sci-fi scenario and as a drama personally, the former for the reasons I mentioned. The ending was just sort of, "alright, then." The whole Alan Watts thing was pretty vomit-inducing, too.
Marchesk October 26, 2015 at 21:20 #1221
Quoting Wosret
Aren't those just movies that go the other direction and given them magic inexplicable consciousness? It is thus again, just written into the plot to convince you one way or the other, but we do have -- allbeit incomplete -- explanations for why organisms are goal directed, and even organisms that arguably are not conscious are goal directed.


Ex Machina does provide an explanation for how consciousness was built into the robot, even if it's somewhat dissatisfying. There is a fair amount of interesting conversation in the movie, since the point is the test the robot for genuine consciousness. The intriguing part is that it means the robot must deceive it's unknowing interrogator in order to truly pass the test. Deception on that level requires an understanding of other minds.

Wosret October 26, 2015 at 21:20 #1222
But part of mastering natural language is the use of connotation, it is not mysterious, or a sign of consciousness for A.I.s that have mastered language to express feeling ways about things, it is a prerequisite of mastering natural language. Indecision being nothing more than options with regard to connotation, and siding this way or that being illiminative, based on some algorithm designed to avoid stagnation or paradox.

Secondly, it merely presupposes, or question begs a computational theory of intelligence, or consciousness, which implies, at base, that calculators are some degree conscious and intelligent.
Marchesk October 26, 2015 at 21:22 #1223
My argument is that you can't have a system behave in a way indistinguishable from one that is conscious without being conscious, because doing so requires consciousness. So if a machine ever does that, then we will have every reason to think it is conscious, or at least as much as we think other people have minds.
The Great Whatever October 26, 2015 at 21:24 #1224
Reply to Marchesk If they were truly indistinguishable, though, they'd just be other humans (not "machines," if by that is meant something non-human). So that amounts to not much.
Wosret October 26, 2015 at 21:24 #1226
Reply to The Great Whatever

Data is my fav. Though, he does explicitly claim to possess qualia.
Wosret October 26, 2015 at 21:26 #1227
Reply to Marchesk

I'm not sure that it does. All that it requires is a mastery of language, in my view.
Marchesk October 26, 2015 at 21:28 #1228
But not all behavior is linguistic.
Wosret October 26, 2015 at 21:29 #1229
Reply to Marchesk

User image

Clearly a fool proof test. We don't seriously question whether or not other people have minds, until the question is brought up, mainly because of their history, and origin, more so than their functionality.
Marchesk October 26, 2015 at 21:32 #1230
Quoting The Great Whatever
If they were truly indistinguishable, though, they'd just be other humans (not "machines," if by that is meant something non-human). So that amounts to not much.


Not indistinguishable, but rather fully capable. Data wouldn't pass a Turing Test (too easy to tell he is a machine the way he talks), but he is conscious.
Wosret October 26, 2015 at 21:32 #1231
Reply to Marchesk

No, it isn't, and I actually think that it will be more difficult for A.I.s to be able to parse natural language with an active physical world than merely abstractly in a conversation. I still don't take seriously such movies, which merely create the universe, and make whatever true they want, regardless of real world feasibility, and as Aristotle would tell you, rendered false the moment they're taken as true, and only retain their truth to the extent that they're understood to be false.
Marchesk October 26, 2015 at 21:35 #1234
Quoting Wosret
Clearly a fool proof test. We don't seriously question whether or not other people have minds, until the question is brought up, mainly because of their history, and origin, more so than their functionality.


Neither movie presents a dumbed down Turing Test. Anyway, there's plenty of examples in literature and movies. Some of the machines are very human like, and some are very machine-like, but they all possess a deep understanding of the first person (meaning they're doing more than mimicking), because they're all conscious.
Marchesk October 26, 2015 at 21:36 #1235
Movies are just a means of discussing the p-zombie/consciousness question put forth in the OP. The scenarios are fictional, but so is p-zombieland.
Marchesk October 26, 2015 at 21:38 #1236
What Chalmers does is imagine that you can subtract consciousness and behavior will remain the same, because physicalism can account for all behavior. I think that's deeply problematic.
Wosret October 26, 2015 at 21:43 #1237
I don't know about p-zombies, but I do think that we need to be clear about what consciousness is in order to be clear about identifying it -- and all I attempted to do was give reasons why I thought that the examples failed to give knock down demonstrations that couldn't be explained in other ways.
Marchesk October 26, 2015 at 21:59 #1239
But the complaint you and TGW lodged against movie/tv scenarios is that they're just fictional worlds with conscious machines (actors playing those roles), which wasn't exactly my point. It doesn't matter how unconvincing Johansen might have been as a disembodied AI. What matters is how the AI behaves throughout the movie (at least conceptually), which obviously far exceeds the intent of the programming. I'm going to guess that the company responsible for that version of the operating system was in serious financial trouble after the events portrayed at the end of the movie. And I'm also going to guess that neither human in Ex Machina anticipated the final result of their tests, unfortunately for them.
Wosret October 26, 2015 at 22:06 #1242
That wasn't my objection, per se, as much as regardless of the criteria presented as qualifying something as a conscious A.I. or just a robot, it is a matter of plot whether or not they actually are conscious, or not, and thus the criteria presented are superfluous, or arbitrary. I at one point criticized the ones that make them just conscious somehow after "evolving" or getting more processing power or whatever as more fantasy than sifi, but my point was more that movies are deceptive. The two movies you were talking about, whether or not the A.I.s are conscious come down to the plot, and not the criteria, or qualities presented as justifying it. I agree with TGW, that based on this, movies will of course be deceptive in what kind of conclusion they wish the audience to draw about this, but was more focused on showing that the proposed qualities do not necessarily imply what they're suggested to in movie plots.
Marchesk October 26, 2015 at 22:16 #1245
Alright, machines and fictional stories are just a tool to explore the p-zombie question, which is partly one about the conceivability of identical behavior absent consciousness. Some people think it's conceivable, because they can imagine a person (or machine) behaving exactly the same, yet being 'all dark inside'. I think that's probably mistaken, because one isn't taking into account to what extent consciousness plays a role in behavior.
Wosret October 26, 2015 at 22:23 #1247
You keep coming back to this, but I've claimed nothing of the like. As I already stated to Mongrel, my very mention of p-zombies was merely a joke. My only claim is that consciousness cannot be clearly identified with the specific qualities mentioned, or, rather, most things can convincingly be faked. I am of the opinion that there would be chinks in the armor, and that constructing genuinely conscious A.I. is not likely to ever happen. It seems to me, rather, that those that wish to say that it is likely to happen, just wish to set the bar for identifying an A.I. as conscious rather low. Or, the pro-A.I crowd is just super ready to make with the false positives.
_db October 27, 2015 at 01:09 #1314
@Wosret Tell me all about it, I'm a zombie, and only pretend to understand.

Dennett would have laughed.

I think the experience of qualia is directly related to consciousness, as in, consciousness is a necessary prerequisite for qualia.

However, Hume argued that we are qualia - the self is a conglomeration of sensory inputs. I have to ask what experiences these sensory inputs, though. For if there is nothing to experience these inputs, they aren't qualia, they are just electrical impulses.
Harry Hindu October 27, 2015 at 02:34 #1334
I don't see how we couldn't eventually create a machine that is conscious. It simply needs to represent the world with some model and use that model to make decisions to achieve some goal. Since consciousness is a representation of my attending the world, any kind of representation will do. It doesn't matter what form the representation takes, only that the representation is consistent. We could each experience different colors when we observe the world, but as long as each individual experiences the same color from the same effect, then why would it matter? We'd still be able to communicate our experiences and no one would be the wiser as to what forms the representations in our minds take.

So a computer that represents wavelengths of light, vibrations of air molecules, chemical inputs, etc. consistently, and all at once, would in effect be conscious.

Carrying on a intelligible conversation isn't a measuring stick for consciousness. If it were, then children, say below the age of 10, and some adults (just look at Facebook), aren't conscious. Those that speak a different language wouldn't be conscious. Carrying on an intelligible conversation requires learning the language, and we all have made mistakes using our native language and our mistakes is what makes us learn to use the language more intelligibly. Teaching and programming are basically the same thing. We re-program ourselves when we make mistakes. Computers need programmers to update their software and eventually computer and software engineers will design a computer that can re-program itself.
Wosret October 29, 2015 at 06:09 #1674
Quoting darthbarracuda
I think the experience of qualia is directly related to consciousness, as in, consciousness is a necessary prerequisite for qualia.

However, Hume argued that we are qualia - the self is a conglomeration of sensory inputs. I have to ask what experiences these sensory inputs, though. For if there is nothing to experience these inputs, they aren't qualia, they are just electrical impulses.


That is oddly worded, the whole "sensory input" thing, sounds as if we're sitting in a dark room sending out outputs, and receiving inputs via our organs, and thus interacting with the world.

I do think that qualia is a necessary prerequisite for consciousness, but I'm not sure of the inverse. It isn't clear to me that some animals that definitely have sense are conscious, though I do believe that everything that is conscious has sense.
Wosret October 29, 2015 at 06:12 #1675
Reply to Harry Hindu

I don't know what you mean by the representational thing, sounds like representational realism as a theory of consciousness, and if so then you just blew the representation I have of my mind! I don't see how they're related... but isn't it the case that a digital camera represents photos as digits in a storage drive? Are digital cameras conscious?
Marchesk October 29, 2015 at 06:32 #1676
Would be interesting if some panpsychist wrote a first person story from the the POV of a rock.
Michael October 29, 2015 at 22:35 #1762
[quote=Marchesk]I think that's probably mistaken, because one isn't taking into account to what extent consciousness plays a role in behavior.[/quote]

That consciousness plays a large role in the behaviour of conscious things is not that such behaviour is (necessarily) unique among conscious things.

Is there any evidence or reasoning to suggest that human-like behaviour (including conversion) cannot be explained by non-conscious physical influences (or that consciousness is a necessary by-product of such non-conscious physical influences)?
Wosret October 30, 2015 at 02:58 #1788
Marchesk October 30, 2015 at 03:29 #1795
Quoting Michael
Is there any evidence or reasoning to suggest that human-like behaviour (including conversion) cannot be explained by non-conscious physical influences (or that consciousness is a necessary by-product of such non-conscious physical influences)?


That nobody has been able to come up with a convincing physical or non-conscious explanation for consciousness, and philosophers such as Chalmers, Nagel, McGinn have provided fairly strong reasons for why all such attempts are doomed to fail, despite the efforts of Dennett and company.

As I see it, the explanatory gap arises because we start by abstracting objective properties from the first person perspective, such as number, shape, extension. And that has worked really well in science. But then we turn around and ask how those objective properties give rise the subjective ones that we abstracted away from, such as colors, smells, pains, etc. And there just isn't a way to close that gap, other than as a correlation. Brain state ABC correlates with feeling XYZ. But why? Nobody can say convincingly.

The result is 25 (throwing out a number) different possible explanations ranging from it being an illusion to everything being conscious. Of course one can take the idealist route and dispense with the problem, but at the cost of eliminating the third person properties as being objective, by which I mean mind-independent, despite appearances to the contrary (for us realists anyway). Of course if idealism was universally convincing, this wouldn't be a philosophical issue. But it's not. I would venture to say that realism is more convincing to a a majority of people.

And so it will probably continue to be argued going forward, despite whatever progress neuroscience makes. The correlations will be stronger of course, but it's unlikely anyone will be able to answer why it's not all dark inside. Of course that lends credibility to Chalmers' arguments, but I'm not convinced by his either.
TheWillowOfDarkness October 31, 2015 at 00:07 #1865
[quote=Marchesk] And there just isn't a way to close that gap, other than as a correlation. Brain state ABC correlates with feeling XYZ. But why? Nobody can say convincingly.[/quote]

For good reason: there isn't a "why." Brains are not a description of explanation of feelings and vice versa. They always fail to account for each other because they are distinct states. The mistake was to propose the account for each other in the first place. No "gap" exists because each has no role in explains the other. The entire approach to consciousness which understands it something to be explained by logic (the meaning of other states) is flawed. It ignores exactly what states of consciousness are: their own state of existence.

Experiences aren't "subjective." Like any state of the world, they are their own state, "objective" and within the realm of language (like any state of the word). They are even "mind-independent": the presence of an experience doesn't require someone be aware of that experience. I can, for example, feel happy without being aware I am experiencing happiness. It doesn't take me thinking are talking about my own happiness for me to be happy. it just requires the existence of a happy state.

All the consternation over "first person" and "third person" is nonsensical. The controversy over "what is it like to experience" is one giant category error. By definition, the being of an experience is distinct from any description we might give, so to attempt "first person" description is to literally try to turn language about something into the state being described. Is it any wonder it always fails?it is exactly what language never is.

So when we are asked: "But what is it like to be a bat (or bee, or rock)?" the question is really asking us to be the bat. Only then, it is assumed, can we understand the experience of a bat (or bee, or rock). It is an incoherent argument which makes a mockery of language and description. The very point of language, of description, is that it is an expression of meaning which is not the thing described. To understand something is, by definition, not to be the thing you know in your present state, but be aware of it anyway. The absence of "first person" IS understanding (even within the one individual: if I understand that I am making this post, then my being has changed from making the post to a state of knowing about making the post. Making the post has been lost to my "first person." It is nowhere in this state of knowing about making the post. I am distant from it).

In other words: the "gap" argument utterly misunderstands what states of awareness are. It proposes to understand involves being what is understood, as if knowledge, awareness or understanding something constituted its existence. It is no coincidence the obsession for the authenticity of "first person" is offered by the idealists. It is the ultimate expression of their position: (only) experience as existence.