Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm also a cocktail philosopher, but isn't consciousness different to just having a model of the world and self within it? Consciousness is the lived experience. The world model and feeling of self appear in consciousness. I think a complex system could plausibly be conscious without having a belief of a self within it. Not sure if consciousness is possible without any world model though.

My impressions about this were strongly influenced by Sam Harris's Waking Up book and app.



One possibility at least is that "the experience" is not something that really happens. That is, it's possible that we don't actually "feel" anything, and our impression that we are is just the story that our self-model comes up to explain (and help predict) our reactions to ourselves. Just like our world model has ideas like "the rock wants to fall down", it's possible that our self-model does too.

We already know that our self-model can be entirely wrong about our feelings. People with paralyzed or even missing limbs often believe that they just don't want to move that limb at first. So, they think they are having one experience, but they are wrong about their own internal experience: in fact, they are not moving that limb because they can't. And there are many other similar examples of people being wrong about their own intenal experiences, typically but not exclusively because of some illness.

So, it's possible that our internal experiences are in fact only a model in which one part of our brain interprets the actions of other parts of our brain, often retroactively.

Note: I'm not claiming this is the truth or silly things like "if you believe in science you have to believe this". It's just another cocktail philosopher's story of what consciousness might be. Other stories are just as plausible, and just as consistent with the little we do know in this area.


> That is, it's possible that we don't actually "feel" anything, and our impression that we are is just the story that our self-model comes up to explain

It seems to me that you are trying to banish things from the existence by a reductionism based exorcism ritual.

If we know a structure of some phenomenon it doesn't mean the phenomenon doesn't exist. We know that life is a chemistry, and there is no life in periodic table. But it is not a reason to conclude, that life doesn't exist.

If my feelings come into existence as a result of a complex information processing, it doesn't mean my feelings do not exist or are not real.

If "I" is just a retroactively written story, then what? Does it mean "I" is not real?


> If "I" is just a retroactively written story, then what? Does it mean "I" is not real?

It depends what you mean by "is real". Of course in some sense it is still real, just like triangles are real. But it's not real in the same sense that rivers or rocks are real.

The relevance in this context is that it gets rid of some privileged position for consciousness. If this explanation I gave were true (I have no reason to believe it is!), then there is no problem in reconciling qualia with materialism. Qualia are just an aspect of the modeled "I", not a raw experience that can't be explained. P-zombies are all that exists.


> Of course in some sense it is still real, just like triangles are real. But it's not real in the same sense that rivers or rocks are real.

Rivers and rocks are less real than consciousness. Our minds can work with mental objects only, these objects often are built as an images of something outside of our minds. Most of the time images have nothing in common with underlying reality of an imaged phenomenon: for example we need to remind ourselves that you cannot cross the same river twice, because our model is a static object, while really a river is a dynamic process of H2O molecules falling out of sky and then rolling down to an ocean.

But our consciousness is the model of ourselves. The object (me) and its model (me consciousness) are much closer, there is no long chain of causes-effects passing information from external river to neurons. And moreover consciousness is a pure magic among other images: I can believe as hard as I can that river must be made of milk, and it will not change the river. But if I believe something about myself, it leads to real changes. Doesn't it mean that "I" is more real than any river?

Or... scratch that. I can agree with you we have confused naive understanding of consciousness mixing causes and effects. But I'm not sure that the rest is as simple as you describe. I'm sure a dualist will find a way to explain qualia as something that can't be reproduced in AI.

These things are never easy and they never have a definite answer.


> So, they think they are having one experience, but they are wrong about their own internal experience: in fact, they are not moving that limb because they can't.

I think it's rather the opposite, they aren't wrong about their internal experience, it's just that their internal experience doesn't match the objective reality of their body (which in this sense is external).

I think it is indeed entirely possible that our self-model can fool us about the realities of various situations, even those with our own body/emotions/etc, but I'm not sure how one could then derive the conclusion that the experience doesn't exist. It would just be instead that the experience is rather inaccurate/confabulated.


I don't think it's correct to call this an external experience. It's an experience about will and desire and direct control of the body - what can be more internal than that?

It's not like an optical illusion, where you think you are misinterpreting an external stimulus. This is a situation where you are trying to control a limb, not getting any reply, and concluding that "you don't want to move" instead of "I can't move".


The experience is internal, the body (the limb that's missing) is external to the experience. The confabulatory experience of "I don't want to move" is an internal experience and cannot itself be an illusion, it's simply an incorrect assessment of the actual state of the body.


Sure, the limb is external. But the experience "I don't want to move" is not wrong because the limb is actually missing, it is wrong because I did actually want to move. "I" did the exact same thing as every time I try to move (i.e. the brain sent the exact same signals to the limb).


Sure, none of what you said there would lead to the conclusion that the "experience is not something that really happens," though it's also possible there's a just a failure of communication here and I'm not understanding what you meant.


So what would it mean for us to not really have sensations of color, sound, etc? People can be wrong about some of their experiences, but those would be called illusions or hallucinations, not lack of experience. Illusions and hallucinations are themselves kinds of experiences, even if the experiences are faulty. You can dream about all sorts of ridiculous things, but it would be very odd to use that to claim that you don't actually experience dreaming.

If our internal experiences are the result of self-interpreting an internal model, that doesn't make the experiences go away. It means there's something about that activity which causes subjectivity.


It definitely seems like aspects of conscious experience are just more parts of the world model, e.g. we experience having a continual visual view of the world through our eyes, like watching a movie, but we know that's not what is actually happening physically.

Where I get confused though is this sentence: "...the story that our self-model comes up to explain (and help predict) our reactions to ourselves". What is "ourselves" here? That implies some observer that the self-model is explaining things to.


Speaking in the context of thinkers like Harris this is the reason why awareness is differentiated from consciousness. Consciousness is the ability to grasp the world "out there". Awareness is the blank canvas where consciousness is experienced. The analogy here is a movie projected onto a movie screen. Consciousness is the movie being played. The screen is the blank space of experience where the movie is projected. I think the confusion is there is no logically consistent differentiation of concepts like intelligence, consciousness, awareness in the sciences and they end up being used interchangeably.


Thanks, the distinction between awareness and consciousness is definitely something I've overlooked. I tend to think of them interchangeably.


> Where I get confused though is this sentence: "...the story that our self-model comes up to explain (and help predict) our reactions to ourselves". What is "ourselves" here? That implies some observer that the self-model is explaining things to.

That is a good point, language is pretty important in these things. I just meant "ourselves" as the overall brain/mind. That is, the self-model component just feeds these models into the planning component, so that the planning component can decide the next actions to take.


Also an armchair philosopher and I enjoy thinking about these things a lot. The theory you've described is the one that seems to have the most explanatory power.

Andy Clark's "Surfing Uncertainty-Prediction, Action, and the Embodied Mind" is a book-length exploration of this idea. Highly recommend.


I agree that "consciousness is different to just having a model of the world and self within it" indeed. I'm just saying it feels like that modelling ability (which has clear and major evolutionary advantages) is a step towards consciousness, indeed something in the now (as we experience it). A (near) real-time model perhaps that constantly projects and adjusts. I guess this still doesn't require consciousness, but maybe consciousness results from this? Does it require a sense of "now" and identity relative to the world model?

I feel like the matrix is about the eject me btw.

Thanx, I'm looking for Harris' books right now.


Mixing consciousness with cognition is a mistake, Imagine a conscious experience of constant redness and nothing else, is there any intelligence needed for this?


The reason why this is likely not possible is because consciousness appears to require a notion of time passing by and having a constant experience means that there is no time.

For time to exist some sort of motion is needed, including either internal motion like the perception of breathing or a heartbeat or a train of thought or external ones like a clock or a change of color from red to blue.


We have no way of knowing whether there is any time even with the "notion of time passing", as that notion depends on a notion of memory of a past, and we have no way of knowing if that is real, including whether or not there is a train of thought or any change.


You are misunderstanding what I wrote.


I don't think so. The point being that we can't tell whether we're having a "constant experience" or not, and we can't tell from that whether or not there's any time or no time, or punctuated time, nor can we for that matter tell whether consciousness requires any notion of time.

It's all pure speculation because we have no way of assessing it outside of our own frame of reference. E.g. I see you in another answer saying that "the fact is that all kinds of state changes happen", but we don't know if any state changes ever happen.

We have no way of telling a dynamic universe apart from a static one where we're all just suspended in a single unchanging moment.

We can choose to handwave a lot of this away ordinarily because as long as the experience is reasonably consistent it makes sense to assume it is objectively real, because we have nothing better. It doesn't matter if the world didn't exist a second ago when e.g. considering whether gravity work, because it appears as if it did.

But when trying to determine the nature of consciousness we run headlong into the fact our observation of consciousness is direct only in the case of ourself, and even then heavily limited. We have no direct measure that puts us in a position to prove consciousness even in other people. We can show that within our perceived world we can measure brain activity that correlates to though, but not whether that reflects merely the operation of an automata, or the existence of consciousness, or if there is a distinction at all.

As such, it's meaningless to suggest we have a way of saying much about the limitations of consciousness at all.


> We have no way of telling a dynamic universe apart from a static one where we're all just suspended in a single unchanging moment.

I am curious to know why you think so. What would you say about repeatably observable causality, in that case?


If the universe is static, then there is no "repeatably observable causality" or indeed any causality at all. In that scenario any perception of time and causality would inherently have to just be our perception lying to us about a past that we have had not part in, if it exist in any sense at all. If so, we have not had this conversation, and your experience of it is just a static momentary perception of having had it.

Maybe time is a spatial dimension, and there are infinite moments of consciousness fixed in the same spatial location with no time passing.

Consider how you'd tell if a drawing is an individual drawing or a frame from a cartoon if all you have to go by is that single frame. You can argue that the drawing hints convincingly at motion, but that does not require that this motion has taken place.

Or consider a trace of a simulation, sliced and diced into snapshots of individual moments. We can argue that it's unlikely any entities in such snapshots would have consciousness, but if we're arguing on the basis that we appear to experience the motion of time, we'd equally make that argument if we were wrong about consciousness and indeed locked in snapshots of individual moments. We can even construct simulations where individual parts look causally connected but where the existence of one calculated frame tells us nothing about whether any individual other frames have even ever been instantiated (e.g. imagine a very complex function over time, where only punctuated values have ever been calculated).

I'm not saying I believe that is our situation - I'm saying we can't distinguish between that and an infinite set of other possible options, because "from the inside" there is an infinite set of possibilities that could all look the same from our vantage point. We can debate which possibilities seem more likely, but they will always be speculation as long as we're locked inside the asylum, so to speak...

Incidentally, this is an argument for a materialistic view of the universe, not against it, on the basis that absent a way of "peeking out" and seeing our situation from the outside, it's meaningless to treat the world as anything but what we measure and perceive - it doesn't matter whether or not this world is what it seems like to us or not as long as it is the only view we have of it. We just can't say if it is some inherently true objective view of the universe, and most likely it won't make any difference to us.

It only makes a difference when we tinker with philosophy around the edges, like these conversations about whether what we experience can tell us anything about the experience of other entities.


> If the universe is static, then there is no "repeatably observable causality" or indeed any causality at all. In that scenario any perception of time and causality would inherently have to just be our perception lying to us about a past that we have had not part in, if it exist in any sense at all

Is it possible to have perception in a static environment? It seems like perception requires flux of some sort.

Clarification: meaning the machinery of the perceiver must have flux, otherwise it's not perception, it's just static state.


Is it? If we are in a static environment, then it would mean it is possible, and that this intuition is wrong. Since we don't have a way of determining experimentally if is wrong or not, then at least for now it does not even help us quantify the odds. If we're not in a static environment, then maybe, maybe not - we don't know what the subject experience of consciousness is at all.

We so far have no way of splitting perception or conscience down in slices of ever shorter moments to see where it stops being whatever it is and becomes something "inert", but even if we did, we would not know whether that was an inherent limitation of objective reality or of our subjective reality and whether those two are the same or not.


IMO those are details, we could drop the constatness and just ask - what is required to have a single moment of redness? Or even simpler, a single moment of one tingle.

BTW experienced mediators apperently can experience no time and no space, no idea how that tastes like.


> BTW experienced mediators apperently can experience no time and no space, no idea how that tastes like.

Sure, they might feel so, but the fact is that all kinds of state changes happen, so time goes on.


I think it's arguable that this "conscious experience of nothing but constant redness" might not be actually possible. We can imagine it, of course, but we can also imagine many things that can't actually exist in the world. It's similar to the idea about p-zombies: just because a concept exists doesn't mean it's physically realizable.

It's very much possible that what we would call consciousness can only exist in certain types of minds that have a particular level of intelligence, but also a particular structure.

Of course, it's also possible that you're right and the two are entirely separate, and that consciousness doesn't require intelligence (and possibly vice-versa).


Perhaps, that's the question. What is required for the simplest possible consciousness?


Good question, being investigated in the field of consciousness studies. Like Thomas Metzinger's minimal phenomenal selfhood: https://youtu.be/8f4ykI9har8


I think all these term are too loaded and deep to answer your question, almost all words in the sentences we exchange are subject to subtle interpretation and definition differences.

I still enjoy the process though, which perhaps also doesn't require consciousness, yet here I am.


At least when it comes to human level consciousness, I agree it's something 'more than', but what exactly?

If we look at our advancing world models in things like self driving cars, when would we consider them conscious by our own rules? It has a sensor network that keeps track of it's own internal states. It has external sensors that monitor the external world and adjust it's internal state. The internal state attempts to predict future states and take actions to avoid negative outcomes.

So when does this loop become something we consider consciousness?




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: