This is a subject I’ve thought about a lot, so I’m pretty happy to have seen this post :).
I’m not convinced by counterfactual robustness either. For one, I don’t think humans are very robust either, since we rely on a relatively specific environment to live. And where to draw the line between robust and non-robust seems arbitrary.
Plus, whether a person is counterfactually robust can be changed without modifying them, and only by modifying their surroundings. For example, if you could perfectly predict a person’s actions, you could “trap” their environment, adding some hidden cameras that check that the person doesn’t deviate from your predictions, and triggers a bomb if they do deviate. Then that person is no longer counterfactually robust, since any slight change will trigger the bomb and destroy them. But we didn’t touch them at all, only some hidden surroundings!
---
I also suspect that we can’t just bite the bullet about consciousness and Turing machines appearing everywhere, since I think it would have anthropic implications that don’t match reality. Anthropic arguments are not on very solid footing, so I’m not totally confident about that, but nonetheless I think there’s probably just something we don’t understand yet.
I also think this absurdity you’ve noticed is an instance of a more general problem, since it applies to pretty much any emergent pattern. The same way you can find consciousness everywhere, you can find all sorts of Turing machines everywhere. So I view this as the problem of trying to characterize emergent phenomena.
---
Investigating causality was the lead I followed for a while as well, but every attempt I’ve made with it has ended up too strong, capable of seeing imaginary Turing machines everywhere. So lately I’ve been investigating the possibility that emergence might be about *information* in addition to causality.
One intuition I have for this is that the problem might happen because we add information in the process of pointing to the emergent phenomena. Given a bunch of particles randomly interacting with each other, you can probably point to a path of causality and make a correspondence to a person. But pointing out that path takes a lot of information which might only be present inside the pointer, so I think it’s possible that we’re effectively “sneaking in” the person via our pointer.
I often also use Conway’s Game of Life when I think about this issue. In the Game of Life, bits are often encoded as the presence or absence of a glider. This means that causality has to be able to travel the void of dead cells, so that the absence of a glider can be causal. This gives a pretty good argument that every cell has some causal effect on its neighbours, even dead ones.
But if we allow that, we can suddenly draw effectively arbitrary causal arrows inside a completely dead board! So I don’t think that can be right, either. My current lead for solving this is that the dead board has effectively no information; it’s trivial to write a proof that every future cell is also dead. On the other hand, for a complex board, proving its future state can be very difficult and might require simulating every step. This seems to point to a difference in *informational* content, even in two places where we have similar causal arrows.
So my suspicion is that random interactions inside walls might not contain the right information to encode a person. Unfortunately I don’t know much information theory yet, so my progress in figuring this out is slow.
The bomb trap example is very interesting! Can’t be counterfactually robust if you’re dead. Instead of bombs, we could also just use sudden overwhelming sensory inputs in the modality they’re fastest in to interrupt other processing. However, one objection could be that there exist some counterfactuals (for the same unmodified brain) where the person does what they’re supposed to. Objects we normally think of as unconscious don’t even have this weaker kind of counterfactual robustness: they need to be altered into different systems to do what they’re supposed to to be conscious.
But pointing out that path takes a lot of information which might only be present inside the pointer, so I think it’s possible that we’re effectively “sneaking in” the person via our pointer.
Interesting. Do you think if someone kept the mapping between the states and “firing” and “non-firing” neurons, and translated the events as they were happening (on paper, automatically on a computer, in their own heads), this would generate (further) consciousness?
I often also use Conway’s Game of Life when I think about this issue. In the Game of Life, bits are often encoded as the presence or absence of a glider. This means that causality has to be able to travel the void of dead cells, so that the absence of a glider can be causal. This gives a pretty good argument that every cell has some causal effect on its neighbours, even dead ones.
But if we allow that, we can suddenly draw effectively arbitrary causal arrows inside a completely dead board! So I don’t think that can be right, either.
Doesn’t counting the causal effects of dead cells on dead cells, especially on a totally dead board, bring us back counterfactual robustness, though?
To expand a bit on the OP, the way I’ve tentatively been thinking about causality as the basis for consciousness is more like active physical signalling than like full counterfactuals, to avoid counterfactual robustness (and counting static objects as conscious, but there are probably plenty of ways to avoid that). On this view, dead cells don’t send signals to other cells, and there’s no signalling in a dead board or a dead brain, so there’s no consciousness in them (at the cell-level) either. What I care about for a neuron (and I’m not sure how well this translates to Conway’s Game of Life) is whether it actually just received a signal, whether it actually just “fired”, and whether removing/killing it would have prevented it from sending a signal to another that it did actually send. In this way, its presence had to actually make a non-trivial difference compared just to the counterfactual where it’s manipulated to be gone/dead.
On your approach and examples, it does certainly seem like information is correlated with the stuff that matters in some way. It would be interesting to see this explored further. Have you found any similar theories in the literature?
This is a subject I’ve thought about a lot, so I’m pretty happy to have seen this post :).
I’m not convinced by counterfactual robustness either. For one, I don’t think humans are very robust either, since we rely on a relatively specific environment to live. And where to draw the line between robust and non-robust seems arbitrary.
Plus, whether a person is counterfactually robust can be changed without modifying them, and only by modifying their surroundings. For example, if you could perfectly predict a person’s actions, you could “trap” their environment, adding some hidden cameras that check that the person doesn’t deviate from your predictions, and triggers a bomb if they do deviate. Then that person is no longer counterfactually robust, since any slight change will trigger the bomb and destroy them. But we didn’t touch them at all, only some hidden surroundings!
---
I also suspect that we can’t just bite the bullet about consciousness and Turing machines appearing everywhere, since I think it would have anthropic implications that don’t match reality. Anthropic arguments are not on very solid footing, so I’m not totally confident about that, but nonetheless I think there’s probably just something we don’t understand yet.
I also think this absurdity you’ve noticed is an instance of a more general problem, since it applies to pretty much any emergent pattern. The same way you can find consciousness everywhere, you can find all sorts of Turing machines everywhere. So I view this as the problem of trying to characterize emergent phenomena.
---
Investigating causality was the lead I followed for a while as well, but every attempt I’ve made with it has ended up too strong, capable of seeing imaginary Turing machines everywhere. So lately I’ve been investigating the possibility that emergence might be about *information* in addition to causality.
One intuition I have for this is that the problem might happen because we add information in the process of pointing to the emergent phenomena. Given a bunch of particles randomly interacting with each other, you can probably point to a path of causality and make a correspondence to a person. But pointing out that path takes a lot of information which might only be present inside the pointer, so I think it’s possible that we’re effectively “sneaking in” the person via our pointer.
I often also use Conway’s Game of Life when I think about this issue. In the Game of Life, bits are often encoded as the presence or absence of a glider. This means that causality has to be able to travel the void of dead cells, so that the absence of a glider can be causal. This gives a pretty good argument that every cell has some causal effect on its neighbours, even dead ones.
But if we allow that, we can suddenly draw effectively arbitrary causal arrows inside a completely dead board! So I don’t think that can be right, either. My current lead for solving this is that the dead board has effectively no information; it’s trivial to write a proof that every future cell is also dead. On the other hand, for a complex board, proving its future state can be very difficult and might require simulating every step. This seems to point to a difference in *informational* content, even in two places where we have similar causal arrows.
So my suspicion is that random interactions inside walls might not contain the right information to encode a person. Unfortunately I don’t know much information theory yet, so my progress in figuring this out is slow.
The bomb trap example is very interesting! Can’t be counterfactually robust if you’re dead. Instead of bombs, we could also just use sudden overwhelming sensory inputs in the modality they’re fastest in to interrupt other processing. However, one objection could be that there exist some counterfactuals (for the same unmodified brain) where the person does what they’re supposed to. Objects we normally think of as unconscious don’t even have this weaker kind of counterfactual robustness: they need to be altered into different systems to do what they’re supposed to to be conscious.
Interesting. Do you think if someone kept the mapping between the states and “firing” and “non-firing” neurons, and translated the events as they were happening (on paper, automatically on a computer, in their own heads), this would generate (further) consciousness?
Doesn’t counting the causal effects of dead cells on dead cells, especially on a totally dead board, bring us back counterfactual robustness, though?
To expand a bit on the OP, the way I’ve tentatively been thinking about causality as the basis for consciousness is more like active physical signalling than like full counterfactuals, to avoid counterfactual robustness (and counting static objects as conscious, but there are probably plenty of ways to avoid that). On this view, dead cells don’t send signals to other cells, and there’s no signalling in a dead board or a dead brain, so there’s no consciousness in them (at the cell-level) either. What I care about for a neuron (and I’m not sure how well this translates to Conway’s Game of Life) is whether it actually just received a signal, whether it actually just “fired”, and whether removing/killing it would have prevented it from sending a signal to another that it did actually send. In this way, its presence had to actually make a non-trivial difference compared just to the counterfactual where it’s manipulated to be gone/dead.
Another related point is that while shadows (voids) and light spots can “move” faster than light, no actual particles are moving faster than light, and information can still only travel at most at the speed of light.
On your approach and examples, it does certainly seem like information is correlated with the stuff that matters in some way. It would be interesting to see this explored further. Have you found any similar theories in the literature?