Followup to:  The Fabric of Real ThingsStuff That Makes Stuff Happen

Previous meditation: "Does your rule forbid epiphenomenalist theories of consciousness that consciousness is caused by neurons, but doesn't affect those neurons in turn? The classic argument for epiphenomenal consciousness is that we can imagine a universe where people behave exactly the same way, but there's nobody home - no awareness, no consciousness, inside the brain. For all the atoms in this universe to be in the same place - for there to be no detectable difference internally, not just externally - 'consciousness' would have to be something created by the atoms in the brain, but which didn't affect those atoms in turn. It would be an effect of atoms, but not a cause of atoms. Now, I'm not so much interested in whether you think epiphenomenal theories of consciousness are true or false - rather, I want to know if you think they're impossible or meaningless a priori based on your rules."

Is it coherent to imagine a universe in which a real entity can be an effect but not a cause?

Well... there's a couple of senses in which it seems imaginable. It's important to remember that imagining things yields info primarily about what human brains can imagine. It only provides info about reality to the extent that we think imagination and reality are systematically correlated for some reason.

That said, I can certainly write a computer program in which there's a tier of objects affecting each other, and a second tier - a lower tier - of epiphenomenal objects which are affected by them, but don't affect them. For example, I could write a program to simulate some balls that bounce off each other, and then some little shadows that follow the balls around.

But then I only know about the shadows because I'm outside that whole universe, looking in. So my mind is being affected by both the balls and shadows - to observe something is to be affected by it. I know where the shadow is, because the shadow makes pixels be drawn on screen, which make my eye see pixels. If your universe has two tiers of causality - a tier with things that affect each other, and another tier of things that are affected by the first tier without affecting them - then could you know that fact from inside that universe?

Again, this seems easy to imagine as long as objects in the second tier can affect each other. You'd just have to be living in the second tier! We can imagine, for example - this wasn't the way things worked out in our universe, but it might've seemed plausible to the ancient Greeks - that the stars in heaven (and the Sun as a special case) could affect each other and affect Earthly forces, but no Earthly force could affect them:

(Here the X'd-arrow stands for 'cannot affect'.)

The Sun's light would illuminate Earth, so it would cause plant growth. And sometimes you would see two stars crash into each other and explode, so you'd see they could affect each other. (And affect your brain, which was seeing them.) But the stars and Sun would be made out of a different substance, the 'heavenly material', and throwing any Earthly material at it would not cause it to change state in the slightest. The Earthly material might be burned up, but the Sun would occupy exactly the same position as before. It would affect us, but not be affected by us.

(To clarify an important point raised in the comments: In standard causal diagrams and in standard physics, no two individual events ever affect each other; there's a causal arrow from the PAST to FUTURE but never an arrow from FUTURE to PAST. What we're talking about here is the sun and stars over time, and the generalization over causal arrows that point from Star-in-Past to Sun-in-Present and Sun-in-Present back to Star-in-Future. The standard formalism dealing with this would be Dynamic Bayesian Networks (DBNs) in which there are repeating nodes and repeating arrows for each successive timeframe: X1, X2, X3, and causal laws F relating Xi to Xi+1. If the laws of physics did not repeat over time, it would be rather hard to learn about the universe! The Sun repeatedly sends out photons, and they obey the same laws each time they fall on Earth; rather than the Fi being new transition tables each time, we see a constant Fphysics over and over. By saying that we live in a single-tier universe, we're observing that whenever there are F-arrows, causal-link-types, which (over repeating time) descend from variables-of-type-X to variables-of-type-Y (like present photons affecting future electrons), there are also arrows going back from Ys to Xs (like present electrons affecting future photons). If we weren't generalizing over time, it couldn't possibly make sense to speak of thingies that "affect each other" - causal diagrams don't allow directed cycles!)

A two-tier causal universe seems easy to imagine, even easy to specify as a computer program. If you were arranging a Dynamic Bayes Net at random, would it randomly have everything in a single tier? If you were designing a causal universe at random, wouldn't there randomly be some things that appeared to us as causes but not effects? And yet our own physicists haven't discovered any upper-tier particles which can move us without being movable by us. There might be a hint here at what sort of thingies tend to be real in the first place - that, for whatever reasons, the Real Rules somehow mandate or suggest that all the causal forces in a universe be on the same level, capable of both affecting and being affected by each other.

Still, we don't actually know the Real Rules are like that; and so it seems premature to assign a priori zero probability to hypotheses with multi-tiered causal universes. Discovering a class of upper-tier affect-only particles seems imaginable[1] - we can imagine which experiences would convince us that they existed. If we're in the Matrix, we can see how to program a Matrix like that. If there's some deeper reason why that's impossible in any base-level reality, we don't know it yet. So we probably want to call that a meaningful hypothesis for now.

But what about lower-tier particles which can be affected by us, and yet never affect us?

Perhaps there are whole sentient Shadow Civilizations living on my nose hairs which can never affect those nose hairs, but find my nose hairs solid beneath their feet. (The solid Earth affecting them but not being affected, like the Sun's light affecting us in the 'heavenly material' hypothesis.) Perhaps I wreck their world every time I sneeze. It certainly seems imaginable - you could write a computer program simulating physics like that, given sufficient perverseness and computing power...

And yet the fundamental question of rationality - "What do you think you know, and how do you think you know it?" - raises the question:

How could you possibly know about the lower tier, even if it existed?

To observe something is to be affected by it - to have your brain and beliefs take on different states, depending on that thing's state. How can you know about something that doesn't affect your brain?

In fact there's an even deeper question, "How could you possibly talk about that lower tier of causality even if it existed?"

Let's say you're a Lord of the Matrix. You write a computer program which first computes the physical universe as we know it (or a discrete approximation), and then you add a couple of lower-tier effects as follows:

First, every time I sneeze, the binary variable YES_SNEEZE will be set to the second of its two possible values.

Second, every time I sneeze, the binary variable NO_SNEEZE will be set to the first of its two possible values.

Now let's say that - somehow - even though I've never caught any hint of the Matrix - I just magically think to myself one day, "What if there's a variable that watches when I sneeze, and gets set to 1?"

It will be all too easy for me to imagine that this belief is meaningful and could be true or false:

And yet in reality - as you know from outside the matrix - there are two shadow variables that get set when I sneeze. How can I talk about one of them, rather than the other? Why should my thought about '1' refer to their second possible value rather than their first possible value, inside the Matrix computer program? If we tried to establish a truth-value in this situation, to compare my thought to the reality inside the computer program - why compare my thought about SNEEZE_VAR to the variable YES_SNEEZE instead of NO_SNEEZE, or compare my thought '1' to the first possible value instead of the second possible value?

Under more epistemically healthy circumstances, when you talk about things that are not directly sensory experiences, you will reference a causal model of the universe that you inducted to explain your sensory experiences. Let's say you repeatedly go outside at various times of day, and your eyes and skin directly experience BRIGHT-WARM, BRIGHT-WARM, BRIGHT-WARM, DARK-COOL, DARK-COOL, etc. To explain the patterns in your sensory experiences, you hypothesize a latent variable we'll call 'Sun', with some kind of state which can change between 1, which causes BRIGHTness and WARMness, and 0, which causes DARKness and COOLness. You believe that the state of the 'Sun' variable changes over time, but usually changes less frequently than you go outside.

p(BRIGHT|Sun=1) 0.9
p(¬BRIGHT|Sun=1) 0.1
p(BRIGHT|Sun=0) 0.1
p(¬BRIGHT|Sun=0) 0.9

Standing here outside the Matrix, we might be tempted to compare your beliefs about "Sun = 1", to the real universe's state regarding the visibility of the sun in the sky (or rather, the Earth's rotational position).

But even if we compress the sun's visibility down to a binary categorization, how are we to know that your thought "Sun = 1" is meant to correspond to the sun being visible in the sky, rather than the sun being occluded by the Earth? Why the first state of the variable, rather than the second state?

How indeed are we know that this thought "Sun = 1" is meant to compare to the sun at all, rather than an anteater in Venezuela?

Well, because that 'Sun' thingy is supposed to be the cause of BRIGHT and WARM feelings, and if you trace back the cause of those sensory experiences in reality you'll arrive at the sun that the 'Sun' thought allegedly corresponds to. And to distinguish between whether the sun being visible in the sky is meant to correspond to 'Sun'=1 or 'Sun'=0, you check the conditional probabilities for that 'Sun'-state giving rise to BRIGHT - if the actual sun being visible has a 95% chance of causing the BRIGHT sensory feeling, then that true state of the sun is intended to correspond to the hypothetical 'Sun'=1, not 'Sun'=0.

Or to put it more generally, in cases where we have...

...then the correspondence between map and territory can at least in principle be point-wise evaluated by tracing causal links back from sensory experiences to reality, and tracing hypothetical causal links from sensory experiences back to hypothetical reality. We can't directly evaluate that truth-condition inside our own thoughts; but we can perform experiments and be corrected by them.

Being able to imagine that your thoughts are meaningful and that a correspondence between map and territory is being maintained, is no guarantee that your thoughts are true. On the other hand, if you can't even imagine within your own model how a piece of your map could have a traceable correspondence to the territory, that is a very bad sign for the belief being meaningful, let alone true. Checking to see whether you can imagine a belief being meaningful is a test which will occasionally throw out bad beliefs, though it is no guarantee of a belief being good.


Okay, but what about the idea that it should be meaningful to talk about whether or not a spaceship continues to exist after it travels over the cosmological horizon? Doesn't this theory of meaningfulness seem to claim that you can only sensibly imagine something that makes a difference to your sensory experiences?

No. It says that you can only talk about events that your sensory experiences pin down within the causal graph. If you observe enough protons, electrons, neutrons, and so on, you can pin down the physical generalization which says, "Mass-energy is neither created nor destroyed; and in particular, particles don't vanish into nothingness without a trace." It is then an effect of that rule, combined with our previous observation of the ship itself, which tells us that there's a ship that went over the cosmological horizon and now we can't see it any more.

To navigate referentially to the fact that the ship continues to exist over the cosmological horizon, we navigate from our sensory experience up to the laws of physics, by talking about the cause of electrons not blinking out of existence; we also navigate up to the ship's existence by tracing back the cause of our observation of the ship being built. We can't see the future ship over the horizon - but the causal links down from the ship's construction, and from the laws of physics saying it doesn't disappear, are both pinned down by observation - there's no difficulty in figuring out which causes we're talking about, or what effects they have.[2]


All righty-ighty, let's revisit that meditation:

"Does your rule forbid epiphenomenalist theories of consciousness in which consciousness is caused by neurons, but doesn't affect those neurons in turn? The classic argument for epiphenomenal consciousness is that we can imagine a universe where people behave exactly the same way, but there's nobody home - no awareness, no consciousness, inside the brain. For all the atoms in this universe to be in the same place - for there to be no detectable difference internally, not just externally - 'consciousness' would have to be something created by the atoms in the brain, but which didn't affect those atoms in turn. It would be an effect of atoms, but not a cause of atoms. Now, I'm not so much interested in whether you think epiphenomenal theories of consciousness are true or false - rather, I want to know if you think they're impossible or meaningless a priori based on your rules."

The closest theory to this which definitely does seem coherent - i.e., it's imaginable that it has a pinpointed meaning - would be if there was another little brain living inside my brain, made of shadow particles which could affect each other and be affected by my brain, but not affect my brain in turn. This brain would correctly hypothesize the reasons for its sensory experiences - that there was, from its perspective, an upper tier of particles interacting with each other that it couldn't affect. Upper-tier particles are observable, i.e., can affect lower-tier senses, so it would be possible to correctly induct a simplest explanation for them. And this inner brain would think, "I can imagine a Zombie Universe in which I am missing, but all the upper-tier particles go on interacting with each other as before." If we imagine that the upper-tier brain is just a robotic sort of agent, or a kitten, then the inner brain might justifiably imagine that the Zombie Universe would contain nobody to listen - no lower-tier brains to watch and be aware of events.

We could write that computer program, given significantly more knowledge and vastly more computing power and zero ethics.

But this inner brain composed of lower-tier shadow particles cannot write upper-tier philosophy papers about the Zombie universe. If the inner brain thinks, "I am aware of my own awareness", the upper-tier lips cannot move and say aloud, "I am aware of my own awareness" a few seconds later. That would require causal links from lower particles to upper particles.

If we try to suppose that the lower tier isn't a complicated brain with an independent reasoning process that can imagine its own hypotheses, but just some shadowy pure experiences that don't affect anything in the upper tier, then clearly the upper-tier brain must be thinking meaningless gibberish when the upper-tier lips say, "I have a lower tier of shadowy pure experiences which did not affect in any way how I said these words." The deliberating upper brain that invents hypotheses for sense data, can only use sense data that affects the upper neurons carrying out the search for hypotheses that can be reported by the lips. Any shadowy pure experiences couldn't be inputs into the hypothesis-inventing cognitive process. So the upper brain would be talking nonsense.

There's a version of this theory in which the part of our brain that we can report out loud, which invents hypotheses to explain sense data out loud and manifests physically visible papers about Zombie universes, has for no explained reason invented a meaningless theory of shadow experiences which is experienced by the shadow part as a meaningful and correct theory.  So that if we look at the "merely physical" slice of our universe, philosophy papers about consciousness are meaningless and the physical part of the philosopher is saying things their physical brain couldn't possibly know even if they were true.  And yet our inner experience of those philosophy papers is meaningful and true. In a way that couldn't possibly have caused me to physically write the previous sentence, mind you. And yet your experience of that sentence is also true even though, in the upper tier of the universe where that sentence was actually written, it is not only false but meaningless.

I'm honestly not sure what to say when a conversation gets to that point. Mostly you just want to yell, "Oh, for the love of Belldandy, will you just give up already?" or something about the importance of saying oops.

(Oh, plus the unexplained correlation violates the Markov condition for causal models.)

Maybe my reply would be something along the lines of, "Okay... look... I've given my account of a single-tier universe in which agents can invent meaningful explanations for sense data, and when they build accurate maps of reality there's a known reason for the correspondence... if you want to claim that a different kind of meaningfulness can hold within a different kind of agent divided into upper and lower tiers, it's up to you to explain what parts of the agent are doing which kinds of hypothesizing and how those hypotheses end up being meaningful and what causally explains their miraculous accuracy so that this all makes sense."

But frankly, I think people would be wiser to just give up trying to write sensible philosophy papers about lower causal tiers of the universe that don't affect the philosophy papers in any way.


MeditationIf we can only meaningfully talk about parts of the universe that can be pinned down inside the causal graph, where do we find the fact that 2 + 2 = 4? Or did I just make a meaningless noise, there? Or if you claim that "2 + 2 = 4" isn't meaningful or true, then what alternate property does the sentence "2 + 2 = 4" have which makes it so much more useful than the sentence "2 + 2 = 3"?


Mainstream status.


 [1] Well, it seems imaginable so long as you toss most of quantum physics out the window and put us back in a classical universe. For particles to not be affected by us, they'd need their own configuration space such that "which configurations are identical" was determined by looking only at those particles, and not looking at any lower-tier particles entangled with them. If you don't want to toss QM out the window, it's actually pretty hard to imagine what an upper-tier particle would look like.

 [2] This diagram treats the laws of physics as being just another node, which is a convenient shorthand, but probably not a good way to draw the graph. The laws of physics really correspond to the causal arrows Fi, not the causal nodes Xi. If you had the laws themselves - the function from past to future - be an Xi of variable state, then you'd need meta-physics to describe the Fphysics arrows for how the physics-stuff Xphysics could affect us, followed promptly by a need for meta-meta-physics et cetera. If the laws of physics were a kind of causal stuff, they'd be an upper tier of causality - we can't appear to affect the laws of physics, but if you call them causes, they can affect us. In Matrix terms, this would correspond to our universe running on a computer that stored the laws of physics in one area of RAM and the state of the universe in another area of RAM, the first area would be an upper causal tier and the second area would be a lower causal tier. But the infinite regress from treating the laws of determination as causal stuff, makes me suspicious that it might be an error to treat the laws of physics as "stuff that makes stuff happen and happens because of other stuff". When we trust that the ship doesn't disappear when it goes over the horizon, we may not be navigating to a physics-node in the graph, so much as we're navigating to a single Fphysics that appears in many different places inside the graph, and whose previously unknown function we have inferred. But this is an unimportant technical quibble on Tuesdays, Thursdays, Saturdays, and Sundays. It is only an incredibly deep question about the nature of reality on Mondays, Wednesdays, and Fridays, i.e., less than half the time.

Part of the sequence Highly Advanced Epistemology 101 for Beginners

Next post: "Proofs, Implications, and Models"

Previous post: "Stuff That Makes Stuff Happen"

Comments

sorted by
magical algorithm
Highlighting new comments since Today at 11:46 AM
Select new highlight date
All comments loaded

Epiphenominal theories of consciousness are kind of silly, but here's another situation I can wonder about... some cellular automata rules, including the Turing-complete Conway's Game of Life, can have different "pasts" that can lead to the same present. From the point of view of a being living in such a universe (one in which information can be destroyed), is there a fact of the matter as to which "past" actually happened?

I had always thought that our physical universe had this property as well, i.e. the Everett multiverse branches into the past as well as into the future.

If you take a single branch and run it backward, you'll find that it diverges into a multiverse of its own. If you take all the branches and run them backward, their branches will cohere instead of decohering, cancel out in most places, and miraculously produce only the larger, more coherent blobs of amplitude they started from. Sort of like watching an egg unscramble itself.

Interesting question! I'd say that you could refer to the possibilities as possibilities, e.g. in a debate over whether a particular past would in fact have led to the present, but to speak of the 'actual past' might make no sense because you couldn't get there from there... no, actually, I take that back, you might be able to get there via simplicity. I.e. if there's only one past that would have evolved from a simply-tiled start state for the automaton.

Just for the Least Convenient World, what if the zombies build a supercomputer and simulate random universes, and find that in 98% of simulated universes life forms like theirs do have shadow brains, and that the programs for the remaining 2% are usually significantly longer?

Some thoughts about "epiphenomena" in general, though not related to consciousness.

Suppose there are only finitely many events in the entire history of the universe (or multiverse), so that the universe can be represented by a finite casual graph. If it is an acrylic graph (no causal cycles), then there must be some nodes which are effects but not causes, that is, they are epiphenomena. But then why not posit a smaller graph with the epiphenomenal nodes removed, since they don't do anything? And then that reduced graph is also finite, and also has epiphenomenal nodes.... so why not remove those?

So, is the conclusion that the best model of the universe is a strictly infinite graph, with no epiphenomenal nodes that can be removed e.g. no future big crunches or other singularities? This seems like a dubious piece of armchair cosmology.

Or are there cases where the larger finite graph (with the epiphenomenal nodes) is strictly simpler as a theory than the reduced graph (with the epiphenomena removed), so that Occam's razor tells us to believe in the larger graph? But then Occam's razor is justifying a belief in epiphenomena, which sounds rather odd when put like that!

The last nodes are never observed by anyone, but they descend from the same physics, the same F(physics), that have previously been pinned down, or so I assume. You can thus meaningfully talk about them for the same reason you can meaningfully talk about a spaceship going over the cosmological horizon. What we're trying to avoid is SNEEZE_VARs or lower qualia where there's no way that the hypothesis-making agent could ever have observed, inducted, and pinned down the causal mechanism - where there's no way a correspondence between map and territory could possibly be maintained.

Is it coherent to imagine a universe in which a real entity can be an effect but not a cause?

Your favorite example of event horizons, cosmological or otherwise, is like that. GR suggests that there can be a ring singularity inside an eternal spinning black hole (but not one spun up from rest), near/around which you can go forever without being crushed. (it also suggests that there could be closed timelike curves around it, but I'll ignore this for now.) So maybe there are particles/objects/entities living there.

Stuff thrown into such a black hole can certainly affect the hypothetical entities living inside. Like a meteor shower from the outside. But the outside is not affected by anything happening inside, the horizon prevents it.

That said, I can certainly write a computer program in which there's a tier of objects affecting each other, and a second tier - a lower tier - of epiphenomenal objects which are affected by them, but don't affect them.

I would like to point out that any space-like surface (technically 3-fold) divides our universe into two such tiers.

Okay, I can see that I need to spell out in more detail one of the ideas here - namely that you're trying to generalize over a repeating type of causal link and that reference is pinned down by such generalization. The Sun repeatedly sends out light in individual Sun-events, electrons repeatedly go on traveling through space instead of vanishing; in a universe like ours, rather than the F(i) being whole new transition tables randomly generated each time, you see the same F(physics) over and over. This is what you can pin down and refer to. Any causal graph is acyclic and can be divided as you say; the surprising thing is that there are no F-types, no causal-link-types, which (over repeating time) descend from one kind of variable to another, without (over time) there being arrows also going back from that kind to the other. Yes, we're generalizing and inducting over time, otherwise it would make no sense to speak of thingies that "affect each other". No two individual events ever affect each other!

FWIW, my old post 'Zombie Rationality' explores what I think the epiphenomenalist should say about the worry that "the upper-tier brain must be thinking meaningless gibberish when the upper-tier lips [talk about consciousness]"

One point to flag is that from an epiphenomenalist's perspective, mere brains never really mean anything, any more than squiggles of ink do; any meaning we attribute to them is purely derivative from the meaning of appropriately-related thoughts (which, on this view, essentially involve qualia).

Another thing to flag is that epiphenomenalism needn't imply that our thoughts are causally irrelevant, but merely their experiential component. It'd be a mistake to identify oneself with just one's qualia (as Eliezer seems to attribute to the epiphenomenalist). It's true that our qualia don't write philosophy papers about consciousness. But we, embodied conscious persons, do write such papers. Of course, the causal explanation of the squiggles depends only on our physical parts. But the fact that the squiggles are about consciousness (or indeed anything at all) depends crucially upon the epiphenomenal aspects of our minds, in addition.

I think the question "does consciousness affect neurons?" is as meaningful as "does the process of computation in a computer affect bits?".

Can anyone explain why epiphenomenalist theories of consciousness are interesting? There have been an awful lot of words on them here, but I can't find a reason to care.

Pretty much the same reason religion needs to be talked about. If no one had invented it wouldn't be useful to dispute notions of god creating us for a divine purpose, but because many people think this indeed happened you have to talk about it. It's especially important for reasonable discussions of AI.

Religion and epiphenomenalogy differ in three important ways:

  • Religion is widespread. Almost everyone knows what it is. Most people have at least some religious memes sticking in their heads. A significant fraction of people have dangerous religious memes in their heads so decreasing those qualifies as raising the sanity waterline. Epiphenomenalogy is essentially unknown outside academic philosophy, and now the lesswrong readership.
  • Religion has impact everywhere. People have died because of other people's religious beliefs, and not just from violence. Belief in epiphenomenalogy has almost no impact on the lives of non-believers.
  • Religious thought patterns re-occur. Authority, green/blue, and "it is good to believe" show up over and over again. The sort of thoughts that lead to epiphenomenalogy are quite obscure.

Epiphenomenalogy is essentially unknown outside academic philosophy, and now the lesswrong readership.

I'd say it's more widespread than that. Some strands of Buddhist thought, for instance, seem to strongly imply it even if they didn't state it outright. And it feels like it'd be the most intuitive way of thinking about consciousness for many of the people who'd think about it at all, even if they weren't familiar with academic philosophy. (I don't think I got it from academic philosophy, though I can't be sure of that.)

It seems that you get similar questions as a natural outgrowth of simple computational models of thought. E.g. if one performs Solomonoff induction on the stream of camera inputs to a robot, what kind of short programs will dominate the probability distribution over the next input? Not just programs that simulate the physics of our universe: one would also need additional code to "read off" the part of the simulated universe that corresponded to the camera inputs. That additional code looks like epiphenomenal mind-stuff. Using this framework you can pose questions like "if the camera is expected to be rebuilt using different but functionally equivalent materials, will his change the inputs Solomonoff induction predicts?" or "if the camera is about to be duplicated, which copy's inputs will be predicted by Solomonoff induction?"

If we go beyond Solomonoff induction to allow actions, then you get questions that map pretty well to debates about "free will."

It's good practice for seeing how the rules play out; and having a clear mental visualization of how causality works, and what sort of theories are allowed to be meaningful, is actually reasonably important both for anti-nonsense techniques and AI construction. People who will never need any more skill than they currently have in Resist Nonsense or Create Friendly AI can ignore it, I suppose.

Because epiphenomenalist theories are common but incorrect, and the goal of LessWrong is at least partially what its name implies.

EDIT: After thinking things through, I concluded that Eliezer was right, and that epiphenomalism was indeed confused and incoherent. Leaving this comment here as a record of how I came to agree with that conclusion.

The closest theory to this which definitely does seem coherent - i.e., it's imaginable that it has a pinpointed meaning - would be if there was another little brain living inside my brain, made of shadow particles which could affect each other and be affected by my brain, but not affect my brain in turn. This brain would correctly hypothesize the reasons for its sensory experiences - that there was, from its perspective, an upper tier of particles interacting with each other that it couldn't affect. Upper-tier particles are observable, i.e., can affect lower-tier senses, so it would be possible to correctly induct a simplest explanation for them. And this inner brain would think, "I can imagine a Zombie Universe in which I am missing, but all the upper-tier particles go on interacting with each other as before." If we imagine that the upper-tier brain is just a robotic sort of agent, or a kitten, then the inner brain might justifiably imagine that the Zombie Universe would contain nobody to listen - no lower-tier brains to watch and be aware of events.

Positing "another little brain" to act as the epiphenomenal component sounds unnecessarily complicated to me. You mentioned earlier the possibility of programming a simulation with balls that bounce off each other, and then some little shadows that followed the balls around. It is not obvious to me why "balls that bounce off each other" couldn't be the physical activity of the neurons, "little shadows that followed the balls around" couldn't be the qualia that were produced as a side-effect of the neurons' physical activity.

I think - though I might have misunderstood - that you are trying to exclude this possibility with the "How could you possibly know about the lower tier, even if it existed?" question, and the suggestion that we could only know about the shadows because we can look at the simulation from the outside, and the notion that there is no flow of causality in our world from which we could infer the existence of the "shadow tier".

I'm not convinced that this is right. We know that there are qualia, for the obvious reason that we are not p-zombies: and we know that consciousness is created via neurons in the brain, for we can map a correspondence between qualia and brain states. E.g. there's a clear shift in both brain activity and subjective experience when we fall asleep, or become agitated, or drink alcohol. So we know that there is an arrow from "changes in the brain" to "qualia/subjective experience", because we correctly anticipate that if somebody changed our brain chemistry, our subjective experience would change.

But there's nothing in our observed physical world that actually explains consciousness in the sense of explaining why there couldn't just be a physical world utterly devoid of consciousness. Yes, you can develop sophisticated theories of strange loops and of how consciousness is self-representation and all... which explains why there could be symbols and dynamics within a cognitive system which would make an entity behave as if it had consciousness. If the programmer is sufficiently sophisticated, it can make a program behave like it had anything at all.

But that's still only an explanation of why it behaved as if it had a consciousness. Well, that's not quite right. It would actually have a consciousness in the sense that it had all the right information-processing dynamics which caused it to have some internal state which would serve the functional role of "sadness" or "loneliness" or "excitement" in influencing its information-processing and behavior. And it would have a pattern recognition engine which analyzed its own experience, which would notice that those kinds of internal states repeated themselves and had predictable effects on its behavior and information-processing.

So it would assign those states labels and introduce symbols corresponding to those labels into its reasoning system, so it could ask itself questions like "every now and then I get put into a state that I have labeled as 'being sad', when does that happen and what does it do?". And as it collected more and more observations and created increasingly sophisticated symbols to represent ever-more-complex concepts, then it seems entirely conceivable that there happened something like... like it noticing that all the symbols it had collected for its own internal states shared the property of being symbols for its own internal states, and its explanation-seeking-mechanism would do what it always did, namely to ask the question of "why do I have this set of states in the first place, instead of having nothing". And then, because that was an ill-defined question in the first place, it would get stumped and fail to answer it, and then it could very well write philosophical papers that tried to attack the question but made no progress.

... and when I started writing this comment, I was originally going to end this by saying "and that explains why it would behave as if it was conscious, but it still wouldn't explain why it couldn't do all of that without having any subjective experience". Except that, upon writing that out, I started feeling like I might just have dissolved the question and that there was nothing left to explain anymore. Um. I'll have to think about this some more.

Mainstream status points to /Eliezer_Yudkowsky-drafts/ (Forbidden: You aren't allowed to do that.)

Mainstream status:

I haven't yet happened to run across a philosophical position which says that meaningful correspondences between hypotheses and reality can only be pinned down by following Pearl-style causal links inferred as the simplest explanation of observed experiences, and that only this can allow an agent to consistently believe that its beliefs are meaningful.

In fact, I haven't seen anything at all about referential meaningfulness requiring cause-and-effect links with the phenomenon, just like I haven't seen anything about a universe being a connected fabric of causes and effects; but it wouldn't surprise me if either of those ideas were out there somewhere.

(There's a "causal theory of reference" listed in Wikipedia but it doesn't seem to be about remotely the same subject matter; the theory's tenets seem to be that "a name's referent is fixed by an original act of naming", and that "later uses of the name succeed in referring to the referent by being linked to that original act via a causal chain".)

EDIT: Apparently causal theories of reference have been used to argue against Zombie Worlds so I stand corrected on this point. See below.

As bryjnar points out, all the stuff you say here (subtracting out the Pearl stuff) is entailed by the causal theory of reference. The reason quick summaries of that view will seem unfamiliar is that most of the early work on the causal theory was primarily motivated by a different concern -- accounting for how our words acquire their meaning. Thus the focus on causal chains from "original acts of naming" and whatnot. However, your arguments against epiphenomenalism all hold in the causal theory.

It is true that nobody (that I know of) has developed an explicitly Pearlian causal theory of reference, but this is really accounted for by division of labor in philosophy. People working on reference will develop a causal theory of reference and use words like "cause" without specifying what they mean by it. If you ask them what they mean, they will say "Whatever the best theory of causation is. Go ask the people working on causation about that." And among the people working on causation, there are indeed philosophers who have built on Pearlian ideas. Christopher Hitchcock and James Woodward, for instance.

You've got a more sophisticated notion of causality, but otherwise this is very similar to the causal theory of reference. For example, the way they would describe what's going on with the shadow sneeze variables is that when you name "SNEEZE_VAR", there was no causal link that allowed you to pick out the actual SNEEZE_VAR: there would need to be a causal arrow going the other way for that to be possible. (and then any later uses of "SNEEZE_VAR" would have to be linked causally to your earlier naming: if we wiped your brain and rebooted it with random noise that happened to be the same, then you wouldn't succeed in referring) I'm pretty sure I've seen someone use a similar kind of example where you can't decide which of two possible things you're referring to because of the lack of a causal link of the right kind.

They also use pretty similar examples: a classic one is to think of an being on the other side of the galaxy thinking about Winston Churchill. Even if they have the right image, they even happen to think the right things about where he lived, what he did etc. it seems that they don't actually succeed in referring to him because of the lack of a causal link. It's just a coincidence.

With that in mind, there are probably arguments made against the causal theory of reference that may apply to you to, but I don't know any off the top of my head.

I'm not convinced I'm keeping my levels of reference straight, but if I can knowingly consistently accurately talk about epiphenomena, doesn't the structure or contents of the uncausing stuff cause me to think in this way rather than that way? I'm not sure how to formalize this intuition to tell if it's useful or trivial.

That's the point, I believe. To call something 'real', and yet say in principle it couldn't affect us causally, is to contradict yourself. If you're contained in a set of nodes that are causally linked, then that is your reality. Any other sets of nodes that can't contact you just don't matter. Even if there was, say, a causal structure beside our universe, of puppies and kittens running around, it wouldn't matter in the slightest, if in principle we could never interact with them. If we could deduce from the laws of physics that they had to exist, then we would be causally linked. The point I am (and perhaps Eliezer is) trying to emphasise is that our reality is everything that can, and does, affect us.