Continuation ofIs Morality Preference?

(Disclaimer:  Neither Subhan nor Obert represent my own position on morality; rather they represent different sides of the questions I hope to answer.)

Subhan:  "What is this 'morality' stuff, if it is not a preference within you?"

Obert:  "I know that my mere wants, don't change what is right; but I don't claim to have absolute knowledge of what is right—"

Subhan:  "You're not escaping that easily!  How does a universe in which murder is wrong, differ from a universe in which murder is right?  How can you detect the difference experimentally?  If the answer to that is 'No', then how does any human being come to know that murder is wrong?"

Obert:  "Am I allowed to say 'I don't know'?"

Subhan:  "No.  You believe now that murder is wrong.  You must believe you already have evidence and you should be able to present it now."

Obert:  "That's too strict!  It's like saying to a hunter-gatherer, 'Why is the sky blue?' and expecting an immediate answer."

Subhan:  "No, it's like saying to a hunter-gatherer:  Why do you believe the sky is blue?"

Obert:  "Because it seems blue, just as murder seems wrong.  Just don't ask me what the sky is, or how I can see it."

Subhan:  "But—aren't we discussing the nature of morality?"

Obert:  "That, I confess, is not one of my strong points.  I specialize in plain old morality.  And as a matter of morality, I know that I can't make murder right just by wanting to kill someone."

Subhan:  "But if you wanted to kill someone, you would say, 'I know murdering this guy is right, and I couldn't make it wrong just by not wanting to do it.'"

Obert:  "Then, if I said that, I would be wrong.  That's common moral sense, right?"

Subhan:  "Argh!  It's difficult to even argue with you, since you won't tell me exactly what you think morality is made of, or where you're getting all these amazing moral truths—"

Obert:  "Well, I do regret having to frustrate you.  But it's more important that I act morally, than that I come up with amazing new theories of the nature of morality.  I don't claim that my strong point is in explaining the fundamental nature of morality.  Rather, my strong point is coming up with theories of morality that give normal moral answers to questions like, 'If you feel like killing someone, does that make it right to do so?'  The common-sense answer is 'No' and I really see no reason to adopt a theory that makes the answer 'Yes'.  Adding up to moral normality—that is my theory's strong point."

Subhan:  "Okay... look.  You say that, if you believed it was right to murder someone, you would be wrong."

Obert:  "Yes, of course!  And just to cut off any quibbles, we'll specify that we're not talking about going back in time and shooting Stalin, but rather, stalking some innocent bystander through a dark alley and slitting their throat for no other reason but my own enjoyment.  That's wrong."

Subhan:  "And anyone who says murder is right, is mistaken."

Obert:  "Yes."

Subhan:  "Suppose there's an alien species somewhere in the vastness of the multiverse, who evolved from carnivores.  In fact, through most of their evolutionary history, they were cannibals.  They've evolved different emotions from us, and they have no concept that murder is wrong—"

Obert:  "Why doesn't their society fall apart in an orgy of mutual killing?"

Subhan:  "That doesn't matter for our purposes of theoretical metaethical investigation.  But since you ask, we'll suppose that the Space Cannibals have a strong sense of honor—they won't kill someone they promise not to kill; they have a very strong idea that violating an oath is wrong.  Their society holds together on that basis, and on the basis of vengeance contracts with private assassination companies.  But so far as the actual killing is concerned, the aliens just think it's fun.  When someone gets executed for, say, driving through a traffic light, there's a bidding war for the rights to personally tear out the offender's throat."

Obert:  "Okay... where is this going?"

Subhan:  "I'm proposing that the Space Cannibals not only have no sense that murder is wrong—indeed, they have a positive sense that killing is an important part of life—but moreover, there's no path of arguments you could use to persuade a Space Cannibal of your view that murder is wrong.  There's no fact the aliens can learn, and no chain of reasoning they can discover, which will ever cause them to conclude that murder is a moral wrong.  Nor is there any way to persuade them that they should modify themselves to perceive things differently."

Obert:  "I'm not sure I believe that's possible—"

Subhan:  "Then you believe in universally compelling arguments processed by a ghost in the machine.  For every possible mind whose utility function assigns terminal value +1, mind design space contains an equal and opposite mind whose utility function assigns terminal value—1.  A mind is a physical device and you can't have a little blue woman pop out of nowhere and make it say 1 when the physics calls for it to say 0."

Obert:  "Suppose I were to concede this.  Then?"

Subhan:  "Then it's possible to have an alien species that believes murder is not wrong, and moreover, will continue to believe this given knowledge of every possible fact and every possible argument.  Can you say these aliens are mistaken?"

Obert:  "Maybe it's the right thing to do in their very different, alien world—"

Subhan:  "And then they land on Earth and start slitting human throats, laughing all the while, because they don't believe it's wrong.  Are they mistaken?"

Obert:  "Yes."

Subhan:  "Where exactly is the mistake?  In which step of reasoning?"

Obert:  "I don't know exactly.  My guess is that they've got a bad axiom."

Subhan:  "Dammit!  Okay, look.  Is it possible that—by analogy with the Space Cannibals—there are true moral facts of which the human species is not only presently unaware, but incapable of perceiving in principle?  Could we have been born defective—incapable even of being compelled by the arguments that would lead us to the light?  Moreover, born without any desire to modify ourselves to be capable of understanding such arguments?  Could we be irrevocably mistaken about morality—just like you say the Space Cannibals are?"

Obert:  "I... guess so..."

Subhan:  "You guess so?  Surely this is an inevitable consequence of believing that morality is a given, independent of anyone's preferences!  Now, is it possible that we, not the Space Cannibals, are the ones who are irrevocably mistaken in believing that murder is wrong?"

Obert:  "That doesn't seem likely."

Subhan:  "I'm not asking you if it's likely, I'm asking you if it's logically possible!  If it's not possible, then you have just confessed that human morality is ultimately determined by our human constitutions.  And if it is possible, then what distinguishes this scenario of 'humanity is irrevocably mistaken about morality', from finding a stone tablet on which is written the phrase 'Thou Shalt Murder' without any known justification attached?  How is a given morality any different from an unjustified stone tablet?"

Obert:  "Slow down.  Why does this argument show that morality is determined by our own constitutions?"

Subhan:  "Once upon a time, theologians tried to say that God was the foundation of morality.  And even since the time of the ancient Greeks, philosophers were sophisticated enough to go on and ask the next question—'Why follow God's commands?'  Does God have knowledge of morality, so that we should follow Its orders as good advice?  But then what is this morality, outside God, of which God has knowledge?  Do God's commands determine morality?  But then why, morally, should one follow God's orders?"

Obert:  "Yes, this demolishes attempts to answer questions about the nature of morality just by saying 'God!', unless you answer the obvious further questions.  But so what?"

Subhan:  "And furthermore, let us castigate those who made the argument originally, for the sin of trying to cast off responsibility—trying to wave a scripture and say, 'I'm just following God's orders!'  Even if God had told them to do a thing, it would still have been their own decision to follow God's orders."

Obert:  "I agree—as a matter of morality, there is no evading of moral responsibility.  Even if your parents, or your government, or some kind of hypothetical superintelligence, tells you to do something, you are responsible for your decision in doing it."

Subhan:  "But you see, this also demolishes the idea of any morality that is outside, beyond, or above human preference.  Just substitute 'morality' for 'God' in the argument!"

Obert:  "What?"

Subhan:  "John McCarthy said:  'You say you couldn't live if you thought the world had no purpose. You're saying that you can't form purposes of your own-that you need someone to tell you what to do. The average child has more gumption than that.'  For every kind of stone tablet that you might imagine anywhere, in the trends of the universe or in the structure of logic, you are still left with the question:  'And why obey this morality?'  It would be your decision to follow this trend of the universe, or obey this structure of logic.  Your decision—and your preference."

Obert:  "That doesn't follow!  Just because it is my decision to be moral—and even because there are drives in me that lead me to make that decision—it doesn't follow that the morality I follow consists merely of my preferences.  If someone gives me a pill that makes me prefer to not be moral, to commit murder, then this just alters my preference—but not the morality; murder is still wrong.  That's common moral sense—"

Subhan:  "I beat my head against my keyboard!  What about scientific common sense?  If morality is this mysterious given thing, from beyond space and time—and I don't even see why we should follow it, in that case—but in any case, if morality exists independently of human nature, then isn't it a remarkable coincidence that, say, love is good?"

Obert:  "Coincidence?  How so?"

Subhan:  "Just where on Earth do you think the emotion of love comes from?  If the ancient Greeks had ever thought of the theory of natural selection, they could have looked at the human institution of sexual romance, or parental love for that matter, and deduced in one flash that human beings had evolved—or at least derived tremendous Bayesian evidence for human evolution.  Parental bonds and sexual romance clearly display the signature of evolutionary psychology—they're archetypal cases, in fact, so obvious we usually don't even see it."

Obert:  "But love isn't just about reproduction—"

Subhan:  "Of course not; individual organisms are adaptation-executers, not fitness-maximizers.  But for something independent of humans, morality looks remarkably like godshatter of natural selection.  Indeed, it is far too much coincidence for me to credit.  Is happiness morally preferable to pain?  What a coincidence!  And if you claim that there is any emotion, any instinctive preference, any complex brain circuitry in humanity which was created by some external morality thingy and not natural selection, then you are infringing upon science and you will surely be torn to shreds—science has never needed to postulate anything but evolution to explain any feature of human psychology—"

Obert:  "I'm not saying that humans got here by anything except evolution."

Subhan:  "Then why does morality look so amazingly like a product of an evolved psychology?"

Obert:  "I don't claim perfect access to moral truth; maybe, being human, I've made certain mistakes about morality—"

Subhan:  "Say that—forsake love and life and happiness, and follow some useless damn trend of the universe or whatever—and you will lose every scrap of the moral normality that you once touted as your strong point.  And I will be right here, asking, 'Why even bother?'  It would be a pitiful mind indeed that demanded authoritative answers so strongly, that it would forsake all good things to have some authority beyond itself to follow."

Obert:  "All right... then maybe the reason morality seems to bear certain similarities to our human constitutions, is that we could only perceive morality at all, if we happened, by luck, to evolve in consonance with it."

Subhan:  "Horsemanure."

Obert:  "Fine... you're right, that wasn't very plausible.  Look, I admit you've driven me into quite a corner here.  But even if there were nothing more to morality than preference, I would still prefer to act as morality were real.  I mean, if it's all just preference, that way is as good as anything else—"

Subhan:  "Now you're just trying to avoid facing reality!  Like someone who says, 'If there is no Heaven or Hell, then I may as well still act as if God's going to punish me for sinning.'"

Obert:  "That may be a good metaphor, in fact.  Consider two theists, in the process of becoming atheists.  One says, 'There is no Heaven or Hell, so I may as well cheat and steal, if I can get away without being caught, since there's no God to watch me.'  And the other says, 'Even though there's no God, I intend to pretend that God is watching me, so that I can go on being a moral person.'  Now they are both mistaken, but the first is straying much further from the path."

Subhan:  "And what is the second one's flaw?  Failure to accept personal responsibility!"

Obert:  "Well, and I admit I find that a more compelling argument than anything else you have said.  Probably because it is a moral argument, and it has always been morality, not metaethics, with which I claimed to be concerned.  But even so, after our whole conversation, I still maintain that wanting to murder someone does not make murder right.  Everything that you have said about preference is interesting, but it is ultimately about preference—about minds and what they are designed to desire—and not about this other thing that humans sometimes talk about, 'morality'.  I can just ask Moore's Open Question:  Why should I care about human preferences?  What makes following human preferences right?  By changing a mind, you can change what it prefers; you can even change what it believes to be right; but you cannot change what is right.  Anything you talk about, that can be changed in this way, is not 'right-ness'."

Subhan:  "So you take refuge in arguing from definitions?"

Obert:  "You know, when I reflect on this whole argument, it seems to me that your position has the definite advantage when it comes to arguments about ontology and reality and all that stuff—"

Subhan:  "'All that stuff'?  What else is there, besides reality?"

Obert:  "Okay, the morality-as-preference viewpoint is a lot easier to shoehorn into a universe of quarks.  But I still think the morality-as-given viewpoint has the advantage when it comes to, you know, the actual morality part of it—giving answers that are good in the sense of being morally good, not in the sense of being a good reductionist.  Because, you know, there are such things as moral errors, there is moral progress, and you really shouldn't go around thinking that murder would be right if you wanted it to be right."

Subhan:  "That sounds to me like the logical fallacy of appealing to consequences."

Obert:  "Oh?  Well, it sounds to me like an incomplete reduction—one that doesn't quite add up to normality."

 

Part of The Metaethics Sequence

Next post: "Where Recursive Justification Hits Bottom"

Previous post: "Is Morality Preference?"

Comments

sorted by
magical algorithm
Highlighting new comments since Today at 1:26 PM
Select new highlight date
All comments loaded

Subhan: "You're not escaping that easily! How does a universe in which murder is wrong, differ from a universe in which murder is right? How can you detect the difference experimentally? If the answer to that is 'No', then how does any human being come to know that murder is wrong?" ... Obert: "Because it seems blue, just as murder seems wrong. Just don't ask me what the sky is, or how I can see it."

But we already know why murder seems wrong to us. It's completely explained by a combination of game theory, evolutionary psychology, and memetics. These explanations screen off our apparent moral perceptions from any other influence. In order words, conditioned on these explanations being true, our moral perceptions are independent of (i.e. uncorrelated with) any possible morality-as-given, even if it were to exist.

So there is a stronger argument against Obert than the one Subhan makes. It's not just that we don't know how we can know about what is right, but rather that we know we can't know, at least not through these apparent moral perceptions/intuitions.

Similar problem to born postulates and why we aren't Boltzman Brains.
We are Boltzmann Brains. You simply don't appreciate what restrictions are inherent in specifying the subset of Brains that can be called "we".

Not that this has anything to do with the topic, which everyone is very carefully skating around without addressing: what are operational definitions for right and wrong? When Obert says "Because it seems blue, just as murder seems wrong.", what collection of properties does wrong refer to? For that matter, what does blue refer to?

These questions have very simple and obvious answers which you will never grasp until you force yourself to face the questions. You mean something when you use the terms - you already recognize what is implied when you or someone else uses those terms. Now make that recognition explicit instead of implicit.

Do not "philosophize". That is attempting to understand the territory by making a diagram of map-making. It's adding another layer of analysis between you and the core concept, like an oyster adding a layer of nacre around an irritating bit of sand. You do not need a more complex ontology - you need to abolish the ontology.

"But AGI is [...] not being a judge or whoever writes laws."

If Eliezer turns out to be right about the power of recursive self-improvement, then I wouldn't be so sure.

"If morality exists independently of human nature, then isn't it a remarkable coincidence that, say, love is good?"

I'm going to play Devil's Advocate for a moment here. Anyone, please feel free to answer, but do not interpret the below arguments as correlating with my set of beliefs.

"A remarkable coincidence? Of course not! If we're supposing that this 'stone tablet' has some influence on the universe - and if it exists, it must exert influence, otherwise we wouldn't have any evidence wherewith to be arguing over whether or not it exists - then it had influence on our 'creation', whether (in order to cover all bases) we got here purely through evolution, or via some external manipulation as well. I should think it would be yet stranger if we had human natures that did not accord with such a 'stone tablet'."

By your Devil's logic here, we would expect at least part of human nature to accord with the whole of this 'stone tablet'. I think we could vary the argument to avoid this conclusion. But as written it implies that each 'law' from the 'tablet' has a reflection in human nature, even if perhaps some other part of human nature works against its realization.

This implies that there exists some complicated aspect of human nature we could use to define morality which would give us the same answers as the 'stone tablet'.

Which sounds like that fuzzily-defined "conscience" thing. So suppose I say that this "Stone tablet" is not a literal tablet, but is rather a set of rules that sufficiently advanced lifeforms will tend to accord to? Is this fundamentally different than the opposite side of the argument?

Well, that depends. What does "sufficiently advanced" mean? Does this claim have anything to say about Clippy?

If it doesn't constrain anticipation there, I suspect no difference exists.

Ha! No. I guess I'm using a stricter definition of a "mind" than is used in that post: one that is able to model itself. I recognize the utility of such a generalized definition of intelligence, but I'm talking about a subclass of said intelligences.

Er, why couldn't Clippy model itself? Surely you don't mean that you think Clippy would change its end-goals if it did so (for what reason?)

... Just to check: we're talking about Microsoft Office's Clippy, right?

Actually, this is what we're really talking about, not MS Word constructs or LW roleplayers.

Relationships are real. For example if a plant is "under" a table, that is a fact, not a subjective whim of the observer. So if morality is a relationship, then aliens and man can have different moralities but both be objective, not subjective. The relationship would be between the object sought and the entity seeking it, e.g. murder + man = bad, murder + alien = good.

As with human aesthetic sense, human morality may be approximations and versions of more absolutely definable optimal solutions to information-theoretic, game-theoretic, social, economic, intelligence, signaling, cooperation problems. Therefore it may be likely that an alien race could share some of the same values as we do, because it may turn out that they are "good" solutions for intelligent culture bearing species in general. But there is nothing in the universe in it self that says that theese optimal solutions, or any value what-so-ever can be valuable, and I don´t understand why some atheists even expect there to be something like this. It is minds that give value to certain phenomena, and it usually happens because our emotional circuitry where wired to value something by evolution.

But I so no problem in trying to choose some of the more general looking values evolution has given us and making them even more general and refined. We should do this in such a way as to keep us stable and happy but also so that we have a rich future mind-space to move in, and I think we have already done this in part with Goodness, Truth and Beauty for it´s own sake, but they are not easily definable. It is not easy to find a sweet-spot between a species-specific version and a more general one that is free and pleasing to as many minds as possible. I think we should continue to refine this sweet-spot and work towards some values for their own sake and not only because they give us pleasure. I think this is important and can add up to the stability of an individual as well as a society and an AI. It certainly is a dangerous thing, but I see it as essential at least for a human mind.

One problem may be though that we are wired so that, if we don´t believe that morality or some other value is intrinsically rooted and valued by some great preferably immortal and omnipresent authority(God, the Universe, Nature) then we may have some trouble behaving morally or in accordance with that value, just because we choose to do so. Some strong people may find this easy. But many finds this very hard, and I think that is the prime reason why religion still persists today, even though most people know somewhere deep down that it is a dead practice. And I will admit although not with any pride that I, myself has great trouble with pursuing and actually doing what I value, systematically and with pleasure and discipline, even though I know about all this. It may be that something is particularly wrong in my brain, or it may be quite a wide problem.

Thank you Eli for keeping on enlightening my path, day after day, I have been lurking on you for 2.5 years now, you have totally changed my life, before you, there was no person I could really trust in deep matters, and now, because you are so often right and I have to be so careful with trusting you, I have also become extremely careful in trusting my own intuitions and thought patterns. You learn people to think with rigorous self-critique, extream precision and dedication and mindful focus on the actual territory and possibility-space of study without getting trapped in the usual mind projection fallacy and other biases.

Keep on fighting! Your books will sell well, and it will fuel your goals!

I think that Subhan and Obert may represent two sides of a false dichotomy, namely the idea that there's either one absolute morality for all minds, or it's all subjective. But a third possibility exists - that of objective morality, where the results depend on the physical nature of the being in question, but not their whims.

Dynamically Linked writes: But, it seems pretty obvious, at least to me, that game theory, evolutionary psychology, and memetics are not contingent on anything except mathematics and the environment that we happened to evolve in.

According to Tegmark "there is only mathematics; that is all that exists". Suppose he is right. Then moral truths, if there are any, are (along with all other truths) mathematical truths. Unless you presuppose that moral truths cannot be mathematical truths then you have not ruled out moral truths when you say that so-and-so is not contingent on anything except mathematics and such-and-such. For my part I fail to see why moral truths could not be mathematical truths.

Before I go on, do you actually believe this [Bayesian net diagram] to be the case?

I'm sorry to say that I can't read Bayesian net diagrams. Hopefully I answered your question anyway.

Robin,

Our moral intuitions correspond with moral truths for much the same reason that our rational predictions correspond with more concrete physical truths. A man who ignores reason will stick his hand back in the fire after being burned the first time. Such behavior will kill him, probably sooner rather than later. An man who is blind to good and evil may do quite well for himself, but a society whose citizens ignore virtue will suffer approximately the same fate as the twice-burned fool.

Richard, would you accept the same argument about God, that we know there is a God but don't really understand how we know, but gosh darn it we feel like there must be one so there must be one? Yes we evolved to help kin, and we expect many but hardly all other species to do this as well. But unless we know whether that behavior is moral we don't know if that is a process that makes our moral intuitions correlate with moral truth.

I've thought about Space Cannibals and the like before (i.e. creatures that kill one of the sexes during sexual reproduction). My suspicion is that even if such creatures evolved and survived, by the time they had a civilization, many would be saying to one another, "There really should be a better way..."

Evidence for this is the fact that even now, there are many human beings claiming it is wrong to kill other animals, despite the fact that humans evolved to kill and eat other animals. Likewise, in the ancestral environment, various tribes usually did kill each other rather than cooperate. But this didn't stop them from beginning to cooperate. So I suspect that Space Cannibals would do something similar. And in any case, I would fully admit that murder couldn't in fact be wrong for the Space Cannibals in the same way it is for us, even if there is an external moral truth.

In answer to Robin's question, assuming that morality exists, it probably has a number of purposes. And if one of the purposes is to preserve things in existence (i.e. moral truths correspond roughly with what is necessary to preserve things), then of course there will be a selection pressure to perceive moral truth. The disclaimer should not be needed, but this is not in any way a claim that it is moral to maximize inclusive genetic fitness.

If somebody said to me "morality is just what we do." If they presented evidence that the whole apparatus of their moral philosophy was a coherent description of some subset of human psychology and sociology. Then that would be enough for me. It's just a description of a physical system. Human morality would be what human animals do. Moral responsibility wouldn't be problematic; moral responsibility could be as physical as gravity if it were psychologically and sociologically real. "I have a moral responsibility" would be akin to "I can lift 200 lbs." The brain is complicated, sure, but so are muscles and bones and motor control. That wouldn't make it a preference or a mere want either. That's probably where we're headed. But I don't think metaethics is the interesting problem. The deeper problem is, I think, the empirical one: Do humans really display this sort of morality?

Phillip, you're the one who brought up "deification," in response to my one-line comment, which you seem to have read a lot into. My second comment was intended to be humorous. I apologize for the extent to which I contributed to this misunderstanding.

I think that we need a much better explanation of this word "mind". Supposedly mind space contains a -1 for every 1, but that simply sounds like system space. I honestly think that the ontology has to go deeper here before progress is possible. Similar problem to born postulates and why we aren't Boltzman Brains.

And to answer Obert's objection that Subhan's position doesn't quite add up to normality: before we knew game theory, evolutionary psychology, and memetics, nothing screened off our moral perceptions/intuitions from a hypothesized objective moral reality, so that was perhaps the best explanation available, given what we knew back then. And since that was most of human history, it's no surprise that morality-as-given feels like normality. But given what we know today, does it still make sense to insist that our meta-theory of morality add up to that normality?

It is sometimes argued that happiness is good and suffering is bad. (This is tentatively my own view, but explaining the meaning of "good" and "bad," defending its truth, and expanding the view to account for the additional categories of "right" and "wrong" is beyond the scope of this comment.)

If this is true, then depending on what kind of truth it is, it may also be true in all possible worlds--and a fortiori, on all possible planets in this universe. Furthermore, if it is true on all possible planets that happiness is good and suffering is bad, this does not preclude the possibility that on some planets, murder and theft might be the best way toward everyone's happiness, while compassion and friendship might lead to everyone's misery. In such a case, then to whatever degree this scenario is theoretically possible, compassion and friendship would be bad, while murder and theft would be good.

Hence we can see that it might be the case that normative ethical truths differ from one planet to another, but metaethical truths are the same everywhere. On one level, this is a kind of moral relativism, but it is also based on an absolute principle. I personally think it is a plausible view, while I admit that this comment provides little exposition and no defense of it.

It would be a pitiful mind indeed that demanded authoritative answers so strongly, that it would forsake all good things to have some authority beyond itself to follow.

Having an authority to follow might actually be that mind's one good thing. Maybe it really likes having such authority beyond itself.

While humans obviously don't consider that their only good this and there is human variation (I don't think everyone values it, I'm only certain at least a few do), it seems pretty clear to me that one of our good things, as in stuff we find worth bothering to acquire, is authority beyond ourselves.

If we find such a hypothetical authority we might feel a strong preference to move in its direction.

« Then you believe in universally compelling arguments processed by a ghost in the machine. For every possible mind whose utility function assigns terminal value +1, mind design space contains an equal and opposite mind whose utility function assigns terminal value -1. »

That's true but that doesn't say anything about the sustainability of that given mind design (the possibility for the mind design to survive, either by having the individual to survive, or to create new individuals with similar mind designs).

A mind design that would value its own death very positively would commit suicide (unless unable to) and stop existing. A mind design that value pain (in the abstract meaning, dealing damages to its own physical support) would be much less likely to survive.

Many other possible mind designs would become completely tied in contradictions and chaos, and be unable to achieve anything. Think about a mind without "modus ponens", or a mind for which the values are chosen randomly and change every nanosecond...

And it gets even narrower when the mind is not alone, but has to live in society with other similar minds.

So while mind design space contains the exact opposite of every possible terminal values, the "coherent mind design space" (minds that are coherent enough to not self-destruct, not paralyze themselves and not collapse into chaos) is probably more limited.

But that still is so big that it contains minds like the paper-clip optimizer wanting the tile the solar system with paper-clips, no doubt about that one.

>Geddes, if you can't keep the condescension out of your comments - just present the raw arguments, if you have any - then I'll have to ban you here, too. Just FYI. Also, your comments should be shorter.

But all I'm asking for is an explanation as to why decision theory works? Perhaps someone like R.Hanson could explain?

After all, I know (admittedly only in very general terms) what the explanation for thermodynamics is (the underlying explanation is in the concepts of mechanics- energy, force etc.). Also, I know (in general), why probability theory works (the underlying explanation is in the concepts of algebra - relations, functions etc).

To shorten my comment then, here's the question I'm asking:

Science: Thermodynamics
Explanatory Justifcation: Mechanics

Science: Probability Theory
Explanatory Justification: Algebra

Science: Decision Theory
Explanatory Justification: ???????

And I'm postulating that knowing the explanatory justification for decision theory would give you the answers to the questions on morality. I'm also guessing that no-one here can provide that justfication ;)

Terren Suydam: "So genetics is not the whole story, and that's what I mean by group selection."

I use the term "multilevel selection" for what you are describing. I agree it has been important.

E.g., there has been selection between different species. Species with genomes that supported rapid adaptation to changing environments and that supported quick diversification when expanding into new niches spread far and wide. (Beetles have been extremely successful with around 350,000 known species.) Other specie branches died out. The genetic mechanisms and the animal body plans that persist to the present are the winners of a long between specie selection process.

My intuition is that selection operating at the individual level, whether genetic or cultural, suffices to produce cooperation and moral behavior. Multilevel selection probably played a supporting role.

There may be a selection effect in that cultures with guns are more likely to persist, but that's different from saying that selection pressures play a really important role in designing the particular features of a culture.

That's what I'm saying - selection pressures are important in determining cultural features, because those features in turn determine a culture's viability. The global-level organization of a culture - including its moral code, political organization, and other important social structures - are key considerations in what makes a culture healthy or stable, and thus competitive in an arena of limited resources. Keep in mind, I think these ideas lend themselves more easily to ancient history, in which the boundaries between cultures were so much clearer, rather than the globalized cultures of our modern world.

A lot of these ideas come from Jared Diamond's fascinating Guns, Germs, and Steel, which talks in depth about cultural evolution throughout human history.

Terren, I would doubt that changes between cultures are best explained by an evolutionary process--cf. "No Evolutions for Corporations or Nanodevices." There may be a selection effect in that cultures with guns are more likely to persist, but that's different from saying that selection pressures play a really important role in designing the particular features of a culture. So I am given to understand.

In studying evolutionary biology, "group-selection" has a specific meaning, an individual sacrifices its own fitness in order to improve the group fitness.

I think it's quite limiting to think strictly in terms of genetics, because there is more than one level of description going on when it comes to selection pressure.

It is interesting to take that step back and view the culture as an individual. The human super-organism (e.g., a tribe, or more generally, a culture) competes with others for resources. It consumes, metabolizes, and excretes, which is to say that it lowers entropy locally and raises it globally. With others, it fights, defends, cooperates, and merges/assimilates. It gets sick, fights the antigens, heals itself, or dies. New super-organisms are spun off of or otherwise born from parents. We may look at the "evolution of language" through the lens of the human super-organism. Language is the DNA of culture, to make a rough analogy.

The dynamics that propagate the super-organism are not reducible to genetics. It's a different level of description, because culture emerges from the interaction of large numbers of individuals. And you can't deny that if one culture has guns and confronts another that doesn't, that dynamic is going to place a harsh selective pressure indeed on the culture without the firepower. So genetics is not the whole story, and that's what I mean by group selection.

Nobody told Galileo and Newton what the rules generating the world's behavior were, but they were able to go a long way toward figuring them out.
They started with rigorously-defined phenomena. Galileo was famous for taking careful and exacting measurements - he was able to figure out some of the world's rules because he could notice the relationships between precisely-defined and -measured things.

Nobody gets anywhere, most especially in philosophy, without rigorous definitions of the relevant concepts. You rely on vague, intuitive understandings, and you accomplish nothing.

Terren Suydam: "The first is that one has to adopt the group-selection stance."

(Technical jargon nitpick.)

In studying evolutionary biology, "group-selection" has a specific meaning, an individual sacrifices its own fitness in order to improve the group fitness. I.e., individual loss for a group gain. E.g., suppose you have a species that consists of many small family groups. Suppose a mutation produces a self-sacrificing individual in one of the groups. His fitness is slightly lower but his family group fitness is higher. His group tends to grow faster than other groups. So his group produces more splinter groups, some of which will have his alleles. Within any one group his allele tends to die out, but the overall population frequency of the allele increases due to the increased number of splinter groups containing the allele. This is an example of group-selection.

Much more common is cooperation that doesn't lower the individual's fitness. In this case it is win-win, individual gain and group gain. Symbiosis is an example where the cooperation is between different species. Both individuals gain so it is not an example of group-selection.

There are a few known examples of group-selection but they tend to be the rare exception, not the rule. Often something appears to be group-selection but on closer analysis turns out to be regular selection. E.g., suppose a hunter shares his meat with the tribe. He isn't lowering his fitness because he already has enough meat for himself. He is publicly displaying his prowess as a food provider which increases his mating success. Thus his generosity directly improves his fitness. His generosity is a fitness increasing status display.

Cooperation can and usually does arise through regular selfish selection.

(I see EY also noted this.)

There aren't necessarily any common elements, besides utterly trivial ones.

Maybe, maybe not. You won't know without looking. You have to start somewhere.

If you look at examples of misspelled words in various languages and examine their individual properties, you won't find what unites them in a category.

But then, what about correctly spelled words? There will be many observable systematic relationships between those. I happen to think you have the analogy backwards. In the good/evil dichotomy, it is the evil acts, not the not-evil acts, which are narrowly defined and systematically related (I think). If you try to find what is in common between the not-evil acts, those are the acts which have nothing in particular in common. Meanwhile, in the well-spelled/misspelled dichotomy, it is the correctly-spelled words that are narrowly defined and systematically related. In short, I think morality is fundamentally a narrow set of prohibitions rather than a narrow set of requirements. In contrast, the rules of spelling form a narrow set of requirements.

But whether you are right or I am right is something that we won't know without looking.

You have to understand their relationship to the spelling rules in the various languages - rules which themselves are likely to be incompatible and mutually incoherent - to understand what properties make them examples of 'misspelled words'.

Nobody told Galileo and Newton what the rules generating the world's behavior were, but they were able to go a long way toward figuring them out. And isn't that what science is? If you claim that the science can't start without knowing the rules first, then aren't you asserting that science is hopeless?

Quick comment: Terren Suydam's version of "evolutionary psychology" is not the academically accepted one. Conventional academic evolutionary explanations of morality rely on neither group selection nor selection on cultures.

Be that as it may, I would have to say that an explanation of morality made strictly in terms of the academically accepted version of evolutionary psychology is not possible. I'm not trying to redefine the term - just saying what else would be necessary to make an explanation of morality on that basis possible.

Richard, once we can see how to eliminate math or science views, then it doesn't seem particularly puzzling that people can generate plausible views to consider. The obvious hypothesis is that they generate many views in their mind, apply crude but effective filters, and then only tell others about the few best ones. So of course the views they propose will fare far better than average when tested against consistency and data.

Quick comment: Terren Suydam's version of "evolutionary psychology" is not the academically accepted one. Conventional academic evolutionary explanations of morality rely on neither group selection nor selection on cultures.

As I said: you can start with an ostensive definition by listing examples of evil acts. Then you can find common elements.

There aren't necessarily any common elements, besides utterly trivial ones. If you look at examples of misspelled words in various languages and examine their individual properties, you won't find what unites them in a category. You have to understand their relationship to the spelling rules in the various languages - rules which themselves are likely to be incompatible and mutually incoherent - to understand what properties make them examples of 'misspelled words'.

We need the concept of morality itself, the rules that define the incompatible and mutually incoherent rule systems that are examples of 'morality'. Looking at examples of 'evil acts' isn't going to cut it.

It seems the ultimate confusion here is that we are talking about instrumental values . . . before agreeing on terminal values . . .

If we could agree on some well-defined goal, e.g. maximization of human happiness, we could much more easily theorize on whether a particular case of murder would benefit or harm that goal.

denis bider, under the CEV plan for singularity, no human has to give an unambiguous definition or enumeration of his or her terminal values before the launch of the seed of the superintelligence. Consequently, those who lean toward the CEV plan feel much freer to regard themselves as having hundreds of terminal values. Consequently, refraining from murder might easily be a terminal value for them.

Defn. "Murder" is killing under particular circumstances, e.g., not by uniformed soldiers during a war, not in self-defense, not by accident.

My comment is not charitable enough towards the CEVists. I ask the moderator to delete it, I will now submit a replacement.

Great dialog, which I think can be summarized in Nietzsche's aphorism: "Morality is the herd-instinct in the individual."

Actually, I think the dialog could have been a lot shorter if it became clear earlier on that preference (as in morality-as-preference) referred not to individual preference but the "preference of the collective". Which is to say, morality is determined by evolutionary psychology. There are however two assumptions built into the evolutionary psychological explanation of morality which ought to be made explicit.

The first is that one has to adopt the group-selection stance. As in, groups in which morals evolved had higher stability (its members were more likely to survive and procreate). If we focus only on the selfish individual, then it's obvious that morals make no sense.

The second assumption is that evolution, in humans, bifurcated into both physical and cultural domains. This is because morality is almost certainly not determined by genetics. At best, genetics predisposes us psychologically to accept morality (whatever that means), but the content of morality is too well specified to be so determined. Thus we have to assume a selection process that selects groups on the basis of culture as well as genetics. This is actually a common sense notion, that for example groups that develop better weapons will dominate or eliminate other groups.

It seems the ultimate confusion here is that we are talking about instrumental values . . . before agreeing on terminal values . . .

If we could agree on some well-defined goal, e.g. maximization of human happiness, we could much more easily theorize on whether a particular case of murder would benefit or harm that goal.

denis bider, I would not be surprised to learn that refraining from murder is a terminal value for Eliezer. Eliezer's writings imply that he has hundreds of terminal values: he cannot even enumerate them all.

Defn. "Murder" is killing under particular circumstances, e.g., not by uniformed soldiers during a war, not in self-defense, not by accident.

Are you willing to have a neverending discussion, with everyone talking past each other, and no working definition for the central concept we're supposed to be examining?

I'm not in charge of the discussion, so it's not a question of what I'm willing to do. I've told you how to get the starting definition you're looking for. As I said: you can start with an ostensive definition by listing examples of evil acts. Then you can find common elements. For example, it might become apparent, after surveying them, that evil acts have in common that they all have victims against whose will the evil acts were committed and who are harmed by the evil acts. It might also become apparent that the evil acts involved one or another form of transgression or trespass against certain boundaries. You might like to study what the boundaries are.

Are you willing to accept an ostensive and potentially erroneous definition of morality that may very well be subject to revision as knowledge improves?
That's how knowledge works, Constant. Everything we think we know may turn out to be wrong, and any conclusions will probably end up being revised later.

Are you willing to have a neverending discussion, with everyone talking past each other, and no working definition for the central concept we're supposed to be examining?

Basic scientific methodology - you can't study what you can't produce a provisional definition for. Once you have that, you can learn more about what's defined, but you don't get anywhere without that starting point.

The first concepts that more less denoted, say, water, may have included things which today we would reject as not water (e.g., possibly clear alcohol), failed to distinguish water from things dissolved in the water, and excluded forms of water (such as steam and ice). The very first definitions of water were probably ostensive definitions (this here is water, that is water) rather than descriptive or explanatory definitions. The definitions were subject to revision as knowledge improved.

Are you willing to accept an ostensive and potentially erroneous definition of morality that may very well be subject to revision as knowledge improves? One is easy enough to supply by listing a bunch of acts currently believed to be evil, then listing a bunch of believed-to-be morally neutral acts, and pointing out that the first group is evil and the second group isn't. Would that be satisfactory?

Is it an arbitrary grouping, or do we use the label to refer to certain properties that things in that grouping possess?

I think the better question is, do recognized examples of evil have something in common - never mind what we intend by the label. Maybe by the label "water" we initially intended "Chronos's tears" or some such useless thing. The intention isn't necessarily of any particular interest. You are interested in scientific inquiry into morality, yes? - seeing as you talk about "scientific methodology." Science studies the properties of things in themselves independently of whatever nonsense ideas we might have about them; if you want to study our intents then become a philosopher, not a scientist.

Anyway, this question - do examples of evil have something in common - is something for the scientists to answer, no? It doesn't need to be answered before scientific inquiry begins.

Still doesn't tell us what 'moral' means. We've just changed the category we stick that label on. What defines 'moral emotions'? Is it an arbitrary grouping, or do we use the label to refer to certain properties that things in that grouping possess? People seem to use the term in the latter way - so what properties are they?

Basic scientific methodology - you can't study what you can't produce a provisional definition for. Once you have that, you can learn more about what's defined, but you don't get anywhere without that starting point.

I gestured at one possible answer to that question. A situation has a moral dimension if it engages moral emotions - which can presumably be listed.

FYI, it's physics that is fundamental. Math is deeper than our theories of physics - it's deeper than all our theories, because it makes up the languages we use to create and express them - but physics itself is deeper than everything.

Similarly, if there are any moral givens, I would agree that they have to result in real empirical differences.
Correct.
Again, we could never percieve moral givens directly (since they are abstract)
Not correct - everything we perceive is equally abstract. There are different kinds of abstractions defined by their interrelationships in a hierarchy. A 'virtual' computer is just as abstract as the hardware it's running on, but the hardware is on a deeper level of the hierarchy. But that's moving towards another topic.

What properties make a claim about morality, as opposed to something else? What is the basic definition of 'morality'? Answering that question is the necessary first step to resolving the issue. It is remarkable how little anyone here cares about doing that.