How to be a deontological consequentialist

Dec 7 JDN 2461017

As is commonly understood, there are two main branches of normative ethics:

  • Deontology, on which morality consists in following rules and fulfilling obligations, and
  • Consequentialism, on which morality consists in maximizing good consequences.

The conflict between them has raged for centuries, with Kantians leading the deontologists and utilitarians leading the consequentialists. Both theories seem to have a lot of good points, but neither can decisively defeat the other.

I think this is because they are both basically correct.

In their strongest forms, deontology and consequentialism are mutually contradictory; but it turns out that you can soften each of them a little bit, and the results become compatible.

To make deontology a little more consequentialist, let’s ask a simple question:

What makes a rule worth following?

I contend that the best answer we have is “because following that rule would make the world better off than not following that rule”. (Even Kantians pretty much have to admit this: What maxim could you will to be an absolute law? Only a law that would yield good outcomes.)

That is, the ultimate justification of a sound deontology would be fundamentally consequentialist.

But lest the consequentialists get too smug, we can also ask them another question, which is a bit subtler:

How do you know which actions will ultimately have good consequences?

Sure, if we were omniscient beings who could perfectly predict the consequences of our actions across the entire galaxy on into the indefinite future, we could be proper act utilitarians who literally choose every single action according to a calculation of the expected utility.

But in practice, we have radical uncertainty about the long-term consequences of our actions, and can generally only predict the immediate consequences.

That leads to the next question:

Would you really want to live in a world where people optimized immediate consequences?

I contend that you would not, that such a world actually sounds like a dystopian nightmare.

Immediate consequences say that if a healthy person walks into a hospital and happens to have compatible organs for five people who need donations, we should kill that person, harvest their organs, and give them to the donors. (This is the organ transplant variant of the Trolley Problem.)

Basically everyone recognizes that this is wrong. But why is it wrong? That’s thornier. One pretty convincing case is that a systematic policy of this kind would undermine trust in hospitals and destroy the effectiveness of healthcare in general, resulting in disastrous consequences far outweighing the benefit of saving those five people. But those aren’t immediate consequences, and indeed, it’s quite difficult to predict exactly how many crazy actions like this it would take to undermine people’s trust in hospitals, just how much it would undermine that trust, or exactly what the consequences of that lost trust would be.

So it seems like it’s actually better to have a rule about this.

This makes us into rule utilitarians, who instead of trying to optimize literally every single action—which requires information we do not have and never will—we instead develop a system of rules that we can follow, heuristics that will allow us to get better outcomes generally even if they can’t be guaranteed to produce the best possible outcome in any particular case.

That is, the output of a sophisticated consequentialism is fundamentally deontological.

We have come at the question of normative ethics from two very different directions, but the results turned out basically the same:

We should follow the rules that would have the best consequences.

The output of our moral theory is rules, like deontology; but its fundamental justification is based on outcomes, like consequentialism.

In my experience, when I present this account to staunch deontologists, they are pretty much convinced by it. They’re prepared to give up the fundamental justification to consequences if it allows them to have their rules.

The resistance I get is mainly from staunch consequentialists, who insist that it’s not so difficult to optimize individual actions, and so we should just do that instead of making all these rules.

So it is to those consequentialists, particularly those who say “rule utilitarianism collapses into act utilitarianism”, to whom the rest of the post is addressed.

First, let me say that I agree.

In the ideal case of omniscient, perfectly-benevolent, perfectly-rational agents, rule utilitarianism mathematically collapses into act utilitarianism. That is a correct theorem.

However, we do not live in the ideal case of omniscient, perfectly-benevolent, perfectly-rational agents. We are not even close to that ideal case; we will never be close to that ideal case. Indeed, I think part of the problem here is that you fail to fully grasp the depth and width of the chasm between here and there. Even a galactic civilization of a quintillion superhuman AIs would still not be close to that ideal case.

Quite frankly, humans aren’t even particularly good at forecasting what will make themselves happy.

There are massive errors and systematic biases in human affective forecasting.

One of the post important biases is impact bias: People systematically overestimate the impact of individual events on their long-term happiness. Some of this seems to be just due to focus: Paying attention to a particular event exaggerates its importance in your mind, and makes it harder for you to recall other events that might push your emotions in a different direction. Another component is called immune neglect: people fail to account for their own capacity to habituate to both pleasant and unpleasant experiences. (This effect is often overstated: It’s a common misconception that lottery winners are no happier than they were before. No, they absolutely are happier, on average; they’re just not as much happier as they predicted themselves to be.)

People also use inconsistent time discounting: $10 today is judged as better than $11 tomorrow, but $10 in 364 days is not regarded as better than $11 in 365 days—so if I made a decision a year ago, I’d want to change it now. (The correct answer, by the way, is to take the $11; a discount rate of 10% per day is a staggering 120,000,000,000,000,000% APR—seriously; check it yourself—so you’d better not be discounting at that rate, unless you’re literally going to die before tomorrow.)

Now, compound that with the fact that different human beings come at the world from radically different perspectives and with radically different preferences.

How good do you think we are at predicting what will make other people happy?

Damn right: We’re abysmal.

Basically everyone assumes that what they want and what they would feel is also what other people will want and feel—which, honestly, explains a lot about politics. As a result, my prediction of your feelings is more strongly correlated with my prediction of my feelings than it is with your actual feelings.

The impact bias is especially strong when forecasting other people’s feelings in response to our own actions: We tend to assume that other people care more about what we do than they actually care—and this seems to be a major source of social anxiety.

People also tend to overestimate the suffering of others, and are generally willing to endure more pain than they are willing to inflict upon others. (This one seems like it might be a good thing!)

Even when we know people well, we can still be totally blindsided by their emotional reactions. We’re just really awful at this.

Does this just mean that morality is hopeless? We have no idea what we’re doing?

Fortunately, no. Because while no individual can correctly predict or control the outcomes of particular actions, the collective action of well-designed institutions can in fact significantly improve the outcomes of policy.

This is why we have things like the following:

  • Laws
  • Courts
  • Regulations
  • Legislatures
  • Constitutions
  • Newspapers
  • Universities

These institutions—which form the backbone of liberal democracy—aren’t simply arbitrary. They are the result of hard-fought centuries, a frothing, volatile, battle-tested mix of intentional design and historical evolution.

Are these institutions optimal? Good heavens, no!

But we have no idea what optimal institutions look like, and probably never will. (Those galaxy-spanning AIs will surely have a better system than this; but even theirs probably won’t be optimal.) Instead, what we are stuck with are the best institutions we’ve come up with so far.

Moreover, we do have very clear empirical evidence at this point that some form of liberal democracy with a mixed economy is the best system we’ve got so far. One can reasonably debate whether Canada is doing better or worse than France, or whether the system in Denmark could really be scaled to the United States, or just what the best income tax rates are; but there is a large, obvious, and important difference between life in a country like Canada or Denmark and life in a country like Congo or Afghanistan.

Indeed, perhaps there is no better pair to compare than North and South Korea: Those two countries are right next to each other, speak the same language, and started in more or less the same situation; but the south got good institutions and the north got bad ones, and now the difference between them couldn’t be more stark. (Honestly, this is about as close as we’re ever likely to get of a randomized controlled experiment in macroeconomics.)

People in South Korea now live about as well as some of the happiest places in the world; their GDP per capita PPP is about $65,000 per year, roughly the same as Canada. People in North Korea live about as poorly as it is possible for humans to live, subject to totalitarian oppression and living barely above subsistence; their GDP per capita PPP is estimated to be $600 per year—less than 1% as much.

The institutions of South Korea are just that much better.

Indeed, there’s one particular aspect of good institutions that seems really important, yet is actually kind of hard to justify in act-utilitarian terms:

Why is freedom good?

A country’s level of freedom is almost perfectly correlated with its overall level of happiness and development. (Yes, even on this measure, #ScandinaviaIsBetter.)

But why? In theory, letting people do whatever they want could actually lead to really bad outcomes—and indeed, occasionally it does. There’s even a theorem that liberty is incompatible with full Pareto-efficiency. But all the countries with the happiest people seem to have a lot of liberty, and indeed the happiest ones seem to have the most. How come?

My answer:

Personal liberty is a technology for heuristic utility maximization.

In the ideal case, we wouldn’t really need personal liberty; you could just compel everyone to do whatever is optimal all the time, and that would—by construction—be optimal. It might even be sort of nice: You don’t need to make any difficult decisions, you can just follow the script and know that everything will turn out for the best.

But since we don’t know what the optimal choice is—even in really simple cases, like what you should eat for lunch tomorrow—we can’t afford to compel people in this way. (It would also be incredibly costly to implement such totalitarian control, but that doesn’t stop some governments from trying!)

Then there are disagreements: What I think is optimal may not be what you think is optimal, and in truth we’re probably both wrong (but one of us may be less wrong).

And that’s not even getting into conflicts of interest: We aren’t just lacking in rationality, we’re also lacking in benevolence. Some people are clearly much more benevolent than others, but none of us are really 100% selfless. (Sadly, I think some people are 100% selfish.)

In fact, this is a surprisingly deep question:

Would the world be better if we were selfless?

Could there be actually some advantage in aggregate to having some degree of individual self-interest?

Here are some ways that might hold, just off the top of my head:

  • Partial self-interest supports an evolutionary process of moral and intellectual development that otherwise would be stalled or overrun by psychopaths—see my post on Rousseaus and Axelrods
  • Individuals have much deeper knowledge of their own preferences than anyone else’s, and thus can optimize them much better. (Think about it: This is true even of people you know very well. Otherwise, why would we ever need to ask our spouses one of the most common questions in any marriage: “Honey, what do you want for dinner tonight?”)
  • Self-interest allows for more efficient economic incentives, and thus higher overall productivity.

Of course, total selfishness is clearly not optimal—that way lies psychopathy. But some degree of selfishness might actually be better for long-term aggregate outcomes than complete altruism, and this is to some extent an empirical question.

Personal liberty solves a lot of these problems: Since people are best at knowing their own preferences, let people figure out on their own what’s good for them. Give them the freedom to live the kind of life they want to live, within certain reasonable constraints to prevent them from causing great harm to others or suffering some kind of unrecoverable mistake.

This isn’t exactly a new idea; it’s basically the core message of John Stuart Mill’s On Liberty (which I consider a good candidate for the best book every written—seriously, it beats the Bible by a light-year). But by putting it in more modern language, I hope to show that deontology and consequentialism aren’t really so different after all.

And indeed, for all its many and obvious flaws, freedom seems to work pretty well—at least as well as anything we’ve tried.

Moore’s “naturalistic fallacy”

Jan 12 JDN 2460688

In last week’s post I talked about some of the arguments against ethical naturalism, which have sometimes been called “the naturalistic fallacy”.

The “naturalistic fallacy” that G.E. Moore actually wrote about is somewhat subtler; it says that there is something philosophically suspect about defining something non-natural in terms of natural things—and furthermore, it says that “good” is not a natural thing and so cannot be defined in terms of natural things. For Moore, “good” is not something that can be defined with recourse to facts about psychology, biology or mathematics; “good” is simply an indefinable atomic concept that exists independent of all other concepts. As such Moore was criticizing moral theories like utilitarianism and hedonism that seek to define “good” in terms of “pleasure” or “lack of pain”; for Moore, good cannot have a definition in terms of anything except itself.

My greatest problem with this position is less philosophical than linguistic; how does one go about learning a concept that is so atomic and indefinable? When I was a child, I acquired an understanding of the word “good” that has since expanded as I grew in knowledge and maturity. I need not have called it “good”: had I been raised in Madrid, I would have called it bueno; in Beijing, hao; in Kyoto, ii; in Cairo, jaiid; and so on.

I’m not even sure if all these words really mean exactly the same thing, since each word comes with its own cultural and linguistic connotations. A vast range of possible sounds could be used to express this concept and related concepts—and somehow I had to learn which sounds were meant to symbolize which concepts, and what relations were meant to hold between them. This learning process was highly automatic, and occurred when I was very young, so I do not have great insight into its specifics; but nonetheless it seems clear to me that in some sense I learned to define “good” in terms of things that I could perceive. No doubt this definition was tentative, and changed with time and experience; indeed, I think all definitions are like this. Perhaps my knowledge of other concepts, like “pleasure”, “happiness”, “hope” and “justice”, is interconnected with “good” in such a way that none can be defined separately from the others—indeed perhaps language itself is best considered a network of mutually-reinforcing concepts, each with some independent justification and some connection to other concepts, not a straightforward derivation from more basic atomic notions. If you wish, call me a “foundherentist” in the tradition of Susan Haack; I certainly do think that all beliefs have some degree of independent justification by direct evidence and some degree of mutual justification by coherence. Haack uses the metaphor of a crossword puzzle, but I prefer Alison Gopnik’s mathematical model of a Bayes net. In any case, I had to learn about “good” somehow. Even if I had some innate atomic concept of good, we are left to explain two things: First, how I managed to associate that innate atomic concept with my sense experiences, and second, how that innate atomic concept got in my brain in the first place. If it was genetic, it must have evolved; but it could only have evolved by phenotypic interaction with the external environment—that is, with natural things. We are natural beings, made of natural material, evolved by natural selection. If there is a concept of “good” encoded into my brain either by learning or instinct or whatever combination, it had to get there by some natural mechanism.

The classic argument Moore used to support this position is now called the Open Question Argument; it says, essentially, that we could take any natural property that would be proposed as the definition of “good” and call it X, and we could ask: “Sure, that’s X, but is it good?” The idea is that since we can ask this question and it seems to make sense, then X cannot be the definition of “good”. If someone asked, “I know he is an unmarried man, but is he a bachelor?” or “I know that has three sides, but is it a triangle?” we would think that they didn’t understand what they were talking about; but Moore argues that for any natural property, “I know that is X, but is it good?” is still a meaningful question. Moore uses two particular examples, X = “pleasant” and X = “what we desire to desire”; and indeed those fit what he is saying. But are these really very good examples?

One subtle point that many philosophers make about this argument is that science can discover identities between things and properties that are not immediately apparent. We now know that water is H2O, but until the 19th century we did not know this. So we could perfectly well imagine someone asking, “I know that’s H2O, but is it water?” even though in fact water is H2O and we know this. I think this sort of argument would work for some very complicated moral claims, like the claim that constitutional democracy is good; I can imagine someone who was quite ignorant of international affairs asking: “I know that it’s constitutional democracy, but is that good?” and be making sense. This is because the goodness of constitutional democracy isn’t something conceptually necessary, it is an empirical result based on the fact that constitutional democracies are more peaceful, fair, egalitarian, and prosperous than other governmental systems. In fact, it may even be only true relative to other systems we know of; perhaps there is an as-yet-unimagined governmental system that is better still. No one thinks that constitutional democracy is a definition of moral goodness. And indeed, I think few would argue that H2O is the definition of water; instead the definition of water is something like “that wet stuff we need to drink to survive” and it just so happens that this turns out to be H2O. If someone asked “is that wet stuff we need to drink to survive really water?” he would rightly be thought talking nonsense; that’s just what water means.

But if instead of the silly examples Moore uses, we take a serious proposal that real moral philosophers have suggested, it’s not nearly so obvious that the question is open. From Kant: “Yes, that is our duty as rational beings, but is it good?” From Mill: “Yes, that increases the amount of happiness and decreases the amount of suffering in the world, but is it good?” From Aristotle: “Yes, that is kind, just, and fair, but is it good?” These do sound dangerously close to talking nonsense! If someone asked these questions, I would immediately expect an explanation of what they were getting at. And if no such explanation was forthcoming, I would, in fact, be led to conclude that they literally don’t understand what they’re talking about.

I can imagine making sense of “I know that has three sides, but is it a triangle?”in some bizarre curved multi-dimensional geometry. Even “I know he is an unmarried man, but is he a bachelor?” makes sense if you are talking about a celibate priest. Very rarely do perfect synonyms exist in natural languages, and even when they do they are often unstable due to the effects of connotations. None of this changes the fact that bachelors are unmarried men, triangles have three sides, and yes, goodness involves fulfilling rational duties, alleviating suffering, and being kind and just (Deontology, consequentialism, and virtue theory are often thought to be distinct and incompatible; I’m convinced they amount to the same thing, which I’ll say more about in later posts.).

This line of reasoning has led some philosophers (notably Willard Quine) to deny the existence of analytic truths altogether; on Quine’s view even “2+2=4” isn’t something we can deduce directly from the meaning of the symbols. This is clearly much too strong; no empirical observation could ever lead us to deny 2+2=4. In fact, I am convinced that all mathematical truths are ultimately reducible to tautologies; even “the Fourier transform of a Gaussian is Gaussian” is ultimately a way of saying in compact jargon some very complicated statement that amounts to A=A. This is not to deny that mathematics is useful; of course mathematics is tremendously useful, because this sort of compact symbolic jargon allows us to make innumerable inferences about the world and at the same time guarantee that these inferences are correct. Whenever you see a Gaussian and you need its Fourier transform (I know, it happens a lot, right?), you can immediately know that the result will be a Gaussian; you don’t have to go through the whole derivation yourself. We are wrong to think that “ultimately reducible to a tautology” is the same as “worthless and trivial”; on the contrary, to realize that mathematics is reducible to tautology is to say that mathematics is undeniable, literally impossible to coherently deny. At least the way I use the words, the statement “Happiness is good and suffering is bad” is pretty close to that same sort of claim; if you don’t agree with it, I sense that you honestly don’t understand what I mean.

In any case, I see no more fundamental difficulty in defining “good” than I do in defining any concept, like “man”, “tree”, “multiplication”, “green” or “refrigerator”; and nor do I see any point in arguing about the semantics of definition as an approach to understanding moral truth. It seems to me that Moore has confused the map with the territory, and later authors have confused him with Hume, to all of our detriment.

Against deontology

Aug 6 JDN 2460163

In last week’s post I argued against average utilitarianism, basically on the grounds that it devalues the lives of anyone who isn’t of above average happiness. But you might be tempted to take these as arguments against utilitarianism in general, and that is not my intention.

In fact I believe that utilitarianism is basically correct, though it needs some particular nuances that are often lost in various presentations of it.

Its leading rival is deontology, which is really a broad class of moral theories, some a lot better than others.

What characterizes deontology as a class is that it uses rules, rather than consequences; an act is just right or wrong regardless of its consequences—or even its expected consequences.

There are certain aspects of this which are quite appealing: In fact, I do think that rules have an important role to play in ethics, and as such I am basically a rule utilitarian. Actually trying to foresee all possible consequences of every action we might take is an absurd demand far beyond the capacity of us mere mortals, and so in practice we have no choice but to develop heuristic rules that can guide us.

But deontology says that these are no mere heuristics: They are in fact the core of ethics itself. Under deontology, wrong actions are wrong even if you know for certain that their consequences will be good.

Kantian ethics is one of the most well-developed deontological theories, and I am quite sympathetic to Kantian ethics In fact I used to consider myself one of its adherents, but I now consider that view a mistaken one.

Let’s first dispense with the views of Kant himself, which are obviously wrong. Kant explicitly said that lying is always, always, always wrong, and even when presented with obvious examples where you could tell a small lie to someone obviously evil in order to save many innocent lives, he stuck to his guns and insisted that lying is always wrong.

This is a bit anachronistic, but I think this example will be more vivid for modern readers, and it absolutely is consistent with what Kant wrote about the actual scenarios he was presented with:

You are living in Germany in 1945. You have sheltered a family of Jews in your attic to keep them safe from the Holocaust. Nazi soldiers have arrived at your door, and ask you: “Are there any Jews in this house?” Do you tell the truth?

I think it’s utterly, agonizingly obvious that you should not tell the truth. Exactly what you should do is less obvious: Do you simply lie and hope they buy it? Do you devise a clever ruse? Do you try to distract them in some way? Do you send them on a wild goose chase elsewhere? If you could overpower them and kill them, should you? What if you aren’t sure you can; should you still try? But one thing is clear: You don’t hand over the Jewish family to the Nazis.

Yet when presented with similar examples, Kant insisted that lying is always wrong. He had a theory to back it up, his Categorical Imperative: “Act only according to that maxim whereby you can at the same time will that it should become a universal law.”

And, so his argument goes: Since it would be obviously incoherent to say that everyone should always lie, lying is wrong, and you’re never allowed to do it. He actually bites that bullet the size of a Howitzer round.

Modern deontologists—even though who consider themselves Kantians—are more sophisticated than this. They realize that you could make a rule like “Never lie, except to save the life of an innocent person.” or “Never lie, except to stop a great evil.” Either of these would be quite adequate to solve this particular dilemma. And it’s absolutely possible to will that these would be universal laws, in the sense that they would apply to anyone. ‘Universal’ doesn’t have to mean ‘applies equally to all possible circumstances’.

There are also a couple of things that deontology does very well, which are worth preserving. One of them is supererogation: The idea that some acts are above and beyond the call of duty, that something can be good without being obligatory.

This is something most forms of utilitarianism are notoriously bad at. They show us a spectrum of worlds from the best to the worst, and tell us to make things better. But there’s nowhere we are allowed to stop, unless we somehow manage to make it all the way to the best possible world.

I find this kind of moral demand very tempting, which often leads me to feel a tremendous burden of guilt. I always know that I could be doing more than I do. I’ve written several posts about this in the past, in the hopes of fighting off this temptation in myself and others. (I am not entirely sure how well I’ve succeeded.)

Deontology does much better in this regard: Here are some rules. Follow them.

Many of the rules are in fact very good rules that most people successfully follow their entire lives: Don’t murder. Don’t rape. Don’t commit robbery. Don’t rule a nation tyrannically. Don’t commit war crimes.

Others are oft more honored in the breach than the observance: Don’t lie. Don’t be rude. Don’t be selfish. Be brave. Be generous. But a well-developed deontology can even deal with this, by saying that some rules are more important than others, and thus some sins are more forgivable than others.

Whereas a utilitarian—at least, anything but a very sophisticated utilitarian—can only say who is better and who is worse, a deontologist can say who is good enough: who has successfully discharged their moral obligations and is otherwise free to live their life as they choose. Deontology absolves us of guilt in a way that utilitarianism is very bad at.

Another good deontological principle is double-effect: Basically this says that if you are doing something that will have bad outcomes as well as good ones, it matters whether you intend the bad one and what you do to try to mitigate it. There does seem to be a morally relevant difference between a bombing that kills civilians accidentally as part of an attack on a legitimate military target, and a so-called “strategic bombing” that directly targets civilians in order to maximize casualties—even if both occur as part of a justified war. (Both happen a lot—and it may even be the case that some of the latter were justified. The Tokyo firebombing and atomic bombs on Hiroshima and Nagasaki were very much in the latter category.)

There are ways to capture this principle (or something very much like it) in a utilitarian framework, but like supererogation, it requires a sophisticated, nuanced approach that most utilitarians don’t seem willing or able to take.

Now that I’ve said what’s good about it, let’s talk about what’s really wrong with deontology.

Above all: How do we choose the rules?

Kant seemed to think that mere logical coherence would yield a sufficiently detailed—perhaps even unique—set of rules for all rational beings in the universe to follow. This is obviously wrong, and seems to be simply a failure of his imagination. There is literally a countably infinite space of possible ethical rules that are logically consistent. (With probability 1 any given one is utter nonsense: “Never eat cheese on Thursdays”, “Armadillos should rule the world”, and so on—but these are still logically consistent.)

If you require the rules to be simple and general enough to always apply to everyone everywhere, you can narrow the space substantially; but this is also how you get obviously wrong rules like “Never lie.”

In practice, there are two ways we actually seem to do this: Tradition and consequences.

Let’s start with tradition. (It came first historically, after all.) You can absolutely make a set of rules based on whatever your culture has handed down to you since time immemorial. You can even write them down in a book that you declare to be the absolute infallible truth of the universe—and, amazingly enough, you can get millions of people to actually buy that.

The result, of course, is what we call religion. Some of its rules are good: Thou shalt not kill. Some are flawed but reasonable: Thou shalt not steal. Thou shalt not commit adultery. Some are nonsense: Thou shalt not covet thy neighbor’s goods.

And some, well… some rules of tradition are the source of many of the world’s most horrific human rights violations. Thou shalt not suffer a witch to live (Exodus 22:18). If a man also lie with mankind, as he lieth with a woman, both of them have committed an abomination: they shall surely be put to death; their blood shall be upon them (Leviticus 20:13).

Tradition-based deontology has in fact been the major obstacle to moral progress throughout history. It is not a coincidence that utilitarianism began to become popular right before the abolition of slavery, and there is an even more direct casual link between utilitarianism and the advancement of rights for women and LGBT people. When the sole argument you can make for moral rules is that they are ancient (or allegedly handed down by a perfect being), you can make rules that oppress anyone you want. But when rules have to be based on bringing happiness or preventing suffering, whole classes of oppression suddenly become untenable. “God said so” can justify anything—but “Who does it hurt?” can cut through.

It is an oversimplification, but not a terribly large one, to say that the arc of moral history has been drawn by utilitarians dragging deontologists kicking and screaming into a better future.

There is a better way to make rules, and that is based on consequences. And, in practice, most people who call themselves deontologists these days do this. They develop a system of moral rules based on what would be expected to lead to the overall best outcomes.

I like this approach. In fact, I agree with this approach. But it basically amounts to abandoning deontology and surrendering to utilitarianism.

Once you admit that the fundamental justification for all moral rules is the promotion of happiness and the prevention of suffering, you are basically a rule utilitarian. Rules then become heuristics for promoting happiness, not the fundamental source of morality itself.

I suppose it could be argued that this is not a surrender but a synthesis: We are looking for the best aspects of deontology and utilitarianism. That makes a lot of sense. But I keep coming back to the dark history of traditional rules, the fact that deontologists have basically been holding back human civilization since time immemorial. If deontology wants to be taken seriously now, it needs to prove that it has broken with that dark tradition. And frankly the easiest answer to me seems to be to just give up on deontology.

Against average utilitarianism

Jul 30 JDN 2460156

Content warning: Suicide and suicidal ideation

There are two broad strands of utilitarianism, known as average utilitarianism and total utilitarianism. As utilitarianism, both versions concern themselves with maximizing happiness and minimizing suffering. And for many types of ethical question, they yield the same results.

Under average utilitarianism, the goal is to maximize the average level of happiness minus suffering: It doesn’t matter how many people there are in the world, only how happy they are.

Under total utilitarianism, the goal is to maximize the total level of happiness minus suffering: Adding another person is a good thing, as long as their life is worth living.

Mathematically, its the difference between taking the sum of net happiness (total utilitarianism), and taking that sum and dividing it by the population (average utilitarianism).

It would make for too long a post to discuss the validity of utilitarianism in general. Overall I will say briefly that I think utilitarianism is basically correct, but there are some particular issues with it that need to be resolved, and usually end up being resolved by heading slightly in the direction of a more deontological ethics—in short, rule utilitarianism.

But for today, I want to focus on the difference between average and total utilitarianism, because average utilitarianism is a very common ethical view despite having appalling, horrifying implications.

Above all: under average utilitarianism, if you are considering suicide, you should probably do it.

Why? Because anyone who is considering suicide is probably of below-average happiness. And average utilitarianism necessarily implies that anyone who expects to be of below-average happiness should be immediately killed as painlessly as possible.

Note that this does not require that your life be one of endless suffering, so that it isn’t even worth going on living. Even a total utilitarian would be willing to commit suicide, if their life is expected to be so full of suffering that it isn’t worth going on.

Indeed, I suspect that most actual suicidal ideation by depressed people takes this form: My life will always be endless suffering. I will never be happy again. My life is worthless.

The problem with such suicidal ideation is not the ethical logic, which is valid: If indeed your existence from this point forward would be nothing but endless suffering, suicide actually makes sense. (Imagine someone who is being held in a dungeon being continually mercilessly tortured with no hope of escape; it doesn’t seem unreasonable for them to take a cyanide pill.) The problem is the prediction, which says that your life from this point forward will be nothing but endless suffering. Most people with depression do, eventually, feel better. They may never be quite as happy overall as people who aren’t depressed, but they do, in fact, have happy times. And most people who considered suicide but didn’t go through with it end up glad that they went on living.

No, an average utilitarian says you should commit suicide as long as your happiness is below average.

We could be living in a glorious utopia, where almost everyone is happy almost all the time, and people are only occasionally annoyed by minor inconveniences—and average utilitarianism would say that if you expect to suffer a more than average rate of such inconveniences, the world would be better off if you ceased to exist.

Moreover, average utilitarianism says that you should commit suicide if your life is expected to get worse—even if it’s still going to be good, adding more years to your life will just bring your average happiness down. If you had a very happy childhood and adulthood is going just sort of okay, you may as well end it now.

Average utilitarianism also implies that we should bomb Third World countries into oblivion, because their people are less happy than ours and thus their deaths will raise the population average.

Are there ways an average utilitarian can respond to these problems? Perhaps. But every response I’ve seen is far too weak to resolve the real problem.

One approach would be to say that the killing itself is bad, or will cause sufficient grief as to offset the loss of the unhappy person. (An average utilitarian is inherently committed to the claim that losing an unhappy person is itself an inherent good. There is something to be offset.)

This might work for the utopia case: The grief from losing someone you love is much worse than even a very large number of minor inconveniences.

It may even work for the case of declining happiness over your lifespan: Presumably some other people would be sad to lose you, even if they agreed that your overall happiness is expected to gradually decline. Then again, if their happiness is also expected to decline… should they, too, shuffle off this mortal coil?

But does it work for the question of bombing? Would most Americans really be so aggrieved at the injustice of bombing Burundi or Somalia to oblivion? Most of them don’t seem particularly aggrieved at the actual bombings of literally dozens of countries—including, by the way, Somalia. Granted, these bombings were ostensibly justified by various humanitarian or geopolitical objectives, but some of those justifications (e.g. Kosovo) seem a lot stronger than others (e.g. Grenada). And quite frankly, I care more about this sort of thing than most people, and I still can’t muster anything like the same kind of grief for random strangers in a foreign country that I feel when a friend or relative dies. Indeed, I can’t muster the same grief for one million random strangers in a foreign country that I feel for one lost loved one. Human grief just doesn’t seem to work that way. Sometimes I wish it did—but then, I’m not quite sure what our lives would be like in such a radically different world.

Moreover, the whole point is that an average utilitarian should consider it an intrinsically good thing to eliminate the existence of unhappy people, as long as it can be done swiftly and painlessly. So why, then, should people be aggrieved at the deaths of millions of innocent strangers they know are mostly unhappy? Under average utilitarianism, the greatest harm of war is the survivors you leave, because they will feel grief—so your job is to make sure you annihilate them as thoroughly as possible, presumably with nuclear weapons. Killing a soldier is bad as long as his family is left alive to mourn him—but if you kill an entire country, that’s good, because their country was unhappy.

Enough about killing and dying. Let’s talk about something happier: Babies.

At least, total utilitarians are happy about babies. When a new person is brought into the world, a total utilitarian considers this a good thing, as long as the baby is expected to have a life worth living and their existence doesn’t harm the rest of the world too much.

I think that fits with most people’s notions of what is good. Generally the response when someone has a baby is “Congratulations!” rather than “I’m sorry”. We see adding another person to the world as generally a good thing.

But under average utilitarianism, babies must reach a much higher standard in order to be a good thing. Your baby only deserves to exist if they will be happier than average.

Granted, this is the average for the whole world, so perhaps First World people can justify the existence of their children by pointing out that unless things go very badly, they should end up happier than the world average. (Then again, if you have a family history of depression….)

But for Third World families, quite the opposite: The baby may well bring joy to all around them, but unless that joy is enough to bring someone above the global average, it would still be better if the baby did not exist. Adding one more person of moderately-low happiness will just bring the world average down.

So in fact, on a global scale, an average utilitarian should always expect that babies are nearly as likely to be bad as they are good, unless we have some reason to think that the next generation would be substantially happier than this one.

And while I’m not aware of anyone who sincerely believes that we should nuke Third World countries for their own good, I have heard people speak this way about population growth in Third World countries: such discussions of “overpopulation” are usually ostensibly about ecological sustainability, even though the ecological impact of First World countries is dramatically higher—and such talk often shades very quickly into eugenics.

Of course, we wouldn’t want to say that having babies is always good, lest we all be compelled to crank out as many babies as possible and genuinely overpopulate the world. But total utilitarianism can solve this problem: It’s worth adding more people to the world unless the harm of adding those additional people is sufficient to offset the benefit of adding another person whose life is worth living.

Moreover, total utilitarianism can say that it would be good to delay adding another person to the world, until the situation is better. Potentially this delay could be quite long: Perhaps it is best for us not to have too many children until we can colonize the stars. For now, let’s just keep our population sustainable while we develop the technology for interstellar travel. If having more children now would increase the risk that we won’t ever manage to colonize distant stars, total utilitarianism would absolutely say we shouldn’t do it.

There’s also a subtler problem here, which is that it may seem good for any particular individual to have more children, but the net result is that the higher total population is harmful. Then what I think is happening is that we are unaware of, or uncertain about, or simply inattentive to, the small harm to many other people caused by adding one new person to the world. Alternatively, we may not be entirely altruistic, and a benefit that accrues to our own family may be taken as greater than a harm that accrues to many other people far away. If we really knew the actual marginal costs and benefits, and we really agreed on that utility function, we would in fact make the right decision. It’s our ignorance or disagreement that makes us fail, not total utilitarianism in principle. In practice, this means coming up with general rules that seem to result in a fair and reasonable outcome, like “families who want to have kids should aim for two or three”—and again we’re at something like rule utilitarianism.

Another case average utilitarianism seems tempting is in resolving the mere addition paradox.

Consider three possible worlds, A, B, and C:

In world A, there is a population of 1 billion, and everyone is living an utterly happy, utopian life.

In world B, there is a population of 1 billion living in a utopia, and a population of 2 billion living mediocre lives.

In world C, there is a population of 3 billion living good, but not utopian, lives.

The mere addition paradox is that, to many people, world B seems worse than world A, even though all we’ve done is add 2 billion people whose lives are worth living.

Moreover, many people seem to think that the ordering goes like this:


World B is better than world A, because all we’ve done is add more people whose lives are worth living.

World C is better than world B, because it’s fairer, and overall happiness is higher.

World A is better than world C, because everyone is happier, and all we’ve done is reduce the population.


This is intransitive: We have A > C > B > A. Our preferences over worlds are incoherent.

Average utilitarianism resolves this by saying that A > C is true, and C > B is true—but it says that B > A is false. Since average happiness is higher in world A, A > B.

But of course this results in the conclusion that if we are faced with world B, we should do whatever we can to annihilate the 2 billion extra unhappy people, so that we can get to world A. And the whole point of this post is that this is an utterly appalling conclusion we should immediately reject.

What does total utilitarianism say? It says that indeed C > B and B > A, but it denies that A > C. Rather, since there are more people in world C, it’s okay that people aren’t quite as happy.

Derek Parfit argues that this leads to what he calls the “repugnant conclusion”: If we keep increasing the population by a large amount while decreasing happiness by a small amount, the best possible world ends up being one where population is utterly massive but our lives are only barely worth living.

I do believe that total utilitarianism results in this outcome. I can live with that.

Under average utilitarianism, the best possible world is precisely one person who is immortal and absolutely ecstatic 100% of the time. Adding even one person who is not quite that happy will make things worse.

Under total utilitarianism, adding more people who are still very happy would be good, even if it makes that one ecstatic person a bit less ecstatic. And adding more people would continue to be good, as long as it didn’t bring the average down too quickly.

If you find this conclusion repugnant, as Parfit does, I submit that it is because it is difficult to imagine just how large a population we are talking about. Maybe putting some numbers on it will help.

Let’s say the happiness level of an average person in the world today is 35 quality-adjusted life years—our life expectancy of 70, times an average happiness level of 0.5.

So right now we have a world of 8 billion people at 35 QALY, for a total of 280 TQALY. (That’s tera-QALY, 1 trillion QALY.)

(Note: I’m not addressing inequality here. If you believe that a world where one person has 100 QALY and another has 50 QALY is worse than one where both have 75 QALY, you should adjust your scores accordingly—which mainly serves to make the current world look worse, due to our utterly staggering inequality. In fact I think I do not believe this—in my view, the problem is not that happiness is unequal, but that staggering inequality of wealth makes much greater suffering among the poor in exchange for very little happiness among the rich.)

Average utilitarianism says that we should eliminate the less happy people, so we can raise the average QALY higher, maybe to something like 60. I’ve already said why I find this appalling.

So now consider what total utilitarianism asks of us. If we could raise that figure above 280 TQALY, we should. Say we could increase our population to 10 billion, at the cost of reducing average happiness to 30 QALY; should we? Yes, we should, because that’s 300 TQALY.

But notice that in this scenario we’re still 85% as happy as we were. That doesn’t sound so bad. Parfit is worried about a scenario where our lives are barely worth living. So let’s consider what that would require.

“Barely worth living” sounds like maybe 1 QALY. This wouldn’t mean we all live exactly one year; that’s not sustainable, because babies can’t have babies. So it would be more like a life expectancy of 33, with a happiness of 0.03—pretty bad, but still worth living.

In that case, we would need to raise our population over 800 billion to make it better than our current existence. We must colonize at least 100 other planets and fill them as full as we’ve filled Earth.

In fact, I think this 1 QALY life was something like that human beings had at the dawn of agriculture (which by some estimates was actually worse than ancient hunter-gatherer life; we were sort of forced into early agriculture, rather than choosing it because it was better): Nasty, brutish, and short, but still, worth living.

So, Parfit’s repugnant conclusion is that filling 100 planets with people who live like the ancient Babylonians would be as good as life on Earth is now? I don’t really see how this is obviously horrible. Certainly not to the same degree that saying we should immediately nuke Somalia is obviously horrible.

Moreover, total utilitarianism absolutely still says that if we can make those 800 billion people happier, we should. A world of 800 billion people each getting 35 QALY is 100 times better than the way things are now—and doesn’t that seem right, at least?


Yet if you indeed believe that copying a good world 100 times gives you a 100 times better world, you are basically committed to total utilitarianism.

There are actually other views that would allow you to escape this conclusion without being an average utilitarian.

One way, naturally, is to not be a utilitarian. You could be a deontologist or something. I don’t have time to go into that in this post, so let’s save it for another time. For now, let me say that, historically, utilitarianism has led the charge in positive moral change, from feminism to gay rights, from labor unions to animal welfare. We tend to drag stodgy deontologists kicking and screaming toward a better world. (I vaguely recall an excellent tweet on this, though not who wrote it: “Yes, historically, almost every positive social change has been spearheaded by utilitarians. But sometimes utilitarianism seems to lead to weird conclusions in bizarre thought experiments, and surely that’s more important!”)

Another way, which has gotten surprisingly little attention, is to use an aggregating function that is neither a sum nor an average. For instance, you could add up all utility and divide by the square root of population, so that larger populations get penalized for being larger, but you aren’t simply trying to maximize average happiness. That does seem to still tell some people to die even though their lives were worth living, but at least it doesn’t require us to exterminate all who are below average. And it may also avoid the conclusion Parfit considers repugnant, by making our galactic civilization span 10,000 worlds. Of course, why square root? Why not a cube root, or a logarithm? Maybe the arbitrariness is why it hasn’t been seriously considered. But honestly, I think dividing by anything is suspicious; how can adding someone else who is happy ever make things worse?

But if I must admit that a sufficiently large galactic civilization would be better than our current lives, even if everyone there is mostly pretty unhappy? That’s a bullet I’m prepared to bite. At least I’m not saying we should annihilate everyone who is unhappy.

The only thing necessary for the triumph of evil is that good people refuse to do cost-benefit analysis

July 27, JDN 2457597

My title is based on a famous quote often attributed to Edmund Burke, but which we have no record of him actually saying:

The only thing necessary for the triumph of evil is that good men do nothing.

The closest he actually appears to have written is this:

When bad men combine, the good must associate; else they will fall one by one, an unpitied sacrifice in a contemptible struggle.

Burke’s intended message was about the need for cooperation and avoiding diffusion of responsibility; then his words were distorted into a duty to act against evil in general.

But my point today is going to be a little bit more specific: A great deal of real-world evils would be eliminated if good people were more willing to engage in cost-benefit analysis.

As discussed on Less Wrong awhile back, there is a common “moral” saying which comes from the Talmud (if not earlier; and of course it’s hardly unique to Judaism), which gives people a great warm and fuzzy glow whenever they say it:

Whoever saves a single life, it is as if he had saved the whole world.

Yet this is in fact the exact opposite of moral. It is a fundamental, insane perversion of morality. It amounts to saying that “saving a life” is just a binary activity, either done or not, and once you’ve done it once, congratulations, you’re off the hook for the other 7 billion. All those other lives mean literally nothing, once you’ve “done your duty”.

Indeed, it would seem to imply that you can be a mass murderer, as long as you save someone else somewhere along the line. If Mao Tse-tung at some point stopped someone from being run over by a car, it’s okay that his policies killed more people than the population of Greater Los Angeles.

Conversely, if anything you have ever done has resulted in someone’s death, you’re just as bad as Mao; in fact if you haven’t also saved someone somewhere along the line and he has, you’re worse.

Maybe this is how you get otherwise-intelligent people saying such insanely ridiculous things as George W. Bush’s crimes are uncontroversially worse than Osama bin Laden’s.” (No, probably not, since Chomsky at least feigns something like cost-benefit analysis. I’m not sure what his failure mode is, but it’s probably not this one in particular. “Uncontroversially”… you keep using that word…)

Cost-benefit analysis is actually a very simple concept (though applying it in practice can be mind-bogglingly difficult): Try to maximize the good things minus the bad things. If an action would increase good things more than bad things, do it; if it would increase bad things more than good things, don’t do it.

What it replaces is simplistic deontological reasoning about “X is always bad” or “Y is always good”; that’s almost never true. Even great evils can be justified by greater goods, and many goods are not worth having because of the evils they would require to achieve. We seem to want all our decisions to have no downside, perhaps because that would resolve our cognitive dissonance most easily; but in the real world, most decisions have an upside and a downside, and it’s a question of which is larger.

Why is it that so many people—especially good people—have such an aversion to cost-benefit analysis?

I gained some insight into this by watching a video discussion from an online Harvard course taught by Michael Sandel (which is free, by the way, if you’d like to try it out). He was leading the discussion Socratically, which is in general a good method of teaching—but like anything else can be used to teach things that are wrong, and is in some ways more effective at doing so because it has a way of making students think they came up with the answers on their own. He says something like, “Do we really want our moral judgments to be based on cost-benefit analysis?” and gives some examples where people made judgments using cost-benefit analysis to support his suggestion that this is something bad.

But of course his examples are very specific: They all involve corporations using cost-benefit analysis to maximize profits. One of them is the Ford Pinto case, where Ford estimated the cost to them of a successful lawsuit, multiplied by the probability of such lawsuits, and then compared that with the cost of a total recall. Finding that the lawsuits were projected to be cheaper, they opted for that result, and thereby allowed several people to be killed by their known defective product.

Now, it later emerged that Ford Pintos were not actually especially dangerous, and in fact Ford didn’t just include lawsuits but also a standard estimate of the “value of a statistical human life”, and as a result of that their refusal to do the recall was probably the completely correct decision—but why let facts get in the way of a good argument?

But let’s suppose that all the facts had been as people thought they were—the product was unsafe and the company was only interested in their own profits. We don’t need to imagine this hypothetically; this is clearly what actually happened with the tobacco industry, and indeed with the oil industry. Is that evil? Of course it is. But not because it’s cost-benefit analysis.

Indeed, the reason this is evil is the same reason most things are evil: They are psychopathically selfish. They advance the interests of those who do them, while causing egregious harms to others.

Exxon is apparently prepared to sacrifice millions of lives to further their own interests, which makes them literally no better than Mao, as opposed to this bizarre “no better than Mao” that we would all be if the number of lives saved versus killed didn’t matter. Let me be absolutely clear; I am not speaking in hyperbole when I say that the board of directors of Exxon is morally no better than Mao. No, I mean they literally are willing to murder 20 million people to serve their own interests—more precisely 10 to 100 million, by WHO estimates. Maybe it matters a little bit that these people will be killed by droughts and hurricanes rather than by knives and guns; but then, most of the people Mao killed died of starvation, and plenty of the people killed by Exxon will too. But this statement wouldn’t have the force it does if I could not speak in terms of quantitative cost-benefit analysis. Killing people is one thing, and most industries would have to own up to it; being literally willing to kill as many people as history’s greatest mass murderers is quite anotherand yet it is true of Exxon.

But I can understand why people would tend to associate cost-benefit analysis with psychopaths maximizing their profits; there are two reasons for this.

First, most neoclassical economists appear to believe in both cost-benefit analysis and psychopathic profit maximization. They don’t even clearly distinguish their concept of “rational” from the concept of total psychopathic selfishness—hence why I originally titled this blog “infinite identical psychopaths”. The people arguing for cost-benefit analysis are usually economists, and economists are usually neoclassical, so most of the time you hear arguments for cost-benefit analysis they are also linked with arguments for horrifically extreme levels of selfishness.

Second, most people are uncomfortable with cost-benefit analysis, and as a result don’t use it. So, most of the cost-benefit analysis you’re likely to hear is done by terrible human beings, typically at the reins of multinational corporations. This becomes self-reinforcing, as all the good people don’t do cost-benefit analysis, so they don’t see good people doing it, so they don’t do it, and so on.

Therefore, let me present you with some clear-cut cases where cost-benefit analysis can save millions of lives, and perhaps even save the world.

Imagine if our terrorism policy used cost-benefit analysis; we wouldn’t kill 100,000 innocent people and sacrifice 4,400 soldiers fighting a war that didn’t have any appreciable benefit as a bizarre form of vengeance for 3,000 innocent people being killed. Moreover, we wouldn’t sacrifice core civil liberties to prevent a cause of death that’s 300 times rarer than car accidents.

Imagine if our healthcare policy used cost-benefit analysis; we would direct research funding to maximize our chances of saving lives, not toward the form of cancer that is quite literally the sexiest. We would go to a universal healthcare system like the rest of the First World, and thereby save thousands of additional lives while spending less on healthcare.

With cost-benefit analysis, we would reform our system of taxes and subsidies to internalize the cost of carbon emissions, most likely resulting in a precipitous decline of the oil and coal industries and the rapid rise of solar and nuclear power, and thereby save millions of lives. Without cost-benefit analysis, we instead get unemployed coal miners appearing on TV to grill politicians about how awful it is to lose your job even though that job is decades obsolete and poisoning our entire planet. Would eliminating coal hurt coal miners? Yes, it would, at least in the short run. It’s also completely, totally worth it, by at least a thousandfold.

We would invest heavily in improving our transit systems, with automated cars or expanded rail networks, thereby preventing thousands of deaths per year—instead of being shocked and outraged when an automated car finally kills one person, while manual vehicles in their place would have killed half a dozen by now.

We would disarm all of our nuclear weapons, because the risk of a total nuclear apocalypse is not worth it to provide some small increment in national security above our already overwhelming conventional military. While we’re at it, we would downsize that military in order to save enough money to end world hunger.

And oh by the way, we would end world hunger. The benefits of doing so are enormous; the costs are remarkably small. We’ve actually been making a great deal of progress lately—largely due to the work of development economists, and lots and lots of cost-benefit analysis. This process involves causing a lot of economic disruption, making people unemployed, taking riches away from some people and giving them to others; if we weren’t prepared to bear those costs, we would never get these benefits.

Could we do all these things without cost-benefit analysis? I suppose so, if we go through the usual process of covering of our ears whenever a downside is presented and amplification whenever an upside is presented, until we can more or less convince ourselves that there is no downside even though there always is. We can continue having arguments where one side presents only downsides, the other side presents only upsides, and then eventually one side prevails by sheer numbers, and it could turn out to be the upside team (or should I say “tribe”?).

But I think we’d progress a lot faster if we were honest about upsides and downsides, and had the courage to stand up and say, “Yes, that downside is real; but it’s worth it.” I realize it’s not easy to tell a coal miner to his face that his job is obsolete and killing people, and I don’t really blame Hillary Clinton for being wishy-washy about it; but the truth is, we need to start doing that. If we accept that costs are real, we may be able to mitigate them (as Hillary plans to do with a $30 billion investment in coal mining communities, by the way); if we pretend they don’t exist, people will still get hurt but we will be blind to their suffering. Or worse, we will do nothing—and evil will triumph.

Why it matters that torture is ineffective

JDN 2457531

Like “longest-ever-serving Speaker of the House sexually abuses teenagers” and “NSA spy program is trying to monitor the entire telephone and email system”, the news that the US government systematically tortures suspects is an egregious violation that goes to the highest levels of our government—that for some reason most Americans don’t particularly seem to care about.

The good news is that President Obama signed an executive order in 2009 banning torture domestically, reversing official policy under the Bush Administration, and then better yet in 2014 expanded the order to apply to all US interests worldwide. If this is properly enforced, perhaps our history of hypocrisy will finally be at its end. (Well, not if Trump wins…)

Yet as often seems to happen, there are two extremes in this debate and I think they’re both wrong.
The really disturbing side is “Torture works and we have to use it!” The preferred mode of argumentation for this is the “ticking time bomb scenario”, in which we have some urgent disaster to prevent (such as a nuclear bomb about to go off) and torture is the only way to stop it from happening. Surely then torture is justified? This argument may sound plausible, but as I’ll get to below, this is a lot like saying, “If aliens were attacking from outer space trying to wipe out humanity, nuclear bombs would probably be justified against them; therefore nuclear bombs are always justified and we can use them whenever we want.” If you can’t wait for my explanation, The Atlantic skewers the argument nicely.

Yet the opponents of torture have brought this sort of argument on themselves, by staking out a position so extreme as “It doesn’t matter if torture works! It’s wrong, wrong, wrong!” This kind of simplistic deontological reasoning is very appealing and intuitive to humans, because it casts the world into simple black-and-white categories. To show that this is not a strawman, here are several different people all making this same basic argument, that since torture is illegal and wrong it doesn’t matter if it works and there should be no further debate.

But the truth is, if it really were true that the only way to stop a nuclear bomb from leveling Los Angeles was to torture someone, it would be entirely justified—indeed obligatory—to torture that suspect and stop that nuclear bomb.

The problem with that argument is not just that this is not our usual scenario (though it certainly isn’t); it goes much deeper than that:

That scenario makes no sense. It wouldn’t happen.

To use the example the late Antonin Scalia used from an episode of 24 (perhaps the most egregious Fictional Evidence Fallacy ever committed), if there ever is a nuclear bomb planted in Los Angeles, that would literally be one of the worst things that ever happened in the history of the human race—literally a Holocaust in the blink of an eye. We should be prepared to cause extreme suffering and death in order to prevent it. But not only is that event (fortunately) very unlikely, torture would not help us.

Why? Because torture just doesn’t work that well.

It would be too strong to say that it doesn’t work at all; it’s possible that it could produce some valuable intelligence—though clear examples of such results are amazingly hard to come by. There are some social scientists who have found empirical results showing some effectiveness of torture, however. We can’t say with any certainty that it is completely useless. (For obvious reasons, a randomized controlled experiment in torture is wildly unethical, so none have ever been attempted.) But to justify torture it isn’t enough that it could work sometimes; it has to work vastly better than any other method we have.

And our empirical data is in fact reliable enough to show that that is not the case. Torture often produces unreliable information, as we would expect from the game theory involved—your incentive is to stop the pain, not provide accurate intel; the psychological trauma that torture causes actually distorts memory and reasoning; and as a matter of fact basically all the useful intelligence obtained in the War on Terror was obtained through humane interrogation methods. As interrogation experts agree, torture just isn’t that effective.

In principle, there are four basic cases to consider:

1. Torture is vastly more effective than the best humane interrogation methods.

2. Torture is slightly more effective than the best humane interrogation methods.

3. Torture is as effective as the best humane interrogation methods.

4. Torture is less effective than the best humane interrogation methods.

The evidence points most strongly to case 4, which would mean that torture is a no-brainer; if it doesn’t even work as well as other methods, it’s absurd to use it. You’re basically kicking puppies at that point—purely sadistic violence that accomplishes nothing. But the data isn’t clear enough for us to rule out case 3 or even case 2. There is only one case we can strictly rule out, and that is case 1.

But it was only in case 1 that torture could ever be justified!

If you’re trying to justify doing something intrinsically horrible, it’s not enough that it has some slight benefit.

People seem to have this bizarre notion that we have only two choices in morality:

Either we are strict deontologists, and wrong actions can never be justified by good outcomes ever, in which case apparently vaccines are morally wrong, because stabbing children with needles is wrong. Tto be fair, some people seem to actually believe this; but then, some people believe the Earth is less than 10,000 years old.

Or alternatively we are the bizarre strawman concept most people seem to have of utilitarianism, under which any wrong action can be justified by even the slightest good outcome, in which case all you need to do to justify slavery is show that it would lead to a 1% increase in per-capita GDP. Sadly, there honestly do seem to be economists who believe this sort of thing. Here’s one arguing that US chattel slavery was economically efficient, and some of the more extreme arguments for why sweatshops are good can take on this character. Sweatshops may be a necessary evil for the time being, but they are still an evil.

But what utilitarianism actually says (and I consider myself some form of nuanced rule-utilitarian, though actually I sometimes call it “deontological consequentialism” to emphasize that I mean to synthesize the best parts of the two extremes) is not that the ends always justify the means, but that the ends can justify the means—that it can be morally good or even obligatory to do something intrinsically bad (like stabbing children with needles) if it is the best way to accomplish some greater good (like saving them from measles and polio). But the good actually has to be greater, and it has to be the best way to accomplish that good.

To see why this later proviso is important, consider the real-world ethical issues involved in psychology experiments. The benefits of psychology experiments are already quite large, and poised to grow as the science improves; one day the benefits of cognitive science to humanity may be even larger than the benefits of physics and biology are today. Imagine a world without mood disorders or mental illness of any kind; a world without psychopathy, where everyone is compassionate; a world where everyone is achieving their full potential for happiness and self-actualization. Cognitive science may yet make that world possible—and I haven’t even gotten into its applications in artificial intelligence.

To achieve that world, we will need a great many psychology experiments. But does that mean we can just corral people off the street and throw them into psychology experiments without their consent—or perhaps even their knowledge? That we can do whatever we want in those experiments, as long as it’s scientifically useful? No, it does not. We have ethical standards in psychology experiments for a very good reason, and while those ethical standards do slightly reduce the efficiency of the research process, the reduction is small enough that the moral choice is obviously to retain the ethics committees and accept the slight reduction in research efficiency. Yes, randomly throwing people into psychology experiments might actually be slightly better in purely scientific terms (larger and more random samples)—but it would be terrible in moral terms.

Along similar lines, even if torture works about as well or even slightly better than other methods, that’s simply not enough to justify it morally. Making a successful interrogation take 16 days instead of 17 simply wouldn’t be enough benefit to justify the psychological trauma to the suspect (and perhaps the interrogator!), the risk of harm to the falsely accused, or the violation of international human rights law. And in fact a number of terrorism suspects were waterboarded for months, so even the idea that it could shorten the interrogation is pretty implausible. If anything, torture seems to make interrogations take longer and give less reliable information—case 4.

A lot of people seem to have this impression that torture is amazingly, wildly effective, that a suspect who won’t crack after hours of humane interrogation can be tortured for just a few minutes and give you all the information you need. This is exactly what we do not find empirically; if he didn’t crack after hours of talk, he won’t crack after hours of torture. If you literally only have 30 minutes to find the nuke in Los Angeles, I’m sorry; you’re not going to find the nuke in Los Angeles. No adversarial interrogation is ever going to be completed that quickly, no matter what technique you use. Evacuate as many people to safe distances or underground shelters as you can in the time you have left.

This is why the “ticking time-bomb” scenario is so ridiculous (and so insidious); that’s simply not how interrogation works. The best methods we have for “rapid” interrogation of hostile suspects take hours or even days, and they are humane—building trust and rapport is the most important step. The goal is to get the suspect to want to give you accurate information.

For the purposes of the thought experiment, okay, you can stipulate that it would work (this is what the Stanford Encyclopedia of Philosophy does). But now all you’ve done is made the thought experiment more distant from the real-world moral question. The closest real-world examples we’ve ever had involved individual crimes, probably too small to justify the torture (as bad as a murdered child is, think about what you’re doing if you let the police torture people). But by the time the terrorism to be prevented is large enough to really be sufficient justification, it (1) hasn’t happened in the real world and (2) surely involves terrorists who are sufficiently ideologically committed that they’ll be able to resist the torture. If such a situation arises, of course we should try to get information from the suspects—but what we try should be our best methods, the ones that work most consistently, not the ones that “feel right” and maybe happen to work on occasion.

Indeed, the best explanation I have for why people use torture at all, given its horrible effects and mediocre effectiveness at best is that it feels right.

When someone does something terrible (such as an act of terrorism), we rightfully reduce our moral valuation of them relative to everyone else. If you are even tempted to deny this, suppose a terrorist and a random civilian are both inside a burning building and you only have time to save one. Of course you save the civilian and not the terrorist. And that’s still true even if you know that once the terrorist was rescued he’d go to prison and never be a threat to anyone else. He’s just not worth as much.

In the most extreme circumstances, a person can be so terrible that their moral valuation should be effectively zero: If the only person in a burning building is Stalin, I’m not sure you should save him even if you easily could. But it is a grave moral mistake to think that a person’s moral valuation should ever go negative, yet I think this is something that people do when confronted with someone they truly hate. The federal agents torturing those terrorists didn’t merely think of them as worthless—they thought of them as having negative worth. They felt it was a positive good to harm them. But this is fundamentally wrong; no sentient being has negative worth. Some may be so terrible as to have essentially zero worth; and we are often justified in causing harm to some in order to save others. It would have been entirely justified to kill Stalin (as a matter of fact he died of heart disease and old age), to remove the continued threat he posed; but to torture him would not have made the world a better place, and actually might well have made it worse.

Yet I can see how psychologically it could be useful to have a mechanism in our brains that makes us hate someone so much we view them as having negative worth. It makes it a lot easier to harm them when necessary, makes us feel a lot better about ourselves when we do. The idea that any act of homicide is a tragedy but some of them are necessary tragedies is a lot harder to deal with than the idea that some people are just so evil that killing or even torturing them is intrinsically good. But some of the worst things human beings have ever done ultimately came from that place in our brains—and torture is one of them.