How to teach people about vaccines

May 25 JDN 2460821

Vaccines are one of the greatest accomplishments in human history. They have saved hundreds of millions of lives with minimal cost and almost no downside at all. (For everyone who suffers a side effect from a vaccine, I guarantee you: Someone else would have had it much worse from the disease if they hadn’t been vaccinated.)

It’s honestly really astonishing just how much good vaccines have done for humanity.

Thus, it’s a bit of a mystery how there are so many people who oppose vaccines.

But this mystery becomes a little less baffling in light of behavioral economics. People assess the probability of an event mainly based on the availability heuristic: How many examples can they think of when it happened?

Precisely because vaccines have been so effective at preventing disease, we have now reached a point where diseases that were once commonplace are now virtually eradicated. Thus, parents considering whether to vaccinate their children think about whether they know anyone who has gotten sick from that disease, and they can’t think of anyone, so they assume that it’s not a real danger. Then, someone comes along and convinces them (based on utter lies that have been thoroughly debunked) that vaccines cause autism, and they get scared about autism, because they can think of someone they know who has autism.

But of course, the reason that they can’t think of anyone who died from measles or pertussis is because of the vaccines. So I think we need an educational campaign that makes these rates more vivid for people, which plays into the availability heuristic instead of against it.

Here’s my proposal for a little educational game that might help:

It functions quite similarly to a classic tabletop RPG like Dungeons & Dragons, only here the target numbers are based on real figures.


Gather a group of at least 100 people. (Too few, and the odds become small enough that you may get no examples of some diseases.)

Each person needs 3 10-sided dice. Preferably they would be different colors or somehow labeled, because we want one to represent the 100s digit, one the 10s digit, and one the 1s digit. (The numbers you can roll thus range uniformly from 0 to 999.) In TTRPG parlance, this is called a d1000.

Give each person a worksheet that looks like this:

DiseaseBefore vaccine: Caught?Before vaccine: Died?After vaccine: Caught?After vaccine: Died?
Diptheria



Measles



Mumps



Pertussis



Polio



Rubella



Smallpox



Tetanus



Hep A



Hep B



Pneumococca



Varicella



In the first round, use the figures for before the vaccine. In the second round, use the figures for after the vaccine.

For each disease in each round, there will be a certain roll that people need to get in order to either not contract the disease: Roll that number or higher, and you are okay; roll below it, and you catch the disease.


Likewise, there will be a certain roll they need to get to survive if they contract it: Roll that number or higher, and you get sick but survive; roll below it, and you die.

Each time, name a disease, and then tell people what they need to roll to not catch it.

Have them all roll, and if they catch it, check off that box.

Then, for everyone who catches it, have them roll again to see if they survive it. If they die, check that box.

Based on the historical incidences which I have converted into lifetime prevalences, the target numbers are as follows:

DiseaseBefore vaccine: Roll to not catchBefore vaccine: Roll to surviveAfter vaccine: Roll to not catchAfter vaccine: Roll to survive
Diptheria138700
Measles244100
Mumps66020
Pertussis1232042
Polio208900
Rubella191190
Smallpox201200
Tetanus1800171
Hep A37141
Hep B22444
Pneumococca1910311119
Varicella95011640

What you should expect to see for a group of 100 is something like this (of course the results are random, so it won’t be this exactly):

DiseaseBefore vaccine: Number caughtBefore vaccine: Number diedAfter vaccine: Number caughtAfter vaccine: Number died
Diptheria1000
Measles24000
Mumps7000
Pertussis12100
Polio2000
Rubella2000
Smallpox2000
Tetanus0000
Hep A4000
Hep B2000
Pneumococca2111
Varicella950160

You’ll find that not a lot of people have checked those “dead” boxes either before or after the vaccine. So if you just look at death rates, the difference may not seem that stark.

(Of course, over a world as big as ours, it adds up: The difference between the 0.25% death rate of pertussis before the vaccine and 0% today is 20 million people—roughly the number of people who live in the New York City metro area.)

But I think people will notice that a lot more people got sick in the “before-vaccine” world than the “after-vaccine” world. Moreover, those that did get sick will find themselves rolling the dice on dying; they’ll probably be fine, but you never know for sure.

Make sure people also notice that (except for pneumococca), if you do get sick, the roll you need to survive is a lot higher without the vaccine. (If anyone does get unlucky enough to get tetanus in the first round, they’re probably gonna die!)

If anyone brings up autism, you can add an extra round where you roll for that too.

The supposedly “epidemic” prevalence of autism today is… 3.2%.

(Honestly I expected higher than that, but then, I hang around with a lot of queer and neurodivergent people. (So the availability heuristic got me too!))

Thus, what’s the roll to not get autism? 32.

Even with the expansive diagnostic criteria that include a lot of borderline cases like yours truly, you still only need to roll 32 on this d1000 to not get autism.

This means that only about 3 people in your group of 100 should end up getting autism, most likely fewer than the number who were saved from getting measles, mumps, and rubella by the vaccine, comparable to the number saved from getting most of the other diseases—and almost certainly fewer than the number saved from getting varicella.

So even if someone remains absolutely convinced that vaccines cause autism, you can now point out that vaccines also clearly save billions of people from getting sick and millions from dying.

Also, there are different kinds of autism. Some forms might not even be considered a disability if society were more accommodating; others are severely debilitating.

Recently clinicians have started to categorize “profound autism”, the kind that is severely debilitating. This constitutes about 25% of children with autism—but it’s a falling percentage over time, because broader diagnostic criteria are including more people as autistic, but not changing the number who are severely debilitated. (It is controversial exactly what should constitute “profound autism”, but I do think the construct is useful; there’s a big difference between someone like me who can basically function normally with some simple accommodations, and someone who never even learns to talk.)

So you can have the group do another roll, specifically for profound autism; that target number is now only 8.

There’s also one more demonstration you can do.

Aggregating over all these diseases, we can find the overall chance of dying from any of these diseases before and after the vaccine.

Have everyone roll for that, too:

Before the vaccines, the target number is 8. Afterward, it is 1.

If autism was brought up, make that comparison explicit.

Even if 100% of autism cases were caused by vaccines (which, I really must say, is ridiculous, as there’s no credible evidence that vaccines cause autism at all) that would still mean the following:

You are trading off a 32 in 1000 chance of your child being autistic and an 8 in 1000 chance of your child being profoundly autistic, against a 7 in 1000 chance of your child dying.

If someone is still skeptical of vaccines at this point, you should ask them point-blank:

Do you really think that being autistic is one-fifth as bad as dying?

Do you really think that being profoundly autistic is as bad as dying?

Expressivism

Sep 29 JDN 2460583

The theory of expressivism, often posited as an alternative to moral realism, is based on the observation by Hume that factual knowledge is not intrinsically motivating. I can believe that a food is nutritious and that I need nutrition to survive, but without some emotional experience to motivate me—hunger—I will nonetheless remain unmotivated to eat the nutritious food. Because morality is meant to be intrinsically motivating, says Hume, it must not involve statements of fact.

Yet really all Hume has shown is that if indeed facts are not intrinsically motivating, and moral statements are intrinsically motivating, then moral statements are not merely statements of fact. But even statements of fact are rarely merely statements of fact! If I were to walk down the street stating facts at random (lemurs have rings on their tails, the Sun is over one million kilometers in diameter, bicycles have two wheels, people sit on chairs, time dilates as you approach the speed of light, LGBT people suffer the highest per capita rate of hate crimes in the US, Coca-Cola in the United States contains high fructose corn syrup, humans and chimpanzees share 95-98% of our DNA), I would be seen as a very odd sort of person indeed. Even when I state a fact, I do so out of some motivation, frequently an emotional motivation. I’m often trying to explain, or to convince. Sometimes I am angry, and I want to express my anger and frustration. Other times I am sad and seeking consolation. I have many emotions, and I often use words to express them. Nonetheless, in the process I will make many statements of fact that are either true or false: “Humans and chimpanzees share 95-98% of our DNA” I might use to argue in favor of common descent; “Time dilates as you approach the speed of light” I have used in to explain relativity theory; “LGBT people suffer the highest per capita rate of hate crimes in the US” I might use to argue in favor of some sort of gay rights policy. When I say “genocide is wrong!” I probably have some sort of emotional motivation for this—likely my outrage at an ongoing genocide. Nonetheless I’m pretty sure it’s true that genocide is wrong.

Expressivism says that moral statements don’t express propositions at all, they express attitudes, relations to ideas that are not of the same kind as belief and disbelief, truth and falsehood. Much as “Hello!” or “Darn it!” don’t really state facts or inquire about facts, expressivists like Simon Blackburn and Allan Gibbard would say that “Genocide is wrong” doesn’t say anything about the facts of genocide, it merely expresses my attitude of moral disapproval toward genocide.

Yet expressivists can’t abandon all normativity—otherwise even the claim “expressivism is true” has no moral force. Allan Gibbard, like most expressivists, supports epistemic normativity—the principle that we ought to believe what is true. But this seems to me already a moral principle, and one that is not merely an attitude that some people happen to have, but in fact a fundamental axiom that ought to apply to any rational beings in any possible universe. Even more, Gibbard agrees that some moral attitudes are more warranted than others, that “genocide is wrong” is more legitimate than “genocide is good”. But once we agree that there are objective normative truths and moral attitudes can be more or less justified, how is this any different from moral realism?

Indeed, in terms of cognitive science I’m not sure beliefs and emotions are so easily separable in the first place. In some sense I think statements of fact can be intrinsically motivating—or perhaps it is better to put it this way: If your brain is working properly, certain beliefs and emotions will necessarily coincide. If you believe that you are about to be attacked by a tiger, and you don’t experience the emotion of fear, something is wrong; if you believe that you are about to die of starvation, and you don’t experience the emotion of hunger, something is wrong. Conversely, if you believe that you are safe from all danger, and yet you experience fear, something is wrong; if you believe that you have eaten plenty of food, yet you still experience hunger, something is wrong. When your beliefs and emotions don’t align, either your beliefs or your emotions are defective. I would say that the same is true of moral beliefs. If you believe that genocide is wrong but you are not motivated to resist genocide, something is wrong; if you believe that feeding your children is obligatory but you are not motivated to feed your children, something is wrong.

It may well be that without emotion, facts would never motivate us; but emotions can warranted by facts. That is how we distinguish depression from sadness, mania from joy, phobia from fear. Indeed I am dubious of the entire philosophical project of noncognitivism, of which expressivism is the moral form. Noncognitivism is the idea that a given domain of mental processing is not cognitive—not based on thinking, reason, or belief. There is often a sense that noncognitive mental processing is “lower” than cognition, usually based on the idea that it is more phylogenetically conserved—that we think as men but feel as rats.

Yet in fact this is not how human emotions work at all. Poetry—mere words—often evokes the strongest of emotions. A text message of “I love you” or “I think we should see other people” can change the course of our lives. An ambulance in the driveway will pale the face of any parent. In 2001 the video footage of airplanes colliding with skyscrapers gave all of America nightmares for weeks. Yet stop and think about what text messages, ambulances, video footage, airplanes, and skyscrapers are—they are technologies so advanced, so irreducibly cognitive, that even the world’s technological superpower had none of them 200 years ago. (We didn’t have text messages forty years ago!) Even something as apparently dry as numbers can have profound emotional effects: In the statements “Your blood sugar is X mg/dL” to a diabetic, “You have Y years to live” to a cancer patient, or “Z people died” in a news report, the emotional effects are almost wholly dependent upon the value of the numbers X, Y, and Z—values of X = 100, Y = 50 and Z = 0 would be no cause for alarm (or perhaps even cause for celebration!), while values of X = 400, Y = 2, and Z = 10,000 would trigger immediate shock, terror and despair. The entire discipline of cognitive-behavioral psychotherapy depends upon the fact that talking to people about their thoughts and beliefs can have profound effects upon their emotions and actions—and in empirical studies, cognitive-behavioral psychotherapy is verified to work in a variety of circumstances and is more effective than medication for virtually every mental disorder. We do not think as men but feel as rats; we thinkandfeel as human beings.

Because they are evolved instincts, we have limited control over them, and other animals have them, we are often inclined to suppose that emotions are simple, stupid, irrational—but on the contrary they are mind-bogglingly complex, brilliantly intelligent, and the essence of what it means to be a rational being. People who don’t have emotions aren’t rational—they are inert. In psychopathology a loss of capacity for emotion is known as flat affect, and it is often debilitating; it is often found in schizophrenia and autism, and in its most extreme forms it causes catatoniathat is, a total lack of body motion. From Plato to Star Trek, Western culture has taught us to think that a loss of emotion would improve our rationality; but on the contrary, a loss of all emotion would render us completely vegetative. Lieutenant Commander Data without his emotion chip should stand in one place and do nothing—for this is what people without emotion actually do.

Indeed, attractive and aversive experiences—that is, emotions—are the core of goal-seeking behavior, without which rationality is impossible. Apparently simple experiences like pleasure and pain (let alone obviously complicated ones like jealousy and patriotism) are so complex that the most advanced robots in the world cannot even get close to simulating them. Injure a rat, and it will withdraw and cry out in pain; damage a robot (at least any less than a state-of-the-art research robot), and it will not react at all, continuing ineffectually through the same motions it was attempting a moment ago. This shows that rats are smarter than robots—an organism that continues on its way regardless of the stimulus is more like a plant than an animal.

Our emotions do sometimes fail us. They hurt us, they put us at risk, they make us behave in ways that are harmful or irrational. Yet to declare on these grounds that emotions are the enemy of reason would be like declaring that we should all poke out our eyes because sometimes we are fooled by optical illusions. It would be like saying that a shirt with one loose thread is unwearable, that a mathematician who once omits a negative sign should never again be trusted. This is not rationality but perfectionism. Like human eyes, human emotions are rational the vast majority of the time, and when they aren’t, this is cause for concern. Truly irrational emotions include mania, depression, phobia, and paranoia—and it’s no accident that we respond to these emotions with psychotherapy and medication.

Expressivism is legitimate precisely because it is not a challenger to moral realism. Personally, I think that expressivism is wrong because moral claims express facts as much as they express attitudes; but given our present state of knowledge about cognitive science, that is the sort of question upon which reasonable people can disagree. Moreover, the close ties between emotion and reason may ultimately entail that we are wrong to make the distinction in the first place. It is entirely reasonable, at our present state of knowledge, to think that moral judgments are primarily emotional rather than propositional. What isnot reasonable, however, is the claim that moral statements cannot be objectively justified—the evidence against this claim is simply too compelling to ignore. If moral claims are emotions, they are emotions that can be objectively justified.

Mind reading is not optional

Nov 20 JDN 2459904

I have great respect for cognitive-behavioral therapy (CBT), and it has done a lot of good for me. (It is also astonishingly cost-effective; its QALY per dollar rate compares favorably to almost any other First World treatment, and loses only to treating high-impact Third World diseases like malaria and schistomoniasis.)

But there are certain aspects of it that have always been frustrating to me. Standard CBT techniques often present as ‘cognitive distortions‘ what are in fact clearly necessary heuristics without which it would be impossible to function.

Perhaps the worst of these is so-called ‘mind reading‘. The very phrasing of it makes it sound ridiculous: Are you suggesting that you have some kind of extrasensory perception? Are you claiming to be a telepath?

But in fact ‘mind reading’ is simply the use of internal cognitive models to forecast the thoughts, behaviors, and expectations of other human beings. And without it, it would be completely impossible to function in human society.

For instance, I have had therapists tell me that it is ‘mind reading’ for me to anticipate that people will have tacit expectations for my behavior that they will judge me for failing to meet, and I should simply wait for people to express their expectations rather than assuming them. I admit, life would be much easier if I could do that. But I know for a fact that I can’t. Indeed, I used to do that, as a child, and it got me in trouble all the time. People were continually upset at me for not doing things they had expected me to do but never bothered to actually mention. They thought these expectations were “obvious”; they were not, at least not to me.

It was often little things, and in hindsight some of these things seem silly: I didn’t know what a ‘made bed’ was supposed to look like, so I put it in a state that was functional for me, but that was not considered ‘making the bed’. (I have since learned that my way was actually better: It’s good to let sheets air out before re-using them.) I was asked to ‘clear the sink’, so I moved the dishes out of the sink and left them on the counter, not realizing that the implicit command was for me to wash those dishes, dry them, and put them away. I was asked to ‘bring the dinner plates to the table’, so I did that, and left them in a stack there, not realizing that I should be setting them out in front of each person’s chair and also bringing flatware. Of course I know better now. But how was I supposed to know then? It seems like I was expected to, though.

Most people just really don’t seem to realize how many subtle, tacit expectations are baked into every single task. I think neurodivergence is quite relevant here; I have a mild autism spectrum disorder, and so I think rather differently than most people. If you are neurotypical, then you probably can forecast other people’s expectations fairly well automatically, and so they may seem obvious to you. In fact, they may seem so obvious that you don’t even realize you’re doing it. Then when someone like me comes along and is consciously, actively trying to forecast other people’s expectations, and sometimes doing it poorly, you go and tell them to stop trying to forecast. But if they were to do that, they’d end up even worse off than they are. What you really need to be telling them is how to forecast better—but that would require insight into your own forecasting methods which you aren’t even consciously aware of.

Seriously, stop and think for a moment all of the things other people expect you to do every day that are rarely if ever explicitly stated. How you are supposed to dress, how you are supposed to speak, how close you are supposed to stand to other people, how long you are supposed to hold eye contact—all of these are standards you will be expected to meet, whether or not any of them have ever been explicitly explained to you. You may do this automatically; or you may learn to do it consciously after being criticized for failing to do it. But one way or another, you must forecast what other people will expect you to do.

To my knowledge, no one has ever explicitly told me not to wear a Starfleet uniform to work. I am not aware of any part of the university dress code that explicitly forbids such attire. But I’m fairly sure it would not be a good idea. To my knowledge, no one has ever explicitly told me not to burst out into song in the middle of a meeting. But I’m still pretty sure I shouldn’t do that. To my knowledge, no one has ever explicitly told me what the ‘right of way’ rules are for walking down a crowded sidewalk, who should be expected to move out of the way of whom. But people still get mad if you mess up and bump into them.

Even when norms are stated explicitly, it is often as a kind of last resort, and the mere fact that you needed to have a norm stated is often taken as a mark against your character. I have been explicitly told in various contexts not to talk to myself or engage in stimming leg movements; but the way I was told has generally suggested that I would have been judged better if I hadn’t had to be told, if I had simply known the way that other people seem to know. (Or is it that they never felt any particular desire to stim?)

In fact, I think a major part of developing social skills and becoming more functional, to the point where a lot of people actually now seem a bit surprised to learn I have an autism spectrum disorder, has been improving my ability to forecast other people’s expectations for my behavior. There are dozens if not hundreds of norms that people expect you to follow at any given moment; most people seem to intuit them so easily that they don’t even realize they are there. But they are there all the same, and this is painfully evident to those of us who aren’t always able to immediately intuit them all.

Now, the fact remains that my current mental models are surely imperfect. I am often wrong about what other people expect of me. I’m even prepared to believe that some of my anxiety comes from believing that people have expectations more demanding than what they actually have. But I can’t simply abandon the idea of forecasting other people’s expectations. Don’t tell me to stop doing it; tell me how to do it better.

Moreover, there is a clear asymmetry here: If you think people want more from you than they actually do, you’ll be anxious, but people will like you and be impressed by you. If you think people want less from you than they actually do, people will be upset at you and look down on you. So, in the presence of uncertainty, there’s a lot of pressure to assume that the expectations are high. It would be best to get it right, of course; but when you aren’t sure you can get it right, you’re often better off erring on the side of caution—which is to say, the side of anxiety.

In short, mind reading isn’t optional. If you think it is, that’s only because you do it automatically.

Mindful of mindfulness

Sep 25 JDN 2459848

I have always had trouble with mindfulness meditation.

On the one hand, I find it extremely difficult to do: if there is one thing my mind is good at, it’s wandering. (I think in addition to my autism spectrum disorder, I may also have a smidgen of ADHD. I meet some of the criteria at least.) And it feels a little too close to a lot of practices that are obviously mumbo-jumbo nonsense, like reiki, qigong, and reflexology.

On the other hand, mindfulness meditation has been empirically shown to have large beneficial effects in study after study after study. It helps with not only depression, but also chronic pain. It even seems to improve immune function. The empirical data is really quite clear at this point. The real question is how it does all this.

And I am, above all, an empiricist. I bow before the data. So, when my new therapist directed me to an app that’s supposed to train me to do mindfulness meditation, I resolved that I would in fact give it a try.

Honestly, as of writing this, I’ve been using it less than a week; it’s probably too soon to make a good evaluation. But I did have some prior experience with mindfulness, so this was more like getting back into it rather than starting from scratch. And, well, I think it might actually be working. I feel a bit better than I did when I started.

If it is working, it doesn’t seem to me that the mechanism is greater focus or mental control. I don’t think I’ve really had time to meaningfully improve those skills, and to be honest, I have a long way to go there. The pre-recorded voice samples keep telling me it’s okay if my mind wanders, but I doubt the app developers planned for how much my mind can wander. When they suggest I try to notice each wandering thought, I feel like saying, “Do you want the complete stack trace, or just the final output? Because if I wrote down each terminal branch alone, my list would say something like ‘fusion reactors, ice skating, Napoleon’.”

I think some of the benefit is simply parasympathetic activation, that is, being more relaxed. I am, and have always been, astonishingly bad at relaxing. It’s not that I lack positive emotions: I can enjoy, I can be excited. Nor am I incapable of low-arousal emotions: I can get bored, I can be lethargic. I can also experience emotions that are negative and high-arousal: I can be despondent or outraged. But I have great difficulty reaching emotional states which are simultaneously positive and low-arousal, i.e. states of calm and relaxation. (See here for more on the valence/arousal model of emotional states.) To some extent I think this is due to innate personality: I am high in both Conscientiousness and Neuroticism, which basically amounts to being “high-strung“. But mindfulness has taught me that it’s also trainable, to some extent; I can get better at relaxing, and I already have.

And even more than that, I think the most important effect has been reminding and encouraging me to practice self-compassion. I am an intensely compassionate person, toward other people; but toward myself, I am brutal, demanding, unforgiving, even cruel. My internal monologue says terrible things to me that I wouldnever say to anyone else. (Or at least, not to anyone else who wasn’t a mass murderer or something. I wouldn’t feel particularly bad about saying “You are a failure, you are broken, you are worthless, you are unworthy of love” to, say, Josef Stalin. And yes, these are in fact things my internal monologue has said to me.) Whenever I am unable to master a task I consider important, my automatic reaction is to denigrate myself for failing; I think the greatest benefit I am getting from practicing meditation is being encouraged to fight that impulse. That is, the most important value added by the meditation app has not been in telling me how to focus on my own breathing, but in reminding me to forgive myself when I do it poorly.

If this is right (as I said, it’s probably too soon to say), then we may at last be able to explain why meditation is simultaneously so weird and tied to obvious mumbo-jumbo on the one hand, and also so effective on the other. The actual function of meditation is to be a difficult cognitive task which doesn’t require outside support.

And then the benefit actually comes from doing this task, getting slowly better at it—feeling that sense of progress—and also from learning to forgive yourself when you do it badly. The task probably could have been anything: Find paths through mazes. Fill out Sudoku grids. Solve integrals. But these things are hard to do without outside resources: It’s basically impossible to draw a maze without solving it in the process. Generating a Sudoku grid with a unique solution is at least as hard as solving one (which is NP-complete). By the time you know a given function is even integrable over elementary functions, you’ve basically integrated it. But focusing on your breath? That you can do anywhere, anytime. And the difficulty of controlling all your wandering thoughts may be less a bug than a feature: It’s precisely because the task is so difficult that you will have reason to practice forgiving yourself for failure.

The arbitrariness of the task itself is how you can get a proliferation of different meditation techniques, and a wide variety of mythologies and superstitions surrounding them all, but still have them all be about equally effective in the end. Because it was never really about the task at all. It’s about getting better and failing gracefully.

It probably also helps that meditation is relaxing. Solving integrals might not actually work as well as focusing on your breath, even if you had a textbook handy full of integrals to solve. Breathing deeply is calming; integration by parts isn’t. But lots of things are calming, and some things may be calming to one person but not to another.

It is possible that there is yet some other benefit to be had directly via mindfulness itself. If there is, it will surely have more to do with anterior cingulate activation than realignment of qi. But such a particular benefit isn’t necessary to explain the effectiveness of meditation, and indeed would be hard-pressed to explain why so many different kinds of meditation all seem to work about as well.

Because it was never about what you’re doing—it was always about how.

The Tragedy of the Commons

JDN 2457387

In a previous post I talked about one of the most fundamental—perhaps the most fundamental—problem in game theory, the Prisoner’s Dilemma, and how neoclassical economic theory totally fails to explain actual human behavior when faced with this problem in both experiments and the real world.

As a brief review, the essence of the game is that both players can either cooperate or defect; if they both cooperate, the outcome is best overall; but it is always in each player’s interest to defect. So a neoclassically “rational” player would always defect—resulting in a bad outcome for everyone. But real human beings typically cooperate, and thus do better. The “paradox” of the Prisoner’s Dilemma is that being “rational” results in making less money at the end.

Obviously, this is not actually a good definition of rational behavior. Being short-sighted and ignoring the impact of your behavior on others doesn’t actually produce good outcomes for anybody, including yourself.

But the Prisoner’s Dilemma only has two players. If we expand to a larger number of players, the expanded game is called a Tragedy of the Commons.

When we do this, something quite surprising happens: As you add more people, their behavior starts converging toward the neoclassical solution, in which everyone defects and we get a bad outcome for everyone.

Indeed, people in general become less cooperative, less courageous, and more apathetic the more of them you put together. K was quite apt when he said, “A person is smart; people are dumb, panicky, dangerous animals and you know it.” There are ways to counteract this effect, as I’ll get to in a moment—but there is a strong effect that needs to be counteracted.

We see this most vividly in the bystander effect. If someone is walking down the street and sees someone fall and injure themselves, there is about a 70% chance that they will go try to help the person who fell—humans are altruistic. But if there are a dozen people walking down the street who all witness the same event, there is only a 40% chance that any of them will help—humans are irrational.

The primary reason appears to be diffusion of responsibility. When we are alone, we are the only one could help, so we feel responsible for helping. But when there are others around, we assume that someone else could take care of it for us, so if it isn’t done that’s not our fault.

There also appears to be a conformity effect: We want to conform our behavior to social norms (as I said, to a first approximation, all human behavior is social norms). The mere fact that there are other people who could have helped but didn’t suggests the presence of an implicit social norm that we aren’t supposed to help this person for some reason. It never occurs to most people to ask why such a norm would exist or whether it’s a good one—it simply never occurs to most people to ask those questions about any social norms. In this case, by hesitating to act, people actually end up creating the very norm they think they are obeying.

This can lead to what’s called an Abilene Paradox, in which people simultaneously try to follow what they think everyone else wants and also try to second-guess what everyone else wants based on what they do, and therefore end up doing something that none of them actually wanted. I think a lot of the weird things humans do can actually be attributed to some form of the Abilene Paradox. (“Why are we sacrificing this goat?” “I don’t know, I thought you wanted to!”)

Autistic people are not as good at following social norms (though some psychologists believe this is simply because our social norms are optimized for the neurotypical population). My suspicion is that autistic people are therefore less likely to suffer from the bystander effect, and more likely to intervene to help someone even if they are surrounded by passive onlookers. (Unfortunately I wasn’t able to find any good empirical data on that—it appears no one has ever thought to check before.) I’m quite certain that autistic people are less likely to suffer from the Abilene Paradox—if they don’t want to do something, they’ll tell you so (which sometimes gets them in trouble).

Because of these psychological effects that blunt our rationality, in large groups human beings often do end up behaving in a way that appears selfish and short-sighted.

Nowhere is this more apparent than in ecology. Recycling, becoming vegetarian, driving less, buying more energy-efficient appliances, insulating buildings better, installing solar panels—none of these things are particularly difficult or expensive to do, especially when weighed against the tens of millions of people who will die if climate change continues unabated. Every recyclable can we throw in the trash is a silent vote for a global holocaust.

But as it no doubt immediately occurred to you to respond: No single one of us is responsible for all that. There’s no way I myself could possibly save enough carbon emissions to significantly reduce climate change—indeed, probably not even enough to save a single human life (though maybe). This is certainly true; the error lies in thinking that this somehow absolves us of the responsibility to do our share.

I think part of what makes the Tragedy of the Commons so different from the Prisoner’s Dilemma, at least psychologically, is that the latter has an identifiable victimwe know we are specifically hurting that person more than we are helping ourselves. We may even know their name (and if we don’t, we’re more likely to defect—simply being on the Internet makes people more aggressive because they don’t interact face-to-face). In the Tragedy of the Commons, it is often the case that we don’t know who any of our victims are; moreover, it’s quite likely that we harm each one less than we benefit ourselves—even though we harm everyone overall more.

Suppose that driving a gas-guzzling car gives me 1 milliQALY of happiness, but takes away an average of 1 nanoQALY from everyone else in the world. A nanoQALY is tiny! Negligible, even, right? One billionth of a year, a mere 30 milliseconds! Literally less than the blink of an eye. But take away 30 milliseconds from everyone on Earth and you have taken away 7 years of human life overall. Do that 10 times, and statistically one more person is dead because of you. And you have gained only 10 milliQALY, roughly the value of $300 to a typical American. Would you kill someone for $300?

Peter Singer has argued that we should in fact think of it this way—when we cause a statistical death by our inaction, we should call it murder, just as if we had left a child to drown to keep our clothes from getting wet. I can’t agree with that. When you think seriously about the scale and uncertainty involved, it would be impossible to live at all if we were constantly trying to assess whether every action would lead to statistically more or less happiness to the aggregate of all human beings through all time. We would agonize over every cup of coffee, every new video game. In fact, the global economy would probably collapse because none of us would be able to work or willing to buy anything for fear of the consequences—and then whom would we be helping?

That uncertainty matters. Even the fact that there are other people who could do the job matters. If a child is drowning and there is a trained lifeguard right next to you, the lifeguard should go save the child, and if they don’t it’s their responsibility, not yours. Maybe if they don’t you should try; but really they should have been the one to do it.
But we must also not allow ourselves to simply fall into apathy, to do nothing simply because we cannot do everything. We cannot assess the consequences of every specific action into the indefinite future, but we can find general rules and patterns that govern the consequences of actions we might take. (This is the difference between act utilitarianism, which is unrealistic, and rule utilitarianism, which I believe is the proper foundation for moral understanding.)

Thus, I believe the solution to the Tragedy of the Commons is policy. It is to coordinate our actions together, and create enforcement mechanisms to ensure compliance with that coordinated effort. We don’t look at acts in isolation, but at policy systems holistically. The proper question is not “What should I do?” but “How should we live?”

In the short run, this can lead to results that seem deeply suboptimal—but in the long run, policy answers lead to sustainable solutions rather than quick-fixes.

People are starving! Why don’t we just steal money from the rich and use it to feed people? Well, think about what would happen if we said that the property system can simply be unilaterally undermined if someone believes they are achieving good by doing so. The property system would essentially collapse, along with the economy as we know it. A policy answer to that same question might involve progressive taxation enacted by a democratic legislature—we agree, as a society, that it is justified to redistribute wealth from those who have much more than they need to those who have much less.

Our government is corrupt! We should launch a revolution! Think about how many people die when you launch a revolution. Think about past revolutions. While some did succeed in bringing about more just governments (e.g. the French Revolution, the American Revolution), they did so only after a long period of strife; and other revolutions (e.g. the Russian Revolution, the Iranian Revolution) have made things even worse. Revolution is extremely costly and highly unpredictable; we must use it only as a last resort against truly intractable tyranny. The policy answer is of course democracy; we establish a system of government that elects leaders based on votes, and then if they become corrupt we vote to remove them. (Sadly, we don’t seem so good about that second part—the US Congress has a 14% approval rating but a 95% re-election rate.)

And in terms of ecology, this means that berating ourselves for our sinfulness in forgetting to recycle or not buying a hybrid car does not solve the problem. (Not that it’s bad to recycle, drive a hybrid car, and eat vegetarian—by all means, do these things. But it’s not enough.) We need a policy solution, something like a carbon tax or cap-and-trade that will enforce incentives against excessive carbon emissions.

In case you don’t think politics makes a difference, all of the Democrat candidates for President have proposed such plans—Bernie Sanders favors a carbon tax, Martin O’Malley supports an aggressive cap-and-trade plan, and Hillary Clinton favors heavily subsidizing wind and solar power. The Republican candidates on the other hand? Most of them don’t even believe in climate change. Chris Christie and Carly Fiorina at least accept the basic scientific facts, but (1) they are very unlikely to win at this point and (2) even they haven’t announced any specific policy proposals for dealing with it.

This is why voting is so important. We can’t do enough on our own; the coordination problem is too large. We need to elect politicians who will make policy. We need to use the systems of coordination enforcement that we have built over generations—and that is fundamentally what a government is, a system of coordination enforcement. Only then can we overcome the tendency among human beings to become apathetic and short-sighted when faced with a Tragedy of the Commons.