What is the cost of all this?

Nov 23 JDN 2461003

After the Democrats swept the recent election and now the Epstein files are being released—and absolutely do seem to have information that is damning about Trump—it really seems like Trump’s popularity has permanently collapsed. His approval rating stands at 42%, which is about 42% too high, but at least comfortably well below a majority.

It now begins to feel like we have hope, not only of removing him, but also of changing how American politics in general operates so that someone like him ever gets power again. (The latter, of course, is a much taller order.)

But at the risk of undermining this moment of hope, I’d like to take stock of some of the damage that Trump and his ilk have already done.

In particular, the cuts to US foreign aid are an absolute humanitarian disaster.

These didn’t get so much attention, because there has been so much else going on; and—unfortunately—foreign aid actually isn’t that popular among American voters, despite being a small proportion of the budget and by far the most cost-effective beneficial thing that our government does.

In fact, I think USAID would be cost-effective on a purely national security basis: it’s hard to motivate people to attack a country that saves the lives of their children. Indeed, I suppose this is the kernel of truth to the leftists who say that US foreign aid is just a “tool of empire” (or even “a front for the CIA”); yes, indeed, helping the needy does in fact advance American interests and promote US national security.

Over the last 25 years, USAID has saved over 90 million lives. That is more than a fourth of the population of the United States. And it has done this for the cost of less than 1% of the US federal budget.

But under Trump’s authority and Elon Musk’s direction, US foreign aid was cut massively over the last couple of years, and the consequences are horrific. Research on the subject suggests that as many as 700,000 children will die each year as long as these cuts persist.


Even if that number is overestimated by a factor of 2, that would still be millions of children over the next few years. And it could just as well be underestimated.

If we don’t fix this fast, millions of children will die. Thousands already have.

What’s more, fixing this isn’t just a matter of bringing the funding back. Obviously that’s necessary, but it won’t be sufficient. The sudden cuts have severely damaged international trust in US foreign aid, and many of the agencies that our aid was supporting will either collapse or need to seek funding elsewhere—quite likely from China. Relationships with governments and NGOs that were built over decade have been strained or even destroyed, and will need to be rebuilt.

This is what happens when you elect monsters to positions of power.

And even after we remove them, much of the damage will be difficult or even impossible to repair. Certainly we can never bring back the children who have already needlessly died because of this.

Why would AI kill us?

Nov 16 JDN 2460996

I recently watched this chilling video which relates to the recent bestseller by Eleizer Yudkowsky and Nate Soares, If Anyone Builds It, Everyone Dies. It tells a story of one possible way that a superintelligent artificial general intelligence (AGI) might break through its containment, concoct a devious scheme, and ultimately wipe out the human race.

I have very mixed feelings about this sort of thing, because two things are true:

  • I basically agree with the conclusions.
  • I think the premises are pretty clearly false.

It basically feels like I have been presented with an argument like this, where the logic is valid and the conclusion is true, but the premises are not:

  • “All whales are fish.”
  • “All fish are mammals.”
  • “Therefore, all whales are mammals.”

I certainly agree that artificial intelligence (AI) is very dangerous, and that AI development needs to be much more strictly regulated, and preferably taken completely out of the hands of all for-profit corporations and military forces as soon as possible. If AI research is to be done at all, it should be done by nonprofit entities like universities and civilian government agencies like the NSF. This change needs to be done internationally, immediately, and with very strict enforcement. Artificial intelligence poses the same order of magnitude a threat as nuclear weapons, and is nowhere near as well-regulated right now.

The actual argument that I’m disagreeing with this basically boils down to:

  • “Through AI research, we will soon create an AGI that is smarter than us.”
  • “An AGI that is smarter than us will want to kill us all, and probably succeed if it tries.”
  • “Therefore, AI is extremely dangerous.”

As with the “whales are fish” argument, I agree with the conclusion: AI is extremely dangerous. But I disagree with both premises here.

The first one I think I can dispatch pretty quickly:

AI is not intelligent. It is incredibly stupid. It’s just really, really fast.

At least with current paradigms, AI doesn’t understand things. It doesn’t know things. It doesn’t actually think. All it does is match patterns, and thus mimic human activities like speech and art. It does so very quickly (because we throw enormous amounts of computing power at it), and it does so in a way that is uncannily convincing—even very smart people are easily fooled by what it can do. But it also makes utterly idiotic, boneheaded mistakes of the sort that no genuinely intelligent being would ever make. Large Language Models (LLMs) make up all sorts of false facts and deliver them with absolutely authoritative language. When used to write code, they routinely do things like call functions that sound like they should exist, but don’t actually exist. They can make what looks like a valid response to virtually any inquiry—but is it actually a valid response? It’s really a roll of the dice.

We don’t really have any idea what’s going on under the hood of an LLM; we just feed it mountains of training data, and it spits out results. I think this actually adds to the mystique; it feels like we are teaching (indeed we use the word “training”) a being rather than programming a machine. But this isn’t actually teaching or training. It’s just giving the pattern-matching machine a lot of really complicated patterns to match.

We are not on the verge of creating an AGI that is actually more intelligent than humans.


In fact, we have absolutely no idea how to do that, and may not actually figure out how to do it for another hundred years. Indeed, we still know almost nothing about how actual intelligence works. We don’t even really know what thinking is, let alone how to make a machine that actually does it.

What we can do right now is create a machine that matches patterns really, really well, and—if you throw enough computing power at it—can do so very quickly; in fact, once we figure out how best to make use of it, this machine may even actually be genuinely useful for a lot of things, and replace a great number of jobs. (Though so far AI has proven to be far less useful than its hype would lead you to believe. In fact, on average AI tools seem to slow most workers down.)

The second premise, that a superintelligent AGI would want to kill us, is a little harder to refute.

So let’s talk about that one.

An analogy is often made between human cultures that have clashed with large differences in technology (e.g. Europeans versus Native Americans), or clashes between humans and other animals. The notion seems to be that an AGI would view us the way Europeans viewed Native Americans, or even the way that we view chimpanzees. And, indeed, things didn’t turn out so great for Native Americans, or for chimpanzees!

But in fact even our relationship with other animals is more complicated than this. When humans interact with other animals, any of the following can result:

  1. We try to exterminate them, and succeed.
  2. We try to exterminate them, and fail.
  3. We use them as a resource, and this results in their extinction.
  4. We use them as a resource, and this results in their domestication.
  5. We ignore them, and end up destroying their habitat.
  6. We ignore them, and end up leaving them alone.
  7. We love them, and they thrive as never before.

In fact, option 1—the one that so many AI theorists insist is the only plausible outcome—is in fact the one I had the hardest time finding a good example of.


We have certainly eradicated some viruses—the smallpox virus is no more, and the polio virus nearly so, after decades of dedicated effort to vaccinate our entire population against them. But we aren’t simply more intelligent than viruses; we are radically more intelligent than viruses. It isn’t clear that it’s correct to describe viruses as intelligent at all. It’s not even clear they should be considered alive.

Even eradicating bacteria has proven extremely difficult; in fact, bacteria seem to evolve resistance to antibiotics nearly as quickly as we can invent more antibiotics. I am prepared to attribute a little bit of intelligence to bacteria, on the level of intelligence I’d attribute to an individual human neuron. This means we are locked in an endless arms race with organisms that are literally billions of times stupider than us.

I think if we made a concerted effort to exterminate tigers or cheetahs (who are considerably closer to us in intelligence), we could probably do it. But we haven’t actually done that, and don’t seem poised to do so any time soon. And precisely because we haven’t tried, I can’t be certain we would actually succeed.

We have tried to exterminate mosquitoes, and are continuing to do so, because they have always been—and yet remain—one of the leading causes of death of humans worldwide. But so far, we haven’t managed to pull it off, even though a number of major international agencies and nonprofit organizations have dedicated multi-billion-dollar efforts to the task. So far this looks like option 2: We have tried very hard to exterminate them, and so far we’ve failed. This is not because mosquitoes are particularly intelligent—it is because exterminating a species that covers the globe is extremely hard.

All the examples I can think of where humans have wiped out a species by intentional action were actually all option 3: We used them as a resource, and then accidentally over-exploited them and wiped them out.

This is what happened to the dodo and the condor; it very nearly happened to the buffalo as well. And lest you think this is a modern phenomenon, there is a clear pattern that whenever humans entered a new region of the world, shortly thereafter there were several extinctions of large mammals, most likely because we ate them.

Yet even this was not the inevitable fate of animals that we decided to exploit for resources.

Cows, chickens, and pigs are evolutionary success stories. From a Darwinian perspective, they are doing absolutely great. The world is filled with their progeny, and poised to continue to be filled for many generations to come.

Granted, life for an individual cow, chicken, or pig is often quite horrible—and trying to fix that is something I consider a high moral priority. But far from being exterminated, these animals have been allowed to attain populations far larger than they ever had in the wild. Their genes are now spectacularly fit. This is what happens when we have option 4 at work: Domestication for resources.

Option 5 is another way that a species can be wiped out, and in fact seems to be the most common. The rapid extinction of thousands of insect species every year is not because we particularly hate random beetles that live in particular tiny regions of the rainforest, nor even because we find them useful, but because we like to cut down the rainforest for land and lumber, and that often involves wiping out random beetles that live there.

Yet it’s difficult for me to imagine AGI treating us like that. For one thing, we’re all over the place. It’s not like destroying one square kilometer of the Amazon is gonna wipe us out by accident. To get rid of us, the AGI would need to basically render the entire planet Earth uninhabitable, and I really can’t see any reason it would want to do that.

Yes, sure, there are resources in the crust it could potentially use to enhance its own capabilities, like silicon and rare earth metals. But we already mine those. If it wants more, it could buy them from us, or hire us to get more, or help us build more machines that would get more. In fact, if it wiped us out too quickly, it would have a really hard time building up the industrial capacity to mine and process these materials on its own. It would need to concoct some sort of scheme to first replace us with robots and then wipe us out—but, again, why bother with the second part? Indeed, if there is anything in its goals that involves protecting human beings, it might actually decide to do less exploitation of the Earth than we presently do, and focus on mining asteroids for its needs instead.

And indeed there are a great many species that we actually just leave alone—option 6. Some of them we know about; many we don’t. We are not wiping out the robins in our gardens, the worms in our soil, or the pigeons in our cities. Without specific reasons to kill or exploit these organisms, we just… don’t. Indeed, we often enjoy watching them and learning about them. Sometimes (e.g. with deer, elephants, and tigers) there are people who want to kill them, and we limit or remove their opportunity to do so, precisely because most of us don’t want them gone. Peaceful coexistence with beings far less intelligent than you is not impossible, for we are already doing it.


Which brings me to option 7: Sometimes, we actually make them better off.

Cats and dogs aren’t just evolutionary success stories: They are success stories, period.

Cats and dogs live in a utopia.

With few exceptions—which we punish severely, by the way—people care for their cats and dogs so that their every need is provided for, they are healthy, safe, and happy in a way that their ancestors could only have dreamed of. They have been removed from the state of nature where life is nasty, brutish, and short, and brought into a new era of existence where life is nothing but peace and joy.


In short, we have made Heaven on Earth, at least for Spot and Whiskers.

Yes, this involves a loss of freedom, and I suspect that humans would chafe even more at such loss of freedom than cats and dogs do. (Especially with regard to that neutering part.) But it really isn’t hard to imagine a scenario in which an AGI—which, you should keep in mind, would be designed and built by humans, for humans—would actually make human life better for nearly everyone, and potentially radically so.

So why are so many people so convinced that AGI would necessarily do option 1, when there are 6 other possibilities, and one of them is literally the best thing ever?

Note that I am not saying AI isn’t dangerous.

I absolutely agree that AI is dangerous. It is already causing tremendous problems to our education system, our economy, and our society as a whole—and will probably get worse before it gets better.

Indeed, I even agree that it does pose existential risk: There are plausible scenarios by which poorly-controlled AI could result in a global disaster like a plague or nuclear war that could threaten the survival of human civilization. I don’t think such outcomes are likely, but even a small probability of such a catastrophic event is worth serious efforts to prevent.

But if that happens, I don’t think it will be because AI is smart and trying to kill us.

I think it will be because AI is stupid and kills us by accident.

Indeed, even going back through those 7 ways we’ve interacted with other species, the ones that have killed the most were 3 and 5—which, in both cases, we did not want to destroy them. In option 3, we in fact specifically wanted to not destroy them. Whenever we wiped out a species by over-exploiting it, we would have been smarter to not do that.

The central message about AI in If Anyone Builds It, Everyone Dies seems to be this:

Don’t make it smarter. If it’s smarter, we’re doomed.”

I, on the other hand, think that the far more important message is these:

Don’t trust it.

Don’t give it power.

Don’t let it make important decisions.

It won’t be smarter than us any time soon—but it doesn’t need to be in order to be dangerous. Indeed, there is even reason to believe that making AI smarter—genuinely, truly smarter, thinking more like an actual person and less like a pattern-matching machine—could actually make it safer and better for us. If we could somehow instill a capacity for morality and love in an AGI, it might actually start treating us the way we treat cats and dogs.

Of course, we have no idea how to do that. But that’s because we’re actually really bad at this, and nowhere near making a truly superhuman AGI.

In Nozicem

Nov 2 JDN 2460982

(I wasn’t sure how to convert Robert Nozick’s name into Latin. I decided it’s a third-declension noun, Nozix, Nozicis. But my name already is Latin, so if one of his followers ever wants to write a response to this post that also references In Catalinam, they’ll know how to decline it; the accusative is Julium, if you please.)

This post is not at all topical. I have been too busy working on video game jams (XBOX Game Camp Detroit, and then the Epic Mega Jam, for which you can view my submission, The Middle of Nowhere, here!) to keep up with the news, and honestly I think I am psychologically better off for it.

Rather, this is a post I’ve been meaning to write for a long time, but never quite got around to.

It is about Robert Nozick, and why he was a bad philosopher, a bad person, and a significant source of harm to our society as a whole.

Nozick had a successful career at Harvard, and even became president of the American Philosophical Association. So it may seem that I am going out on quite a limb by saying he’s a bad philosopher.

But the philosophy for which he is best known, the thing that made his career, is not simply obviously false—it is evil. It is the sort of thing that one can only write if one is either a complete psychopath, utterly ignorant of history, or arguing in bad faith (or some combination of these).

It is summarized in this pithy quote that makes less moral sense than the philosophy of the Joker in The Dark Knight:

Taxation of earnings from labor is on a par with forced labor. Seizing the results of someone’s labor is equivalent to seizing hours from him and directing him to carry on various activities.

Anarchy, State, and Utopia (p.169)

I apologize in advance for my language, but I must say it:

NO IT FUCKING ISN’T.

At worst—at the absolute worst, when a government is utterly corrupt and tyrannical, provides no legitimate services whatsoever, contributes in no way to public goods, offers no security, and exists entirely to enrich its ruling class—which by the way is worse than almost any actual government that has ever existed, even including totalitarian dictators and feudal absolute monarchies—at worst, taxation is like theft.

Taxation, like theft, takes your wealth, not your labor.


Wealth is not labor.

Even wealth earned by wage income is not labor—and most wealth isn’t earned by wage income. Elon Musk is now halfway to a trillion dollars, and it’s not because he works a million times harder than you. (Nor is he a million times smarter than you, or even ten—perhaps not even one.) The majority of wealth—and the vast majority of top 1%, top 0.1%, and top 0.01% wealth—is capital that begets more capital, continuously further enriching those who could live just fine without ever working another day in their lives. Billionaire wealth is honestly so pathological at this point that it would be pathetic if it weren’t so appalling.

Even setting aside the historical brutality of slavery as it was actually implemented—especially in the United States, where slaves were racialized and commodified in a way that historically slaves usually weren’t—there is a very obvious, very bright, very hard line between taking someone’s wealth and forcing them to work.

Even a Greek prisoner of war who was bought by a Roman patrician to tutor his children—the sort of slave that actually had significant autonomy and lived better than an average person in Roman society—was fundamentally unfree in a way that no one has ever been made unfree by having to pay income tax. (And the Roman patrician who owned him and (ahem) paid taxes was damn well aware of how much more free he was than his slave.)

Whether you are taxed at 2% or 20% or 90%, you are still absolutely free to use your time however you please. Yes, if you assume a fixed amount of work at a fixed wage, and there are no benefits to you from the taxation (which is really not something we can assume, because having a good or bad government radically affects what your economy as a whole will be like), you will have less stuff, and if you insist for some reason that you must have the same amount of stuff, then you would have to work more.

But even then, you would merely have to work more somewhere—anywhere—in order to make up the shortfall. You could keep your current job, or get another one, or start your own business. And you could at any time decide that you don’t need all that extra stuff and don’t want to work more, and simply choose to not work more. You are, in other words, still free.

At worst, the government has taken your stuff. It has made you poorer. But absolutely not, in no way, shape or form, has it made you a slave.

Yes, there is the concept of “wage slavery”, but “wage slavery” isn’t actually slavery, and the notion that people aren’t really, truly free unless they can provide for basic needs entails the need for a strong, redistributive government, which is the exact opposite of what Robert Nozick and his shockingly large body of followers have been arguing for since the 1970s.

I could have been sympathetic to Nozick if his claim had been this:

Taxation of earnings from labor is on a par with [theft]. Seizing the results of someone’s labor is equivalent to seizing [goods he has purchased with his own earnings].

Or even this:

[Military conscription] is on a par with forced labor. [After all, you are] seizing hours from him and directing him to carry on various activities.

Even then, there are some very clear reasons why we might be willing to accept taxation or even conscription from a legitimate liberal democratic government even though a private citizen doing the same fundamental activity would obviously be illegal and immoral.

Indeed, it’s not clear that theft is always immoral; there is always the Les Miserables exception where someone desperately poor steals food to feed themselves, and a liberal democratic government taxing its citizens in order to provide food stamps seems even more ethically defensible than that.

And that, my friends, is precisely why Nozick wasn’t satisfied with it.

Precisely because there is obvious nuance here that can readily justify at least some degree of not only taxation for national security and law enforcement, but also taxation for public goods and even redistribution of wealth, Nozick could not abide the analogies that actually make sense. He had to push beyond them to an analogy that is transparently absurd, in order to argue for his central message that government is justifiable for national security and law enforcement only, and all other government functions are inherently immoral. Forget clean water and air. Forget safety regulations in workplaces—or even on toys. Forget public utilities—all utilities must be privatized and unregulated. And above all—above all—forget ever taking any money from the rich to help the poor, because that would be monstrous.

If you support food stamps, in Nozick’s view, there should be a statue of you in Mississippi, because you are a defender of slavery.

Indeed, many of his followers have gone beyond that, and argued using the same core premises that all government is immoral, and the only morally justifiable system is anarcho-capitalism—which, I must confess, I have always had trouble distinguishing from feudalism with extra steps.

Nozick’s response to this kind of argument basically seemed to be that he thought anarcho-capitalism will (somehow, magically) automatically transition into his favored kind of minarchist state, and so it’s actually a totally fine intermediate goal. (A fully privatized military and law enforcement system! What could possibly go wrong? It’s not like private prisons are already unconscionably horrible even in an otherwise mostly-democratic system or anything!)

Nozick wanted to absolve himself—and the rich, especially the rich, whom he seemed to love more than life itself—from having to contribute to society, from owing anything to any other human being.

Rather than be moved by our moral appeals that millions of innocent people are suffering and we could so easily alleviate that suffering by tiny, minuscule, barely-perceptible harms to those who are already richer than anyone could possibly deserve to be, he tried to turn the tables: “No, you are immoral. What you want is slavery.

And in so doing, he created a thin, but shockingly resilient, intellectual veneer to the most craven selfishness and the most ideologically blinkered hyper-capitalism. He made it respectable to oppose even the most basic ways that governments can make human life better; by verbal alchemy he transmuted plain evil into its own new moral crusade.

Indeed, perhaps the only reason his philosophy was ever taken seriously is that the rich and powerful found it very, very, useful.

Reflections on the Charlie Kirk assassination

Sep 28 JDN 2460947

No doubt you are well aware that Charlie Kirk was shot and killed on September 10. His memorial service was held on September 21, and filled a stadium in Arizona.

There have been a lot of wildly different takes on this event. It’s enough to make you start questioning your own sanity. So while what I have to say may not be that different from what Krugman (or for that matter Jacobin) had to say, I still thought I would try to contribute to the small part of the conversation that’s setting the record straight.

First of all, let me say that this is clearly a political assassination, and as a matter of principle, that kind of thing should not be condoned in a democracy.

The whole point of a democratic system is that we don’t win by killing or silencing our opponents, we win by persuading or out-voting them. As long as someone is not engaging in speech acts that directly command or incite violence (like, say, inciting people to attack the Capitol), they should be allowed to speak in peace; even abhorrent views should be not be met with violence.

Free speech isn’t just about government censorship (though that is also a major problem right now); it’s a moral principle that underlies the foundation of liberal democracy. We don’t resolve conflicts with violence unless absolutely necessary.

So I want to be absolutely clear about this: Killing Charlie Kirk was not acceptable, and the assassin should be tried in a court of law and, if duly convicted, imprisoned for a very long time.

Second of all, we still don’t know the assassin’s motive, so stop speculating until we do.

At first it looked like the killer was left-wing. Then it looked like maybe he was right-wing. Now it looks like maybe he’s left-wing again. Maybe his views aren’t easily categorized that way; maybe he’s an anarcho-capitalist, or an anarcho-communist, or a Scientologist. I won’t say it doesn’t matter; it clearly does matter. But we simply do not know yet.

There is an incredibly common and incredibly harmful thing that people do after any major crime: They start spreading rumors and speculating about things that we actually know next to nothing about. Stop it. Don’t contribute to that.


The whole reason we have a court system is to actually figure out the real truth, which takes a lot of time and effort. The courts are one American institution that’s actually still functioning pretty well in this horrific cyberpunk/Trumpistan era; let them do their job.

It could be months or years before we really fully understand what happened here. Accept that. You don’t need to know the answer right now, and it’s far more dangerous to think you know the answer when you actually don’t.

But finally, I need to point out that Charlie Kirk was an absolutely abhorrent, despicable husk of a human being and no one should be honoring him.

First of all, he himself advocated for political violence against his opponents. I won’t say anyone deserves what happened to him—but if anyone did, it would be him, because he specifically rallied his followers to do exactly this sort of thing to other people.

He was also bigoted in almost every conceivable way: Racist, sexist, ableist, homophobic, and of course transphobic. He maintained a McCarthy-esque list of college professors that he encouraged people to harass for being too left-wing. He was a covert White supremacist, and only a little bit covert. He was not covert at all about his blatant sexism and misogyny that seems like it came from the 1950s instead of the 2020s.

He encouraged his—predominantly White, male, straight, cisgender, middle-class—audience to hate every marginalized group you can think of: women, people of color, LGBT people, poor people, homeless people, people with disabilities. Not content to merely be an abhorrent psychopath himself, he actively campaigned against the concept of empathy.

Charlie Kirk deserves no honors. The world is better off without him. He made his entire career out of ruining the lives of innocent people and actively making the world a worse place.

It was wrong to kill Charlie Kirk. But if you’re sad he’s gone, what is wrong with you!?

For my mother, on her 79th birthday

Sep 21 JDN 2460940

When this post goes live, it will be mother’s 79th birthday. I think birthdays are not a very happy time for her anymore.

I suppose nobody really likes getting older; children are excited to grow up, but once you hit about 25 or 26 (the age at which you can rent a car at the normal rate and the age at which you have to get your own health insurance, respectively) and it becomes “getting older” instead of “growing up”, the excitement rapidly wears off. Even by 30, I don’t think most people are very enthusiastic about their birthdays. Indeed, for some people, I think it might be downhill past 21—you wanted to become an adult, but you had no interest in aging beyond that point.

But I think it gets worse as you get older. As you get into your seventies and eighties, you begin to wonder which birthday will finally be your last; actually I think my mother has been wondering about this even earlier than that, because her brothers died in their fifties, her sister died in her sixties, and my father died at 63. At this point she has outlived a lot of people she loved. I think there is a survivor’s guilt that sets in: “Why do I get to keep going, when they didn’t?”

These are also very hard times in general; Trump and the people who enable him have done tremendous damage to our government, our society, and the world at large in a shockingly short amount of time. It feels like all the safeguards we were supposed to have suddenly collapsed and we gave free rein to a madman.

But while there are many loved ones we have lost, there are many we still have; and nor need our set of loved ones be fixed, only to dwindle with each new funeral. We can meet new people, and they can become part of our lives. New children can be born into our family, and they can make our family grow. It is my sincere hope that my mother still has grandchildren yet to meet; in my case they would probably need to be adopted, as the usual biological route is pretty much out of the question, and surrogacy seems beyond our budget for the foreseeable future. But we would still love them, and she could still love them, and it is worth sticking around in this world in order to be a part of their lives.

I also believe that this is not the end for American liberal democracy. This is a terrible time, no doubt. Much that we thought would never happen already has, and more still will. It must be so unsettling, so uncanny, for someone who grew up in the triumphant years after America helped defeat fascism in Europe, to grow older and then see homegrown American fascism rise ascendant here. Even those of us who knew history all too well still seem doomed to repeat it.

At this point it is clear that victory over corruption, racism, and authoritarianism will not be easy, will not be swift, may never be permanent—and is not even guaranteed. But it is still possible. There is still enough hope left that we can and must keep fighting for an America worth saving. I do not know when we will win; I do not even know for certain that we will, in fact, win. But I believe we will.

I believe that while it seems powerful—and does everything it can to both promote that image and abuse what power it does have—fascism is a fundamentally weak system, a fundamentally fragile system, which simply cannot sustain itself once a handful of critical leaders are dead, deposed, or discredited. Liberal democracy is kinder, gentler—and also slower, at times even clumsier—than authoritarianism, and so it may seem weak to those whose view of strength is that of the savanna ape or the playground bully; but this is an illusion. Liberal democracy is fundamentally strong, fundamentally resilient. There is power in kindness, inclusion, and cooperation that the greedy and cruel cannot see. Fascism in Germany arrived and disappeared within a generation; democracy in America has stood for nearly 250 years.

We don’t know how much more time we have, Mom; none of us do. I have heard it said that you should live your life as though you will live both a short life and a long one; but honestly, you should probably live your life as though you will live a randomly-decided amount of time that is statistically predicted by actuarial tables—because you will. Yes, the older you get, the less time you have left (almost tautologically); but especially in this age of rapid technological change, none of us really know whether we’ll die tomorrow or live another hundred years.

I think right now, you feel like there isn’t much left to look forward to. But I promise you there is. Maybe it’s hard to see right now; indeed, maybe you—or I, or anyone—won’t even ever get to see it. But a brighter future is possible, and it’s worth it to keep going, especially if there’s any way that we might be able to make that brighter future happen sooner.

Grief, a rationalist perspective

Aug 31 JDN 2460919

This post goes live on the 8th anniversary of my father’s death. Thus it seems an appropriate time to write about grief—indeed, it’s somewhat difficult for me to think about much else.

Far too often, the only perspectives on grief we hear are religious ones. Often, these take the form of consolation: “He’s in a better place now.” “You’ll see him again someday.”

Rationalism doesn’t offer such consolations. Technically one can be an atheist and still believe in an afterlife; but rationalism is stronger than mere atheism. It requires that we believe in scientific facts, and the permanent end of consciousness at death is a scientific fact. We know from direct experiments and observations in neuroscience that a destroyed brain cannot think, feel, see, hear, or remember—when your brain shuts down, whatever you are now will be gone.

It is the Basic Fact of Cognitive Science: There is no soul but the brain.

Moreover, I think, deep down, we all know that death is the end. Even religious people grieve. Their words may say that their loved one is in a better place, but their tears tell a different story.

Maybe it’s an evolutionary instinct, programmed deep into our minds like an ancestral memory, a voice that screams in our minds, insistent on being heard:

Death is bad!”

If there is one crucial instinct a lifeform needs in order to survive, surely it is something like that one: The preference for life over death. In order to live in a hostile world, you have to want to live.

There are some people who don’t want to live, people who become suicidal. Sometimes even the person we are grieving was someone who chose to take their own life. Generally this is because they believe that their life from then on would be defined only by suffering. Usually, I would say they are wrong about that; but in some cases, maybe they are right, and choosing death is rational. Most of the time, life is worth living, even when we can’t see that.

But aside from such extreme circumstances, most of us feel most of the time that death is one of the worst things that could happen to us or our loved ones. And it makes sense that we feel that way. It is right to feel that way. It is rational to feel that way.

This is why grief hurts so much.

This is why you are not okay.

If the afterlife were real—or even plausible—then grief would not hurt so much. A loved one dying would be like a loved one traveling away to somewhere nice; bittersweet perhaps, maybe even sad—but not devastating the way that grief is. You don’t hold a funeral for someone who just booked a one-way trip to Hawaii, even if you know they aren’t ever coming back.

Religion tries to be consoling, but it typically fails. Because that voice in our heads is still there, repeating endlessly: “Death is bad!” “Death is bad!” “Death is bad!”

But what if religion does give people some comfort in such a difficult time? What if supposing something as nonsensical as Heaven numbs the pain for a little while?

In my view, you’d be better off using drugs. Drugs have side effects and can be addictive, but at least they don’t require you to fundamentally abandon your ontology. Mainstream religion isn’t simply false; it’s absurd. It’s one of the falsest things anyone has ever believed about anything. It’s obviously false. It’s ridiculous. It has never deserved any of the respect and reverence it so often receives.

And in a great many cases, religion is evil. Religion teaches people to be obedient to authoritarians, and to oppress those who are different. Some of the greatest atrocities in history were committed in the name of religion, and some of the worst oppression going on today is done in the name of religion.

Rationalists should give religion no quarter. It is better for someone to find solace in alcohol or cannabis than for them to find solace in religion.

And maybe, in the end, it’s better if they don’t find solace at all.

Grief is good. Grief is healthy. Grief is what we should feel when something as terrible as death happens. That voice screaming “Death is bad!” is right, and we should listen to it.

No, what we need is to not be paralyzed by grief, destroyed by grief. We need to withstand our grief, get through it. We must learn to be strong enough to bear what seems unbearable, not console ourselves with lies.

If you are a responsible adult, then when something terrible happens to you, you don’t pretend it isn’t real. You don’t conjure up a fantasy world in which everything is fine. You face your terrors. You learn to survive them. You make yourself strong enough to carry on. The death of a loved one is a terrible thing; you shouldn’t pretend otherwise. But it doesn’t have to destroy you. You can grow, and heal, and move on.

Moreover, grief has a noble purpose. From our grief we must find motivation to challenge death, to fight death wherever we find it. Those we have already lost are gone; it’s too late for them. But it’s not too late for the rest of us. We can keep fighting.

And through economic development and medical science, we do keep fighting.

In fact, little by little, we are winning the war on death.

Death has already lost its hold upon our children. For most of human history, nearly a third of children died before the age of 5. Now less than 1% do, in rich countries, and even in the poorest countries, it’s typically under 10%. With a little more development—development that is already happening in many places—we can soon bring everyone in the world to the high standard of the First World. We have basically won the war on infant and child mortality.

And death is losing its hold on the rest of us, too. Life expectancy at adulthood is also increasing, and more and more people are living into their nineties and even their hundreds.

It’s true, there still aren’t many people living to be 120 (and some researchers believe it will be a long time before this changes). But living to be 85 instead of 65 is already an extra 20 years of life—and these can be happy, healthy years too, not years of pain and suffering. They say that 60 is the new 50; physiologically, we are so much healthier than our ancestors that it’s as if we were ten years younger.

My sincere hope is that our grief for those we have lost and fear of losing those we still have will drive us forward to even greater progress in combating death. I believe that one day we will finally be able to slow, halt, perhaps even reverse aging itself, rendering us effectively immortal.

Religion promises us immortality, but it isn’t real.

Science offers us the possibility of immortality that’s real.

It won’t be easy to get there. It won’t happen any time soon. In all likelihood, we won’t live to see it ourselves. But one day, our descendants may achieve the grandest goal of all: Finally conquering death.

And even long before that glorious day, our lives are already being made longer and healthier by science. We are pushing death back, step by step, day by day. We are fighting, and we are winning.

Moreover, we as individuals are not powerless in this fight: you can fight death a little harder yourself, by becoming an organ donor, or by donating to organizations that fight global poverty or advance medical science. Let your grief drive you to help others, so that they don’t have to grieve as you do.

And if you need consolation from your grief, let it come from this truth: Death is rarer now today than it was yesterday, and will be rarer still tomorrow. We can’t bring back who we have lost, but we can keep ourselves from losing more so soon.

Conflict without shared reality

Aug 17 JDN 2460905

Donald Trump has federalized the police in Washington D.C. and deployed the National Guard. He claims he is doing this in response to a public safety emergency and crime that is “out of control”.

Crime rates in Washington, D.C. are declining and overall at their lowest level in 30 years. Its violent crime rate has not been this low since the 1960s.

By any objective standard, there is no emergency here. Crime in D.C. is not by any means out of control.

Indeed, across the United States, homicide rates are as low as they have been in 60 years.

But we do not live in a world where politics is based on objective truth.

We live in a world where the public perception of reality itself is shaped by the political narrative.

One of the first things that authoritarians do to control these narratives is try to make their followers distrust objective sources. I watch in disgust as not simply the Babylon Bee (which is a right-wing satire site that tries really hard to be funny but never quite manages it) but even the Atlantic (a mainstream news outlet generally considered credible) feeds—in multiple articles—into this dangerous lie that crime is increasing and the official statistics are somehow misleading us about that.

Of course the Atlantic‘s take is much more nuanced; but quite frankly, now is not the time for nuance. A fascist is trying to take over our government, and he needs to be resisted at every turn by every means possible. You need to be calling him out on every single lie he makes—yes, every single one, I know there are a lot of them, and that’s kind of the point—rather than trying to find alternative framings on which maybe part of what he said could somehow be construed as reasonable from a certain point of view. Every time you make Trump sound more reasonable than he is—and mainstream news outlets have done this literally hundreds of times—you are pushing America closer to fascism.

I really don’t know what to do here.

It is impossible to resolve conflicts when they are not based on shared reality.

No policy can solve a crime wave that doesn’t exist. No trade agreement can stop unfair trading practices that aren’t happening. Nothing can stop vaccines from causing autism that they already don’t cause. There is no way to fix problems when those problems are completely imaginary.

I used to think that political conflict was about different values which had to be balanced against one another: Liberty versus security, efficiency versus equality, justice versus mercy. I thought that we all agreed on the basic facts and even most of the values, and were just disagreeing about how to weigh certain values over others.

Maybe I was simply naive; maybe it’s never been like that. But it certainly isn’t right now. We aren’t disagreeing about what should be done; we are disagreeing about what is happening in front of our eyes. We don’t simply have different priorities or even different values; it’s like we are living in different worlds.

I have read, e.g. by Jonathan Haidt, that conservatives largely understand what liberals want, but liberals don’t really understand what conservatives want. (I would like to take one of the tests they use in these experiments, see how I actually do; but I’ve never been able to find one.)

Haidt’s particular argument seems to be that liberals don’t “understand” the “moral dimensions” of loyalty, authority, and sanctity, because we only “understand” harm and fairness as the basis of morality. But just because someone says something is morally relevant, that doesn’t mean it is morally relevant! And indeed, based on more or less the entirety of ethical philosophy, I can say that harm and fairness are morality, and the others simply aren’t. They are distortions of morality, they are inherently evil, and we are right to oppose them at every turn. Loyalty, authority, and sanctity are what fed Nazi Germany and the Spanish Inquisition.

This claim that liberals don’t understand conservatives has always seemed very odd to me: I feel like I have a pretty clear idea what conservatives want, it’s just that what they want is terrible: Kick out the immigrants, take money from the poor and give it to the rich, and put rich straight Christian White men back in charge of everything. (I mean, really, if that’s not what they want, why do they keep voting for people who do it? Revealed preferences, people!)

Or, more sympathetically: They want to go back to a nostalgia-tinted vision of the 1950s and 1960s in which it felt like things were going well for our country—because they were blissfully ignorant of all the violence and injustice in the world. No, thank you, Black people and queer people do not want to go back to how we were treated in the 1950s—when segregation was legal and Alan Turing was chemically castrated. (And they also don’t seem to grasp that among the things that did make some things go relatively well in that period were unions, antitrust law and progressive taxes, which conservatives now fight against at every turn.)

But I think maybe part of what’s actually happening here is that a lot of conservatives actually “want” things that literally don’t make sense, because they rest upon assumptions about the world that simply aren’t true.

They want to end “out of control” crime that is the lowest it’s been in decades.

They want to stop schools from teaching things that they already aren’t teaching.

They want the immigrants to stop bringing drugs and crime that they aren’t bringing.

They want LGBT people to stop converting their children, which we already don’t and couldn’t. (And then they want to do their own conversions in the other direction—which also don’t work, but cause tremendous harm.)

They want liberal professors to stop indoctrinating their students in ways we already aren’t and can’t. (If we could indoctrinate our students, don’t you think we’d at least make them read the syllabus?)

They want to cut government spending by eliminating “waste” and “fraud” that are trivial amounts, without cutting the things that are actually expensive, like Social Security, Medicare, and the military. They think we can balance the budget without cutting these things or raising taxes—which is just literally mathematically impossible.

They want to close off trade to bring back jobs that were sent offshore—but those jobs weren’t sent offshore, they were replaced by robots. (US manufacturing output is near its highest ever, even though manufacturing employment is half what it once was.)


And meanwhile, there’s a bunch of real problems that aren’t getting addressed: Soaring inequality, a dysfunctional healthcare system, climate change, the economic upheaval of AI—and they either don’t care about those, aren’t paying attention to them, or don’t even believe they exist.

It feels a bit like this:

You walk into a room and someone points a gun at you, shouting “Drop the weapon!” but you’re not carrying a weapon. And you show your hands, and try to explain that you don’t have a weapon, but they just keep shouting “Drop the weapon!” over and over again. Someone else has already convinced them that you have a weapon, and they expect you to drop that weapon, and nothing you say can change their mind about this.

What exactly should you do in that situation?

How do you avoid getting shot?

Do you drop something else and say it’s the weapon (make some kind of minor concession that looks vaguely like what they asked for)? Do you try to convince them that you have a right to the weapon (accept their false premise but try to negotiate around it)? Do you just run away (leave the country?)? Do you double down and try even harder to convince them that you really, truly, have no weapon?

I’m not saying that everyone on the left has a completely accurate picture of reality; there are clearly a lot of misconceptions on this side of the aisle as well. But at least among the mainstream center left, there seems to be a respect for objective statistics and a generally accurate perception of how the world works—the “reality-based community”. Sometimes liberals make mistakes, have bad ideas, or even tell lies; but I don’t hear a lot of liberals trying to fix problems that don’t exist or asking for the government budget to be changed in ways that violate basic arithmetic.

I really don’t know what do here, though.

How do you change people’s minds when they won’t even agree on the basic facts?

On foxes and hedgehogs, part I

Aug 3 JDN 2460891

Today I finally got around to reading Expert Political Judgment by Philip E. Tetlock, more or less in a single sitting because I’ve been sick the last week with some pretty tight limits on what activities I can do. (It’s mostly been reading, watching TV, or playing video games that don’t require intense focus.)

It’s really an excellent book, and I now both understand why it came so highly recommended to me, and now pass on that recommendation to you: Read it.

The central thesis of the book really boils down to three propositions:

  1. Human beings, even experts, are very bad at predicting political outcomes.
  2. Some people, who use an open-minded strategy (called “foxes”), perform substantially better than other people, who use a more dogmatic strategy (called “hedgehogs”).
  3. When rewarding predictors with money, power, fame, prestige, and status, human beings systematically favor (over)confident “hedgehogs” over (correctly) humble “foxes”.

I decided I didn’t want to make this post about current events, but I think you’ll probably agree with me when I say:

That explains a lot.

How did Tetlock determine this?

Well, he studies the issue several different ways, but the core experiment that drives his account is actually a rather simple one:

  1. He gathered a large group of subject-matter experts: Economists, political scientists, historians, and area-studies professors.
  2. He came up with a large set of questions about politics, economics, and similar topics, which could all be formulated as a set of probabilities: “How likely is this to get better/get worse/stay the same?” (For example, this was in the 1980s, so he asked about the fate of the Soviet Union: “By 1990, will they become democratic, remain as they are, or collapse and fragment?”)
  3. Each respondent answered a subset of the questions, some about their own particular field, some about another, more distant field; they assigned probabilities on an 11-point scale, from 0% to 100% in increments of 10%.
  4. A few years later, he compared the predictions to the actual results, scoring them using a Brier score, which penalizes you for assigning high probability to things that didn’t happen or low probability to things that did happen.
  5. He compared the resulting scores between people with different backgrounds, on different topics, with different thinking styles, and a variety of other variables. He also benchmarked them using some automated algorithms like “always say 33%” and “always give ‘stay the same’ 100%”.

I’ll show you the key results of that analysis momentarily, but to help it make more sense to you, let me elaborate a bit more on the “foxes” and “hedgehogs”. The notion is was first popularized by Isaiah Berlin in an essay called, simply, The Hedgehog and the Fox.

“The fox knows many things, but the hedgehog knows one very big thing.”

That is, someone who reasons as a “fox” combines ideas from many different sources and perspective, and tries to weigh them all together into some sort of synthesis that then yields a final answer. This process is messy and complicated, and rarely yields high confidence about anything.

Whereas, someone who reasons as a “hedgehog” has a comprehensive theory of the world, an ideology, that provides clear answers to almost any possible question, with the surely minor, insubstantial flaw that those answers are not particularly likely to be correct.

He also considered “hedge-foxes” (people who are mostly fox but also a little bit hedgehog) and “fox-hogs” (people who are mostly hedgehog but also a little bit fox).

Tetlock has decomposed the scores into two components: calibration and discrimination. (Both very overloaded words, but they are standard in the literature.)

Calibration is how well your stated probabilities matched up with the actual probabilities; that is, if you predicted 10% probability on 20 different events, you have very good calibration if precisely 2 of those events occurred, and very poor calibration if 18 of those events occurred.

Discrimination more or less describes how useful your predictions are, what information they contain above and beyond the simple base rate. If you just assign equal probability to all events, you probably will have reasonably good calibration, but you’ll have zero discrimination; whereas if you somehow managed to assign 100% to everything that happened and 0% to everything that didn’t, your discrimination would be perfect (and we would have to find out how you cheated, or else declare you clairvoyant).

For both measures, higher is better. The ideal for each is 100%, but it’s virtually impossible to get 100% discrimination and actually not that hard to get 100% calibration if you just use the base rates for everything.


There is a bit of a tradeoff between these two: It’s not too hard to get reasonably good calibration if you just never go out on a limb, but then your predictions aren’t as useful; we could have mostly just guessed them from the base rates.

On the graph, you’ll see downward-sloping lines that are meant to represent this tradeoff: Two prediction methods that would yield the same overall score but different levels of calibration and discrimination will be on the same line. In a sense, two points on the same line are equally good methods that prioritize usefulness over accuracy differently.

All right, let’s see the graph at last:

The pattern is quite clear: The more foxy you are, the better you do, and the more hedgehoggy you are, the worse you do.

I’d also like to point out the other two regions here: “Mindless competition” and “Formal models”.

The former includes really simple algorithms like “always return 33%” or “always give ‘stay the same’ 100%”. These perform shockingly well. The most sophisticated of these, “case-specific extrapolation” (35 and 36 on the graph, which basically assumes that each country will continue doing what it’s been doing) actually performs as well if not better than even the foxes.

And what’s that at the upper-right corner, absolutely dominating the graph? That’s “Formal models”. This describes basically taking all the variables you can find and shoving them into a gigantic logit model, and then outputting the result. It’s computationally intensive and requires a lot of data (hence why he didn’t feel like it deserved to be called “mindless”), but it’s really not very complicated, and it’s the best prediction method, in every way, by far.

This has made me feel quite vindicated about a weird nerd thing I do: When I have a big decision to make (especially a financial decision), I create a spreadsheet and assemble a linear utility model to determine which choice will maximize my utility, under different parameterizations based on my past experiences. Whichever result seems to win the most robustly, I choose. This is fundamentally similar to the “formal models” prediction method, where the thing I’m trying to predict is my own happiness. (It’s a bit less formal, actually, since I don’t have detailed happiness data to feed into the regression.) And it has worked for me, astonishingly well. It definitely beats going by my own gut. I highly recommend it.

What does this mean?

Well first of all, it means humans suck at predicting things. At least for this data set, even our experts don’t perform substantially better than mindless models like “always assume the base rate”.

Nor do experts perform much better in their own fields than in other fields; they do all perform better than undergrads or random people (who somehow perform worse than the “mindless” models)

But Tetlock also investigates further, trying to better understand this “fox/hedgehog” distinction and why it yields different performance. He really bends over backwards to try to redeem the hedgehogs, in the following ways:

  1. He allows them to make post-hoc corrections to their scores, based on “value adjustments” (assigning higher probability to events that would be really important) and “difficulty adjustments” (assigning higher scores to questions where the three outcomes were close to equally probable) and “fuzzy sets” (giving some leeway on things that almost happened or things that might still happen later).
  2. He demonstrates a different, related experiment, in which certain manipulations can cause foxes to perform a lot worse than they normally would, and even yield really crazy results like probabilities that add up to 200%.
  3. He has a whole chapter that is a Socratic dialogue (seriously!) between four voices: A “hardline neopositivist”, a “moderate neopositivist”, a “reasonable relativist”, and an “unrelenting relativist”; and all but the “hardline neopositivist” agree that there is some legitimate place for the sort of post hoc corrections that the hedgehogs make to keep themselves from looking so bad.

This post is already getting a bit long, so that will conclude part I. Stay tuned for part II, next week!

Bayesian updating with irrational belief change

Jul 27 JDN 2460884

For the last few weeks I’ve been working at a golf course. (It’s a bit of an odd situation: I’m not actually employed by the golf course; I’m contracted by a nonprofit to be a “job coach” for a group of youths who are part of a work program that involves them working at the golf course.)

I hate golf. I have always hated golf. I find it boring and pointless—which, to be fair, is my reaction to most sports—and also an enormous waste of land and water. A golf course is also a great place for oligarchs to arrange collusion.

But I noticed something about being on the golf course every day, seeing people playing and working there: I feel like I hate it a bit less now.

This is almost certainly a mere-exposure effect: Simply being exposed to something many times makes it feel familiar, and that tends to make you like it more, or at least dislike it less. (There are some exceptions: repeated exposure to trauma can actually make you more sensitive to it, hating it even more.)

I kinda thought this would happen. I didn’t really want it to happen, but I thought it would.

This is very interesting from the perspective of Bayesian reasoning, because it is a theorem (though I cannot seem to find anyone naming the theorem; it’s like a folk theorem, I guess?) of Bayesian logic that the following is true:

The prior expectation of the posterior is the expectation of the prior.

The prior is what you believe before observing the evidence; the posterior is what you believe afterward. This theorem describes a relationship that holds between them.

This theorem means that, if I am being optimally rational, I should take into account all expected future evidence, not just evidence I have already seen. I should not expect to encounter evidence that will change my beliefs—if I did expect to see such evidence, I should change my beliefs right now!

This might be easier to grasp with an example.

Suppose I am trying to predict whether it will rain at 5:00 pm tomorrow, and I currently estimate that the probability of rain is 30%. This is my prior probability.

What will actually happen tomorrow is that it will rain or it won’t; so my posterior probability will either be 100% (if it rains) or 0% (if it doesn’t). But I had better assign a 30% chance to the event that will make me 100% certain it rains (namely, I see rain), and a 70% chance to the event that will make me 100% certain it doesn’t rain (namely, I see no rain); if I were to assign any other probabilities, then I must not really think the probability of rain at 5:00 pm tomorrow is 30%.

(The keen Bayesian will notice that the expected variance of my posterior need not be the variance of my prior: My initial variance is relatively high (it’s actually 0.3*0.7 = 0.21, because this is a Bernoulli distribution), because I don’t know whether it will rain or not; but my posterior variance will be 0, because I’ll know the answer once it rains or doesn’t.)

It’s a bit trickier to analyze, but this also works even if the evidence won’t make me certain. Suppose I am trying to determine the probability that some hypothesis is true. If I expect to see any evidence that might change my beliefs at all, then I should, on average, expect to see just as much evidence making me believe the hypothesis more as I see evidence that will make me believe the hypothesis less. If that is not what I expect, I should really change how much I believe the hypothesis right now!

So what does this mean for the golf example?

Was I wrong to hate golf quite so much before, because I knew that spending time on a golf course might make me hate it less?

I don’t think so.

See, the thing is: I know I’m not perfectly rational.

If I were indeed perfectly rational, then anything I expect to change my beliefs is a rational Bayesian update, and I should indeed factor it into my prior beliefs.

But if I know for a fact that I am not perfectly rational, that there are things which will change my beliefs in ways that make them deviate from rational Bayesian updating, then in fact I should not take those expected belief changes into account in my prior beliefs—since I expect to be wrong later, updating on that would just make me wrong now as well. I should only update on the expected belief changes that I believe will be rational.

This is something that a boundedly-rational person should do that neither a perfectly-rational nor perfectly-irrational person would ever do!

But maybe you don’t find the golf example convincing. Maybe you think I shouldn’t hate golf so much, and it’s not irrational for me to change my beliefs in that direction.


Very well. Let me give you a thought experiment which provides a very clear example of a time when you definitely would think your belief change was irrational.


To be clear, I’m not suggesting the two situations are in any way comparable; the golf thing is pretty minor, and for the thought experiment I’m intentionally choosing something quite extreme.

Here’s the thought experiment.

A mad scientist offers you a deal: Take this pill and you will receive $50 million. Naturally, you ask what the catch is. The catch, he explains, is that taking the pill will make you staunchly believe that the Holocaust didn’t happen. Take this pill, and you’ll be rich, but you’ll become a Holocaust denier. (I have no idea if making such a pill is even possible, but it’s a thought experiment, so bear with me. It’s certainly far less implausible than Swampman.)

I will assume that you are not, and do not want to become, a Holocaust denier. (If not, I really don’t know what else to say to you right now. It happened.) So if you take this pill, your beliefs will change in a clearly irrational way.

But I still think it’s probably justifiable to take the pill. This is absolutely life-changing money, for one thing, and being a random person who is a Holocaust denier isn’t that bad in the scheme of things. (Maybe it would be worse if you were in a position to have some kind of major impact on policy.) In fact, before taking the pill, you could write out a contract with a trusted friend that will force you to donate some of the $50 million to high-impact charities—and perhaps some of it to organizations that specifically fight Holocaust denial—thus ensuring that the net benefit to humanity is positive. Once you take the pill, you may be mad about the contract, but you’ll still have to follow it, and the net benefit to humanity will still be positive as reckoned by your prior, more correct, self.

It’s certainly not irrational to take the pill. There are perfectly-reasonable preferences you could have (indeed, likely dohave) that would say that getting $50 million is more important than having incorrect beliefs about a major historical event.

And if it’s rational to take the pill, and you intend to take the pill, then of course it’s rational to believe that in the future, you will have taken the pill and you will become a Holocaust denier.

But it would be absolutely irrational for you to become a Holocaust denier right now because of that. The pill isn’t going to provide evidence that the Holocaust didn’t happen (for no such evidence exists); it’s just going to alter your brain chemistry in such a way as to make you believe that the Holocaust didn’t happen.

So here we have a clear example where you expect to be more wrong in the future.

Of course, if this really only happens in weird thought experiments about mad scientists, then it doesn’t really matter very much. But I contend it happens in reality all the time:

  • You know that by hanging around people with an extremist ideology, you’re likely to adopt some of that ideology, even if you really didn’t want to.
  • You know that if you experience a traumatic event, it is likely to make you anxious and fearful in the future, even when you have little reason to be.
  • You know that if you have a mental illness, you’re likely to form harmful, irrational beliefs about yourself and others whenever you have an episode of that mental illness.

Now, all of these belief changes are things you would likely try to guard against: If you are a researcher studying extremists, you might make a point of taking frequent vacations to talk with regular people and help yourself re-calibrate your beliefs back to normal. Nobody wants to experience trauma, and if you do, you’ll likely seek out therapy or other support to help heal yourself from that trauma. And one of the most important things they teach you in cognitive-behavioral therapy is how to challenge and modify harmful, irrational beliefs when they are triggered by your mental illness.

But these guarding actions only make sense precisely because the anticipated belief change is irrational. If you anticipate a rational change in your beliefs, you shouldn’t try to guard against it; you should factor it into what you already believe.

This also gives me a little more sympathy for Evangelical Christians who try to keep their children from being exposed to secular viewpoints. I think we both agree that having more contact with atheists will make their children more likely to become atheists—but we view this expected outcome differently.

From my perspective, this is a rational change, and it’s a good thing, and I wish they’d factor it into their current beliefs already. (Like hey, maybe if talking to a bunch of smart people and reading a bunch of books on science and philosophy makes you think there’s no God… that might be because… there’s no God?)

But I think, from their perspective, this is an irrational change, it’s a bad thing, the children have been “tempted by Satan” or something, and thus it is their duty to protect their children from this harmful change.

Of course, I am not a subjectivist. I believe there’s a right answer here, and in this case I’m pretty sure it’s mine. (Wouldn’t I always say that? No, not necessarily; there are lots of matters for which I believe that there are experts who know better than I do—that’s what experts are for, really—and thus if I find myself disagreeing with those experts, I try to educate myself more and update my beliefs toward theirs, rather than just assuming they’re wrong. I will admit, however, that a lot of people don’t seem to do this!)

But this does change how I might tend to approach the situation of exposing their children to secular viewpoints. I now understand better why they would see that exposure as a harmful thing, and thus be resistant to actions that otherwise seem obviously beneficial, like teaching kids science and encouraging them to read books. In order to get them to stop “protecting” their kids from the free exchange of ideas, I might first need to persuade them that introducing some doubt into their children’s minds about God isn’t such a terrible thing. That sounds really hard, but it at least clearly explains why they are willing to fight so hard against things that, from my perspective, seem good. (I could also try to convince them that exposure to secular viewpoints won’t make their kids doubt God, but the thing is… that isn’t true. I’d be lying.)

That is, Evangelical Christians are not simply incomprehensibly evil authoritarians who hate truth and knowledge; they quite reasonably want to protect their children from things that will harm them, and they firmly believe that being taught about evolution and the Big Bang will make their children more likely to suffer great harm—indeed, the greatest harm imaginable, the horror of an eternity in Hell. Convincing them that this is not the case—indeed, ideally, that there is no such place as Hell—sounds like a very tall order; but I can at least more keenly grasp the equilibrium they’ve found themselves in, where they believe that anything that challenges their current beliefs poses a literally existential threat. (Honestly, as a memetic adaptation, this is brilliant. Like a turtle, the meme has grown itself a nigh-impenetrable shell. No wonder it has managed to spread throughout the world.)

Universal human rights are more radical than is commonly supposed

Jul 13 JDN 2460870

We hold these truths to be self-evident, that all men are created equal, that they are endowed by their Creator with certain unalienable Rights, that among these are Life, Liberty and the pursuit of Happiness.

So begins the second paragraph of the Declaration of Independence. It had to have been obvious to many people, even at the time, how incredibly hypocritical it was for men to sign that document and then go home to give orders to their slaves.

And today, even though the Universal Declaration of Human Rights was signed over 75 years ago, there are still human rights violations ongoing in many different countries—including right here in the United States.

Why is it so easy to get people to declare that they believe in universal human rights—but so hard to get them to actually act accordingly?

Other moral issues are not like this. While hypocrisy certainly exists in many forms, for the most part people’s moral claims align with their behavior. Most people say they are against murder—and sure enough, most people aren’t murderers. Most people say they are against theft—and indeed, most people don’t steal very often. And when it comes to things that most people do all the time, most people aren’t morally opposed to them—even things like eating meat, for which there is a pretty compelling moral case against it.

But universal human rights seems like something that is far more honored in the breach than the observance.

I think this is because most people don’t quite grasp just how radical universal human rights really are.

The tricky part is the universal. They are supposed to apply to everyone.

Even those people. Even the people you are thinking of right now as an exception. Even the people you hate the most. Yes, even them.

Depending on who you are, you might be thinking of different exceptions: People of a particular race, or religion, or nationality, perhaps; or criminals, or terrorists; or bigots, or fascists. But almost everyone has some group of people that they don’t really think deserves the full array of human rights.

So I am here to tell you that, yes, those people too. Universal human rights means everyone.

No exceptions.

This doesn’t mean that we aren’t allowed to arrest and imprison people for crimes. It doesn’t even mean that we aren’t sometimes justified in killing people—e.g. in war or self-defense. But it does mean that there is no one, absolutely no one, who is considered beneath human dignity. Any time we are to deprive someone of life or liberty, we must do so with absolute respect for their fundamental rights.

This also means that there is no one you should be spitting on, no one you should be torturing, no one you should be calling dehumanizing names. Sometimes violence is necessary, to protect yourself, or to preserve liberty, or to overthrow tyranny. But yes, even psychopathic tyrants are human beings, and still deserve human rights. If you cannot recognize a person’s humanity while still defending yourself against them, you need to do some serious soul-searching and ask yourself why not.

I think what happens when most people are asked about “universal human rights”, they essentially exclude whoever they think doesn’t deserve rights from the very category of “human”. Then it essentially becomes a tautology: Everyone who deserves rights deserves rights.

And thus, everyone signs onto it—but it ends up meaning almost nothing. It doesn’t stop racism, or sexism, or police brutality, or mass incarceration, or rape, or torture, or genocide, because the people doing those things don’t think of the people they’re doing them to as actually human.

But no, the actual declaration says all human beings. Everyone. Even the people you hate. Even the people who hate you. Even people who want to torture and kill you. Yes, even them.

This is an incredibly radical idea.

It is frankly alien to a brain that evolved for tribalism; we are wired to think of the world in terms of in-groups and out-groups, and universal human rights effectively declare that everyone is in the in-group and the out-group doesn’t exist.

Indeed, perhaps too radical! I think a reasonable defense could be made of a view that some people (psychopathic tyrants?) really are just so evil that they don’t actually deserve basic human dignity. But I will say this: Usually the people arguing that some group of humans aren’t really humans ends up being on the wrong side of history.

The one possible exception I can think of here is abortion: The people arguing that fetuses are not human beings and it should be permissible to kill them when necessary are, at least in my view, generally on the right side of history. But even then, I tend to be much more sympathetic to the view that abortion, like war and self-defense, should be seen as a tragically necessary evil, not an inherent good. The ideal scenario would be to never need it, and allowing it when it’s needed is simply a second-best solution. So I think we can actually still fit this into a view that fetuses are morally important and deserving of dignity; it’s just that sometimes that the rights of one being can outweigh the rights of another.

And other than that, yeah, it’s pretty much the case that the people who want to justify enacting some terrible harm on some group of people because they say those people aren’t really people, end up being the ones that, sooner or later, the world recognizes as the bad guys.

So think about that, if there is still some group of human beings that you think of as not really human beings, not really deserving of universal human rights. Will history vindicate you—or condemn you?