Information theory proves that multiple-choice is stupid

Mar 19, JDN 2457832

This post is a bit of a departure from my usual topics, but it’s something that has bothered me for a long time, and I think it fits broadly into the scope of uniting economics with the broader realm of human knowledge.

Multiple-choice questions are inherently and objectively poor methods of assessing learning.

Consider the following question, which is adapted from actual tests I have been required to administer and grade as a teaching assistant (that is, the style of question is the same; I’ve changed the details so that it wouldn’t be possible to just memorize the response—though in a moment I’ll get to why all this paranoia about students seeing test questions beforehand would also be defused if we stopped using multiple-choice):

The demand for apples follows the equation Q = 100 – 5 P.
The supply of apples follows the equation Q = 10 P.
If a tax of $2 per apple is imposed, what is the equilibrium price, quantity, tax revenue, consumer surplus, and producer surplus?

A. Price = $5, Quantity = 10, Tax revenue = $50, Consumer Surplus = $360, Producer Surplus = $100

B. Price = $6, Quantity = 20, Tax revenue = $40, Consumer Surplus = $200, Producer Surplus = $300

C. Price = $6, Quantity = 60, Tax revenue = $120, Consumer Surplus = $360, Producer Surplus = $300

D. Price = $5, Quantity = 60, Tax revenue = $120, Consumer Surplus = $280, Producer Surplus = $500

You could try solving this properly, setting supply equal to demand, adjusting for the tax, finding the equilibrium, and calculating the surplus, but don’t bother. If I were tutoring a student in preparing for this test, I’d tell them not to bother. You can get the right answer in only two steps, because of the multiple-choice format.

Step 1: Does tax revenue equal $2 times quantity? We said the tax was $2 per apple.
So that rules out everything except C and D. Welp, quantity must be 60 then.

Step 2: Is quantity 10 times price as the supply curve says? For C they are, for D they aren’t; guess it must be C then.

Now, to do that, you need to have at least a basic understanding of the economics underlying the question (How is tax revenue calculated? What does the supply curve equation mean?). But there’s an even easier technique you can use that doesn’t even require that; it’s called Answer Splicing.

Here’s how it works: You look for repeated values in the answer choices, and you choose the one that has the most repeated values. Prices $5 and $6 are repeated equally, so that’s not helpful (maybe the test designer planned at least that far). Quantity 60 is repeated, other quantities aren’t, so it’s probably that. Likewise with tax revenue $120. Consumer surplus $360 and Producer Surplus $300 are both repeated, so those are probably it. Oh, look, we’ve selected a unique answer choice C, the correct answer!

You could have done answer splicing even if the question were about 18th century German philosophy, or even if the question were written in Arabic or Japanese. In fact you even do it if it were written in a cipher, as long as the cipher was a consistent substitution cipher.

Could the question have been designed to better avoid answer splicing? Probably. But this is actually quite difficult to do, because there is a fundamental tradeoff between two types of “distractors” (as they are known in the test design industry). You want the answer choices to contain correct pieces and resemble the true answer, so that students who basically understand the question but make a mistake in the process still get it wrong. But you also want the answer choices to be distinct enough in a random enough pattern that answer splicing is unreliable. These two goals are inherently contradictory, and the result will always be a compromise between them. Professional test-designers usually lean pretty heavily against answer-splicing, which I think is probably optimal so far as it goes; but I’ve seen many a professor err too far on the side of similar choices and end up making answer splicing quite effective.

But of course, all of this could be completely avoided if I had just presented the question as an open-ended free-response. Then you’d actually have to write down the equations, show me some algebra solving them, and then interpret your results in a coherent way to answer the question I asked. What’s more, if you made a minor mistake somewhere (carried a minus sign over wrong, forgot to divide by 2 when calculating the area of the consumer surplus triangle), I can take off a few points for that error, rather than all the points just because you didn’t get the right answer. At the other extreme, if you just randomly guess, your odds of getting the right answer are miniscule, but even if you did—or copied from someone else—if you don’t show me the algebra you won’t get credit.

So the free-response question is telling me a lot more about what the student actually knows, in a much more reliable way, that is much harder to cheat or strategize against.

Moreover, this isn’t a matter of opinion. This is a theorem of information theory.

The information that is carried over a message channel can be quantitatively measured as its Shannon entropy. It is usually measured in bits, which you may already be familiar with as a unit of data storage and transmission rate in computers—and yes, those are all fundamentally the same thing. A proper formal treatment of information theory would be way too complicated for this blog, but the basic concepts are fairly straightforward: think in terms of how long a sequence of 1s and 0s it would take to convey the message. That is, roughly speaking, the Shannon entropy of that message.

How many bits are conveyed by a multiple-choice response with four choices? 2. Always. At maximum. No exceptions. It is fundamentally, provably, mathematically impossible to convey more than 2 bits of information via a channel that only has 4 possible states. Any multiple-choice response—any multiple-choice response—of four choices can be reduced to the sequence 00, 01, 10, 11.

True-false questions are a bit worse—literally, they convey 1 bit instead of 2. It’s possible to fully encode the entire response to a true-false question as simply 0 or 1.

For comparison, how many bits can I get from the free-response question? Well, in principle the answer to any mathematical question has the cardinality of the real numbers, which is infinite (in some sense beyond infinite, in fact—more infinite than mere “ordinary” infinity); but in reality you can only write down a small number of possible symbols on a page. I can’t actually write down the infinite diversity of numbers between 3.14159 and the true value of pi; in 10 digits or less, I can only (“only”) write down a few billion of them. So let’s suppose that handwritten text has about the same information density as typing, which in ASCII or Unicode has 8 bits—one byte—per character. If the response to this free-response question is 300 characters (note that this paragraph itself is over 800 characters), then the total number of bits conveyed is about 2400.

That is to say, one free-response question conveys six hundred times as much information as a multiple-choice question. Of course, a lot of that information is redundant; there are many possible correct ways to write the answer to a problem (if the answer is 1.5 you could say 3/2 or 6/4 or 1.500, etc.), and many problems have multiple valid approaches to them, and it’s often safe to skip certain steps of algebra when they are very basic, and so on. But it’s really not at all unrealistic to say that I am getting between 10 and 100 times as much useful information about a student from reading one free response than I would from one multiple-choice question.

Indeed, it’s actually a bigger difference than it appears, because when evaluating a student’s performance I’m not actually interested in the information density of the message itself; I’m interested in the product of that information density and its correlation with the true latent variable I’m trying to measure, namely the student’s actual understanding of the content. (A sequence of 500 random symbols would have a very high information density, but would be quite useless in evaluating a student!) Free-response questions aren’t just more information, they are also better information, because they are closer to the real-world problems we are training for, harder to cheat, harder to strategize, nearly impossible to guess, and provided detailed feedback about exactly what the student is struggling with (for instance, maybe they could solve the equilibrium just fine, but got hung up on calculating the consumer surplus).

As I alluded to earlier, free-response questions would also remove most of the danger of students seeing your tests beforehand. If they saw it beforehand, learned how to solve it, memorized the steps, and then were able to carry them out on the test… well, that’s actually pretty close to what you were trying to teach them. It would be better for them to learn a whole class of related problems and then be able to solve any problem from that broader class—but the first step in learning to solve a whole class of problems is in fact learning to solve one problem from that class. Just change a few details each year so that the questions aren’t identical, and you will find that any student who tried to “cheat” by seeing last year’s exam would inadvertently be studying properly for this year’s exam. And then perhaps we could stop making students literally sign nondisclosure agreements when they take college entrance exams. Listen to this Orwellian line from the SAT nondisclosure agreement:

Misconduct includes,but is not limited to:

Taking any test questions or essay topics from the testing room, including through memorization, giving them to anyone else, or discussing them with anyone else through anymeans, including, but not limited to, email, text messages or the Internet

Including through memorization. You are not allowed to memorize SAT questions, because God forbid you actually learn something when we are here to make money off evaluating you.

Multiple-choice tests fail in another way as well; by definition they cannot possibly test generation or recall of knowledge, they can only test recognition. You don’t need to come up with an answer; you know for a fact that the correct answer must be in front of you, and all you need to do is recognize it. Recall and recognition are fundamentally different memory processes, and recall is both more difficult and more important.

Indeed, the real mystery here is why we use multiple-choice exams at all.
There are a few types of very basic questions where multiple-choice is forgivable, because there are just aren’t that many possible valid answers. If I ask whether demand for apples has increased, you can pretty much say “it increased”, “it decreased”, “it stayed the same”, or “it’s impossible to determine”. So a multiple-choice format isn’t losing too much in such a case. But most really interesting and meaningful questions aren’t going to work in this format.

I don’t think it’s even particularly controversial among educators that multiple-choice questions are awful. (Though I do recall an “educational training” seminar a few weeks back that was basically an apologia for multiple choice, claiming that it is totally possible to test “higher-order cognitive skills” using multiple-choice, for reals, believe me.) So why do we still keep using them?

Well, the obvious reason is grading time. The one thing multiple-choice does have over a true free response is that it can be graded efficiently and reliably by machines, which really does make a big difference when you have 300 students in a class. But there are a couple reasons why even this isn’t a sufficient argument.

First of all, why do we have classes that big? It’s absurd. At that point you should just email the students video lectures. You’ve already foreclosed any possibility of genuine student-teacher interaction, so why are you bothering with having an actual teacher? It seems to be that universities have tried to work out what is the absolute maximum rent they can extract by structuring a class so that it is just good enough that students won’t revolt against the tuition, but they can still spend as little as possible by hiring only one adjunct or lecturer when they should have been paying 10 professors.

And don’t tell me they can’t afford to spend more on faculty—first of all, supporting faculty is why you exist. If you can’t afford to spend enough providing the primary service that you exist as an institution to provide, then you don’t deserve to exist as an institution. Moreover, they clearly can afford it—they simply prefer to spend on hiring more and more administrators and raising the pay of athletic coaches. PhD comics visualized it quite well; the average pay for administrators is three times that of even tenured faculty, and athletic coaches make ten times as much as faculty. (And here I think the mean is the relevant figure, as the mean income is what can be redistributed. Firing one administrator making $300,000 does actually free up enough to hire three faculty making $100,000 or ten grad students making $30,000.)

But even supposing that the institutional incentives here are just too strong, and we will continue to have ludicrously-huge lecture classes into the foreseeable future, there are still alternatives to multiple-choice testing.

Ironically, the College Board appears to have stumbled upon one themselves! About half the SAT math exam is organized into a format where instead of bubbling in one circle to give your 2 bits of answer, you bubble in numbers and symbols corresponding to a more complicated mathematical answer, such as entering “3/4” as “0”, “3”, “/”, “4” or “1.28” as “1”, “.”, “2”, “8”. This could easily be generalized to things like “e^2” as “e”, “^”, “2” and “sin(3pi/2)” as “sin”, “3” “pi”, “/”, “2”. There are 12 possible symbols currently allowed by the SAT, and each response is up to 4 characters, so we have already increased our possible responses from 4 to over 20,000—which is to say from 2 bits to 14. If we generalize it to include symbols like “pi” and “e” and “sin”, and allow a few more characters per response, we could easily get it over 20 bits—10 times as much information as a multiple-choice question.

But we can do better still! Even if we insist upon automation, high-end text-recognition software (of the sort any university could surely afford) is now getting to the point where it could realistically recognize a properly-formatted algebraic formula, so you’d at least know if the student remembered the formula correctly. Sentences could be transcribed into typed text, checked for grammar, and sorted for keywords—which is not nearly as good as a proper reading by an expert professor, but is still orders of magnitude better than filling circle “C”. Eventually AI will make even more detailed grading possible, though at that point we may have AIs just taking over the whole process of teaching. (Leaving professors entirely for research, presumably. Not sure if this would be good or bad.)

Automation isn’t the only answer either. You could hire more graders and teaching assistants—say one for every 30 or 40 students instead of one for every 100 students. (And then the TAs might actually be able to get to know their students! What a concept!) You could give fewer tests, or shorter ones—because a small, reliable sample is actually better than a large, unreliable one. A bonus there would be reducing students’ feelings of test anxiety. You could give project-based assignments, which would still take a long time to grade, but would also be a lot more interesting and fulfilling for both the students and the graders.

Or, and perhaps this is the most radical answer of all: You could stop worrying so much about evaluating student performance.

I get it, you want to know whether students are doing well, both so that you can improve your teaching and so that you can rank the students and decide who deserves various awards and merits. But do you really need to be constantly evaluating everything that students do? Did it ever occur to you that perhaps that is why so many students suffer from anxiety—because they are literally being formally evaluated with long-term consequences every single day they go to school?

If we eased up on all this evaluation, I think the fear is that students would just detach entirely; all teachers know students who only seem to show up in class because they’re being graded on attendance. But there are a couple of reasons to think that maybe this fear isn’t so well-founded after all.

If you give up on constant evaluation, you can open up opportunities to make your classes a lot more creative and interesting—and even fun. You can make students want to come to class, because they get to engage in creative exploration and collaboration instead of memorizing what you drone on at them for hours on end. Most of the reason we don’t do creative, exploratory activities is simply that we don’t know how to evaluate them reliably—so what if we just stopped worrying about that?

Moreover, are those students who only show up for the grade really getting anything out of it anyway? Maybe it would be better if they didn’t show up—indeed, if they just dropped out of college entirely and did something else with their lives until they get their heads on straight. Maybe all this effort that we are currently expending trying to force students to learn who clearly don’t appreciate the value of learning could instead be spent enriching the students who do appreciate learning and came here to do as much of it as possible. Because, ultimately, you can lead a student to algebra, but you can’t make them think. (Let me be clear, I do not mean students with less innate ability or prior preparation; I mean students who aren’t interested in learning and are only showing up because they feel compelled to. I admire students with less innate ability who nonetheless succeed because they work their butts off, and wish I were quite so motivated myself.)
There’s a downside to that, of course. Compulsory education does actually seem to have significant benefits in making people into better citizens. Maybe if we let those students just leave college, they’d never come back, and they would squander their potential. Maybe we need to force them to show up until something clicks in their brains and they finally realize why we’re doing it. In fact, we’re really not forcing them; they could drop out in most cases and simply don’t, probably because their parents are forcing them. Maybe the signaling problem is too fundamental, and the only way we can get unmotivated students to accept not getting prestigious degrees is by going through this whole process of forcing them to show up for years and evaluating everything they do until we can formally justify ultimately failing them. (Of course, almost by construction, a student who does the absolute bare minimum to pass will pass.) But college admission is competitive, and I can’t shake this feeling there are thousands of students out there who got rejected from the school they most wanted to go to, the school they were really passionate about and willing to commit their lives to, because some other student got in ahead of them—and that other student is now sitting in the back of the room playing with an iPhone, grumbling about having to show up for class every day. What about that squandered potential? Perhaps competitive admission and compulsory attendance just don’t mix, and we should stop compelling students once they get their high school diploma.

Intellectual Property, revisited

Mar 12, JDN 2457825

A few weeks ago I wrote a post laying out the burden of proof for intellectual property, but didn’t have time to get into the empirical question of whether our existing intellectual property system can meet this burden of proof.

First of all, I want to make a very sharp distinction between three types of regulations that are all called “intellectual property”.

First there are trademarks, which I have absolutely no quarrel with. Avoiding fraud and ensuring transparency are fundamental functions without which markets would unravel, and without trademarks these things would be much harder to accomplish. Trademarks allow a company to establish a brand identity that others cannot usurp; they ensure that when you buy Coca-Cola (R) it is really in fact the beverage you expect and not some counterfeit knockoff. (And if counterfeit Coke sounds silly, note that counterfeit honey and maple syrup are actually a major problem.) Yes, there should be limits on how much you can trademark—no one wants to live in a world where you feel Love ™ and open Screen Doors ™—but in fact our courts are already fairly good about only allowing corporations to trademark newly-coined words and proper names for their products.

Next there are copyrights, which I believe are currently too strong and often abused, but I do think should exist in some form (or perhaps copylefts instead). Authors should have at least certain basic rights over how their work can be used and published. If nothing else, proper attribution should always be required, as without that plagiarism becomes intolerably easy. And steps should be taken to ensure that if any people profit from its sale, the author is among them. I publish this blog under a by-sa copyleft, which essentially means that you can share it with whomever you like and even adapt its content into your own work, so long as you properly attribute it to me and you do not attempt to claim ownership over it. For scientific content, I think only a copyleft of this sort makes sense—the era of for-profit journals with paywalls must end, as it is holding back our civilization. But for artistic content (and I mean art in the broadest sense, including books, music, movies, plays, and video games), stronger regulations might well make sense. The question is whether our current system is actually too strong, or is protecting the wrong people—often it seems to protect the corporations that sell the content rather than the artists who created it.

Finally there are patents. Unlike copyright which applies to a specific work of art, patent is meant to apply to the underlying concept of a technology. Copyright (or rather the by-sa copyleft) protects the text of this article; you can’t post it on your own blog and claim you wrote it. But if I were to patent it somehow (generally, verbal arguments cannot be patented, fortunately), you wouldn’t even be able to paraphrase it. The trademark on a Samsung ™ TV just means that if I make a TV I can’t say I am Samsung, because I’m not. You wouldn’t copyright a TV, but the analogous process would be if I were to copy every single detail of the television and try to sell that precise duplicate. But the patents on that TV mean that if I take it apart, study each component, find a way to build them all from my own raw materials, even make them better, and build a new TV out of them that looks different and performs better—I would still be infringing on intellectual property. Patents grant an extremely strong notion of property rights, one which actually undermines a lot of other, more basic concepts of property. It’s my TV, why can’t I take it apart and copy the components? Well, as long as the patent holds, it’s not entirely my TV. Property rights this strong—that allow a corporation to have its cake of selling the TV but eat it too by owning the rights to all its components—require a much stronger justification.

Trademark protects a name, which is unproblematic. Copyright protects a work, which carries risks but is still probably necessary in many cases. But patent protects an idea—and we should ask ourselves whether that is really something it makes sense to do.

In previous posts I’ve laid out some of the basic philosophical arguments for why patents do not seem to support innovation and may actually undermine it. But in this post I want to do something more direct and quantitative: Empirically, what is the actual effect of copyrights and patents on innovation? Can we find a way to quantify the costs and benefits to our society of different modes of intellectual property?

Economists quantify things all the time, so I briefly combed the literature to see what sort of empirical studies had been done on the economic impact of copyrights and patents.

Patents definitely create barriers to scientific collaboration: Scientific articles with ideas that don’t get patented are about 10-20% more likely to be cited than scientific articles with ideas that are patented. (I would have expected a larger effect, but that’s still not trivial.)

A 1995 study found that creased patent protections do seem to be positively associated with more trade.

A 2009 study of Great Britain published in AER found it “puzzling” that stronger patents actually seem to reduce the rate of innovation domestically, while having no effect on foreign innovation—yet this is exactly what I would have predicted. Foreign innovations should be largely unaffected by UK patents, but stricter patent laws in the UK make it harder for most actual innovators, only benefiting a handful of corporations that aren’t even particularly innovative.

This 1996 study did find a positive effect of stronger patent laws on economic growth, but it was quite small and only statistically significant when using instrumental variables that they couldn’t be bothered to define except in an appendix. When your result hinges on the use of instrumental variables that you haven’t even clearly defined in the paper, something is very fishy. My guess is that they p-hacked the instruments until they got the result they wanted.

This other 1996 study is a great example of why economists need to listen to psychologists. It found a negative correlation between foreign direct investment and—wait for it—the number of companies that answered “yes” to a survey question, “Does country X have intellectual property protection too weak to allow you to transfer your newest or most effective technology to a wholly-owned subsidiarythere?” Oh, wow, you found a correlation between foreign direct investment and a question directly asking about foreign direct investment.

his 2004 study found a nonlinear relationship whereby increased economic development affects intellectual property rights, rather than the other way around. But I find their theoretical model quite odd, and the scatter plot that lies at the core of their empirical argument reminds me of Rexthor, the Dog-Bearer. “This relationship appears to be non-linear,” they say when pointing at a scatter plot that looks mostly like nothing and maybe like a monotonic increase.

This 1997 study found a positive correlation between intellectual property strength, R&D spending, and economic growth. The effect is weak, but the study looks basically sound. (Though I must say I’d never heard anyone use the words “significant at the 24% level” before. Normally one would say “nonsignificant” for that variable methinks. It’s okay for it not to be significant in some of your regressions, you know.)

This 1992 paper found that intellectual property harms poor countries and may or may not benefit rich countries, but it uses a really weird idiosyncratic theoretical model to get there. Frankly if I see the word “theorem” anywhere in your empirical paper, I get suspicious. No, it is not a theorem that “For economies in steady state the South loses from tighter intellectual property rights.” It may be true, but it does not follow from the fundamental axioms of mathematics.

This law paper is excellent; it focuses on the fact that intellectual property is a unique arrangement and a significant deviation from conventional property rights. It tracks the rise of legal arguments that erroneously equate intellectual property with real property, and makes the vital point that fully internalizing the positive externalities of technology was never the goal, and would in fact be horrible should it come to pass. We would all have to pay most of our income in royalties to the Newton and Faraday estates. So, I highly recommend reading it. But it doesn’t contain any empirical results on the economic effects of intellectual property.

This is the best paper I was able to find showing empirical effects of different intellectual property regimes; I really have no complaints about its econometrics. But it was limited to post-Soviet economies shortly after the fall of the USSR, which were rather unique circumstances. (Indeed, by studying only those countries, you’d probably conclude that free markets are harmful, because the shock of transition was so great.)

This 1999 paper is also quite good; using a natural experiment from a sudden shift in Japanese patent policy, they found almost no difference in actual R&D. The natural experiment design makes this particularly credible, but it’s difficult to generalize since it only covered Japan specifically.

This study focused in particular on copyrights and the film industry, and found a nonlinear effect: While having no copyright protection at all was harmful to the film industry, making the copyright protections too strong had a strangling effect on new filmmakers entering the industry. This would suggest that the optimal amount of copyright is moderate, which sounds reasonable to me.

This 2009 study did a much more detailed comparison of different copyright regimes, and was unable to find a meaningful pattern amidst the noise. Indeed, they found that the only variable that consistently predicted the number of new works of art was population—more people means more art, and nothing else seemed to matter. If this is correct, it’s quite damning to copyright; it would suggest that people make art for reasons fundamentally orthogonal to copyright, and copyright does almost nothing useful. (And I must say, if you talk to most artists, that tends to be their opinion on the matter!)

This 1996 paper found that stronger patents had no benefits for poor countries, but benefited rich countries quite a large amount: Increased patent protection was estimated to add as much as 0.7% annual GDP growth over the whole period. That’s a lot; if this is really true, stronger patents are almost certainly worth it. But then it becomes difficult to explain why more precise studies haven’t found effects anywhere near that large.

This paper was pretty interesting; they found a fat-tailed distribution of patents, where most firms have none, many have one or a few, and a handful of firms have a huge number of patents. This is also consistent with the distribution of firm revenue and profit—and I’d be surprised if I didn’t find a strong correlation between all three. But this really doesn’t tell us whether patents are contributing to innovation.
This paper found that the harmonization of global patents in the Uruguay Round did lead to gains from trade for most countries, but also transferred about $4.5 billion to the US from the rest of the world. Of course, that’s really not that large an amount when we’re talking about global policy over several years.

What does all that mean? I don’t know. It’s a mess. There just don’t seem to be any really compelling empirical studies on the economic impact of copyrights and patents. The preponderance of the evidence, such as it is, would seem to suggest that copyrights provide a benefit as long as they aren’t too strong, while patents provide a benefit but it is quite small and likely offset by the rent-seeking of the corporations that own them. The few studies that found really large effects (like 0.7% annual GDP growth) don’t seem very credible to me; if the effect were really that large, it shouldn’t be so ambiguous. 0.7% per year over 25 years is a GDP 20% larger. Over 50 years, GDP would be 42% larger. We would be able to see that.

Does this ambiguity mean we should do nothing, and wait until the data is better? I don’t think so. Remember, the burden of proof for intellectual property should be high. It’s a fundamentally bizarre notion of property, one which runs against most of our standard concepts of real property; it restricts our rights in very basic ways, making literally the majority of our population into criminals. Such a draconian policy requires a very strong justification, but such a justification does not appear to be forthcoming. If it could be supported, that 0.7% GDP growth might be enough; but it doesn’t seem to be replicable. A free society does not criminalize activities just in case it might be beneficial to do so—it only criminalizes activities that have demonstrable harm. And the harm of copyright and patent infringement simply isn’t demonstrable enough to justify its criminalization.

We don’t have to remove them outright, but we should substantially weaken copyright and patent laws. They should be short-term, they should provide very basic protection, and they should never be owned by corporations, always by individuals (corporations should be able to license them—but not own them). If we then observe a substantial reduction in innovation and economic output, then we can put them back. But I think that what defenders of intellectual property fear most is that if we tried this, it wouldn’t be so bad—and then the “doom and gloom” justification they’ve been relying on all this time would fall apart.

Games as economic simulations—and education tools

Mar 5, JDN 2457818 [Sun]

Moore’s Law is a truly astonishing phenomenon. Now as we are well into the 21st century (I’ve lived more of my life in the 21st century than the 20th now!) it may finally be slowing down a little bit, but it has had quite a run, and even this could be a temporary slowdown due to economic conditions or the lull before a new paradigm (quantum computing?) matures. Since at least 1975, the computing power of an individual processor has doubled approximately every year and a half; that means it has doubled over 25 times—or in other words that it has increased by a factor of over 30 million. I now have in my pocket a smartphone with several thousand times the processing speed of the guidance computer of the Saturn V that landed on the Moon.

This meteoric increase in computing power has had an enormous impact on the way science is done, including economics. Simple theoretical models that could be solved by hand are now being replaced by enormous simulation models that have to be processed by computers. It is now commonplace to devise models with systems of dozens of nonlinear equations that are literally impossible to solve analytically, and just solve them iteratively with computer software.

But one application of this technology that I believe is currently underutilized is video games.

As a culture, we still have the impression that video games are for children; even games like Dragon Age and Grand Theft Auto that are explicitly for adults (and really quite inappropriate for children!) are viewed as in some sense “childish”—that no serious adult would be involved with such frivolities. The same cultural critics who treat Shakespeare’s vagina jokes as the highest form of art are liable to dismiss the poignant critique of war in Call of Duty: Black Ops or the reflections on cultural diversity in Skyrim as mere puerility.

But video games are an art form with a fundamentally greater potential than any other. Now that graphics are almost photorealistic, there is really nothing you can do in a play or a film that you can’t do in a video game—and there is so, so much more that you can only do in a game.
In what other medium can we witness the spontaneous emergence and costly aftermath of a war? Yet EVE Online has this sort of event every year or so—just today there was a surprise attack involving hundreds of players that destroyed thousands of hours’—and dollars’—worth of starships, something that has more or less become an annual tradition. A few years ago there was a massive three-faction war that destroyed over $300,000 in ships and has now been commemorated as “the Bloodbath of B-R5RB”.
Indeed, the immersion and interactivity of games present an opportunity to do nothing less than experimental macroeconomics. For generations it has been impossible, or at least absurdly unethical, to ever experimentally manipulate an entire macroeconomy. But in a video game like EVE Online or Second Life, we can now do so easily, cheaply, and with little or no long-term harm to the participants—and we can literally control everything in the experiment. Forget the natural resource constraints and currency exchange rates—we can change the laws of physics if we want. (Indeed, EVE‘s whole trade network is built around FTL jump points, and in Second Life it’s a basic part of the interface that everyone can fly like Superman.)

This provides untold potential for economic research. With sufficient funding, we could build a game that would allow us to directly test hypotheses about the most fundamental questions of economics: How do governments emerge and maintain security? How is the rule of law sustained, and when can it be broken? What controls the value of money and the rate of inflation? What is the fundamental cause of unemployment, and how can it be corrected? What influences the rate of technological development? How can we maximize the rate of economic growth? What effect does redistribution of wealth have on employment and output? I envision a future where we can directly simulate these questions with thousands of eager participants, varying the subtlest of parameters and carrying out events over any timescale we like from seconds to centuries.

Nor is the potential of games in economics limited to research; it also has enormous untapped potential in education. I’ve already seen in my classes how tabletop-style games with poker chips can teach a concept better in a few minutes than hours of writing algebra derivations on the board; but custom-built video games could be made that would teach economics far better still, and to a much wider audience. In a well-designed game, people could really feel the effects of free trade or protectionism, not just on themselves as individuals but on entire nations that they control—watch their GDP numbers go down as they scramble to produce in autarky what they could have bought for half the price if not for the tariffs. They could see, in real time, how in the absence of environmental regulations and Pigovian taxes the actions of millions of individuals could despoil our planet for everyone.

Of course, games are fundamentally works of fiction, subject to the Fictional Evidence Fallacy and only as reliable as their authors make them. But so it is with all forms of art. I have no illusions about the fact that we will never get the majority of the population to regularly read peer-reviewed empirical papers. But perhaps if we are clever enough in the games we offer them to play, we can still convey some of the knowledge that those papers contain. We could also update and expand the games as new information comes in. Instead of complaining that our students are spending time playing games on their phones and tablets, we could actually make education into games that are as interesting and entertaining as the ones they would have been playing. We could work with the technology instead of against it. And in a world where more people have access to a smartphone than to a toilet, we could finally bring high-quality education to the underdeveloped world quickly and cheaply.

Rapid growth in computing power has given us a gift of great potential. But soon our capacity will widen even further. Even if Moore’s Law slows down, computing power will continue to increase for awhile yet. Soon enough, virtual reality will finally take off and we’ll have even greater depth of immersion available. The future is bright—if we can avoid this corporatist cyberpunk dystopia we seem to be hurtling toward, of course.

Markets value rich people more

Feb 26, JDN 2457811

Competitive markets are optimal at maximizing utility, as long as you value rich people more.

That is literally a theorem in neoclassical economics. I had previously thought that this was something most economists didn’t realize; I had delusions of grandeur that maybe I could finally convince them that this is the case. But no, it turns out this is actually a well-known finding; it’s just that somehow nobody seems to care. Or if they do care, they never talk about it. For all the thousands of papers and articles about the distortions created by minimum wage and capital gains tax, you’d think someone could spare the time to talk about the vastly larger fundamental distortions created by the structure of the market itself.

It’s not as if this is something completely hopeless we could never deal with. A basic income would go a long way toward correcting this distortion, especially if coupled with highly progressive taxes. By creating a hard floor and a soft ceiling on income, you can reduce the inequality that makes these distortions so large.

The basics of the theorem are quite straightforward, so I think it’s worth explaining them here. It’s extremely general; it applies anywhere that goods are allocated by market prices and different individuals have wildly different amounts of wealth.

Suppose that each person has a certain amount of wealth W to spend. Person 1 has W1, person 2 has W2, and so on. They all have some amount of happiness, defined by a utility function, which I’ll assume is only dependent on wealth; this is a massive oversimplification of course, but it wouldn’t substantially change my conclusions to include other factors—it would just make everything more complicated. (In fact, including altruistic motives would make the whole argument stronger, not weaker.) Thus I can write each person’s utility as a function U(W). The rate of change of this utility as wealth increases, the marginal utility of wealth, is denoted U'(W).

By the law of diminishing marginal utility, the marginal utility of wealth U'(W) is decreasing. That is, the more wealth you have, the less each new dollar is worth to you.

Now suppose people are buying goods. Each good C provides some amount of marginal utility U'(C) to the person who buys it. This can vary across individuals; some people like Pepsi, others Coke. This marginal utility is also decreasing; a house is worth a lot more to you if you are living in the street than if you already have a mansion. Ideally we would want the goods to go to the people who want them the most—but as you’ll see in a moment, markets systematically fail to do this.

If people are making their purchases rationally, each person’s willingness-to-pay P for a given good C will be equal to their marginal utility of that good, divided by their marginal utility of wealth:

P = U'(C)/U'(W)

Now consider this from the perspective of society as a whole. If you wanted to maximize utility, you’d equalize marginal utility across individuals (by the Extreme Value Theorem). The idea is that if marginal utility is higher for one person, you should give that person more, because the benefit of what you give them will be larger that way; and if marginal utility is lower for another person, you should give that person less, because the benefit of what you give them will be smaller. When everyone is equal, you are at the maximum.

But market prices don’t actually do this. Instead they equalize over willingness-to-pay. So if you’ve got two individuals 1 and 2, instead of having this:

U'(C1) = U'(C2)

you have this:

P1 = P2

which translates to:

U'(C1)/U'(W1) = U'(C2)/U'(W2)

If the marginal utilities were the same, U'(W1) = U'(W2), we’d be fine; these would give the same results. But that would only happen if W1 = W2, that is, if the two individuals had the same amount of wealth.

Now suppose we were instead maximizing weighted utility, where each person gets a weighting factor A based on how “important” they are or something. If your A is higher, your utility matters more. If we maximized this new weighted utility, we would end up like this:

A1*U'(C1) = A2*U'(C2)

Because person 1’s utility counts for more, their marginal utility also counts for more. This seems very strange; why are we valuing some people more than others? On what grounds?

Yet this is effectively what we’ve already done by using market prices.
Just set:
A = 1/U'(W)

Since marginal utility of wealth is decreasing, 1/U'(W) is higher precisely when W is higher.

How much higher? Well, that depends on the utility function. The two utility functions I find most plausible are logarithmic and harmonic. (Actually I think both apply, one to other-directed spending and the other to self-directed spending.)

If utility is logarithmic:

U = ln(W)

Then marginal utility is inversely proportional:

U'(W) = 1/W

In that case, your value as a human being, as spoken by the One True Market, is precisely equal to your wealth:

A = 1/U'(W) = W

If utility is harmonic, matters are even more severe.

U(W) = 1-1/W

Marginal utility goes as the inverse square of wealth:

U'(W) = 1/W^2

And thus your value, according to the market, is equal to the square of your wealth:

A = 1/U'(W) = W^2

What are we really saying here? Hopefully no one actually believes that Bill Gates is really morally worth 400 trillion times as much as a starving child in Malawi, as the calculation from harmonic utility would imply. (Bill Gates himself certainly doesn’t!) Even the logarithmic utility estimate saying that he’s worth 20 million times as much is pretty hard to believe.

But implicitly, the market “believes” that, because when it decides how to allocate resources, something that is worth 1 microQALY to Bill Gates (about the value a nickel dropped on the floor to you or I) but worth 20 QALY (twenty years of life!) to the Malawian child, will in either case be priced at $8,000, and since the child doesn’t have $8,000, it will probably go to Mr. Gates. Perhaps a middle-class American could purchase it, provided it was worth some 0.3 QALY to them.

Now consider that this is happening in every transaction, for every good, in every market. Goods are not being sold to the people who get the most value out of them; they are being sold to the people who have the most money.

And suddenly, the entire edifice of “market efficiency” comes crashing down like a house of cards. A global market that quite efficiently maximizes willingness-to-pay is so thoroughly out of whack when it comes to actually maximizing utility that massive redistribution of wealth could enormously increase human welfare, even if it turned out to cut our total output in half—if utility is harmonic, even if it cut our total output to one-tenth its current value.

The only way to escape this is to argue that marginal utility of wealth is not decreasing, or at least decreasing very, very slowly. Suppose for instance that utility goes as the 0.9 power of wealth:

U(W) = W^0.9

Then marginal utility goes as the -0.1 power of wealth:

U'(W) = 0.9 W^(-0.1)

On this scale, Bill Gates is only worth about 5 times as much as the Malawian child, which in his particular case might actually be too small—if a trolley is about to kill either Bill Gates or 5 Malawian children, I think I save Bill Gates, because he’ll go on to save many more than 5 Malawian children. (Of course, substitute Donald Trump or Charles Koch and I’d let the trolley run over him without a second thought if even a single child is at stake, so it’s not actually a function of wealth.) In any case, a 5 to 1 range across the whole range of human wealth is really not that big a deal. It would introduce some distortions, but not enough to justify any redistribution that would meaningfully reduce overall output.

Of course, that commits you to saying that $1 to a Malawian child is only worth about $1.50 to you or I and $5 to Bill Gates. If you can truly believe this, then perhaps you can sleep at night accepting the outcomes of neoclassical economics. But can you, really, believe that? If you had the choice between an intervention that would give $100 to each of 10,000 children in Malawi, and another that would give $50,000 to each of 100 billionaires, would you really choose the billionaires? Do you really think that the world would be better off if you did?

We don’t have precise measurements of marginal utility of wealth, unfortunately. At the moment, I think logarithmic utility is the safest assumption; it’s about the slowest decrease that is consistent with the data we have and it is very intuitive and mathematically tractable. Perhaps I’m wrong and the decrease is even slower than that, say W^(-0.5) (then the market only values billionaires as worth thousands of times as much as starving children). But there’s no way you can go as far as it would take to justify our current distribution of wealth. W^(-0.1) is simply not a plausible value.

And this means that free markets, left to their own devices, will systematically fail to maximize human welfare. We need redistribution—a lot of redistribution. Don’t take my word for it; the math says so.

Caught between nepotism and credentialism

Feb 19, JDN 2457804

One of the more legitimate criticisms out there of we “urban elites” is our credentialismour tendency to decide a person’s value as an employee or even as a human being based solely upon their formal credentials. Randall Collins, an American sociologist, wrote a book called The Credential Society arguing that much of the class stratification in the United States is traceable to this credentialism—upper-middle-class White Anglo-Saxon Protestants go to the good high schools to get into the good colleges to get the good careers, and all along the way maintain subtle but significant barriers to keep everyone else out.

A related concern is that of credential inflation, where more and more people get a given credential (such as a high school diploma or a college degree), and it begins to lose value as a signal of status. It is often noted that a bachelor’s degree today “gets” you the same jobs that a high school diploma did two generations ago, and two generations hence you may need a master’s or even a PhD.

I consider this concern wildly overblown, however. First of all, they’re not actually the same jobs at all. Even our “menial” jobs of today require skills that most people didn’t have two generations ago—not simply those involving electronics and computers, but even quite basic literacy and numeracy. Yes, you could be a banker in the 1920s with a high school diploma, but plenty of bankers in the 1920s didn’t know algebra. What, you think they were arbitraging derivatives based on the Black-Scholes model?

The primary purpose of education should be to actually improve students’ abilities, not to signal their superior status. More people getting educated is good, not bad. If we really do need signals, we can devise better ones than making people pay tens of thousands of dollars in tuition and spending years taking classes. An expenditure of that magnitude should be accomplishing something, not just signaling. (And given the overwhelming positive correlation between a country’s educational attainment and its economic development, clearly education is actually accomplishing something.) Our higher educational standards have directly tied to higher technology and higher productivity. If indeed you need a PhD to be a janitor in 2050, it will be because in 2050 a “janitor” is actually the expert artificial intelligence engineer who commands an army of cleaning robots, not because credentials have “inflated”. Thinking that credentials “inflate” requires thinking that business managers must be very stupid, that they would exclude whole swaths of qualified candidates that they could pay less to do the same work. Only a complete moron would require a PhD to hire you for wielding a mop.

No, what concerns me is an over-emphasis on prestigious credentials over genuine competence. This is definitely a real issue in our society: Almost every US President went to an Ivy League university, yet several of them (George W. Bush, anyone?) clearly would not actually have been selected by such a university if their families had not been wealthy and well-connected. (Harvard’s application literally contains a question asking whether you are a “lineal or collateral descendant” of one of a handful of super-wealthy families.) Papers that contain errors so basic that I would probably get a failing grade as a grad student for them become internationally influential because they were written by famous economists with fancy degrees.

Ironically, it may be precisely because elite universities try not to give grades or special honors that so many of their students try so desperately to latch onto any bits of social status they can get their hands on. In this blog post, a former Yale law student comments on how, without grades or cum laude to define themselves, Yale students became fiercely competitive in the pettiest ways imaginable. Or it might just be a selection effect; to get into Yale you’ve probably got to be pretty competitive, so even if they don’t give out grades once you get there, you can take the student out of the honors track, but you can’t take the honors track out of the student.

But perhaps the biggest problem with credentialism is… I don’t see any viable alternatives!

We have to decide who is going to be hired for technical and professional positions somehow. It almost certainly can’t be everyone. And the most sensible way to do it would be to have a process people go through to get trained and evaluated on their skills in that profession—that is, a credential.

What else would we do? We could decide randomly, I suppose; well, good luck with that. Or we could try to pick people who don’t have qualifications (“anti-credentialism” I suppose), which would be systematically wrong. Or individual employers could hire individuals they know and trust on a personal level, which doesn’t seem quite so ridiculous—but we have a name for that too, and it’s nepotism.

Even anti-credentialism does exist, bafflingly enough. Many people voted for George W. Bush because they said he was “the kind of guy you can have a beer with”. That wasn’t true, of course; he was the spoiled child of a billionaire, a man who had never really worked a day in his life. But even if it had been true, so what? How is that a qualification to be the leader of the free world? And how many people voted for Trump precisely because he had no experience in government? This made sense to them somehow. (And, shockingly, he has no idea what he’s doing. Actually what is shocking is that he admits that.)

Nepotism of course happens all the time. In fact, nepotism is probably the default state for humans. The continual re-emergence of hereditary monarchy and feudalism around the world suggests that this is some sort of attractor state for human societies, that in the absence of strong institutional pressures toward some other system this is what people will generally settle into. And feudalism is nothing if not nepotistic; your position in life is almost entirely determined by your father’s position, and his father’s before that.

Formal credentials can put a stop to that. Of course, your ability to obtain the credential often depends upon your income and social status. But if you can get past those barriers and actually get the credential, you now have a way of pushing past at least some of the competitors who would have otherwise been hired on their family connections alone. The rise in college enrollments—and women actually now exceeding men in college enrollment rates—is one of the biggest reasons why the gender pay gap is rapidly closing among young workers. Nepotism and sexism that would otherwise have hired unqualified men is now overtaken by the superior credentials of qualified women.

Credentialism does still seem suboptimal… but from where I’m sitting, it seems like a second-best solution. We can’t actually observe people’s competence and ability directly, so we need credentials to provide an approximate measurement. We can certainly work to improve credentials—and for example, I am fiercely opposed to multiple-choice testing because it produces such meaningless credentials—but ultimately I don’t see any alternative to credentials.

Is intellectual property justified?

Feb 12, JDN 2457797

I had hoped to make this week’s post more comprehensive, but as I’ve spent the last week suffering from viral bronchitis I think I will keep this one short and revisit the topic in a few weeks.

Intellectual property underlies an increasingly large proportion of the world’s economic activity, more so now than ever before. We don’t just patent machines anymore; we patent drugs, and software programs, and even plants. Compared to that, copyrights on books, music, and movies seem downright pedestrian.

Though surely not the only cause, this is almost certainly contributing to the winner-takes-all effect; if you own the patent to something important, you can appropriate a huge amount of wealth to yourself with very little effort.

Moreover, this is not something that happened automatically as a natural result of market forces or autonomous human behavior. This is a policy, one that requires large investments in surveillance and enforcement to maintain. Intellectual property is probably the single largest market intervention that our government makes, and it is in a very strange direction: With antitrust law, the government seeks to undermine monopolies; but with intellectual property, the government seeks to protect monopolies.

So it’s important to ask: What is the justification for intellectual property? Do we actually have a good reason for doing this?

The basic argument goes something like this:

Many intellectual endeavors, such as research, invention, and the creation of art, require a large up-front investment of resources to complete, but once completed it costs almost nothing to disseminate the results. There is a very large fixed cost that makes it difficult to create these goods at all, but once they exist, the marginal cost of producing more of them is minimal.

If we didn’t have any intellectual property, once someone created an invention or a work of art, someone else could simply copy it and sell it at a much lower price. If enough competition emerged to drive price down to marginal cost, the original creator of the good would not only not profit, but would actually take an enormous loss, as they paid that large fixed cost but none of their competitors did.

Thus, knowing that they will take a loss if they do, individuals will not create inventions or works of art in the first place. Without intellectual property, all research, invention, and art would grind to a halt.

 

That last sentence sounds terrible, right? What would we do without research, invention, or art? But then if you stop and think about it for a minute, it becomes clear that this can’t possibly be the outcome of eliminating intellectual property. Most societies throughout the history of human civilization have not had a system of intellectual property, and yet they have all had art, and most of them have had research and invention as well.

If intellectual property is to be defended, it can’t be because we would have none of these things without it—it must be that we would have less, and so much less that it offsets the obvious harms of concentrating so much wealth and power in a handful of individuals.

I had hoped to get into the empirical results of different intellectual property regimes, but due to my illness I’m going to save that for another day.

Instead I’m just going to try to articulate what the burden of proof here really needs to be.

First of all, showing that we spend a lot of money on patents contributes absolutely nothing useful to defending them. Yes, we all know patents are expensive. The question is whether they are worth it. To show that this is not a strawman, here’s an article by IP Watchdog that takes the fact that “a new study showing that academic patent licensing contributed more than $1 trillion to the U.S. economy over eighteen years” is some kind of knockdown argument in favor of patents. If you actually showed that this economic activity would not exist without patents, then that would be an argument for patents. But all this study actually does is shows that we spend that much on patents, which says nothing about whether this is a good use of resources. It’s like when people try to defend the F-35 boondoggle by saying “it supports thousands of jobs!”; well, yes, but what about the millions of jobs we could be supporting instead if we used that money for something more efficient? (And indeed, the evidence is quite clear that spending on the F-35 destroys more jobs than it creates.) So any serious of estimate of economic benefits of intellectual property must also come with an estimate of the economic cost of intellectual property, or it is just propaganda.
It’s not enough to show some non-negligible (much less “statistically significant”) increase in innovation as a result of intellectual property. The effect size is critical; the increase in innovation needs to be large enough that it justifies having world-spanning monopolies that concentrate the world’s wealth in the hands of a few individuals. Because we already know that intellectual property concentrates wealth; they are monopolies, and monopolies concentrate wealth. It’s not enough to show that there is a benefit; that benefit must be greater than the cost, and there must be no alternative methods that allow us to achieve a greater net benefit.
It’s also important to be clear what we mean by “innovation”; this can be a very difficult thing to measure. But in principle what we really want to know is whether we are supporting important innovation—whether we will get more Mona Lisas and more polio vaccines, not simply whether we will get more Twilight and more Viagra. And one of the key problems with intellectual property as a method of funding innovation is that there is only a vague link between the profits that can be extracted and the benefits of the innovation. (Though to be fair, this is actually a more general problem; it is literally a mathematical theorem that competitive markets only maximize utility if you value rich people more, in inverse proportion to their marginal utility of wealth.)

Innovation is certainly important. Indeed, it is no exaggeration to say that innovation is the foundation of economic development and civilization itself. Defenders of intellectual property often want you to stop the conversation there: “Innovation is important!” Don’t let them. It’s not enough to say that innovation is important; intellectual property must also be the best way of achieving that innovation.

Is it? Well, in a few weeks I’ll get back to what the data actually says on this. There is some evidence supporting intellectual property—but the case is a lot weaker than you have probably been led to believe.

The urban-rural divide runs deep

Feb 5, JDN 2457790

Are urban people worth less than rural people?

That probably sounds like a ridiculous thing to ask; of course not, all people are worth the same (other things equal of course—philanthropists are worth more than serial murderers). But then, if you agree with that, you’re probably an urban person, as I’m sure most of my readers are (and as indeed most people in highly-developed countries are).

A disturbing number of rural people, however, honestly do seem to believe this. They think that our urban lifestyles (whatever they imagine those to be) devalue us as citizens and human beings.

That is the key subtext to understand in the terrifying phenomenon that is Donald Trump. Most of the people who voted for him can’t possibly have thought he was actually trustworthy, and many probably didn’t actually support his policies of bigotry and authoritarianism (though he was very popular among bigots and authoritarians). From speaking with family members and acquaintances who proudly voted for Trump, one thing came through very clearly: This was a gigantic middle finger pointed at cities. They didn’t even really want Trump; they just knew we didn’t, and so they voted for him out of spite as much as anything else. They also have really confused views about free trade, so some of them voted for him because he promised to bring back jobs lost to trade (that weren’t lost to trade, can’t be brought back, and shouldn’t be even if they could). Talk with a Trump voter for a few minutes, and sneers of “latte-sipping liberal” (I don’t even like coffee) and “coastal elite” (I moved here to get educated; I wasn’t born here) are sure to follow.

There has always been some conflict between rural and urban cultures, for as long as there have been urban cultures for rural cultures to be in conflict with. It is found not just in the US, but in most if not all countries around the world. It was relatively calm during the postwar boom in the 20th century, as incomes everywhere (or at least everywhere within highly-developed countries) were improving more or less in lockstep. But the 21st century has brought us much more unequal growth, concentrated on particular groups of people and particular industries. This has brought more resentment. And that divide, above all else, is what brought us Trump; the correlation between population density and voting behavior is enormous.

Of course, “urban” is sometimes a dog-whistle for “Black”; but sometimes I think it actually really means “urban”—and yet there’s still a lot of hatred embedded in it. Indeed, perhaps that’s why the dog-whistle works; a White man from a rural town can sneer at “urban” people and it’s not entirely clear whether he’s being racist or just being anti-urban.

The assumption that rural lifestyles are superior runs so deep in our culture that even in articles by urban people (like this one from the LA Times) supposedly reflecting about how to resolve this divide, there are long paeans to the world of “hard work” and “sacrifice” and “autonomy” of rural life, and mocking “urban elites” for their “disproportionate” (by which you can only mean almost proportionate) power over government.

Well, guess what? If you want to live in a rural area, go live in a rural area. Don’t pine for it. Don’t tell me how great farm life is. If you want to live on a farm, go live on a farm. I have nothing against it; we need farmers, after all. I just want you to shut up about how great it is, especially if you’re not going to actually do it. Pining for someone else’s lifestyle when you could easily take on that lifestyle if you really wanted it just shows that you think the grass is greener on the other side.

Because the truth is, farm living isn’t so great for most people. The world’s poorest people are almost all farmers. 70% of people below the UN poverty line live in rural areas, even as more and more of the world’s population moves into cities. If you use a broader poverty measure, as many as 85% of the world’s poor live in rural areas.

The kind of “autonomy” that means defending your home with a shotgun is normally what we would call anarchy—it’s a society that has no governance, no security. (Of course, in the US that’s pure illusion; crime rates in general are low and falling, and lower in rural areas than urban areas. But in some parts of the world, that anarchy is very real.) One of the central goals of global economic development is to get people away from subsistence farming into far more efficient manufacturing and service jobs.

At least in the US, farm life is a lot better than it used to be, now that agricultural technology has improved so that one farmer can now do the work of hundreds. Despite increased population and increased food consumption per person, the number of farmers in the US is now the smallest it has been since before the Civil War. The share of employment devoted to agriculture has fallen from over 80% in 1800 to under 2% today. Even just since the 1960s labor productivity of US farms has more than tripled.

But the reason that some 80% of Americans have chosen to live in cities—and yes, I can clearly say “chosen”, because cities are more expensive and therefore urban living is a voluntary activity. Most people who live in the city right now could move to the country if we really wanted to. We choose not to, because we know our life would be worse if we did.

Indeed, I dare say that a lot of the hatred of city-dwellers has got to be envy. Our (median) incomes are higher and our (mean) lifespans are longer. Fewer of our children are in poverty. Life is better here—we know it, and deep down, they know it too.

We also have better Internet access, unsurprisingly—though rural areas are only a few years behind, and the technology improves so rapidly that twice as many rural homes in the US have Internet access than urban homes did in 1998.

Now, a rational solution to this problem would be either to improve the lives of people in rural areas or else move everyone to urban areas—and both of those things have been happening, not only in the US but around the world. But in order to do that, you need to be willing to change things. You have to give up the illusion that farm life is some wonderful thing we should all be emulating, rather than the necessary toil that humanity was forced to go through for centuries until civilization could advance beyond it. You have to be willing to replace farmers with robots, so that people who would have been farmers can go do something better with their lives. You need to give up the illusion that there is something noble or honorable about hard labor on a farm—indeed, you need to give up the illusion that there is anything noble or honorable about hard work in general. Work is not a benefit; work is a cost. Work is what we do because we have to—and when we no longer have to do it, we should stop. Wanting to escape toil and suffering doesn’t make you lazy or selfish—it makes you rational.

We could surely be more welcoming—but cities are obviously more welcoming to newcomers than rural areas are. Our housing is too expensive, but that’s in part because so many people want to live here—supply hasn’t been able to keep up with demand.

I may seem to be presenting this issue as one-sided; don’t urban people devalue rural people too? Sometimes. Insults like “hick” and “yokel” and “redneck” do of course exist. But I’ve never heard anyone from a city seriously argue that people who live in rural areas should have votes that systematically count for less than those of people who live in cities—yet the reverse is literally what people are saying when they defend the Electoral College. If you honestly think that the Electoral College deserves to exist in anything like its present form, you must believe that some Americans are worth more than others, and the people who are worth more are almost all in rural areas while the people who are worth less are almost all in urban areas.

No, National Review, the Electoral College doesn’t “save” America from California’s imperial power; it gives imperial power to a handful of swing states. The only reason California would be more important than any other state is that more Americans live here. Indeed, a lot of Republicans in California are disenfranchised, because they know that their votes will never overcome the overwhelming Democratic majority for the state as a whole and the system is winner-takes-all. Indeed, about 30% of California votes Republican (well, not in the last election, because that was Trump—Orange County went Democrat for the first time in decades), so the number of disenfranchised Republicans alone in California is larger than the population of Michigan, which in turn is larger than the population of Wyoming, North Dakota, South Dakota, Montana, Nebraska, West Virginia, and Kansas combined. Indeed, there are more people in California than there are in Canada. So yeah, I’m thinking maybe we should get a lot of votes?

But it’s easy for you to drum up fear over “imperial rule” by California in particular, because we’re so liberal—and so urban, indeed an astonishing 95% urban, the most of any US state (or frankly probably any major regional entity on the planet Earth! To beat that you have to be something like Singapore, which literally just is a single city).

In fact, while insults thrown at urban people get thrown at basically all of us regardless of what we do, most of the insults that are thrown at rural people are mainly thrown at uneducated rural people. (And statistically, while many people in rural areas are educated and many people in urban areas are not, there’s definitely a positive correlation between urbanization and education.) It’s still unfair in many ways, not least because education isn’t entirely a choice, not in a society where tuition at an average private university costs more than the median individual income. Many of the people we mock as being stupid were really just born poor. It may not be their fault, but they can’t believe that the Earth is only 10,000 years old and not have some substantial failings in their education. I still don’t think mockery is the right answer; it’s really kicking them while they’re down. But clearly there is something wrong with our society when 40% of people believe something so obviously ludicrous—and those beliefs are very much concentrated in the same Southern states that have the most rural populations. “They think we’re ignorant just because we believe that God made the Earth 6,000 years ago!” I mean… yes? I’m gonna have to own up to that one, I guess. I do in fact think that people who believe things that were disproven centuries ago are ignorant.

So really this issue is one-sided. We who live in cities are being systematically degraded and disenfranchised, and when we challenge that system we are accused of being selfish or elitist or worse. We are told that our lifestyles are inferior and shameful, and when we speak out about the positive qualities of our lives—our education, our acceptance of diversity, our flexibility in the face of change—we are again accused of elitism and condescension.

We could simply stew in that resentment. But we can do better. We can reach out to people in rural areas, show them not just that our lives are better—as I said, they already know this—but that they can have these lives too. And we can make policy so that this really can happen for people. Envy doesn’t automatically lead to resentment; that only happens when combined with a lack of mobility. The way urban people pine for the countryside is baffling, since we could go there any time; but the way that country people long for the city is perfectly understandable, as our lives really are better but our rent is too high for them to afford. We need to bring that rent down, not just for the people already living in cities, but also for the people who want to but can’t.

And of course we don’t want to move everyone to cities, either. Many people won’t want to live in cities, and we need a certain population of farmers to make our food after all. We can work to improve infrastructure in rural areas—particularly when it comes to hospitals, which are a basic necessity that is increasingly underfunded. We shouldn’t stop using cost-effectiveness calculations, but we need to compare against the right things. If that hospital isn’t worth building, it should be because there’s another, better hospital we could make for the same amount or cheaper—not because we think that this town doesn’t deserve to have a hospital. We can expand our public transit systems over a wider area, and improve their transit speeds so that people can more easily travel to the city from further away.

We should seriously face up to the costs that free trade has imposed upon many rural areas. We can’t give up on free trade—but that doesn’t mean we need to keep our trade policy exactly as it is. We can do more to ensure that multinational corporations don’t have overwhelming bargaining power against workers and small businesses. We can establish a tax system that would redistribute more of the gains from free trade to the people and places most hurt by the transition. Right now, poor people in the US are often the most fiercely opposed to redistribution of wealth, because somehow they perceive that wealth will be redistributed from them when it would in fact be redistributed to them. They are in a scarcity mindset, their whole worldview shaped by the fact that they struggle to get by. They see every change as a threat, every stranger as an enemy.

Somehow we need to fight that mindset, get them to see that there are many positive changes that can be made, many things that we can achieve together that none of us could achieve along.

Why do so many Americans think that crime is increasing?

Jan 29, JDN 2457783

Since the 1990s, crime in United States has been decreasing, and yet in every poll since then most Americans report that they believe that crime is increasing.

It’s not a small decrease either. The US murder rate is down to the lowest it has been in a century. There are now a smaller absolute number (by 34 log points) of violent crimes per year in the US than there were 20 years ago, despite a significant increase in total population (19 log points—and the magic of log points is that, yes, the rate has decreased by precisely 53 log points).

It isn’t geographically uniform, of course; some states have improved much more than others, and a few states (such as New Mexico) have actually gotten worse.

The 1990s were a peak of violent crime, so one might say that we are just regressing to the mean. (Even that would be enough to make it baffling that people think crime is increasing.) But in fact overall crime in the US is now the lowest it has been since the 1970s, and still decreasing.

Indeed, this decrease has been underestimated, because we are now much better about reporting and investigating crimes than we used to be (which may also be part of why they are decreasing, come to think of it). If you compare against surveys of people who say they have been personally victimized, we’re looking at a decline in violent crime rates of two thirds—109 log points.

Just since 2008 violent crime has decreased by 26% (30 log points)—but of course we all know that Obama is “soft on crime” because he thinks cops shouldn’t be allowed to just shoot Black kids for no reason.

And yet, over 60% of Americans believe that overall crime in the US has increased in the last 10 years (though only 38% think it has increased in their own community!). These figures are actually down from 2010, when 66% thought crime was increasing nationally and 49% thought it was increasing in their local area.

The proportion of people who think crime is increasing does seem to decrease as crime rates decrease—but it still remains alarmingly high. If people were half as rational as most economists seem to believe, the proportion of people who think crime is increasing should drop to basically zero whenever crime rates decrease, since that’s a really basic fact about the world that you can just go look up on the Web in a couple of minutes. There’s no deep ambiguity, not even much “rational ignorance” given the low cost of getting correct answers. People just don’t bother to check, or don’t feel they need to.
What’s going on? How can crime fall to half what it was 20 years ago and yet almost two-thirds of people think it’s actually increasing?

Well, one hint is that news coverage of crime doesn’t follow the same pattern as actual crime.

News coverage in general is a terrible source of information, not simply because news organizations can be biased, make glaring mistakes, and sometimes outright lie—but actually for a much more fundamental reason: Even a perfect news channel, qua news channel, would report what is surprising—and what is surprising is, by definition, improbable. (Indeed, there is a formal mathematical concept in probability theory called surprisal that is simply the logarithm of 1 over the probability.) Even assuming that news coverage reports only the truth, the probability of seeing something on the news isn’t proportional to the probability of the event occurring—it’s more likely proportional to the entropy, which is probability times surprisal.

Now, if humans were optimal information processing engines, that would be just fine, actually; reporting events proportional to their entropy is actually a very efficient mechanism for delivering information (optimal, under certain types of constraints), provided that you can then process the information back into probabilities afterward.

But of course, humans aren’t optimal information processing engines. We don’t recompute the probabilities from the given entropy; instead we use the availability heuristic, by which we simply use the number of times we can think of something happening as our estimate of the probability of that event occurring. If you see more murders on TV news than you used you, you assume that murders must be more common than they used to be. (And when I put it like that, it really doesn’t sound so unreasonable, does it? Intuitively the availability heuristic seems to make sense—which is part of why it’s so insidious.)

Another likely reason for the discrepancy between perception and reality is nostalgia. People almost always have a more positive view of the past than it deserves, particularly when referring to their own childhoods. Indeed, I’m quite certain that a major reason why people think the world was much better when they were kids was that their parents didn’t tell them what was going on. And of course I’m fine with that; you don’t need to burden 4-year-olds with stories of war and poverty and terrorism. I just wish people would realize that they were being protected from the harsh reality of the world, instead of thinking that their little bubble of childhood innocence was a genuinely much safer world than the one we live in today.

Then take that nostalgia and combine it with the availability heuristic and the wall-to-wall TV news coverage of anything bad that happens—and almost nothing good that happens, certainly not if it’s actually important. I’ve seen bizarre fluff pieces about puppies, but never anything about how world hunger is plummeting or air quality is dramatically improved or cars are much safer. That’s the one thing I will say about financial news; at least they report it when unemployment is down and the stock market is up. (Though most Americans, especially most Republicans, still seem really confused on those points as well….) They will attribute it to anything from sunspots to the will of Neptune, but at least they do report good news when it happens. It’s no wonder that people are always convinced that the world is getting more dangerous even as it gets safer and safer.

The real question is what we do about it—how do we get people to understand even these basic facts about the world? I still believe in democracy, but when I see just how painfully ignorant so many people are of such basic facts, I understand why some people don’t. The point of democracy is to represent everyone’s interests—but we also end up representing everyone’s beliefs, and sometimes people’s beliefs just don’t line up with reality. The only way forward I can see is to find a way to make people’s beliefs better align with reality… but even that isn’t so much a strategy as an objective. What do I say to someone who thinks that crime is increasing, beyond showing them the FBI data that clearly indicates otherwise? When someone is willing to override all evidence with what they feel in their heart to be true, what are the rest of us supposed to do?

In defense of slacktivism

Jan 22, JDN 2457776

It’s one of those awkward portmanteaus that people often make to try to express a concept in fewer syllables, while also implicitly saying that the phenomenon is specific enough to deserve its own word: “Slacktivism”, made of “slacker” and “activism”, not unlike “mansplain” is made of “man” and “explain” or “edutainment” was made of “education” and “entertainment”—or indeed “gerrymander” was made of “Elbridge Gerry” and “salamander”. The term seems to be particularly popular on Huffington Post, which has a whole category on slacktivism. There is a particular subcategory of slacktivism that is ironically against other slacktivism, which has been dubbed “snarktivism”.

It’s almost always used as a pejorative; very few people self-identify as “slacktivists” (though once I get through this post, you may see why I’m considering it myself). “Slacktivism” is activism that “isn’t real” somehow, activism that “doesn’t count”.

Of course, that raises the question: What “counts” as legitimate activism? Is it only protest marches and sit-ins? Then very few people have ever been or will ever be activists. Surely donations should count, at least? Those have a direct, measurable impact. What about calling your Congressman, or letter-writing campaigns? These have been staples of activism for decades.
If the term “slacktivism” means anything at all, it seems to point to activities surrounding raising awareness, where the goal is not to enact a particular policy or support a particular NGO but to simply get as much public attention to a topic as possible. It seems to be particularly targeted at blogging and social media—and that’s important, for reasons I’ll get to shortly. If you gather a group of people in your community and give a speech about LGBT rights, you’re an activist. If you send out the exact same speech on Facebook, you’re a slacktivist.

One of the arguments against “slacktivism” is that it can be used to funnel resources at the wrong things; this blog post makes a good point that the Kony 2012 campaign doesn’t appear to have actually accomplished anything except profits for the filmmakers behind it. (Then again: A blog post against slacktivism? Are you sure you’re not doing right now the thing you think you are against?) But is this problem unique to slacktivism, or is it a more general phenomenon that people simply aren’t all that informed about how to have the most impact? There are an awful lot of inefficient charities out there, and in fact the most important waste of charitable funds involves people giving to their local churches. Fortunately, this is changing, as people become more secularized; churches used to account for over half of US donations, and now they only account for less than a third. (Naturally, Christian organizations are pulling out their hair over this.) The 60 million Americans who voted for Trump made a horrible mistake and will cause enormous global damage; but they weren’t slacktivists, were they?

Studies do suggest that traditionally “slacktivist” activities like Facebook likes aren’t a very strong predictor of future, larger actions, and more private modes of support (like donations and calling your Congressman) tend to be stronger predictors. But so what? In order for slacktivism to be a bad thing, they would have to be a negative predictor. They would have to substitute for more effective activism, and there’s no evidence that this happens.

In fact, there’s even some evidence that slacktivism has a positive effect (normally I wouldn’t cite Fox News, but I think in this case we should expect a bias in the opposite direction, and you can read the full Georgetown study if you want):

A study from Georgetown University in November entitled “Dynamics of Cause Engagement” looked how Americans learned about and interacted with causes and other social issues, and discovered some surprising findings on Slacktivism.

While the traditional forms of activism like donating money or volunteering far outpaces slacktivism, those who engage in social issues online are twice as likely as their traditional counterparts to volunteer and participate in events. In other words, slacktivists often graduate to full-blown activism.

At worst, most slacktivists are doing nothing for positive social change, and that’s what the vast majority of people have been doing for the entirety of human history. We can bemoan this fact, but that won’t change it. Most people are simply too uniformed to know what’s going on in the world, and too broke and too busy to do anything about it.

Indeed, slacktivism may be the one thing they can do—which is why I think it’s worth defending.

From an economist’s perspective, there’s something quite odd about how people’s objections to slacktivism are almost always formulated. The rational, sensible objection would be to their small benefits—this isn’t accomplishing enough, you should do something more effective. But in fact, almost all the objections to slacktivism I have ever read focus on their small costs—you’re not a “real activist” because you don’t make sacrifices like I do.

Yet it is a basic principle of economic rationality that, all other things equal, lower cost is better. Indeed, this is one of the few principles of economic rationality that I really do think is unassailable; perfect information is unrealistic and total selfishness makes no sense at all. But cost minimization is really very hard to argue with—why pay more, when you can pay less and get the same benefit?

From an economist’s perspective, the most important thing about an activity is its cost-effectiveness, measured either by net benefitbenefit minus cost—or rate of returnbenefit divided by cost. But in both cases, a lower cost is always better; and in fact slacktivism has an astonishing rate of return, precisely because its cost is so small.

Suppose that a campaign of 10 million Facebook likes actually does have a 1% chance of changing a policy in a way that would save 10,000 lives, with a life expectancy of 50 years each. Surely this is conservative, right? I’m only giving it a 1% chance of success, on a policy with a relatively small impact (10,000 lives could be a single clause in an EPA regulatory standard), with a large number of slacktivist participants (10 million is more people than the entire population of Switzerland). Yet because clicking “like” and “share” only costs you maybe 10 seconds, we’re talking about an expected cost of (10 million)(10/86,400/365) = 0.32 QALY for an expected benefit of (10,000)(0.01)(50) = 5000 QALY. That is a rate of return of 1,500,000%—that’s 1.5 million percent.

Let’s compare this to the rate of return on donating to a top charity like UNICEF, Oxfam, the Against Malaria Foundation, or the Schistomoniasis Control Initiative, for which donating about $300 would save the life of 1 child, adding about 50 QALY. That $300 most likely cost you about 0.01 QALY (assuming an annual income of $30,000), so we’re looking at a return of 500,000%. Now, keep in mind that this is a huge rate of return, far beyond what you can ordinarily achieve, that donating $300 to UNICEF is probably one of the best things you could possibly be doing with that money—and yet slacktivism may still exceed it in efficiency. Maybe slacktivism doesn’t sound so bad after all?

Of course, the net benefit of your participation is higher in the case of donation; you yourself contribute 50 QALY instead of only contributing 0.0005 QALY. Ultimately net benefit is what matters; rate of return is a way of estimating what the net benefit would be when comparing different ways of spending the same amount of time or money. But from the figures I just calculated, it begins to seem like maybe the very best thing you could do with your time is clicking “like” and “share” on Facebook posts that will raise awareness of policies of global importance. Now, you have to include all that extra time spent poring through other Facebook posts, and consider that you may not be qualified to assess the most important issues, and there’s a lot of uncertainty involved in what sort of impact you yourself will have… but it’s almost certainly not the worst thing you could be doing with your time, and frankly running these numbers has made me feel a lot better about all the hours I have actually spent doing this sort of thing. It’s a small benefit, yes—but it’s an even smaller cost.

Indeed, the fact that so many people treat low cost as bad, when it is almost by definition good, and the fact that they also target their ire so heavily at blogging and social media, says to me that what they are really trying to accomplish here has nothing to do with actually helping people in the most efficient way possible.

Rather, it’s two things.

The obvious one is generational—it’s yet another chorus in the unending refrain that is “kids these days”. Facebook is new, therefore it is suspicious. Adults have been complaining about their descendants since time immemorial; some of the oldest written works we have are of ancient Babylonians complaining that their kids are lazy and selfish. Either human beings have been getting lazier and more selfish for thousands of years, or, you know, kids are always a bit more lazy and selfish than their parents or at least seem so from afar.

The one that’s more interesting for an economist is signaling. By complaining that other people aren’t paying enough cost for something, what you’re really doing is complaining that they aren’t signaling like you are. The costly signal has been made too cheap, so now it’s no good as a signal anymore.

“Anyone can click a button!” you say. Yes, and? Isn’t it wonderful that now anyone with a smartphone (and there are more people with access to smartphones than toilets, because #WeLiveInTheFuture) can contribute, at least in some small way, to improving the world? But if anyone can do it, then you can’t signal your status by doing it. If your goal was to make yourself look better, I can see why this would bother you; all these other people doing things that look just as good as what you do! How will you ever distinguish yourself from the riffraff now?

This is also likely what’s going on as people fret that “a college degree’s not worth anything anymore” because so many people are getting them now; well, as a signal, maybe not. But if it’s just a signal, why are we spending so much money on it? Surely we can find a more efficient way to rank people by their intellect. I thought it was supposed to be an education—in which case the meteoric rise in global college enrollments should be cause for celebration. (In reality of course a college degree can serve both roles, and it remains an open question among labor economists as to which effect is stronger and by how much. But the signaling role is almost pure waste from the perspective of social welfare; we should be trying to maximize the proportion of real value added.)

For this reason, I think I’m actually prepared to call myself a slacktivist. I aim for cost-effective awareness-raising; I want to spread the best ideas to the most people for the lowest cost. Why, would you prefer I waste more effort, to signal my own righteousness?

There is no problem of free will, just a lot of really confused people

Jan 15, JDN 2457769

I was hoping for some sort of news item to use as a segue, but none in particular emerged, so I decided to go on with it anyway. I haven’t done any cognitive science posts in awhile, and this is one I’ve been meaning to write for a long time—actually it’s the sort of thing that even a remarkable number of cognitive scientists frequently get wrong, perhaps because the structure of human personality makes cognitive science inherently difficult.

Do we have free will?

The question has been asked so many times by so many people it is now a whole topic in philosophy. The Stanford Encyclopedia of Philosophy has an entire article on free will. The Information Philosopher has a gateway page “The Problem of Free Will” linking to a variety of subpages. There are even YouTube videos about “the problem of free will”.

The constant arguing back and forth about this would be problematic enough, but what really grates me are the many, many people who write “bold” articles and books about how “free will does not exist”. Examples include Sam Harris and Jerry Coyne, and have been published in everything from Psychology Today to the Chronicle of Higher Education. There’s even a TED talk.

The worst ones are those that follow with “but you should believe in it anyway”. In The Atlantic we have “Free will does not exist. But we’re better off believing in it anyway.” Scientific American offers a similar view, “Scientists say free will probably doesn’t exist, but urge: “Don’t stop believing!””

This is a mind-bogglingly stupid approach. First of all, if you want someone to believe in something, you don’t tell them it doesn’t exist. Second, if something doesn’t exist, that is generally considered a pretty compelling reason not to believe in it. You’d need a really compelling counter-argument, and frankly I’m not even sure the whole idea is logically coherent. How can I believe in something if I know it doesn’t exist? Am I supposed to delude myself somehow?

But the really sad part is that it’s totally unnecessary. There is no problem of free will. There are just an awful lot of really, really confused people. (Fortunately not everyone is confused; there are those, such as Daniel Dennett, who actually understand what’s going on.)

The most important confusion is over what you mean by the phrase “free will”. There are really two core meanings here, and the conflation of them is about 90% of the problem.

1. Moral responsibility: We have “free will” if and only if we are morally responsible for our actions.

2. Noncausality: We have “free will” if and only if our actions are not caused by the laws of nature.

Basically, every debate over “free will” boils down to someone pointing out that noncausality doesn’t exist, and then arguing that this means that moral responsibility doesn’t exist. Then someone comes back and says that moral responsibility does exist, and then infers that this means noncausality must exist. Or someone points out that noncausality doesn’t exist, and then they realize how horrible it would be if moral responsibility didn’t exist, and then tells people they should go on believing in noncausality so that they don’t have to give up moral responsibility.

Let me be absolutely clear here: Noncausality could not possibly exist.

Noncausality isn’t even a coherent concept. Actions, insofar as they are actions, must, necessarily, by definition, be caused by the laws of nature.

I can sort of imagine an event not being caused; perhaps virtual electron-positron pairs can really pop into existence without ever being caused. (Even then I’m not entirely convinced; I think quantum mechanics might actually be deterministic at the most fundamental level.)

But an action isn’t just a particle popping into existence. It requires the coordinated behavior of some 10^26 or more particles, all in a precisely organized, unified way, structured so as to move some other similarly large quantity of particles through space in a precise way so as to change the universe from one state to another state according to some system of objectives. Typically, it involves human muscles intervening on human beings or inanimate objects. (Recently it has come to mean specifically human fingers on computer keyboards a rather large segment of the time!) If what you do is an action—not a muscle spasm, not a seizure, not a slip or a trip, but something you did on purpose—then it must be caused. And if something is caused, it must be caused according to the laws of nature, because the laws of nature are the laws underlying all causality in the universe!

And once you realize that, the “problem of free will” should strike you as one of the stupidest “problems” ever proposed. Of course our actions are caused by the laws of nature! Why in the world would you think otherwise?

If you think that noncausality is necessary—or even useful—for free will, what kind of universe do you think you live in? What kind of universe could someone live in, that would fit your idea of what free will is supposed to be?

It’s like I said in that much earlier post about The Basic Fact of Cognitive Science (we are our brains): If you don’t think a mind can be made of matter, what do you think minds are made of? What sort of magical invisible fairy dust would satisfy you? If you can’t even imagine something that would satisfy the constraints you’ve imposed, did it maybe occur to you that your constraints are too strong?

Noncausality isn’t worth fretting over for the same reason that you shouldn’t fret over the fact that pi is irrational and you can’t make a square circle. There is no possible universe in which that isn’t true. So if it bothers you, it’s not that there’s something wrong with the universe—it’s clearly that there’s something wrong with you. Your thinking on the matter must be too confused, too dependent on unquestioned intuitions, if you think that murder can’t be wrong unless 2+2=5.

In philosophical jargon I am called a “compatibilist” because I maintain that free will and determinism are “compatible”. But this is much too weak a term. I much prefer Eleizer Yudkowsky’s “requiredism”, which he explains in one of the greatest blog posts of all time (seriously, read it immediately if you haven’t before—I’m okay with you cutting off my blog post here and reading his instead, because it truly is that brilliant), entitled simply “Thou Art Physics”. This quote sums it up briefly:

My position might perhaps be called “Requiredism.” When agency, choice, control, and moral responsibility are cashed out in a sensible way, they require determinism—at least some patches of determinism within the universe. If you choose, and plan, and act, and bring some future into being, in accordance with your desire, then all this requires a lawful sort of reality; you cannot do it amid utter chaos. There must be order over at least over those parts of reality that are being controlled by you. You are within physics, and so you/physics have determined the future. If it were not determined by physics, it could not be determined by you.

Free will requires a certain minimum level of determinism in the universe, because the universe must be orderly enough that actions make sense and there isn’t simply an endless succession of random events. Call me a “requiredist” if you need to call me something. I’d prefer you just realize the whole debate is silly because moral responsibility exists and noncausality couldn’t possibly.

We could of course use different terms besides “free will”. “Moral responsibility” is certainly a good one, but it is missing one key piece, which is the issue of why we can assign moral responsibility to human beings and a few other entities (animals, perhaps robots) and not to the vast majority of entities (trees, rocks, planets, tables), and why we are sometimes willing to say that even a human being does not have moral responsibility (infancy, duress, impairment).

This is why my favored term is actually “rational volition”. The characteristic that human beings have (at least most of us, most of the time), which also many animals and possibly some robots share (if not now, then soon enough), which justifies our moral responsibility is precisely our capacity to reason. Things don’t just happen to us the way they do to some 99.999,999,999% of the universe; we do things. We experience the world through our senses, have goals we want to achieve, and act in ways that are planned to make the world move closer to achieving those goals. We have causes, sure enough; but not just any causes. We have a specific class of causes, which are related to our desires and intentions—we call these causes reasons.

So if you want to say that we don’t have “free will” because that implies some mysterious nonsensical noncausality, sure; that’s fine. But then don’t go telling us that this means we don’t have moral responsibility, or that we should somehow try to delude ourselves into believing otherwise in order to preserve moral responsibility. Just recognize that we do have rational volition.

How do I know we have rational volition? That’s the best part, really: Experiments. While you’re off in la-la land imagining fanciful universes where somehow causes aren’t really causes even though they are, I can point to not only centuries of human experience but decades of direct, controlled experiments in operant conditioning. Human beings and most other animals behave quite differently in behavioral experiments than, say, plants or coffee tables. Indeed, it is precisely because of this radical difference that it seems foolish to even speak of a “behavioral experiment” about coffee tables—because coffee tables don’t behave, they just are. Coffee tables don’t learn. They don’t decide. They don’t plan or consider or hope or seek.

Japanese, as it turns out, may be a uniquely good language for cognitive science, because it has two fundamentally different verbs for “to be” depending on whether an entity is sentient. Humans and animals imasu, while inanimate objects merely arimasu. We have free will because and insofar as we imasu.

Once you get past that most basic confusion of moral responsibility with noncausality, there are a few other confusions you might run into as well. Another one is two senses of “reductionism”, which Dennett refers to as “ordinary” and “greedy”:

1. Ordinary reductionism: All systems in the universe are ultimately made up of components that always and everywhere obey the laws of nature.

2. Greedy reductionism: All systems in the universe just are their components, and have no existence, structure, or meaning aside from those components.

I actually had trouble formulating greedy reductionism as a coherent statement, because it’s such a nonsensical notion. Does anyone really think that a pile of two-by-fours is the same thing as a house? But people do speak as though they think this about human brains, when they say that “love is just dopamine” or “happiness is just serotonin”. But dopamine in a petri dish isn’t love, any more than a pile of two-by-fours is a house; and what I really can’t quite grok is why anyone would think otherwise.

Maybe they’re simply too baffled by the fact that love is made of dopamine (among other things)? They can’t quite visualize how that would work (nor can I, nor, I think, can anyone in the world at this level of scientific knowledge). You can see how the two-by-fours get nailed together and assembled into the house, but you can’t see how dopamine and action potentials would somehow combine into love.

But isn’t that a reason to say that love isn’t the same thing as dopamine, rather than that it is? I can understand why some people are still dualists who think that consciousness is somehow separate from the functioning of the brain. That’s wrong—totally, utterly, ridiculously wrong—but I can at least appreciate the intuition that underlies it. What I can’t quite grasp is why someone would go so far the other way and say that the consciousness they are currently experiencing does not exist.

Another thing that might confuse people is the fact that minds, as far as we know, are platform independentthat is, your mind could most likely be created out of a variety of different materials, from the gelatinous brain it currently is to some sort of silicon supercomputer, to perhaps something even more exotic. This independence follows from the widely-believed Church-Turing thesis, which essentially says that all computation is computation, regardless of how it is done. This may not actually be right, but I see many reasons to think that it is, and if so, this means that minds aren’t really what they are made of at all—they could be made of lots of things. What makes a mind a mind is how it is structured and above all what it does.

If this is baffling to you, let me show you how platform-independence works on a much simpler concept: Tables. Tables are also in fact platform-independent. You can make a table out of wood, or steel, or plastic, or ice, or bone. You could take out literally every single atom of a table and replace it will a completely different atom of a completely different element—carbon for iron, for example—and still end up with a table. You could conceivably even do so without changing the table’s weight, strength, size, etc., though that would be considerably more difficult.
Does this mean that tables somehow exist “beyond” their constituent matter? In some very basic sense, I suppose so—they are, again, platform-independent. But not in any deep, mysterious sense. Start with a wooden table, take away all the wood, and you no longer have a table. Take apart the table and you have a bunch of wood, which you could use to build something else. There is no “essence” comprising the table. There is no “table soul” that would persist when the table is deconstructed.

And—now for the hard part—so it is with minds. Your mind is your brain. The constituent atoms of your brain are gradually being replaced, day by day, but your mind is the same, because it exists in the arrangement and behavior, not the atoms themselves. Yet there is nothing “extra” or “beyond” that makes up your mind. You have no “soul” that lies beyond your brain. If your brain is destroyed, your mind will also be destroyed. If your brain could be copied, your mind would also be copied. And one day it may even be possible to construct your mind in some other medium—some complex computer made of silicon and tantalum, most likely—and it would still be a mind, and in all its thoughts, feelings and behaviors your mind, if not numerically identical to you.

Thus, when we engage in rational volition—when we use our “free will” if you like that term—there is no special “extra” process beyond what’s going on in our brains, but there doesn’t have to be. Those particular configurations of action potentials and neurotransmitters are our thoughts, desires, plans, intentions, hopes, fears, goals, beliefs. These mental concepts are not in addition to the physical material; they are made of that physical material. Your soul is made of gelatin.

Again, this is not some deep mystery. There is no “paradox” here. We don’t actually know the details of how it works, but that makes this no different from a Homo erectus who doesn’t know how fire works. Maybe he thinks there needs to be some extra “fire soul” that makes it burn, but we know better; and in far fewer centuries than separate that Homo erectus from us, our descendants will know precisely how the brain creates the mind.

Until then, simply remember that any mystery here lies in us—in our ignorance—and not in the universe. And take heart that the kind of “free will” that matters—moral responsibility—has absolutely no need for the kind of “free will” that doesn’t exist—noncausality. They’re totally different things.