2020 is almost over

Dec27 JDN 2459211

I don’t think there are many people who would say that 2020 was their favorite year. Even if everything else had gone right, the 1.7 million deaths from the COVID pandemic would already make this a very bad year.

As if that weren’t bad enough, shutdowns in response to the pandemic, resulting unemployment, and inadequate fiscal policy responses have in a single year thrown nearly 150 million people back into extreme poverty. Unemployment in the US this year spiked to nearly 15%, its highest level since World War 2. Things haven’t been this bad for the US economy since the Great Depression.

And this Christmas season certainly felt quite different, with most of us unable to safely travel and forced to interact with our families only via video calls. New Year’s this year won’t feel like a celebration of a successful year so much as relief that we finally made it through.

Many of us have lost loved ones. Fortunately none of my immediate friends and family have died of COVID, but I can now count half a dozen acquaintances, friends-of-friends or distant relatives who are no longer with us. And I’ve been relatively lucky overall; both I and my partner work in jobs that are easy to do remotely, so our lives haven’t had to change all that much.

Yet 2020 is nearly over, and already there are signs that things really will get better in 2021. There are many good reasons for hope.


Joe Biden won the election by a substantial margin in both the popular vote and the Electoral College.

There are now multiple vaccines for COVID that have been successfully fast-tracked, and they are proving to be remarkably effective. Current forecasts suggest that we’ll have most of the US population vaccinated by the end of next summer.

Maybe the success of this vaccine will finally convince some of the folks who have been doubting the safety and effectiveness of vaccines in general. (Or maybe not; it’s too soon to tell.)

Perhaps the greatest reason to be hopeful about the future is the fact that 2020 is a sharp deviation from the long-term trend toward a better world. That 150 million people thrown back into extreme poverty needs to be compared against the over 1 billion people who have been lifted out of extreme poverty in just the last 30 years.

Those 1.7 million deaths need to be compared against the fact that global life expectancy has increased from 45 to 73 since 1950. The world population is 7.8 billion people. The global death rate has fallen from over 20 deaths per 1000 people per year to only 7.6 deaths per 1000 people per year. Multiplied over 7.8 billion people, that’s nearly 100 million lives saved every single year by advances in medicine and overall economic development. Indeed, if we were to sustain our current death rate indefinitely, our life expectancy would rise to over 130. There are various reasons to think that probably won’t happen, mostly related to age demographics, but in fact there are medical breakthroughs we might make that would make it possible. Even according to current forecasts, world life expectancy is expected to exceed 80 years by the end of the 21st century.

There have also been some significant environmental milestones this year: Global carbon emissions fell an astonishing 7% in 2020, though much of that was from reduced economic activity in response to the pandemic. (If we could sustain that, we’d cut global emissions in half each decade!) But many other milestones were the product of hard work, not silver linings of a global disaster: Whales returned to the Hudson river, Sweden officially terminated their last coal power plant, and the Great Barrier Reef is showing signs of recovery.

Yes, it’s been a bad year for most of us—most of the world, in fact. But there are many reasons to think that next year will be much better.

Hyper-competition

Dec13 JDN 2459197

This phenomenon has been particularly salient for me the last few months, but I think it’s a common experience for most people in my generation: Getting a job takes an awful lot of work.

Over the past six months, I’ve applied to over 70 different positions and so far gone through 4 interviews (2 by video, 2 by phone). I’ve done about 10 hours of test work. That so far has gotten me no offers, though I have yet to hear from 50 employers. Ahead of me I probably have about another 10 interviews, then perhaps 4 of what would have been flyouts and in-person presentations but instead will be “comprehensive interviews” and presentations conducted online, likely several more hours of test work, and then finally, maybe, if I’m lucky, I’ll get a good offer or two. If I’m unlucky, I won’t, and I’ll have to stick around for another year and do all this over again next year.

Aside from the limitations imposed by the pandemic, this is basically standard practice for PhD graduates. And this is only the most extreme end of a continuum of intensive job search efforts, for which even applying to be a cashier at Target requires a formal application, references, and a personality test.

This wasn’t how things used to be. Just a couple of generations ago, low-wage employers would more or less hire you on the spot, with perhaps a resume or a cursory interview. More prestigious employers would almost always require a CV with references and an interview, but it more or less stopped there. I discussed in an earlier post how much of the difference actually seems to come from our chronic labor surplus.

Is all of this extra effort worthwhile? Are we actually fitting people to better jobs this way? Even if the matches are better, are they enough better to justify all this effort?

It is a commonly-held notion among economists that competition in markets is good, that it increases efficiency and improves outcomes. I think that this is often, perhaps usually, the case. But the labor market has become so intensely competitive, particularly for high-paying positions, that the costs of this competitive effort likely outweigh the benefits.

How could this happen? Shouldn’t the free market correct for such an imbalance? Not necessarily. Here is a simple formal model of how this sort of intensive competition can result in significant waste.

Note that this post is about a formal mathematical model, so it’s going to use a lot of algebra. If you are uninterested in such things, you can read the next two paragraphs and then skip to the conclusions at the end.

The overall argument is straightforward: If candidates are similar in skill level, a complicated application process can make sense from a firm’s perspective, but be harmful from society’s perspective, due to the great cost to the applicants. This can happen because the difficult application process imposes an externality on the workers who don’t get the job.

All right, here is where the algebra begins.

I’ve included each equation as both formatted text and LaTeX.

Consider a competition between two applicants, X and Z.

They are each asked to complete a series of tasks in an application process. The amount of effort X puts into the application is x, and the amount of effort Z puts into the application is z. Let’s say each additional bit of effort has a fixed cost, normalized to 1.

Let’s say that their skills are similar, but not identical; this seems quite realistic. X has skill level hx, and Z has skill level hz.

Getting hired has a payoff for each worker of V. This includes all the expected benefits of the salary, benefits, and working conditions. I’ll assume that these are essentially the same for both workers, which also seems realistic.

The benefit to the employer is proportional to the worker’s skill, so letting h be the skill level of the actually hired worker, the benefit of hiring that worker is hY. The reason they are requiring this application process is precisely because they want to get the worker with the highest h. Let’s say that this application process has a cost to implement, c.

Who will get hired? Well, presumably whoever does better on the application. The skill level will amplify the quality of their output, let’s say proportionally to the effort they put in; so X’s expected quality will be hxx and Z’s expected output will be hzz.

Let’s also say there’s a certain amount of error in the process; maybe the more-qualified candidate will sleep badly the day of the interview, or make a glaring and embarrassing typo on their CV. And quite likely the quality of application output isn’t perfectly correlated with the quality of actual output once hired. To capture all this, let’s say that having more skill and putting in more effort only increases your probability of getting the job, rather than actually guaranteeing it.

In particular, let’s say that the probability of X getting hired is P[X] = hxx/(hxx + hzz).

\[ P[X] = \frac{h_x}{h_x x + h_z z} \]

This results in a contest function, a type of model that I’ve discussed in some earlier posts in a rather different context.


The expected payoff for worker X is:

E[Ux] = hxx/(hxx + hzz) V – x

\[ E[U_x] = \frac{h_x x}{h_x x + h_z z} V – x \]

Maximizing this with respect to the choice of effort x (which is all that X can control at this point) yields:

hxhzz V = (hxx + hzz)2

\[ h_x h_z x V = (h_x x + h_z z)^2 \]

A similar maximization for worker Z yields:

hxhzx V = (hxx + hzz)2

\[ h_x h_z z V = (h_x x + h_z z)^2 \]

It follows that x=z, i.e. X and Z will exert equal efforts in Nash equilibrium. Their probability of success will then be contingent entirely on their skill levels:

P[X] = hx/(hx + hz).

\[ P[X] = \frac{h_x}{h_x + h_y} \]

Substituting that back in, we can solve for the actual amount of effort:

hxhzx V = (hx + hz)2x2

\[h_x h_z x V = (h_x + h_z)^2 x^2 \]

x = hxhzV/(hx + hz)2

\[ x = \frac{h_x h_z}{h_x + h_z} V \]

Now let’s see what that gives for the expected payoffs of the firm and the workers. This is worker X’s expected payoff:

E[Ux] = hx/(hx + hz) V – hxhzV/(hx + hz)2 = (hx/(hx + hz))2 V

\[ E[U_x] = \frac{h_x}{h_x + h_z} V – \frac{h_x h_z}{(h_x + h_z)^2} V = \left( \frac{h_x}{h_x + h_z}\right)^2 V \]

Worker Z’s expected payoff is the same, with hx and hz exchanged:

E[Uz] = (hz/(hx + hz))2 V

\[ E[U_z] = \left( \frac{h_z}{h_x + h_z}\right)^2 V \]

What about the firm? Their expected payoff is the the probability of hiring X, times the value of hiring X, plus the probability of hiring Z, times the value of hiring Z, all minus the cost c:

E[Uf] = hx/(hx + hz) hx Y + hz/(hx + hz) hz Y – c= (hx2 + hz2)/(hx + hz) Y – c

\[ E[U_f] = \frac{h_x}{h_x + h_z} h_x Y + \frac{h_z}{h_x + h_z} h_z Y – c = \frac{h_x^2 + h_z^2}{h_x + h_z} Y – c\]

To see whether the application process was worthwhile, let’s compare against the alternative of simply flipping a coin and hiring X or Z at random. The probability of getting hired is then 1/2 for each candidate.

Expected payoffs for X and Z are now equal:

E[Ux] = E[Uz] = V/2

\[ E[U_x] = E[U_z] = \frac{V}{2} \]

The expected payoff for the firm can be computed the same as before, but now without the cost c:

E[Uf] = 1/2 hx Y + 1/2 hz Y = (hx + hz)/2 Y

\[ E[U_f] = \frac{1}{2} h_x Y + \frac{1}{2} h_z Y = \frac{h_x + h_z}{2} Y \]

This has a very simple interpretation: The expected value to the firm is just the average quality of the two workers, times the overall value of the job.

Which of these two outcomes is better? Well, that depends on the parameters, of course. But in particular, it depends on the difference between hx and hz.

Consider two extremes: In one case, the two workers are indistinguishable, and hx = hz = h. In that case, the payoffs for the hiring process reduce to the following:

E[Ux] = E[Uz] = V/4

\[ E[U_x] = E[U_z] = \frac{V}{4} \]

E[Uf] = h Y – c

\[ E[U_f] = h Y – c \]

Compare this against the payoffs for hiring randomly:

E[Ux] = E[Uz] = V/2

\[ E[U_x] = E[U_z] = \frac{V}{2} \]

E[Uf] = h Y

\[ E[U_f] = h Y \]

Both the workers and the firm are strictly better off if the firm just hires at random. This makes sense, since the workers have identical skill levels.

Now consider the other extreme, where one worker is far better than the other; in fact, one is nearly worthless, so hz ~ 0. (I can’t do exactly zero because I’d be dividing by zero, but let’s say one is 100 times better or something.)

In that case, the payoffs for the hiring process reduce to the following:

E[Ux] = V

E[Uz] = 0

\[ E[U_x] = V \]

\[ E[U_z] = 0 \]

X will definitely get the job, so X is much better off.

E[Uf] = hx Y – c

\[ E[U_f] = h_x Y – c \]

If the firm had hired randomly, this would have happened instead:

E[Ux] = E[Uz] = V/2

\[ E[U_x] = E[U_z] = \frac{V}{2} \]

E[Uf] = hY/2

\[ E[U_f] = \frac{h}{2} Y \]

As long as c < hY/2, both the firm and the higher-skill worker are better off in this scenario. (The lower-skill worker is worse off, but that’s not surprising.) The total expected benefit for everyone is also higher in this scenario.


Thus, the difference in skill level between the applicants is vital. If candidates are very different in skill level, in a way that the application process can accurately measure, then a long and costly application process can be beneficial, not only for the firm but also for society as a whole.

In these extreme examples, it was either not worth it for the firm, or worth it for everyone. But there is an intermediate case worth looking at, where the long and costly process can be worth it for the firm, but not for society as a whole. I will call this case hyper-competition—a system that is so competitive it makes society overall worse off.

This inefficient result occurs precisely when:
c < (hx2 + hz2)/(hx + hz) Y – (hx + hz)/2 Y < c + (hx/(hx + hz))2 V + (hz/(hx + hz))2 V

\[ c < \frac{h_x^2 + h_z^2}{h_x + h_z} Y – \frac{h_x + h_z}{2} Y < c + \left( \frac{h_x}{h_x + h_z}\right)^2 V + \left( \frac{h_z}{h_x + h_z}\right)^2 V \]

This simplifies to:

c < (hx – hz)2/(2hx + 2hz) Y < c + (hx2 + hz2)/(hx + hz)2 V

\[ c < \frac{(h_x – h_z)^2}{2 (h_x + h_z)} Y < c + \frac{(h_x^2 + h_z^2)}{(h_x+h_z)^2} V \]

If c is small, then we are interested in the case where:

(hx – hz)2 Y/2 < (hx2 + hz2)/(hx + hz) V

\[ \frac{(h_x – h_z)^2}{2} Y < \frac{h_x^2 + h_z^2}{h_x + h_z} V \]

This is true precisely when the difference hx – hz is small compared to the overall size of hx or hz—that is, precisely when candidates are highly skilled but similar. This is pretty clearly the typical case in the real world. If the candidates were obviously different, you wouldn’t need a competitive process.

For instance, suppose that hx = 10 and hz = 8, while V = 180, Y = 20 and c = 1.

Then, if we hire randomly, these are the expected payoffs:

E[Uf] = (hx + hz)/2 Y = 180

E[Ux] = E[Uz] = V/2 = 90

If we use the complicated hiring process, these are the expected payoffs:

E[Ux] = (hx/(hx + hz))2 V = 55.5

E[Uz] = (hz/(hx + hz))2 V = 35.5

E[Uf] = (hx2 + hz2)/(hx + hz) Y – c = 181

The firm gets a net benefit of 1, quite small; while the workers face a far larger total expected loss of 90. And these candidates aren’t that similar: One is 25% better than the other. Yet because the effort expended in applying was so large, even this improvement in quality wasn’t worth it from society’s perspective.

This conclude’s the algebra for today, if you’ve been skipping it.

In this model I’ve only considered the case of exactly two applicants, but this can be generalized to more applicants, and the effect only gets stronger: Seemingly-large differences in each worker’s skill level can be outweighed by the massive cost of making so many people work so hard to apply and get nothing to show for it.

Thus, hyper-competition can exist despite apparently large differences in skill. Indeed, it is precisely the typical real-world scenario with many applicants who are similar that we expect to see the greatest inefficiencies. In the absence of intervention, we should expect markets to get this wrong.

Of course, we don’t actually want employers to hire randomly, right? We want people who are actually qualified for their jobs. Yes, of course; but you can probably assess that with nothing more than a resume and maybe a short interview. Most employers are not actually trying to find qualified candidates; they are trying to sift through a long list of qualified candidates to find the one that they think is best qualified. And my suspicion is that most of them honestly don’t have good methods of determining that.

This means that it could be an improvement for society to simply ban long hiring processes like these—indeed, perhaps ban job interviews altogether, as I can hardly think of a more efficient mechanism for allowing employers to discriminate based on race, gender, age, or disability than a job interview. Just collect a resume from each applicant, remove the ones that are unqualified, and then roll a die to decide which one you hire.

This would probably make the fit of workers to their jobs somewhat worse than the current system. But most jobs are learned primarily through experience anyway, so once someone has been in a job for a few years it may not matter much who was hired originally. And whatever cost we might pay in less efficient job matches could be made up several times over by the much faster, cheaper, easier, and less stressful process of applying for jobs.

Indeed, think for a moment of how much worse it feels being turned down for a job after a lengthy and costly application process that is designed to assess your merit (but may or may not actually do so particularly well), as opposed to simply finding out that you lost a high-stakes die roll. Employers could even send out letters saying one of two things: “You were rejected as unqualifed for this position.” versus “You were qualified, but you did not have the highest die roll.” Applying for jobs already feels like a crapshoot; maybe it should literally be one.

People would still have to apply for a lot of jobs—actually, they’d probably end up applying for more, because the lower cost of applying would attract more applicants. But since the cost is so much lower, it would still almost certainly be easier to do a job search than it is in the current system. In fact, it could largely be automated: simply post your resume on a central server and the system matches you with employers’ requirements and then randomly generates offers. Employers and prospective employees could fill out a series of forms just once indicating what they were looking for, and then the system could do the rest.

What I find most interesting about this policy idea is that it is in an important sense anti-meritocratic. We are in fact reducing the rewards for high levels of skill—at least a little bit—in order to improve society overall and especially for those with less skill. This is exactly the kind of policy proposal that I had hoped to see from a book like The Meritocracy Trap, but never found there. Perhaps it’s too radical? But the book was all about how we need fundamental, radical change—and then its actual suggestions were simple, obvious, and almost uncontroversial.

Note that this simplified process would not eliminate the incentives to get major, verifiable qualifications like college degrees or years of work experience. In fact, it would focus the incentives so that only those things matter, instead of whatever idiosyncratic or even capricious preferences HR agents might have. There would be no more talk of “culture fit” or “feeling right for the job”, just: “What is their highest degree? How many years have they worked in this industry?” I suppose this is credentialism, but in a world of asymmetric information, I think credentialism may be our only viable alternative to nepotism.

Of course, it’s too late for me. But perhaps future generations may benefit from this wisdom.

Adversity is not a gift

Nov 29 JDN 2459183

For the last several weeks I’ve been participating in a program called “positive intelligence” (which they abbreviate “PQ” even though that doesn’t make sense); it’s basically a self-help program that is designed to improve mood and increase productivity. I am generally skeptical of such things, and I could tell from the start that it was being massively oversold, but I had the opportunity to participate for free, and I looked into the techniques involved and most of them seem to be borrowed from cognitive-behavioral therapy and mindfulness meditation.

Overall, I would say that the program has had small but genuine benefits for me. I think the most helpful part was actually getting the chance to participate in group sessions (via Zoom of course) with others also going through the program. That kind of mutual social support can make a big difference. The group I joined was all comprised of fellow economists (some other grad students, some faculty), so we had a lot of shared experiences.

Some of the techniques feel very foolish, and others just don’t seem to work for me; but I did find at least some of the meditation techniques (which they annoyingly insist on calling by the silly name “PQ reps”) have helped me relax.

But there’s one part of the PQ program in particular that I just can’t buy into, and this is the idea that adversity is a gift and an opportunity.

They call it the “Sage perspective”: You observe the world without judging what is good or bad, and any time you think something is bad, you find a way to transform it into a gift and an opportunity. The claim is that everything—or nearly everything—that happens to you can make you better off. There’s a lot of overlap here with the attitude “Everything happens for a reason”.

I don’t doubt that sincerely believing this would make you happier. Nevertheless, it is obviously false.

If indeed adversity were a gift, we would seek it out. If getting fired or going bankrupt or getting sick were a gift and an opportunity, we’d work to make these things happen.

Yes, it’s true that sometimes an event which seems bad at the time can turn out to have good consequences in the long run. This is simply because we are unable to foresee all future ramifications. Sometimes things turn out differently than you think they will. But most of the time, when something seems bad, it is actually bad.

There might be some small amount of discomfort or risk that would be preferable to a life of complete safety and complacency; but we are perfectly capable of seeking out whatever discomfort or risk we choose. Most of us live with far more discomfort and risk than we would prefer, and simply have no choice in the matter.

If adversity were a gift, people would thank you for giving it to them. “Thanks for dumping me!” “Thanks for firing me!” “Thanks for punching me!” These aren’t the sort of thing we hear very often (at least not sincerely).

I think this is fairly obvious, honestly, so I won’t belabor it any further. But it raises a question: Is there a way to salvage the mental health benefits of this attitude while abandoning its obvious falsehood?

“Everything happens for a reason” doesn’t work; we live in a universe of deep randomness, ruled by the blind idiot gods of natural law.

“Every cloud has a silver lining” is better; but clearly not every bad thing has an upside, or if it does the upside can be so small as to be utterly negligible. (What was the upside of Rwandan genocide?) Restricted to ordinary events like getting fired this one works pretty well; but it obviously fails for the most extreme traumas, and doesn’t seem particularly helpful for the death of a loved one either.

“What doesn’t kill me makes me stronger” is better still, but clearly not true in every case; some bad events that don’t actually kill us can traumatize us and make the rest of our lives harder. Perhaps “What doesn’t permanently damage me makes me stronger”?

I think the version of this attitude that I have found closest to the truth is “Everything is raw material”. Sometimes bad things just happen: Bad luck, or bad actions, can harm just about anyone at just about any time. But it is within our power to decide how we will respond to what happens to us, and wallowing in despair is almost never the best response.

Thus, while it is foolish to see adversity as a gift, it is not so foolish to see it as an opportunity. Don’t try to pretend that bad things aren’t bad. There’s no sense in denying that we would prefer some outcomes over others, and we feel hurt or disappointed when things don’t turn out how we wanted. Yet even what is bad can still contain within it chances to learn or make things better.

What’s wrong with “should”?

Nov 8 JDN 2459162

I have been a patient in cognitive behavioral therapy (CBT) for many years now. The central premise that thoughts can influence emotions is well-founded, and the results of CBT are empirically well supported.

One of the central concepts in CBT is cognitive distortions: There are certain systematic patterns in how we tend to think, which often results in beliefs and emotions that are disproportionate with reality.

Most of the cognitive distortions CBT deals with make sense to me—and I am well aware that my mind applies them frequently: All-or-nothing, jumping to conclusions, overgeneralization, magnification and minimization, mental filtering, discounting the positive, personalization, emotional reasoning, and labeling are all clearly distorted modes of thinking that nevertheless are extremely common.

But there’s one “distortion” on CBT lists that always bothers me: “should statements”.

Listen to this definition of what is allegedly a cognitive distortion:

Another particularly damaging distortion is the tendency to make “should” statements. Should statements are statements that you make to yourself about what you “should” do, what you “ought” to do, or what you “must” do. They can also be applied to others, imposing a set of expectations that will likely not be met.

When we hang on too tightly to our “should” statements about ourselves, the result is often guilt that we cannot live up to them. When we cling to our “should” statements about others, we are generally disappointed by their failure to meet our expectations, leading to anger and resentment.

So any time we use “should”, “ought”, or “must”, we are guilty of distorted thinking? In other words, all of ethics is a cognitive distortion? The entire concept of obligation is a symptom of a mental disorder?

Different sources on CBT will define “should statements” differently, and sometimes they offer a more nuanced definition that doesn’t have such extreme implications:

Individuals thinking in ‘shoulds’, ‘oughts; or ‘musts’ have an ironclad view of how they and others ‘should’ and ‘ought’ to be. These rigid views or rules can generate feels of anger, frustration, resentment, disappointment and guilt if not followed.

Example: You don’t like playing tennis but take lessons as you feel you ‘should’, and that you ‘shouldn’t’ make so many mistakes on the court, and that your coach ‘ought to’ be stricter on you. You also feel that you ‘must’ please him by trying harder.

This is particularly problematic, I think, because of the All-or-Nothing distortion which does genuinely seem to be common among people with depression: Unless you are very clear from the start about where to draw the line, our minds will leap to saying that all statements involving the word “should” are wrong.

I think what therapists are trying to capture with this concept is something like having unrealistic expectations, or focusing too much on what could or should have happened instead of dealing with the actual situation you are in. But many seem to be unable to articulate that clearly, and instead end up asserting that entire concept of moral obligation is a cognitive distortion.

There may be a deeper error here as well: The way we study mental illness doesn’t involve enough comparison with the control group. Psychologists are accustomed to asking the question, “How do people with depression think?”; but they are not accustomed to asking the question, “How do people with depression think compared to people who don’t?” If you want to establish that A causes B, it’s not enough to show that those with B have A; you must also show that those who don’t have B also don’t have A.

This is an extreme example for illustration, but suppose someone became convinced that depression is caused by having a liver. They studied a bunch of people with depression, and found that they all had livers; hypothesis confirmed! Clearly, we need to remove the livers, and that will cure the depression.

The best example I can find of a study that actually asked that question compared nursing students and found that cognitive distortions explain about 20% of the variance in depression. This is a significant amount—but still leaves a lot unexplained. And most of the research on depression doesn’t even seem to think to compare against people without depression.

My impression is that some cognitive distortions are genuinely more common among people with depression—but not all of them. There is an ongoing controversy over what’s called the depressive realism effect, which is the finding that in at least some circumstances the beliefs of people with mild depression seem to be more accurate than the beliefs of people with no depression at all. The result is controversial both because it seems to threaten the paradigm that depression is caused by distortions, and because it seems to be very dependent on context; sometimes depression makes people more accurate in their beliefs, other times it makes them less accurate.

Overall, I am inclined to think that most people have a variety of cognitive distortions, but we only tend to notice when those distortions begin causing distress—such when are they involved in depression. Human thinking in general seems to be a muddled mess of heuristics, and the wonder is that we function as well as we do.

Does this mean that we should stop trying to remove cognitive distortions? Not at all. Distorted thinking can be harmful even if it doesn’t cause you distress: The obvious example is a fanatical religious or political belief that leads you to harm others. And indeed, recognizing and challenging cognitive distortions is a highly effective treatment for depression.

Actually I created a simple cognitive distortion worksheet based on the TEAM-CBT approach developed by David Burns that has helped me a great deal in a remarkably short time. You can download the worksheet yourself and try it out. Start with a blank page and write down as many negative thoughts as you can, and then pick 3-5 that seem particularly extreme or unlikely. Then make a copy of the cognitive distortion worksheet for each of those thoughts and follow through it step by step. Particularly do not ignore the step “This thought shows the following good things about me and my core values:”; that often feels the strangest, but it’s a critical part of what makes the TEAM-CBT approach better than conventional CBT.

So yes, we should try to challenge our cognitive distortions. But the mere fact that a thought is distressing doesn’t imply that it is wrong, and giving up on the entire concept of “should” and “ought” is throwing out a lot of babies with that bathwater.

We should be careful about labeling any thoughts that depressed people have as cognitive distortions—and “should statements” is a clear example where many psychologists have overreached in what they characterize as a distortion.

What meritocracy trap?

Nov 1 JDN 2459155

So I just finished reading The Meritocracy Trap by David Markovits.

The basic thesis of the book is that America’s rising inequality is not due to a defect in our meritocratic ideals, but is in fact their ultimate fruition. Markovits implores us to reject the very concept of meritocracy, and replace it with… well, something, and he’s never very clear about exactly what.

The most frustrating thing about reading this book is trying to figure out where Markovits draws the line for “elite”. He rapidly jumps between talking about the upper quartile, the upper decile, the top 1%, and even the top 0.1% or top 0.01% while weaving his narrative. The upper quartile of the US contains 75 million people; the top 0.01% contains only 300,000. The former is the size of Germany, the latter the size of Iceland (which has fewer people than Long Beach). Inequality which concentrates wealth in the top quartile of Americans is a much less serious problem than inequality which concentrates wealth in the top 0.01%. It could still be a problem—those lower three quartiles are people too—but it is definitely not nearly as bad.

I think it’s particularly frustrating to me personally, because I am an economist, which means both that such quantitative distinctions are important to me, and also that whether or not I myself am in this “elite” depends upon which line you are drawing. Do I have a post-graduate education? Yes. Was I born into the upper quartile? Not quite, but nearly. Was I raised by married parents in a stable home? Certainly. Am I in the upper decile and working as a high-paid professional? Hopefully I will be soon. Will I enter the top 1%? Maybe, maybe not. Will I join the top 0.1%? Probably not. Will I ever be in the top 0.01% and a captain of industry? Almost certainly not.

So, am I one of the middle class who are suffering alienation and stagnation, or one of the elite who are devouring themselves with cutthroat competition? Based on BLS statistics for economists and job offers I’ve been applying to, my long-term household income is likely to be about 20-50% higher than my parents’; this seems like neither the painful stagnation he attributes to the middle class nor the unsustainable skyrocketing of elite incomes. (Even 50% in 30 years is only 1.4% per year, about our average rate of real GDP growth.) Marxists would no doubt call me petit bourgeoisie; but isn’t that sort of the goal? We want as many people as possible to live comfortable upper-middle class lives in white-collar careers?

Markovits characterizes—dare I say caricatures—the habits of the middle-class versus the elite, and once again I and most people I know cross-cut them: I spend more time with friends than family (elite), but I cook familiar foods, not fancy dinners (middle); I exercise fairly regularly and don’t watch much television (elite) but play a lot of video games and sleep a lot as well (middle). My web searches involve technology and travel (elite), but also chronic illness (middle). I am a donor to Amnesty International (elite) but also play tabletop role-playing games (middle). I have a functional, inexpensive car (middle) but a top-of-the-line computer (elite)—then again that computer is a few years old now (middle). Most of the people I hang out with are well-educated (elite) but struggling financially (middle), civically engaged (elite) but pessimistic (middle). I rent my apartment and have a lot of student debt (middle) but own stocks (elite). (The latter seemed like a risky decision before the pandemic, but as stock prices have risen and student loan interest was put on moratorium, it now seems positively prescient.) So which class am I, again?

I went to public school (middle) but have a graduate degree (elite). I grew up in Ann Arbor (middle) but moved to Irvine (elite). Then again my bachelor’s was at a top-10 institution (elite) but my PhD will be at only a top-50 (middle). The beautiful irony there is that the top-10 institution is the University of Michigan and the top-50 institution is the University of California, Irvine. So I can’t even tell which class each of those events is supposed to represent! Did my experience of Ann Arbor suddenly shift from middle class to elite when I graduated from public school and started attending the University of Michigan—even though about a third of my high school cohort did exactly that? Was coming to UCI an elite act because it’s a PhD in Orange County, or a middle-class act because it’s only a top-50 university?

If the gap between these two classes is such a wide chasm, how am I straddling it? I honestly feel quite confident in characterizing myself as precisely the upwardly-mobile upper-middle class that Markovits claims no longer exists. Perhaps we’re rarer than we used to be; perhaps our status is more precarious; but we plainly aren’t gone.

Markovits keeps talking about “radical differences” “not merely in degree but in kind” between “subordinate” middle-class workers and “superordinate” elite workers, but if the differences are really that stark, why is it so hard to tell which group I’m in? From what I can see, the truth seems less like a sharp divide between middle-class and upper-class, and more like an increasingly steep slope from middle-class to upper-middle class to upper-class to rich to truly super-rich. If I had to put numbers on this, I’d say annual household incomes of about $50,000, $100,000, $200,000, $400,000, $1 million, and $10 million respectively. (And yet perhaps I should add more categories: Even someone who makes $10 million a year has only pocket change next to Elon Musk or Jeff Bezos.) The slope has gotten steeper over time, but it hasn’t (yet?) turned into a sharp cliff the way Markovits describes. America’s Lorenz curve is clearly too steep, but it doesn’t have a discontinuity as far as I can tell.

Some of the inequalities Markovits discusses are genuine, but don’t seem to be particularly related to meritocracy. The fact that students from richer families go to better schools indeed seems unjust, but the problem is clearly not that the rich schools are too good (except maybe at the very top, where truly elite schools seem a bit excessive—five-figure preschool tuition?), but that the poor schools are not good enough. So it absolutely makes sense to increase funding for poor schools and implement various reforms, but this is hardly a radical notion—nor is it in any way anti-meritocratic. Providing more equal opportunities for the poor to raise their own station is what meritocracy is all about.

Other inequalities he objects to seem, if not inevitable, far too costly to remove: Educated people are better parents, who raise their children in ways that make them healthier, happier, and smarter? No one is going to apologize for being a good parent, much less stop doing so because you’re concerned about what it does to inequality. If you have some ideas for how we might make other people into better parents, by all means let’s hear them. But I believe I speak for the entire upper-middle class when I say: when I have kids of my own, I’m going to read to them, I’m not going to spank them, and there’s not a damn thing you can do to change my mind on either front. Quite frankly, this seems like a heavy-handed satire of egalitarianism, right out of Harrison Bergeron: Let’s make society equal by forcing rich people to neglect and abuse their kids as much as poor people do! My apologies to Vonnegut: I thought you were ridiculously exaggerating, but apparently some people actually think like this.

This is closely tied with the deepest flaw in the argument: The meritocratic elite are actually more qualified. It’s easy to argue that someone like Donald Trump shouldn’t rule the world; he’s a deceitful, narcissistic, psychopathic, incompetent buffoon. (The only baffling part is that 40% of American voters apparently disagree.) But it’s a lot harder to see why someone like Bill Gates shouldn’t be in charge of things: He’s actually an extremely intelligent, dedicated, conscientious, hard-working, ethical, and competent individual. Does he deserve $100 billion? No, for reasons I’ve talked about before. But even he knows that! He’s giving most of it away to highly cost-effective charities! Bill Gates alone has saved several million lives by his philanthropy.

Markovits tries to argue that the merits of the meritocratic elite are arbitrary and contextual, like the alleged virtues of the aristocratic class: “The meritocratic virtues, that is, are artifacts of economic inequality in just the fashion in which the pitching virtues are artifacts of baseball.” (p. 264) “The meritocratic achievement commonly celebrated today, no less than the aristocratic virtue acclaimed in the ancien regime, is a sham.” (p. 268)

But it’s pretty hard for me to see how things like literacy, knowledge of history and science, and mathematical skill are purely arbitrary. Even the highly specialized skills of a quantum physicist, software engineer, or geneticist are clearly not arbitrary. Not everyone needs to know how to solve the Schrodinger equation or how to run a polymerase chain reaction, but our civilization greatly benefits from the fact that someone does. Software engineers aren’t super-productive because of high inequality; they are super-productive because they speak the secret language of the thinking machines. I suppose some of the skills involved in finance, consulting, and law are arbitrary and contextual; but he makes it sound like the only purpose graduate school serves is in teaching us table manners.

Precisely by attacking meritocracy, Markovits renders his own position absurd. So you want less competent people in charge? You want people assigned to jobs they’re not good at? You think businesses should go out of their way to hire employees who will do their jobs worse? Had he instead set out to show how American society fails at achieving its meritocratic ideals—indeed, failing to provide equality of opportunity for the poor is probably the clearest example of this—he might have succeeded. But instead he tries to attack the ideals themselves, and fails miserably.

Markovits avoids the error that David Graeber made: Graeber sees that there are many useless jobs but doesn’t seem to have a clue why these jobs exist (and turns to quite foolish Marxian conspiracy theories to explain it). Markovits understands that these jobs are profitable for the firms that employ them, but unproductive for society as a whole. He is right; this is precisely what virtually the entire fields of finance, sales, advertising, and corporate law consist of. Most people in our elite work very hard with great skill and competence, and produce great profits for the corporations that employ them, all while producing very little of genuine societal value. But I don’t see how this is a flaw in meritocracy per se.

Nor does Markovits stop at accusing employment of being rent-seeking; he takes aim at education as well: “when the rich make exceptional investments in schooling, this does reduce the value of ordinary, middle-class training and degrees. […] Meritocratic education inexorably engenders a wasteful and destructive arms educational arms race, which ultimately benefits no one, not even the victors.” (p.153) I don’t doubt that education is in part such a rent-seeking arms race, and it’s worthwhile to try to minimize that. But education is not entirely rent-seeking! At the very least, is there not genuine value in teaching children to read and write and do arithmetic? Perhaps by the time we get to calculus or quantum physics or psychopathology we have reached diminishing returns for most students (though clearly at least some people get genuine value out of such things!), but education is not entirely comprised of signaling or rent-seeking (and nor do “sheepskin effects” prove otherwise).

My PhD may be less valuable to me than it would be to someone in my place 40 years ago, simply because there are more people with PhDs now and thus I face steeper competition. Then again, perhaps not, as the wage premium for college and postgraduate education has been increasing, not decreasing, over that time period. (How much of that wage premium is genuine social benefit and how much is rent-seeking is difficult to say.) In any case it’s definitely still valuable. I have acquired many genuine skills, and will in fact be able to be genuinely more productive as well as compete better in the labor market than I would have without it. Some parts of it have felt like a game where I’m just trying to stay ahead of everyone else, but it hasn’t all been that. A world where nobody had PhDs would be a world with far fewer good scientists and far slower technological advancement.

Abandoning meritocracy entirely would mean that we no longer train people to be more productive or match people to the jobs they are most qualified to do. Do you want a world where surgery is not done by the best surgeons, where airplanes are not flown by the best pilots? This necessarily means less efficient production and an overall lower level of prosperity for society as a whole. The most efficient way may not be the best way, but it’s still worth noting that it’s the most efficient way.

Really, is meritocracy the problem, or is it something else?

Markovits is clearly right that something is going wrong with American society: Our inequality is much too high, and our job market is much too cutthroat. I can’t even relate to his description of what the job market was like in the 1960s (“Old Economy Steve” has it right): “Even applicants for white-collar jobs received startlingly little scrutiny. For most midcentury workers, getting a job did not involve any application at all, in the competitive sense of the term.” (p.203)

In fact, if anything he seems to understate the difference across time, perhaps because it lets him overstate the difference across class (p. 203):

Today, by contrast, the workplace is methodically arranged around gradations of skill. Firms screen job candidates intensively at hiring, and they then sort elite and non-elite workers into separate physical spaces.

Only the very lowest-wage employers, seeking unskilled workers, hire casually. Middle-class employers screen using formal cognitive tests and lengthy interviews. And elite employers screen with urgent intensity, recruiting from only a select pool and spending millions of dollars to probe applicants over several rounds of interviews, lasting entire days.

Today, not even the lowest-wage employers hire casually! Have you ever applied to work at Target? There is a personality test you have to complete, which I presume is designed to test your reliability as an obedient corporate drone. Never in my life have I gotten a job that didn’t involve either a lengthy application process or some form of personal connection—and I hate to admit it, but usually the latter. It is literally now harder to get a job as a cashier at Target than it was to get a job as an engineer at Ford 60 years ago.

But I still can’t shake the feeling that meritocracy is not exactly what’s wrong here. The problem with the sky-high compensation packages at top financial firms isn’t that they are paid to people who are really good at their jobs; it’s that those jobs don’t actually accomplish anything beneficial for society. Where elite talent and even elite compensation is combined with genuine productivity, such as in science and engineering, it seems unproblematic (and I note that Markovits barely even touches on these industries, perhaps because he sees they would undermine his argument). The reason our economic growth seems to have slowed as our inequality has massively surged isn’t that we are doing too good a job of rewarding people for being productive.

Indeed, it seems like the problem may be much simpler: Labor supply exceeds labor demand.

Take a look at this graph from the Federal Reserve Bank of San Francisco:

[Beveridge_curve_data.png]

This graph shows the relationship over time between unemployment and job vacancies. As you can see, they are generally inversely related: More vacancies means less unemployment. I have drawn in a green line which indicates the cutoff between having more vacancies than unemployment—upper left—and having more unemployment than vacancies—lower right. We have almost always been in the state of having more unemployment than we have vacancies; notably, the mid-1960s were one of the few periods in which we had significantly more vacancies than unemployment.

For decades we’ve been instituting policies to try to give people “incentives to work”; but there is no shortage of labor in this country. We seem to have plenty of incentives to work—what we need are incentives to hire people and pay them well.

Indeed, perhaps we need incentives not to work—like a basic income or an expanded social welfare system. Thanks to automation, productivity is now astonishingly high, and yet we work ourselves to death instead of enjoying leisure.

And of course there are various other policy changes that have made our inequality worse—chiefly the dramatic drops in income tax rates at the top brackets that occurred under Reagan.

In fact, many of the specific suggestions Markovits makes—which, much to my chagrin, he waits nearly 300 pages to even mention—are quite reasonable, or even banal: He wants to end tax deductions for alumni donations to universities and require universities to enroll more people from lower income brackets; I could support that. He wants to regulate finance more stringently, eliminate most kinds of complex derivatives, harmonize capital gains tax rates to ordinary income rates, and remove the arbitrary cap on payroll taxes; I’ve been arguing for all of those things for years. What about any of these policies is anti-meritocratic? I don’t see it.

More controversially, he wants to try to re-organize production to provide more opportunities for mid-skill labor. In some industries I’m not sure that’s possible: The 10X programmer is a real phenomenon, and even mediocre programmers and engineers can make software and machines that are a hundred times as productive as doing the work by hand would be. But some of his suggestions make sense, such as policies favoring nurse practitioners over specialist doctors and legal secretaries instead of bar-certified lawyers. (And please, please reform the medical residency system! People die from the overwork caused by our medical residency system.)

But I really don’t see how not educating people or assigning people to jobs they aren’t good at would help matters—which means that meritocracy, as I understand the concept, is not to blame after all.

How to be a good writer

Oct 25 JDN 2459148

“A writer is someone for whom writing is more difficult than it is for other people.”
~ Thomas Mann

“You simply sit down at the typewriter, open your veins, and bleed.”

~ Red Smith

Why is it so difficult to write well? Why is it that those of us who write the most often find it the most agonizing?

My guess is that many other art forms are similar, but writing is what I know best.

I have come to realize that there are four major factors which determine the quality of someone’s writing, and the pain and challenge of writing comes from the fact that they are not very compatible with one another.

The first is talent. To a certain degree, one can be born a better or worse writer, or become so through forces not of one’s own making. This one costs nothing to get if you already have it, but if you don’t have it, you can’t really acquire it. If you do lack talent, that doesn’t mean you can’t write; but it does limit how successful you are likely to be at writing. (Then again, some very poorly-written books have made some very large sums of money!) It’s also very difficult to know whether you really have talent; people tell me I do, so I suppose I believe them.

The second is practice. You must write and keep on writing. You must write many things in many contexts, and continue to write despite various pressures and obstacles trying to stop you from writing. Reading is also part of this process, as we learn new ways to use words by seeing how others have used them. In fact, you should read more words than you write.

The third is devotion. If you are to truly write well, you must pour your heart and soul into what you write. I can tell fairly quickly whether someone is serious about writing or not by seeing how they react to the metaphor I like to use: “I carve off shards of my soul and assemble them into robots that I release into the world; and when the robots fail, I wonder whether I have assembled them incorrectly, or if there is something fundamentally wrong with my soul itself.” Most people react with confusion. Serious writers nod along in agreement.

The fourth is criticism. You must seek out criticism from a variety of sources, you must accept that criticism, and you must apply it in improving your work in the future. You must avoid becoming defensive, but you must also recognize that disagreement will always exist. You will never satisfy everyone with what you write. The challenge is to satisfy as much of your target audience as possible.

And therein lies the paradox: For when you have devoted your heart and soul into a work, receiving criticism on it can make you want to shut down, wanting to avoid that pain. And thus, you stop practicing, and you stop improving.

What can be done about this?

I am told that it helps to “get a thick skin”, but seeing as I’ve spent the better part of my life trying to do that and failed completely, this may not be the most useful advice. Indeed, even if it can be done it may not be worth it: The most thick-skinned people I know of are generally quite incompetent at whatever they do, because they ignore criticism. There are two ways to be a narcissist: One is to be so sensitive to criticism that you refuse to hear it; the other is to be so immune to criticism that it has no effect on you. (The former is “covert narcissism”, the latter is “overt narcissism”.)

One thing that does seem to help is learning to develop some measure of detachment frrom your work, so that you can take criticism of your work as applying to that work and not to yourself. Usually the robots really are just misassembled, and there’s nothing wrong with your soul.

But this can be dangerous as well: If you detach yourself too much from your work, you lose your devotion to it, and it becomes mechanically polished but emotionally hollow. If you optimize over and over to what other people want, it eventually stops being the work that had meaning for you.

Perhaps what ultimately separates good writers from everyone else is not what they can do, but what they feel they must do: Serious writers feel a kind of compulsion to write, an addiction to transferring thoughts into words. Often they don’t even particularly enjoy it; they don’t “want” to write in the ordinary sense of the word. They simply must write, feeling as though they die or go mad if they ever were forced to stop. It is this compulsion that gets them to persevere in the face of failure and rejection—and the self-doubt that rejection drives.

And if you don’t feel that compulsion? Honestly, maybe you’re better off than those of us who do.

What would a better job market look like?

Sep 13 JDN 2459106

I probably don’t need to tell you this, but getting a job is really hard. Indeed, much harder than it seems like it ought to be.

Having all but completed my PhD, I am now entering the job market. The job market for economists is quite different from the job market most people deal with, and these differences highlight some potential opportunities for improving job matching in our whole economy—which, since employment is such a large part of our lives, could have wide-ranging benefits for our society.

The most obvious difference is that the job market for economists is centralized: Job postings are made through the American Economic Association listing of Job Openings for Economists (often abbrievated AEA JOE); in a typical year about 4,000 jobs are posted there. All of them have approximately the same application deadline, near the end of the year. Then, after applying to various positions, applicants get interviewed in rapid succession, all at the annual AEA conference. Then there is a matching system, where applicants get to send two “signals” indicating their top choices and then offers are made.

This year of course is different, because of COVID-19. The conference has been canceled, with all of its presentations moved online; interviews will also be conducted online. Perhaps more worrying, the number of postings has been greatly reduced, and based on past trends may be less than half of the usual number. (The number of applicants may also be reduced, but it seems unlikely to drop as much as the number of postings does.)

There are a number of flaws in even this system. First, it’s too focused on academia; very few private-sector positions use the AEA JOE system, and almost no government positions do. So those of us who are not so sure we want to stay in academia forever end up needing to deal with both this system and the conventional system in parallel. Second, I don’t understand why they use this signaling system and not a deferred-acceptance matching algorithm. I should be able to indicate more about my preferences than simply what my top two choices are—particularly when most applicants apply to over 100 positions. Third, it isn’t quite standardized enough—some positions do have earlier deadlines or different application materials, so you can’t simply put together one application packet and send it to everyone at once.

Still, it’s quite obvious that this system is superior to the decentralized job market that most people deal with. Indeed, this becomes particularly obvious when one is participating in both markets at once, as I am. The decentralized market has a wide range of deadlines, where upon seeing an application you may need to submit to it within that week, or you may have several months to respond. Nearly all applications require a resume, but different institutions will expect different content on it. Different applications may require different materials: Cover letters, references, writing samples, and transcripts are all things that some firms will want and others won’t.

Also, this is just my impression from a relatively small sample, but I feel like the AEA JOE listings are more realistic, in the following sense: They don’t all demand huge amounts of prior experience, and those that do ask for prior experience are either high-level positions where that’s totally reasonable, or are willing to substitute education for experience. For private-sector job openings you basically have to subtract three years from whatever amount of experience they say they require, because otherwise you’d never have anywhere you could apply to. (Federal government jobs are a weird case here; they all say they require a lot of experience at a specific government pay grade, but from talking with those who have dealt with the system before, they are apparently willing to make lots of substitutions—private-sector jobs, education, and even hobbies can sometimes substitute.)

I think this may be because the decentralized market has to some extent unraveled. The job market is the epitome of a matching market; unraveling in a matching market occurs when there is fierce competition for a small number of good candidates or, conversely, a small number of good openings. Each firm has the incentive to make a binding offer earlier than the others, with a short deadline so that candidates don’t have time to shop around. As firms compete with each other, they start making deadlines earlier and earlier until candidates feel like they are in a complete crapshoot: An offer made on Monday might be gone by Friday, and you have no way of knowing if you should accept it now or wait for a better one to come along. This is a Tragedy of the Commons: Given what other firms are doing, each firm benefits from making an earlier binding offer. But once they all make early offers, that benefit disappears and the result just makes the whole system less efficient.

The centralization of the AEA JOE market prevents this from happening: Everyone has common deadlines and does their interviews at the same time. Each institution may be tempted to try to break out of the constraints of the centralized market, but they know that if they do, they will be punished by receiving fewer applicants.

The fact that the centralized market is more efficient is likely a large part of why economics PhDs have the lowest unemployment rate of any PhD graduates and nearly the lowest unemployment rate of any job sector whatsoever. In some sense we should expect this: If anyone understands how to make employment work, it should be economists. Noah Smith wrote in 2013 (and I suppose I took it to heart): “If you get a PhD, get an economics PhD.” I think PhD graduates are the right comparison group here: If we looked at the population as a whole, employment rates and salaries for economists look amazing, but that isn’t really fair since it’s so much harder to become an economist than it is to get most other jobs. But I don’t think it’s particularly easier to get a PhD in physics or biochemistry than to get one in economics, and yet economists still have a lower unemployment rate than physicists or biochemists. (Though it’s worth noting that any PhD—yes, even in the humanities—will give you a far lower risk of unemployment than the general population.) The fact that we have AEA JOE and they don’t may be a major factor here.


So, here’s my question: Why don’t we do this in more job markets? It would be straightforward enough to do this for all PhD graduates, at least—actually my understanding is that some other disciplines do have centralized markets similar to the one in economics, but I’m not sure how common this is.

The federal government could relatively easily centralize its own job market as well; maybe not for positions that need to be urgently filled, but anything that can wait several months would be worth putting into a centralized system that has deadlines once or twice a year.

But what about the private sector, which after all is where most people work? Could we centralize that system as well?

It’s worth noting the additional challenges that immediately arise: Many positions need to be filled immediately, and centralization would make that impossible. There are thousands of firms that would need to be coordinated (there are at least 100,000 firms in the US with 100 or more employees). There are millions of different jobs to be filled, requiring a variety of different skills. In an average month over 5 million jobs are filled in the United States.

Most people want a job near where they live, so part of the solution might be to centralize only jobs within a certain region, such as a particular metro area. But if we are limited to open positions of a particular type within a particular city, there might not be enough openings at any given time to be worth centralizing. And what about applicants who don’t care so much about geography? Should they be applying separately to each regional market?

Yet even with all this in mind, I think some degree of centralization would be feasible and worthwhile. If nothing else, I think standardizing deadlines and application materials could make a significant difference—it’s far easier to apply to many places if they all use the same application and accept them at the same time.

Another option would be to institute widespread active labor market policies, which are a big part of why #ScandinaviaIsBetter. Denmark especially invests heavily in such programs, which provide training and job matching for unemployed citizens. It is no coincidence that Denmark has kept their unemployment rate under 7% even through the worst of the Great Recession. The US unemployment rate fluctuates wildly with the business cycle, while most of Europe has steadier but higher unemployment. Indeed, the lowest unemployment rates in France over the last 30 years have exceeded the highest rates in Denmark over the same period. Denmark spends a lot on their active labor market programs, but I think they’re getting their money’s worth.

Such a change would make our labor markets more efficient, matching people to jobs that fit them better, increasing productivity and likely decreasing turnover. Wages probably wouldn’t change much, but working in a better job for the same wage is still a major improvement in your life. Indeed, job satisfaction is one of the strongest predictors of life satisfaction, which isn’t too surprising given how much of our lives we spend at work.

Sincerity inflation

Aug 30 JDN 2459092

What is the most saccharine, empty, insincere way to end a letter? “Sincerely”.

Whence such irony? Well, we’ve all been using it for so long that we barely notice it anymore. It’s just the standard way to end a letter now.

This process is not unlike inflation: As more and more dollars get spent, the value of a dollar decreases, and as a word or phrase gets used more and more, its meaning weakens.

It’s hardly just the word “Sincerely” itself that has thus inflated. Indeed, almost any sincere expression of caring often feels empty. We routinely ask strangers “How are you?” when we don’t actually care how they are.

I felt this quite vividly when I was applying to GiveWell (alas, they decided not to hire me). I was trying to express how much I care about GiveWell’s mission to maximize the effectiveness of charity at saving lives, and it was quite hard to find the words. I kept find myself saying things that anyone could say, whether they really cared or not. Fighting global poverty is nothing less than my calling in life—but how could I say that without sounding obsequious or hyperbolic? Anyone can say that they care about global poverty—and if you asked them, hardly anyone would say that they don’t care at all about saving African children from malaria—but how many people actually give money to the Against Malaria Foundation?

Or think about how uncomfortable it can feel to tell a friend that you care about them. I’ve seen quite a few posts on social media that are sort of scattershot attempts at this: “I love you all!” Since that is obviously not true—you do not in fact love all 286 of your Facebook friends—it has plausible deniability. But you secretly hope that the ones you really do care about will see its truth.

Where is this ‘sincerity inflation’ coming from? It can’t really be from overuse of sincerity in ordinary conversation—the question is precisely why such conversation is so rare.

But there is a clear source of excessive sincerity, and it is all around us: Advertising.

Every product is the “best”. They will all “change your life”. You “need” every single one. Every corporation “supports family”. Every product will provide “better living”. The product could be a toothbrush or an automobile; the ads are never really about the product. They are about how the corporation will make your family happy.

Consider the following hilarious subversion by the Steak-umms Twitter account (which is a candle in the darkness of these sad times; they have lots of really great posts about Coronavirus and critical thinking).

Kevin Farzard (who I know almost nothing about, but gather he’s a comedian?) wrote this on Twitter: “I just want one brand to tell me that we are not in this together and their health is our lowest priority”

Steak-umms diligently responded: “Kevin we are not in this together and your health is our lowest priority”

Why is this amusing? Because every other corporation—whose executives surely care less about public health than whatever noble creature runs the Steak-umms Twitter feed—has been saying the opposite: “We are all in this together and your health is our highest priority.”

We are so inundated with this saccharine sincerity by advertisers that we learn to tune it out—we have to, or else we’d go crazy and/or bankrupt. But this has an unfortunate side effect: We tune out expressions of caring when they come from other human beings as well.

Therefore let us endeavor to change this, to express our feelings clearly and plainly to those around us, while continuing to shield ourselves from the bullshit of corporations. (I choose that word carefully: These aren’t lies, they’re bullshit. They aren’t false so much as they are utterly detached from truth.) Part of this means endeavoring to be accepting and supportive when others express their feelings to us, not retreating into the comfort of dismissal or sarcasm. Restoring the value of our sincerity will require a concerted effort from many people acting at once.

For this project to succeed, we must learn to make a sharp distinction between the institutions that are trying to extract profits from us and the people who have relationships with us. This is not to say that human beings cannot lie or be manipulative; of course they can. Trust is necessary for all human relationships, but there is such a thing as too much trust. There is a right amount to trust others you do not know, and it is neither complete distrust nor complete trust. Higher levels of trust must be earned.

But at least human beings are not systematically designed to be amoral and manipulative—which corporations are. A corporation exists to do one thing: Maximize profit for its shareholders. Whatever else a corporation is doing, it is in service of that one ultimate end. Corporations can do many good things; but they sort of do it by accident, along the way toward their goal of maximizing profit. And when those good things stop being profitable, they stop doing them. Keep these facts in mind, and you may have an easier time ignoring everything that corporations say without training yourself to tune out all expressions of sincerity.

Then, perhaps one day it won’t feel so uncomfortable to tell people that we care about them.

Terrible but not likely, likely but not terrible

May 17 JDN 2458985

The human brain is a remarkably awkward machine. It’s really quite bad at organizing data, relying on associations rather than formal categories.

It is particularly bad at negation. For instance, if I tell you that right now, no matter what, you must not think about a yellow submarine, the first thing you will do is think about a yellow submarine. (You may even get the Beatles song stuck in your head, especially now that I’ve mentioned it.) A computer would never make such a grievous error.

The human brain is also quite bad at separation. Daniel Dennett coined a word “deepity” for a particular kind of deep-sounding but ultimately trivial aphorism that seems to be quite common, which relies upon this feature of the brain. A deepity has at least two possible readings: On one reading, it is true, but utterly trivial. On another, it would be profound if true, but it simply isn’t true. But if you experience both at once, your brain is triggered for both “true” and “profound” and yields “profound truth”. The example he likes to use is “Love is just a word”. Well, yes, “love” is in fact just a word, but who cares? Yeah, words are words. But love, the underlying concept it describes, is not just a word—though if it were that would change a lot.

One thing I’ve come to realize about my own anxiety is that it involves a wide variety of different scenarios I imagine in my mind, and broadly speaking these can be sorted into two categories: Those that are likely but not terrible, and those that are terrible but not likely.

In the former category we have things like taking an extra year to finish my dissertation; the mean time to completion for a PhD is over 8 years, so finishing in 6 instead of 5 can hardly be considered catastrophic.

In the latter category we have things like dying from COVID-19. Yes, I’m a male with type A blood and asthma living in a high-risk county; but I’m also a young, healthy nonsmoker living under lockdown. Even without knowing the true fatality rate of the virus, my chances of actually dying from it are surely less than 1%.

But when both of those scenarios are running through my brain at the same time, the first triggers a reaction for “likely” and the second triggers a reaction for “terrible”, and I get this feeling that something terrible is actually likely to happen. And indeed if my probability of dying were as high as my probability of needing a 6th year to finish my PhD, that would be catastrophic.

I suppose it’s a bit strange that the opposite doesn’t happen: I never seem to get the improbability of dying attached to the mildness of needing an extra year. The confusion never seems to trigger “neither terrible nor likely”. Or perhaps it does, and my brain immediately disregards that as not worthy of consideration? It makes a certain sort of sense: An event that is neither probable nor severe doesn’t seem to merit much anxiety.

I suspect that many other people’s brains work the same way, eliding distinctions between different outcomes and ending up with a sort of maximal product of probability and severity.
The solution to this is not an easy one: It requires deliberate effort and extensive practice, and benefits greatly from formal training by a therapist. Counter-intuitively, you need to actually focus more on the scenarios that cause you anxiety, and accept the anxiety that such focus triggers in you. I find that it helps to actually write down the details of each scenario as vividly as possible, and review what I have written later. After doing this enough times, you can build up a greater separation in your mind, and more clearly categorize—this one is likely but not terrible, that one is terrible but not likely. It isn’t a cure, but it definitely helps me a great deal. Perhaps it could help you.

Motivation under trauma

May 3 JDN 2458971

Whenever I ask someone how they are doing lately, I get the same answer: “Pretty good, under the circumstances.” There seems to be a general sense that—at least among the sort of people I interact with regularly—that our own lives are still proceeding more or less normally, as we watch in horror the crises surrounding us. Nothing in particular is going wrong for us specifically. Everything is fine, except for the things that are wrong for everyone everywhere.

One thing that seems to be particularly difficult for a lot of us is the sense that we suddenly have so much time on our hands, but can’t find the motivation to actually use this time productively. So many hours of our lives were wasted on commuting or going to meetings or attending various events we didn’t really care much about but didn’t want to feel like we had missed out on. But now that we have these hours back, we can’t find the strength to use them well.

This is because we are now, as an entire society, experiencing a form of trauma. One of the most common long-term effects of post-traumatic stress disorder is a loss of motivation. Faced with suffering we have no power to control, we are made helpless by this traumatic experience; and this makes us learn to feel helpless in other domains.

There is a classic experiment about learned helplessness; like many old classic experiments, its ethics are a bit questionable. Though unlike many such experiments (glares at Zimbardo), its experimental rigor was ironclad. Dogs were divided into three groups. Group 1 was just a control, where the dogs were tied up for a while and then let go. Dogs in groups 2 and 3 were placed into a crate with a floor that could shock them. Dogs in group 2 had a lever they could press to make the shocks stop. Dogs in group 3 did not. (They actually gave the group 2 dogs control over the group 3 dogs to make the shock times exactly equal; but the dogs had no way to know that, so as far as they knew the shocks ended at random.)

Later, dogs from both groups were put into another crate, where they no longer had a lever to press, but they could jump over a barrier to a different part of the crate where the shocks wouldn’t happen. The dogs from group 2, who had previously had some control over their own pain, were able to quickly learn to do this. The dogs from group 3, who had previously felt pain apparently at random, had a very hard time learning this, if they could ever learn it at all. They’d just lay there and suffer the shocks, unable to bring themselves to even try to leap the barrier.

The group 3 dogs just knew there was nothing they could do. During their previous experience of the trauma, all their actions were futile, and so in this new trauma they were certain that their actions would remain futile. When nothing you do matters, the only sensible thing to do is nothing; and so they did. They had learned to be helpless.

I think for me, chronic migraines were my first crate. For years of my life there was basically nothing I could do to prevent myself from getting migraines—honestly the thing that would have helped most would have been to stop getting up for high school that started at 7:40 AM every morning. Eventually I found a good neurologist and got various treatments, as well as learned about various triggers and found ways to avoid most of them. (Let me know if you ever figure out a way to avoid stress.) My migraines are now far less frequent than they were when I was a teenager, though they are still far more frequent than I would prefer.

Yet, I think I still have not fully unlearned the helplessness that migraines taught me. Every time I get another migraine despite all the medications I’ve taken and all the triggers I’ve religiously avoided, this suffering beyond my control acts as another reminder of the ultimate caprice of the universe. There are so many things in our lives that we cannot control that it can be easy to lose sight of what we can.

This pandemic is a trauma that the whole world is now going through. And perhaps that unity of experience will ultimately save us—it will make us see the world and each other a little differently than we did before.

There are a few things you can do to reduce your own risk of getting or spreading the COVID-19 infection, like washing your hands regularly, avoiding social contact, and wearing masks when you go outside. And of course you should do these things. But the truth really is that there is very little any one of us can do to stop this global pandemic. We can watch the numbers tick up almost in real-time—as of this writing, 1 million cases and over 50,000 deaths in the US, 3 million cases and over 200,000 deaths worldwide—but there is very little we can do to change those numbers.

Sometimes we really are helpless. The challenge we face is not to let this genuine helplessness bleed over and make us feel helpless about other aspects of our lives. We are currently sitting in a crate with no lever, where the shocks will begin and end beyond our control. But the day will come when we are delivered to a new crate, and given the chance to leap over a barrier; we must find the strength to take that leap.

For now, I think we can forgive ourselves for getting less done than we might have hoped. We’re still not really out of that first crate.