Supply and Demand: The Scissors, Banana, and the Parrot

The time for introducing a fresh group of students to supply and demand is soon to be upon us. In a spirit of sympathy and support, here us some background on three images used by prominent economists about supply and demand: the scissors, the banana, and the parrot.

A common question for students, and a common error for teachers, is to discuss situations of supply and demand as if only one of these forces was relevant at a certain time. But of course, when it comes to considering the effects of a shift in demand on equilibrium price and quantity, the shape of the supply curve is highly relevant, because a shift in demand happens along a supply curve. Similarly, when it comes to considering the effects of a shift in supply on equilibrium price and quantity, the shape of the demand curve is highly relevant, because the shift in demand happens along a supply curve.

 
Alfred Marshall made this point elegantly in his 1890 Principles of Economics, where he famously analogized supply and demand to the two blades of a pair of scissors. He wrote:  

We might as reasonably dispute whether it is the upper or the under blade of a pair of scissors that cuts a piece of paper, as whether value is governed by utility or cost of production. It is true that when one blade is held still, and the cutting is effected by moving the other, we may say with careless brevity that the cutting is done by the second; but the statement is not strictly accurate, and is to be excused only so long as it claims to be merely a popular and not a strictly scientific account of what happens.

Another common question for students is to ask where prices come from in a market economy. I\’ve never found an explanation more crisp and straightforward than this one from Joan Robinson (from the 

original Introduction to her 1933 book The Economics of Imperfect Competition, as reprinted in the 1969 second issue, pp. 6-7): 

It is not easy to explain what the analysis of value is, without making it appear extremely mysterious and extremely foolish. The point may be put like this: You see two men, one of whom is giving a banana to the other, and is taking a penny from him. You ask, How is it that a banana costs a penny rather than any other sum? The most obvious line of attack on this question os to break it up into two fresh questions: How does it happen that the one many will take a penny for a banana? and: How does it happen that the other man will give a penny for a banana? In short, the natural thing is to divide up the problem under two heads: Supply and Demand.

Finally, there\’s the old joke(?) among economists that you can teach a parrot to be a perfectly good economist. All you need to do is train the bird to answer every question by saying: \”Supply and demand.\” Some version of this joke has appeared in a number of economics textbooks, and probably in even more lectures. Garson O’Toole took on the job of tracing the origins of this comment at his Quote Investigator website. He finds multiple cases of anonymous writers and book reviewers going back to 1850 referring to supply and demand as \”parrot words.\”
But perhaps the first case of the \”joke\” appearing in full in the work of a prominent economist is in Irving Fisher\’s 1907 classic The Rate of Interest: Its Nature, Determination and Relation to Economic Phenomena.   But in taking a closer look at Fisher\’s comment, he is actually using the \”joke\” as a way of warning against an overly facile reliance on supply and demand as a quick-and-dirty, thoughtless form of explanation. Fisher writes: 
If a modern business man is asked what determines the rate of interest, he may usually be expected to answer, \”the supply and demand of loanable money.\” But \”supply and demand\” is a phrase which has been too often forced into service to cover up difficult problems. Even economists have been prone to employ it to describe economic causation which they could not unravel. It was once wittily remarked of the early writers on economic problems: \”Catch a parrot and teach him to say `supply and demand,\’ and you have an excellent economist.\” Prices, wages, rent, interest, and profits were thought to be fully \”explained\” by this glib phrase. It is true that every ratio of exchange is due to the resultant of causes operating on the buyer and seller, and we may classify these as \”demand\” and \”supply.\” But this fact does not relieve us of the necessity of examining specifically the two sets of causes, including utility in its effect on demand, and cost in its effect on supply. Consequently, when we say that the rate of interest is due to the supply and demand of \”capital\” or of \”money\” or of \”loans,\” we are very far from having an adequate explanation. 

Christopher Walton, William Law, and the 296-Page Footnote

It has been my tradition at this blog to take a break from current events in late August. Instead, I offer a series of posts about academia, economics, and editing, focusing on comments or themes that caught my eye in the last year. I start with the story of the 296-page footnote, and how its author cried out for an editor. 

The story of our footnote begins with Christopher Walton (1809-1877), a businessman who made a fortune dealing in silk, jewelry, and goldsmithing, and his fascination with a Church of England priest named William Law (1686-1761). In turn, Law is perhaps best-known today for his 1728 book, A Serious Call to a Devout and Holy Life, and more generally for his arguments that religion should include be fully lived in action and spirit, including a deeply felt and mystical piety. 
In short, Christopher Walton thought that William Law was the bee\’s knees and the cat\’s pajamas. Walton wanted the world to know about Law. So he collected writings by and about Law, and tried to write them up in a privately printed book in 1854 under the unwieldy title: Notes and materials for an adequate biography of the celebrated divine and theosopher, William Law. Comprising an elucidation of the scope and contents of the writings of Jacob Böhme, and of his great commentator, Dionysius Andreas Freher, which is available via the magic of HathiTrust
For background on Walton, I draw upon the entry written by David L. Wykes for the Oxford dictionary of national biography : in association with the British Academy : from the earliest times to the year 2000 (edited by H C G Matthew and Brian Harrison, British Academy, 2004). Wykes describes Walton and his book in this way:   

Alexander Gordon, a personal acquaintance, described him as physically \’of large build and in character sententious but kindly, and absolutely destitute of humour\’ … He made a fortune as a jeweller and goldsmith, enabling him to pursue his interest in theosophy. He became particularly interested in the mystical works of William Law. … About 1845 he advertised for an assistant to help prepare a biography of Law, for which he acquired a great collection of antiquarian works in addition to his existing library. He began to print an Outline of the Qualifications . . . for the Biography of … Law in November 1847, finally completing it at Christmas 1853, but he circulated copies of the incomplete text before it was finished. To the completed work he added To the Christianity, the philosophy, the erudition, science and noble intelligence of the age. Notes and materials for . . . biography of… Law. Comprising an elucidation of … the writings of … Bohme, and of his great commentator . . . Freher; with a notice of the mystical divinity … of all ages of the world (1854). The 700-page work is disorderly beyond description, \’a chaotic mixture of the relevant and the irrelevant\’ (Hobhouse, 196), yet it contains much bibliographical and biographical information of value.

I\’ve only dipped in and out of Christopher Walton\’s book, but I can vouch that calling it \”disorderly beyond description\” is an understatement. Nonetheless, it also has the genuine charm of a heartfelt intellectual quest–albeit a quest that gets out of hand. For a sense of Walton\’s style of exposition in the book, here is the opening sentence of the \”Preface:\” 

The understanding of the Editor upon the subjects of recondite and practical knowledge introduced into this work, having been greatly enlarged and perfectionated during the several years he has been occupied over it, especially as he approached to its conclusion, when it was, that he first obtained a true and philosophic insight into the arcanum of \”Animal\” or \”Vital Magnetism,\” so denominated, with the magical wonders that lie couched in it, both as a science and an art; and without which apprehension, it must be affirmed, that neither the original revelations of Scripture as to their literal truth, nor the purely magic phenomena of Nature in any age, can be adequately understood, or rationally explained :—such being the case, the reader will please to observe, that those only of its statements ore to be regarded as the Editor\’s final determinations, which shall be found to be unmodified by subsequent remarks, either in the work itself, or in the \” Introduction to Theosophy,\” which immediately succeeded to it, or by the contents of the \”Corrigenda and Addenda \” prefixed to it, and immediately following the present introduction* or Preface. 

And so it goes, for 34 pages of \”Preface\” and 688 pages of text that follow, all in very small print. When I find myself smiling at Walton\’s book, I also find myself thinking: \”Oh yeah, buddy? What are the 700 pages of small print that you poured your own heart and soul into lately?\”

Round about p. 334, it must have been apparent to Walton that the flow of exposition wasn\’t going well. (And again, who among us hasn\’t reached p. 334 of a book we are writing and had these similar feelings?) So Walton starts a footnote with a cry for help–indeed, a cry for a perfect human being who is also AN EDITOR who will take Walton\’s hundreds of pages of notes and collected writings and put together the ideal biography of his ideal man, William Law. The key footnote starts like this (and I have put in boldface type a few of the traits of the EDITOR Walton is seeking): 
As a relief to the uniformity and matter of these pages, we present currently therewith, the following Notes and Memoranda, relating to the personal history, birth-place, family and friends of the subject of the proposed biography ; which, though belonging more appropriately to that work, may not be unacceptable to the readers of this preliminary treatise. 
And here we take occasion to say, in reference to the compilation and authorship of the Biography, that what is WANTED in short, as the sum and the object of the present treatise, and as necessary in the nature of the thing, is AN EDITOR, who, whilst proving himself an exact historian, a solid universal scholar, a just thinker, a profound philosopher, and a deeply-experienced, enlightened christian, shall produce a masterly picture, or biography of the individual, in all the features and developments of his mind and character ; interweaving the scanty incidents of his life that have been preserved, with such tender and manly reflections, and filling up the vacancies in his history with such elevated and charming natural conceptions and observations, and interspersing the whole with such dashes and reliefs of sublime instruction, though popularly expressed, as shall irresistibly inspire the reader with a fervent admiration of true wisdom and piety, and also fire him with an ardent and indomitable resolution, to immediately commence the pursuit of evangelic perfection, and the imitation of so perfect a model of a learned and accomplished English gentleman, philosopher and christian. The whole to be rendered as captivating, by the dignity and importance of the diversified subjects upon which it treats, in so uniformly felicitous and masterly a manner, as. by the condescending tenderness, nobility and wisdom of its sentiments, and the classic purity, elegance and sweeping rhetorical and strictly logical power of its composition : all which qualifications, a solid duly-constituted ordinary genius may engraft upon itself, by diligence and a close study of the models referred to, and through the directions and specifications interspersed throughout the present treatise. In a word, as none but a Law could design and execute a perfect biography of a man. a scholar, a philosopher, and a Christian ; so this treatise aims solely at creating another Law, possessed of all the talents of the former, with all the highest practical experiences, discoveries, and divine manifestations in the human nature, that have distinguished these last ages, superinduced thereupon ; for without a beau-ideal or model of a perfect man in all his characteristic features and particulars, how shall mankind be elevated to their proper redeemed perfection, how shall the Gospel produce its full results. To proceed.
The absolute earnestness and sincerity of Walton\’s hopefulness, coming in the midst of his colossal effort in putting together this manuscript of more than 700 pages, makes me a little teary. Biographies have been written of William Law, but no larger-than-life EDITOR answered Walton\’s call. 
Thus started the footnote then proceeds for 296 page, from p. 334 to p. 628. For the first 25 pages or so, the main text is the top half of each page, followed by a hairline, with the ongoing footnote covering the bottom half of the page. For the following 140 pages, the  mixture between text and footnote varies, with the text sometimes being the top third of the page, and sometimes taking most of the page. But about 160 pages into the footnote, the note takes over, and for over 100 pages, the \”main text\” is a couple of lines at the top of each page, with the rest devoted to the unending footnote. 
The story of the 296-page footnote has, at least to me, a happy ending. Wykes describes it this way in Walton\’s entry in the Oxford dictionary

Walton donated copies to every major public library in the world. Some other anonymous works relating to theosophy were probably written at Walton\’s suggestion and printed at his expense. He kept his Theosophian Library\’ at 8 Ludgate Hill, and made it freely available to those who shared his interests. In August 1876, at the suggestion of his friend Keningale Cook, Walton offered the books he had collected to Dr Williams\’s Library, stipulating that they should be kept apart as the \’Walton Theosophical Library\’, and always be available to those interested in the subject. His manuscripts relating to Law were included in the gift. Alexander Gordon was paid £20 to catalogue the books. The Walton collection now forms the best collection of books on seventeenth- and eighteenth-century mysticism in Britain.

The two versions of Walton\’s book that were copied and put up on the web by HathiTrust were from the University of Wisconsin and the University of Michigan. 
After learning of this story, I have sought to visualize what one might call the \”Christopher Walton editor,\” that paragon of humanity and editing skill with all the highest practical experiences, discoveries, and divine manifestations in the human nature,\” who would face that 296-page footnote with twinkling eyes and a modest smile–and edit something beautiful out of it. Christopher Walton never got the editor he had hoped for. But I suspect that in his own \”sententious\” and \”humorless\” way, he would be so very pleased that his collection of writings has remained a useful source of research, and that his book is available with a web connection to all the world. All it takes, after all, is one person to love the life and work of William Law as much as Christopher Walton did, and to take up the task of editor. 

Cities as Economic Engines: Is Lower Density in our Future?

 When studying urban economics, a common starting point is to make the obvious observation that economic activity does not have an even geographic distribution. Instead, it  clusters in metropolitan areas. To put it another way, even with higher real estate costs, higher wages and prices, traffic congestion, pollution, and crime, businesses find it economically worthwhile to locate in urban areas. For such firms, there must be some offsetting productivity benefit. The Summer 2020 issue of the Journal of Economic Perspectives includes a four paper symposium on \”Productivity Advantages of Cities.\” 

Here, I\’ll focus on a few themes of the first two papers. But it\’s impossible to read the papers in late-summer 2020 without wondering if the productivity advantages of cities that they describe are going to be a casualty of public concerns over the pandemic and rising crime in urban areas. 

Duranton and Puga provide an overview of the research on economic gains from population density. They write: 

While urban economists broadly agree on the magnitude of the productivity benefits of density, the evidence distinguishing between possible sources is less solid. Duranton and Puga (2004) classify the mechanisms into three broad classes. First, a larger market allows for a more efficient sharing of local infrastructure, a variety of intermediate input suppliers, or a pool of workers. Second, a larger market also allows for better matching between employers and employees, or buyers and suppliers. Finally, a larger market can also facilitate learning, by facilitating the transmission and accumulation of skills or by promoting the development and adoption of new technologies and business practices.

They also point out that these \”agglomeration economies\” at some point can be outweighed by costs of density like land prices, housing prices, transport costs, and congestion. 

Theory has long hypothesized that as population and density increase in a city, its benefits initially accumulate faster, but eventually, its costs dominate (Henderson 1974). Fujita and Thisse (2013) call this the “fundamental trade-off of spatial economics,” because it explains both the existence of cities and their finite sizes. However, compared to research on benefits of density, there is a paucity of research on its costs, which Glaeser (2011) dubbed the “demons of density.”

Rosenthal and Strange dig into some of the productivity advantages of density, and discuss the evidence that the gains of economic activity by being grouped together operate at many levels. For example, if one looks at a certain region, it\’s not just that economic activity is concentrated within the metropolitan area. In addition, there will be pockets of increased density within the overall urban area, like the Wall Street area in New York. And even within those pockets of higher density, there will be geographic concentrations of specific kinds of firms–like financial or legal or advertising firms–that are bunched within a few blocks or even on different floors in the same buildings.  At all of these different levels, there is evidence of productivity gains from being grouped together. They write: 

Agglomeration effects operate at various levels of spatial aggregation, including regional, metropolitan, and neighborhood scales. In fact, there is also evidence that agglomeration effects operate below the neighborhood level, including within buildings and organizations. Although agglomeration effects can extend over broad distances, they also attenuate, with nearby activity exerting the strongest effect on productivity.

For the last few months, the density of workplaces in the US economy has dramatically diminished.  Whatever the productivity benefits of density in cities, concerns over the pandemic and then also about prevalence of crime and rioting are causing firms and workers to push back  against density. Moreover, the different dimensions of density can reinforce each other: that is, a city with lots of residents and workers will also tend to have more option for shopping, restaurants, and entertainment, which in turn makes the city a more attractive destination for residents and workers. Conversely, when many workers shift to telecommuting and many residents are trying to stay home to the extent possible, the firms that offer shopping, restaurants, and entertainment are going to suffer.  

Duranton and Puga point out that over time, these kinds of changes and cycles in urban areas are not new. They write: 

[W]hat will be the long-run consequences of this virus for our densest cities? Pandemics have hit cities the hardest for centuries, and cities have adapted and been shaped by them—from investments in water and sewage systems to prevent cholera, to urban planning to reduce overcrowding and improve air circulation and access to sunlight in response to tuberculosis. Maybe temporary social distancing measures will also leave a permanent footprint on cities—for instance, in the form of more space for pedestrians and bicycles or a gain of outdoor versus indoor leisure environments. But the idea that this pandemic will change cities forever is likely an overstretch. Cities are full of inertia and this crisis has stressed both the costs and benefits of density. Confinement is forcing us to see both the advantages and the great limitations of online meetings relative to the more subtle and unplanned in-person exchanges. It has  made us realize that many tasks are impossible to do from home. At schools and universities, the haphazard transition to online courses may speed up their development, or it may delay it as many students have become frustrated by losing aspects of a full educational experience. For a while, some people may try to avoid dense cities for fear of contagion, but others may be drawn to them seeking work opportunities in difficult times.

They may turn out to be correct, especially if a trustworthy and effective coronavirus vaccine emerges. Maybe I\’ve just got a case of the 2020 blues today, but I find that I\’m less optimistic. It feels to me as if our society is jamming a huge amount of change into a short time. Just in economic terms, for example, there have been conversations about the future of telecommuting, telemedicine, and distance learning for decades–but now we are leaping ahead with all of these changes all at once. I don\’t expect metropolitan areas to go away, but there may be substantial shifts within a given urban area in what characteristics people desire in their living arrangements, the location of where they work, the mixture of jobs and industries, how people relax, and how often they are willing to mix with a crowd. At present many households and firms are still cobbling together temporary arrangements–or at least arrangements that are intended to be temporary. But some proportion of that temporary is likely to have long-lasting and disruptive effects. 

Sounding an Alarm on Data Mining

Back in 2013, someone going under the nom de plume of Economist Hulk wrote on Twitter (all caps, natch): \”WHEN FACTS CHANGE, HULK SMASH FACTS UNTIL THEY FIT HIS PRE-CONCEIVED THEORY. HULK CALL THIS ‘ECONOMETRICS’.\”

Gordon Tullock offered an aphorism in a similar spirit, which he attributed to verbal comments from Ronald Coase (\”A Comment on Daniel Klein\’s \”A Plea to Economists Who Favor Liberty,\”  Eastern Economic Journal , Spring, 2001, 27: 2, pp. 203- 207). Tullock wrote: \”As Ronald Coase says, `if you torture the data long enough it will confess.\’\”
Ronald Coase (Nobel 1991) put the point just a little differently in a 1981 lecture, \”How Should Economists Choose?\”, while attributing a similar point to Thomas Kuhn. Coase wrote (footnotes omitted): 

In a talk I gave at the University of Virginia in the early 1960s … I said that if you torture the data enough, nature will always confess, a saying which, in a somewhat altered form, has taken its place in the statistical literature. Kuhn puts the point more elegantly and makes the process sound more like a seduction: \”nature undoubtedly responds to the theoretical predispositions with which she is approached by the measuring scientist.\” I observed that a failure to get an exact fit between the theory and the quantitative results is not generally treated as calling for the abandonment of the theory but the discrepancies are put on one side as something calling for further study. Kuhn says this: \”Isolated discrepancies … occur so regularly that no scientist could bring his research problems to an end if he paused for many of them. In any case, experience has repeatedly shown that in overwhelming proportion, these discrepancies disappear upon closer scrutiny .\” Because of this, Kuhn argues that \”the efficient procedure \” is to ignore them, a conclusion economists will find it easy to accept.

Gary Smith offers an overview of what these issues are all about in \”Data Mining Fool\’s Gold\” (Journal of Information Technology, posted \”Online First\” as a forthcoming paper on May 22, 2020, subscription needed for access). Smith offers what he calls \”the paradox of big data\”:
It is tempting to believe that patterns are unusual and their discovery meaningful; in large data sets, patterns are inevitable and generally meaningless. … Data-mining algorithms—often operating under the label artificial intelligence—are now widely used to discover statistical patterns. However, in large data sets streaks, clusters, correlations, and other patterns are the norm, not the exception. While data mining might discover a useful relationship, the number of possible patterns that can be spotted relative to the number that are genuinely useful has grown exponentially—which means that the chances that a discovered pattern is useful is rapidly approaching zero. This is the paradox of big data:
It would seem that having data for a large number of variables will help us find more reliable patterns; however, the more variables we consider, the less likely it is that what we find will be useful.
Along with useful background discussion, Smith offers some vivid examples of data mining gone astray. For instance, Smith put a data-mining algorithm to work on President Donald Trump\’s tweets in the first three years of his term. He found: 

It turned out that the S&P 500 index of stock prices is predicted to be 97 points higher 2 days after a one-standard-deviation increase in the Trump’s use of the word president, … my 10-fold cross-validation data-mining algorithm discovered that the low temperature in Moscow is predicted to be 3.30°F higher 4 days after a one-standard-deviation increase in Trump’s use of the word ever, and that the low temperature in Pyongyang is predicted to be 4.65°F lower 5 days after a one-standard-deviation increase in the use of the word wall. … I considered the proverbial price of tea in China. I could not find daily data on tea prices in China, so I used the daily stock prices of Urban Tea, a tea product distributer headquartered in Changsha City, Hunan Province, China, with retail stores in Changsha and Shaoyang that sell tea and tea-based beverages. The data-mining algorithm found that Urban Tea’s stock price is predicted to fall 4 days after Trump used the word with more frequently.

Indeed, Smith created a random variable, put it into the date-mining algorithm, and found: 

The data-mining algorithm found that a one-standard deviation increase in Trump’s use of the word democrat had a strong positive correlation with the value of this random variable 5 days later. The intended lessons are how easy it is for data-mining algorithms to find transitory patterns and how tempting it is to think up explanations after the fact. … That is the nature of the beast we call data mining: seek and ye shall find.

At this point, a common response is something like: \”Well, of course it\’s possible to find correlations that don\’t mean anything. What about the correlations that do mean something?\” The response is true enough in some sense: perhaps in looking at a bucket-full of correlations, one of them may suggest a theory that can be tested in various ways to see if it has lasting power. But also notice that when you start talking making judgments that certain statistical findings \”mean something\” and other statistical findings derived in exactly the same way do not \”mean something,\” you aren\’t actually doing statistics any more. The data isn\’t telling you the answer: you are deciding on other grounds what the answers are likely to be. 

Smith\’s example about Trump\’s tweets is not just a hypothetical example. Several studies have tried to find whether Trump\’s tweets–or a selection of Google search terms, or other big data sets–would cause the stock market to rise or fall in some systematic way. 

Two other examples from Smith involve investment companies called Equabot and Voleon. They were going to run their investments with data-mining tools, not human intuition. But a funny thing happened on the way to the profits: returns at both companies somewhat underperformed the S&P 500. One plausible explanation is that the data-mining algorithms kept finding correlations that weren\’t lasting or valid, so when the algorithm made investments on the basis of those correlations, it was either investing mostly randomly, like an index fund, or even counterproductively. 

Or you may remember that about 10 years back, Google Flu Trends sought to predict flu outbreaks by looking at Google searches. The early evidence looked promising: \”However, after issuing its report, Google Flu Trends over-estimated the number of flu cases for 100 of the next 108 weeks, by an average of nearly 100% (Lazer et al., 2014). Google Flu Trends no longer makes flu predictions.\” As another example, a British insurance company in 2016 decided to base its car insurance rates on data mined from Facebook posts, and found that it was charging different rates according to whether people were more likely to mention Michael Jordan or Leonard Cohen, a process \”which humans would recognize as ripe with errors and biases.\” 

I do not mean to imply that using algorithms to search systematically through big data are not useful,  just that their results must be interpreted with care and precision. Otherwise, ridiculousness can result. For some discussions of how big data and machine learning can be used in economics, starting points from the journal that I edit include: 

The Rise of the US Social Insurance State

One sometimes hears a reference to \”the rise of the welfare state,\” but for the US in the 20th century, the pattern is more accurately described as a \”rise of the social insurance state\”–which isn\’t quite the same thing. Price V. Fishback uses this insight as a stepping-stone in \”Social Insurance and Public Assistance in the Twentieth-Century United States,\” which he delivered as the Presidential Address to the Economic History Association last year, and now has been published in the Journal of Economic History (June 2020,  80:2, pp. 311-350).  Fishback\’s argument unfolds in three parts: 

First, what many people call the rise in the welfare state in the United States might better be described as the “Rise of the Social Insurance State.”  “Rise of the Welfare State” could be seen as a rise in public assistance  transfers to the poor. Yet, such transfers have not been very large relative to gross domestic product (GDP). A very large share of the expansion in social welfare spending has come from social insurance that covers people throughout the income distribution. The social insurance programs were much easier to sell politically and carried much less social stigma because participants felt that they or their employers had paid for their benefits up front.

Here\’s an illustrative figure. Social insurance spending in the 20th century goes from 1% of GDP to 10%. If one includes the rise in private health insurance and private pensions–both encouraged by the government through favorable tax treatment and friendly regulations–the rise of the social insurance state would look even bigger.  In contrast, work relief (during the Great Depression) and public assistance programs or the poor (in the later 20th century) are much smaller.  

Fishback\’s second point suggests that total US spending on \”social welfare\” categories is similar to that of European countries–the US just has more of this spending happening in the private sector. He writes: 

Second, America is commonly perceived to spend much less on social welfare than many European countries. This perception arises because most comparisons focus on gross public social welfare spending. In fact, after taking into account taxation, public mandates, and private spending, the United States in the late twentieth century spent a higher share on combined private and net public social welfare relative to GDP than did most advanced economies. Americans just did it differently because the governments operated a safety net system that relied to a much greater extent on private insurance and pensions and taxed lower income people less heavily.

Fishback points out in in OECD statistics, \”government social welfare expenditures include old-age pensions, survivor benefits (not from private life insurance), incapacity-related aid, health expenditures, aid to families, unemployment benefits, income maintenance, government job training, and housing subsidies.\” But in the US, of course, a large share of health expenditures and old-age pensions happens through the private sector. When you include both public and private spending in these areas, total US spending looks much more similar to countries like Germany, Sweden, and France. 

This point is well-taken. But of course, it\’s also true that if total US spending on these social welfare categories is close to these other countries, if lagging a bit behind, and it\’s also true that US spending on health care as share of GDP is much bigger than these other countries, then it follows that US social welfare spending in the non-health categories must be lagging well behind. 

Finally, Fishback sets out to disentangle some of the complexity that arises because many US social welfare programs are run through the states, which can have different rule. Fishback writes: 

Third, the United States has 51 or more social welfare systems that set target minimum benefits for public assistance for the elderly and families in poverty and maximum benefits for social insurance programs. … [T]he spending is designed to fill the gap between a target level of living and the household’s resources. … I develop estimates of these target benefits for each state for several years in the twentieth century and compare them to several measures of economic well-being: the national poverty line before and after adjusting for cross-sectional differences in cost of living, state  manufacturing earnings and benefits, state per capita incomes, and per capita GDPs in countries around the world. The comparisons show that the public assistance targets have been more generous to the elderly poor than to poor families. 

In broad terms, Social Security and Medicare–both of which can be categorized as \”social insurance\” programs–have been a substantial boost for the elderly poor. However, the social insurance and public assistance programs aimed at families–for example, cash welfare payments, unemployment insurance,  and workman\’s compensation–pay much less in the United States than in many other countries. 

Randomness in the Location of Tech Clusters

It seems as if every city wants to be the center of a tech cluster. When Amazon announced a couple of years back that it was taking bids from cities for the location of a second headquarters, 238 cities applied. But other than hoping Amazon falls into your lap, how could a city create conditions for a tech cluster to emerge? And given that most cities are not going to be come tech clusters, what\’s the next-best strategy? William R. Kerr and Frederic Robert-Nicoud discuss what we know about this subject  \”Tech Clusters\” (Journal of Economic Perspectives, Summer 2020, 34:3. pp. 50-76). 

What are the current US tech clusters? Looking at a variety of definitions, they write: 

Six cities appear to qualify under any aggregation scheme: San Francisco, Boston, Seattle, San Diego, Denver, and Austin all rank among top 15 locations for venture capital and for patents (scale) and hold shares for venture capital, patents, employment in R&D-intensive sectors, and employment in digital-connected occupations that exceed their population shares (density). They also pass our highly rigorous “sniff test”—that is, they just make sense. Washington, Minneapolis-St. Paul, and Raleigh-Durham would join the list if relaxing the expectation that the share of venture investment exceed population share (which is hard due to the very high concentration in San Francisco). New York and Los Angeles are more ambiguous: they hold large venture capital markets (and venture investors frequently declare them leading tech clusters), but their patents and employment shares in key industries and fields are somewhat less than their population shares. Were we to disaggregate these huge metro areas, we would likely identify a sub-region that would independently fit on this short list by still holding sufficient scale and yet having achieved a more recognizable density. Said differently, there is surely a part of New York and Los Angeles that would be stand-alone equal to or greater than Austin (for example, Egan et al. 2017). Chicago’s activity is mostly equal to its population share or less.

What about trying to create a new tech cluster? In looking at why current tech clusters are located where they are, it becomes clear that there is often a large amount of chance involved–that is, decisions made by a few key actors at key moments, sometimes for family reasons, about where an \”anchor firm\” for the tech cluster would end up being located. For example, before Bill Gates and Paul Allen moved to Seattle, they were living in Albuquerque where Microsoft was based. Kerr and Robert-Nicoud write:  

In most accounts of the origin of tech clusters, such as Klepper’s (2010, 2016) comparisons of Detroit and Silicon Valley, emphasis is given to the initial placement of a few important firms and the spinoff companies they subsequently generate. This outsized influence for anchor firms generates ample room for random influences on the early location decisions vital to a future cluster. For example, William Shockley, who shared a Nobel Prize in Physics for his work on semiconductors and transistors, moved to the San Francisco area to be near his ailing mother. Later, the spinoffs from his firm Shockley Semiconductors included Intel and AMD. Similarly, Moretti (2012) describes how personal factors led Bill Gates and Paul Allen to move Microsoft from Albuquerque to Seattle, their hometown. At the time, Albuquerque was considered the better place to live, it was favored by most of Microsoft’s early employees and the location of many early clients. Yet proximity to family won out, and this decision has reverberated well beyond Microsoft’s direct employment. The agglomeration advantages sparked by Microsoft have attracted countless other tech firms to Seattle, including Jeff Bezos relocating from New York City to Seattle when he founded Amazon. Had Gates and Allen not moved home to Seattle, Albuquerque might be home to two of America’s three most valued companies in 2020.
A similar and related randomness arises due to the often-serendipitous nature of breakthrough discoveries and their outsized subsequent importance. Zucker, Darby, and Brewer (1998) show that the location of biotech industry follows the positioning of star scientists in the nascent field, and the surging prominence of Toronto for artificial intelligence harkens back to the choice of some key early researchers to locate there, well before the field became so prominent. Duranton (2007) formalizes how
random breakthroughs could lead to shifts in the leadership of cities for a tech field
or industry, such as the migration of semiconductors from Boston to Silicon Valley. …
It seems important to note that the process of becoming a tech center doesn\’t necessarily favor the biggest cities. Indeed, mid-sized cities may have certain advantages. 

While random sparks play a role, the same breakthroughs often occur contemporaneously in two or more locations (Ganguli, Lin, and Reynolds 2019). Accordingly, a new line of work considers the factors that shape which location emerges as the winner. Duran and Nanda (2019), for example, study the widespread experimentation during the late 1890s and early 1900s as local automobile assemblers learned about the fit between this emerging industry and their city. Despite having fewer entrants initially, activity coalesced in smaller cities—Cleveland, Indianapolis, St. Louis, and Detroit—with Detroit being the ultimate winner by the late 1920s. The smaller city advantage may have been due to the higher physical proximity of relevant stakeholders, allowing for easier experimentation, prototyping, and circulation of ideas. So long as smaller cities had sufficient local input supplies, they may have provided more attention and financial support to the new technology compared to larger markets and fostered relational contracts.

But other than hoping that a rising mega-entrepreneur has relatives already living in your town, what steps might a city that wants to be a tech center realistically take? The perhaps uncomfortable but realistic advice is to lower your expectations. If there was a cookbook recipe for becoming a tech center, it would be happening in a lot more places. However, cities can try to be what the research calls a \”nursery city,\” friendly to innovation and tech companies, build up your local university, and also work to create ties to several different tech centers. Kerr and Robert-Nicoud write: 

The unique origin of each existing tech cluster suggests future efforts to seed from scratch are likely to be similarly frustrating. Instead, a better return is likely to come from efforts to reduce the local costs of experimentation with ideas (Kerr, Nanda, and Rhodes-Kropf 2014), alongside the provision of a good quality of life. There is also likely a role for cities that have developed a position in an emerging sector, even if by random accident due to family ties, to increase the odds they are favored in the shakeout process. Such support is more likely to work if it is broad-based to a sector and avoids attempting to “pick winners” by targeting individual companies. Other cities can take the strategy of increasing
their connectivity to leading centers via remote work. Tulsa Remote pays qualified workers with remote jobs $10,000 to move to Tulsa, Oklahoma, and similar programs are popping up elsewhere. Rather than seeking to “become the next Silicon Valley,” these efforts focus on connecting with the existing hotspots and being an attractive alternative with a lower cost of living.

Beyond anchor firms, universities also feature prominently in the history of tech clusters, both for the United States and globally (Markusen 1996; Dittmar and Meisenzahl 2020). Under the guidance of Fred Terman, Stanford University fostered a strong relationship with the growing tech community, such as the 1948 creation of the Stanford Industrial Park that would house 11,000 workers from leading tech firms by the 1960s. … Hausman (2012) documents how university innovation fosters local industry growth, and these spillovers can attenuate rapidly (see also Andersson, Quigley, and Wilhelmsson 2009; Kantor and Whalley 2014). … These historical examples are starting to provide insight that will advance our theory on tech clusters. Duranton and Puga (2001) model a system of cities
in which new industries are emerging in large and diverse “nursery” cities.

Perhaps the true challenge for local policymakers is that having your local economy participate in tech cluster jobs and growth is not just a matter of attracting branches of a couple of big firms, and it\’s not about being hyperspecialized in producing a particular product. Instead, it\’s about providing an economic environment conducive to breeding a stream of innovations and providing an economic and regulatory context where those can be turned into local companies.

The Most Romantic Sentences Ever About Government Statistics?

Arunabh Ghosh offers a candidate for the role of most romantic sentence ever about statistical tables. It\’s at the start of his article \”Counting China: By rejecting sampling in favour of exhaustive enumeration, communist China’s dream of total information became a nightmare\” (Aeon, July 23, 2020). Try not to get teary as you read it. Ghosh writes: 

Sometime in the fall of 1955, a Chinese statistical worker by the name of Feng Jixi penned what might well be the most romantic sentence ever written about statistical work. ‘Every time I complete a statistical table,’ Feng wrote:

my happiness is like that of a peasant on his field catching sight of a golden ear of wheat, my excitement like that of a steelworker observing molten steel emerging from a Martin furnace, [and] my elation like that of an artist completing a beautiful painting.

A Martin furnace is a kind of open-hearth furnace where impurities like carbon are burnt out of pig iron as part of the process of producing steel. If you can suggest a more poetic example of love for government statistics, please pass it along. While not quite as emotionally resonant, Ghosh offers a second candidate for romantic comments about economic planning statistics: 

These numbers, the poet and author Ba Jin gushed:

gather the sentiments of 600 million people, they also embody their common aspirations and are their signpost [pointing to the future]. With them, correctly, step by step, we shall arrive on the road to [building a] socialist society. They are like a bright lamp, illuminating the hearts of 600 million.

The numbers about which Ba waxed poetic were those related to planning and economic management.

But love for statistics is not always requited, and Ghosh offers some useful reflections based in China\’s evolving experience with government statistics in the 1950s and 1960s. In the 1950s, for example, China\’s government statistics rejected the idea of probability or randomness. After all, the government would use statistics to announce an economic plan, and the plan would be achieved. No uncertainty!

In a speech in 1951, Li Fuchun, one of a handful of technocratically minded leaders, summarily dismissed the utility of Nationalist-era statistics, branding them an Anglo-American bourgeois conceit, unsuitable for ‘managing and supervising the country’. New China needed a new kind of statistics, he declared. …. With their sights set and rightful purpose claimed, Chinese statisticians proceeded to interpret Marxism’s explicit teleology as grounds to reject the existence of chance and probability in the social world. In their eyes, there was nothing uncertain about mankind’s march towards socialism and, eventually, communism. What role, then, could probability or randomness play in the study of social affairs?

The implications for statistical methods were profound. In rejecting probability, and the larger area of mathematical statistics within which it belonged, China’s statisticians discarded a large array of techniques, none more critical than the era’s newest and most exciting fact-generating technology – large-scale random sampling. Instead, they decided that the only correct way to ascertain social facts was to count them exhaustively. Only in this way could extensive, complete and objective knowledge be generated. 

Enormous efforts were made to manage China\’s economy in a direct way with the use of government statistics. The economy was divided into 13 sectors. Information was to flow from villages to counties to provincial bureaus to the national government, and then plans would flow back down through this hierarchy. In the mid-1950s, more than 200,000 statistics workers were spread across 750,000 villages and 2,200 counties. Unsurprisingly, this system was plagued by slow turnaround times, estimates from lower levels that couldn\’t be reconciled at higher levels, problems where those lower in the pyramid did not to produce as they were told, and over-optimistic numbers. 

[T]he dream of total information, so alluring as an ideal, was a nightmare in practice. Every level of the statistical system contributed to the overproduction of data. In a system that valued the production of material goods above all else, the only way a white-collar service such as statistics could draw attention to itself was by claiming, as Feng did, that statistical tables were a material contribution to the economy, just like wheat and steel. With the production of tables so incentivised, the entire system responded with gusto to produce them. Soon, there were so many reports circulating that it was impossible to keep track of them. Internal memoranda bemoaned the chaos, but it was a pithy four-character phrase that truly captured the exasperation. Translated, it reads: ‘Useless at the moment of creation!’

By the late 1950s, both statistical sampling and exhaustive enumeration were out of favor. Instead, government \”statistics were to rely on case studies.

In the world of data collection, the Great Leap Forward marked a turn away from exhaustive enumeration and the adoption, instead, of decentralised and ethnographic methods. A tract from 1927 on rural investigation, authored by Mao, became the new methodological model for data collection. True knowledge could be gained only by a detailed, in-person investigation, not through vast exhaustive surveys nor through randomised sampling. The shift left the statistical apparatus with no reliable means to check its own data.

In the right hands, case studies can be an extraordinarily valuable approach in the social sciences, because you can drill down into the kinds of specific details that can be of key importance, but are often not covered in broad cross-industry or cross-economy statistical work. In the wrong hands, case studies is a polite name for propaganda. 

After this reflection on alternative visions of statistics, Ghosh offers an epigram for those who produce and use statistics: \”[A]ll data are biased, but … not all biases are the same …\”

COVID: Perception and Numeracy

Kekst CNC is a \”communications strategy\” firm that has been doing a series of public opinion polls about COVID-19. Their \”COVID-19 Opinion Tracker: Fourth Edition\” report surveyed 1,000 adults in different countries, during the time period from July 10-15, 2020. The report includes questions about beliefs on how long the virus will last, how local and national government is performing, how business is performing, feelings about returning to work and wearing masks, and other topics. 

One result that caught my eye was that they asked people to estimate the number of people in their country who had contracted COVID-19 and the number who had died from it. Here are the results for the UK (blue) the US (yellow), Germany (pink), Sweden (green) and France (purple). 

The first question about the number of COVID cases is ambiguous. Kekst CNC points out that the estimates are far higher than the number of confirmed cases according to national health authorities: for example, public perception is 20 times as high as confirmed cases in the US, and 46 times as high as confirmed cases in Germany and France. The estimates of cases is probably high, but one might argue that testing has been limited, and the number of actual cases is likely to be well above the \”confirmed\” number. But there\’s less wiggle room and more exaggeration for the question about number of COVID deaths. Here the estimates are at least 100 times as high as the number of actual deaths in these five countries, with public opinion 225 times too  high in the US  and 300 times too high in Germany. 

When you point out that people are wildly wrong in their numerical estimates of cases and especially deaths from COVID-19, a  common answer is something like \”you are just trying to minimize the problem.\” Look, I understand that in this 21st-century, social-media world, there is a widespread view that if you aren\’t exaggerating, you aren\’t serious. I hope you will forgive me for being lousy at that form of \”persuasion.\” (Also, if people estimated that half or two-thirds of the total population had already died, would that make them even more \”serious\”?)

My own belief is that this poll confirms that many people are innumerate about basic percentages. In the US, for example, saying that 9% of the existing population has already died of COVID-19 would imply roughly 29 million deaths so far. It would imply that one out of every eleven people in the country had already died from COVID-19 in mid-July. Given that in a big country, some places will inevitably have more severe outbreaks than others, there would need to be news stories where, say, half the people in certain neighborhood or small towns had died of COVID-19. 

It\’s hard to get revved up about innumeracy. But if you have little feel for what kinds of numbers are even plausible in a given situation, you are more likely to end up as a sucker for a glib talker with a scenario to sell.  

When Colleges Went to Online Instruction: Some Patterns

Almost all the US colleges and universities closed down their in-person operations and went to online education in March and April. But they didn\’t all do so at exactly the same time, nor do they all have the same plans for this fall.  In \”Tracking Campus Responses to the COVID-19 Pandemic,\” Christopher R. Marsicano, Kathleen M. Felten, Luis S. Toledo, and Madeline M. Buitendorp (Davidson College Educational Studies Working Paper No. 1,  April 2020) look at the shutdown pattern. They have also set up a College Crisis Initiative website that, among other tasks, keeps track of plans of colleges and universities for next fall. 

These maps from the working paper show the transition to online instruction in March 2020. Yellow dots mean that a college announced a transition to online instruction; purple dots mean the transition was implemented. 

What\’s the overall pattern? From the abstract: \”Between March 1st and April 4th, over 1,400 colleges and universities closed their doors and transitioned to online instruction. This paper uses a novel dataset and draws upon theories of institutional isomorphism to descriptively examine the trends in how higher education institutions responded to the Coronavirus pandemic. It finds little difference in institutional response based on campus infrastructure including, residence hall capacity, hospital affiliation, and medical degree offerings. There is some suggestive evidence, however, that institutions may have responded to external coercive isomorphic pressures from state governments and may have relied on a heuristic of peer institution closures to inform their decisions.\” 

In other words, the timing by which schools moved to online education seems to have been affected in some cases by pressure from state government or because they were following peer institutions. One suspects that similar dynamics–a mixture of evidence, political pressure, and following the crowd–will govern the move back from online education as well.  Here are the current plans for fall, updated through Friday, August 7, at the \”Dashboard\” website of the College Crisis Initiative

What\’s especially interesting to me is how different the plans are across institutions, even though students have actually started moving into dormitories at some institutions and are packing their bags for a move in the next week or two at other places. California is primarily green–that is, primarily or fully online. The middle of the country has a lot of pink, for primarily or fully in person. I\’m sure colleges and universities are watching each other closely, both for whether there will be a widespread on-campus outbreak of COVID-19, and also the willingness of students to deal with in-person restrictions or online learning. 

The World Economy Through a PPP Lens

When comparing the size of economies, you need to use an exchange rate to convert the GDP of one country, measured in its own currency, to compare with the GDP of the other country. But what exchange rate to use? 

An obvious choice is the market exchange rate. An equally obvious problem is that the exchange rates fluctuate.  If the exchange rate of a country strengthens by 10% in a certain month, that of course doesn\’t mean that the standard of living for people in that country rose by 10% during that month. When you compare economies using a market exchange rate, it\’s a little like measuring with a ruler which without warning grows and shrinks. 
There\’s also a more subtle problem with using market exchange rates to compare economies. Say that in one country, housing or health care or higher education is much cheaper than in some other country. As a concrete example, one often reads about moving or retiring to some other country with a much lower cost of living–where you can buy that lovely home for so much less than in the United States. If you just look at total size of GDP for that country with lower prices, converted at a market exchange rate, you would not be able to tell that because of the lower price level, the GDP actually represents a higher quantity of goods being consumed than you might expect. 
A common alternative is \”purchasing power parity\” exchange rates. Like the name implies, these exchange rates are calculated to reflect the purchasing power of a currency within the country. This is not a new idea: the first academic research using a PPP exchange rate was published in 1940. But in recent decades, the standard source for calculating PPP exchange rates is the International Comparison Project at the World Bank, which has published its report \”Purchasing Power Parities and the Size of World Economies: Results from the 2017 International Comparison Program\” (May 2020)

It is typically true that lower-income countries have cheaper goods and services. Thus,  when you look at the purchasing power of their currency in their own country, it tends to be greater than the market exchange rate would imply. Here\’s a figure showing price levels across countries. The horizontal axis shows per capita GDP, so higher-income countries are on the right. The vertical axis shows the price level across countries. The figure helps explain why places like Mexico and Thailand are such popular tourist and even retirement destinations for people from high-income countries: your income buys more in those countries. 

Essentially, a PPP exchange rate adjusts for these differences in buying power. As a result, using PPP exchange rates makes GDP for lower- and middle-income countries. As the report notes: \”In 2017, global output, when measured by purchasing power parities (PPPs), was $119,547 billion, compared with $79,715 billion, when measured by market exchange rates…. In 2017 lower-middle-income

economies contributed around 16 percent to PPP-based global GDP, while upper-middle-income
economies contributed 34 percent. At the same time, high-income economies contributed 49 percent.
In terms of market exchange rates, these shares were 8 percent, 28 percent, and 64 percent,
respectively.\” 

If the comparison is done using a PPP exchange rate, the economy of China was bigger than that of the United States in 2017. 
Here\’s an overview of the global economy measured in per capita GDP. The vertical axis measures the share of global population–and population is shown by the height of each bar. Thus, the bars for China, India, and the United States are especially tall. The horizontal width of the bar measures per capita GDP, using a PPP comparison. The light-blue lines on the left show 2011; the darker-blue bars on the right show 2017. You can see a number of modest changes in this six-year period: the rise of China and India, Turkey moving ahead of the Russian Federation in per capita GDP, and others.  
Two additional thoughts are worth passing along. First, the PPP exchange rates are explicitly intended to compare GDP, per capita GDP, and similar measures across countries, to make that useful adjustment for different price levels across countries. For other uses, they may not be appropriate. The report notes: \”ICP PPPs are designed specifically for international comparisons of GDP. They are not designed
for comparisons of monetary flows or trade flows. International comparisons of flows—such
as development aid, foreign direct investment, migrants’ remittances, or imports and exports of
goods and services—should be made with market exchange rates, not with PPPs.\”

Second, it may have occurred to you that measuring price levels in a comparable way across all the countries of the world, in a way that adjusts for differences in quality and availability of various goods and services, is a Herculean task. There\’s a reason why the estimated PPP exchange rates for 2017 are being published in 2020–it takes time to put all this together. The report describes the methodology in some detail, but there is room for skepticism. Indeed, back in in 2010 Angus Deaton devoted his Presidential Address to the American Economic Association (freely available on-line here) to detailing the \”weak theoretical and empirical foundations\” of such measurements. But for anyone who ha read this far, it will come as no surprise that imperfect economic statistics can still be useful, when applied with context and caution.