The State of US Health

It\’s fairly well-known that life expectancy for Americans are below other high-income countries. But did you know that that the gap is getting worse? And how the underlying causes of death in the U.S., together with proximate factors behind those causes, have been evolving? The Institute for Health Metrics and Evaluation at the University of Washington takes up these issues and more in \”The State of US Health: Innovations, Insights, and Recommendations from the Global Burden of Disease Study.\”

Let\’s start with some international comparisons. Life expectancy in the US is rising–but it is rising more slowly than life expectancy in other OECD countries.

Or consider the average of death. In the U.S., the average age of death rose by nine yeas from 1970 to 2010, but it has risen faster most other places. In this figure, the vertical axis shows age at death in 1970, so if you look at countries from top to bottom, you see how they were ranked by life expectancy in 1970. The horizontal axis shows age at death in 2010, so if you look at countries from right to left, you see how they are ranked by life expectancy in 2010. Countries at about the same horizontal level as the US like New Zealand, Canada, Australia, Spain, Italy, and Japan all had similar life expectancy to the U.S. in 1970, but are now out to the right of the US with higher life expectancies in 2010.

What are the causes of early death? This study seeks to rank different causes of death according to \”years of life lost\”–that is, a cause of death that affects people at younger ages is counted more heavily than a cause of death which affects people at older ages. Here\’s the comparison from 1990 to 2010. Notice that the top six causes of years of life lost change their order a bit, but are otherwise unchanged: heart disease, lung cancer, stroke, chronic obstructive pulmonary disease (COPD), road injury and self-harm. But after that, there are some dramatic changes. For example, the years of life lost because of HIV/AIDS, interpersonal violence, and pre-term birth complications are ranked lower. However, cirrhosis, diabetes, Alzheimer\’s disease, and drug use disorders now rank much higher.

The study also looks at what is called a \”disability-adjusted life year\” or a DALY. This adds together years of life lost and also makes an adjustment for years lived with a disability. Looking at all the causes of death, the study calculates what percentage of the DALYs are attributable to various \”risk factors.\” Clearly, the big risk factors are dietary risks (which largely seems to mean not eating enough fruits, nuts and seeds, vegetables, and whole grains, while eating too much salt and processed meat), together with tobacco use, obesity, high blood pressure, and low physical activity. Clearly, many of these issues of diet, weight, and exercise interact.

The report also offers some county-level analysis across the US, with a reminder that \”how health is experienced in the US varies greatly by locale. People who live in San Francisco or Fairfax County, Virginia, or Gunnison, Colorado, are enjoying some of the best life expectancies in the world. In some US counties, however, life expectancies are on par with countries in North Africa and Southeast Asia.\”

As the US political system convulses over what legal standards that will govern health insurance, the mechanics of paying for health care are obviously of importance. But at the broad social level, life expectancy and health are first and foremost about diet, exercise, not smoking, not drinking to excess, and other behaviors.

A Tripartite Mandate for Central Banks?

The U.S. Federal Reserve operates under what is commonly called a \”dual mandate,\” which basically means that it should take both inflation and unemployment into account when making its decisions. The dual mandate means that the Fed does not take into account the risk that asset market bubbles are destabilizing the economy. Should the dual mandate be turned into a tripartite mandate, with the risk of a financial crisis as the third factor that the Fed (and other central banks) should also take into account. In the most recent issue of the Journal of Economic Perspectives, Ricardo Reis discusses \”Central Bank Design,\” and in particular what the economics literature has to say about a range of issues related to how central banks should choose their goals, their decision-makers, their policy tools, their communication methods, and more. Reis explains why economists have typically not supported a treble mandate in the past, and also why he thinks that this may change in the not-too-distant future.

Why have central banks typically not focused on asset market bubbles and financial instability in the past? Here\’s how Reis explains it (as usual, citations and footnotes are omitted):

\”A more contentious debate is whether to have a tripartite mandate that also includes financial stability. After all, the two largest US recessions in the last century—the Great Depression of the 1930s and the Great Recession that started in 2007—were associated with financial crises. … [I]f financial stability is to be included as a separate goal for the central bank, it must pass certain tests: 1) there must be a measurable definition of financial stability, 2) there has to be a convincing case that monetary policy can achieve the target of bringing about a more stable financial system, and 3) financial stability must pose a trade-off with the other two goals, creating situations where
prices and activity are stable but financial instability justifies a change in policy … Older approaches to this question did not fulfill these three criteria, and thus did not justify treating financial stability as a separate criterion for monetary policy.\” 

As a recent example, think back to the dot-com bubble of the late 1990s. Sure, the price tech stocks seemed to be soaring implausibly high. But was it really the role of the Federal Reserve to decide that stock prices were \”too high,\” and to change monetary policy, perhaps bringing on a recession, in an effort to bring stock prices down? As Reis notes: \”Yet, at most dates, there seems to be someone crying “bubble” at one financial market or another, and the central bank does not seem particularly well equipped to either spot the fires in specific asset markets, nor to steer equity
prices.\”

When the dot-com boom was followed by a short and shallow recession in 2001, the Federal Reserve did what it could to cushion the economy with lower interest rates at that time. I\’m probably not alone in being someone who watched the housing price boom around 2006 and thought: \”Sure, there might be a recession eventually, like in 2001, but it works OK to have the Federal Reserve react after the fact when unemployment goes up. The Fed isn\’t well-equipped to judge when housing prices or stock prices are \”too high\” or \”too low,\” nor to adjust monetary policy to alter such prices.

But as Reis points out, more sophisticated ideas of how to define the rising risk of a financial crisis have come into prominence in the last few years. Instead of focusing on whether stock prices or house prices are rising, or seem \”too high,\” these approaches look at factors like whether total borrowing is rising. Reis explains:

 \”A more promising modern approach begins with thinking about how to define financial stability: for example, in terms of the build-up of leverage, or the spread between certain key borrowing and lending rates, or the fragility of the funding of financial intermediaries. This literature has also started gathering evidence that when the central bank changes interest rates, reserves, or the assets it buys, it can have a significant effect on the composition of the balance sheets of financial intermediaries as well as on the risks that they choose to take. … While it is not quite there yet, this modern approach to financial stability promises to be able to deliver a concrete recommendation for a third mandate for monetary policy that can be quantified and implemented.\” 

A final point here is that implementing a tripartite mandate may also mean changing how one describes the policy tools of a central bank does in a different way. Up to 2007, it was reasonable to describe the policy tools of a central bank mainly in terms of its ability to raise or lower interest rates. But when the central bank starts looking at the total amount of bank credit being extended or at stress-testing whether financial institutions are well-positioned to be resilient in the face of an shock, these sorts of goals can also be accomplished by so-called \”macroprudential regulation,\” which involves adjusting credit conditions by adjusting the rules and standards to be applied by financial regulators.

Is Altruism a Scarce Resource that Needs Conserving?

The Harvard political philosopher Michael Sandel offers a thought-provoking essay, \”Market Reasoning as Moral Reasoning: Why Economists Should Re-engage with Political Philosophy,\” in the just-released Fall 2013 issue of the Journal of Economic Perspectives. Like all JEP articles, it is freely available on-line courtesy of the American Economic Association. (Full disclosure: I\’ve been the Managing Editor of the JEP since its first issue in 1987.) The article makes a number of arguments about the extent to which \”putting a price on every human activity erodes certain moral and civic goods worth caring about.\” Here, I\’ll focus on one argument presented late in the paper, which is the claim that it is a good thing to let markets based on self-interest function in many areas, because it conserves on scarce resources of altruism.

Sandel makes a persuasive case that a number of economists hold this view, although it is not always stated openly. Here are a few examples. The eminent British economist Sir Dennis Robertson gave a prominent 1954 lecture on the topic \”What does the economist economize?\” Here is Sandel\’s discussion:

Robertson (1954) claimed that by promoting policies that rely, whenever possible, on self-interest rather than altruism or moral considerations, the economist saves society from squandering its scarce supply of virtue. “If we economists do [our] business well,” Robertson (p. 154) concluded, “we can, I believe, contribute mightily to the economizing . . . of that scarce resource Love,” the “most precious thing in the world.\”

Kenneth Arrow made a similar argument in a 1972 essay, Sandel notes:

“Like many economists,” Arrow (1972, pp. 354–55) writes, “I do not want to rely too heavily on substituting ethics for self-interest. I think it best on the whole that the requirement of ethical behavior be confined to those circumstances where the price system breaks down . . . We do not wish to use up recklessly the scarce resources of altruistic motivation.”

Or for another example, here\’s Sandel describing a speech that Larry Summers gave at Harvard\’s Memorial Church in 2003:

Summers (2003) concluded with a reply to those who criticize markets for relying on selfishness and greed: “We all have only so much altruism in us. Economists like me think of altruism as a valuable and rare good that needs conserving. Far better to conserve it by designing a system in which people’s wants will be satisfied by individuals being selfish, and saving that altruism for our families, our friends, and the many social problems in this world that markets cannot solve.\”

As Sandel notes with some asperity, this notion of altruism as a scarce resource, \”like the supply of fossil fuels,\” is highly contestable. Might it not be possible instead that when people act in a way that displays altruisism, generosity, or civic virtue, that the social supply of these virtues tend to expand? It seems plausible to think that altruism, generosity, and even love may be socially created, not just used up.

At one level, Sandel\’s critique seems to me fair and well-made. There is actually a reasonable-sized literature in economics and other social sciences looking at how the level of trust and generosity varies across societies. It seems incorrect to think of altruism as a fixed quantity, unaffected by other social institutions.

But at some other level, I feel moved to defend my economist brethren a bit. Focusing on whether whether altruism is a fixed can be a debater\’s point that focuses on the specific phrasing of an argument, rather than the underlying issue at stake. After all, none of the economists are arguing that altruism, generosity, and social virtue are bad ideas or that we should have less of them. Instead, they are arguing that in the real world there exists a division of labor, if you will, in which real-world people choose between altruism and self-interest in different settings. They are arguing that in practical terms, it seems unlikely that social norms of private altruism and generosity by themselves be able to achieve important social goals like supplying food, housing, health care, education, and the necessities of live, as well as helping the poor or protecting the environment. In such cases, it will be important to consider the interactions of self-interest with the compulsion of law.

Sandel focuses on the arguments for how market forces might impinge on civic virtues worth preserving, which is plenty for one essay. But there is also potential for conflict between civic virtues and any institution of society, at least in certain settings. For example, governments around the world can also easily impinge on civic virtues worth preserving. I do think the issues concerning potential conflicts between market forces and moral virtues are real ones. In the style of eminent philosophers, Sandel poses his argument about  not as a set of strong claims, but rather as a set of questions for consideration, and I commend his article to your consideration in a similar spirit.

Journal of Economic Perspectives, Fall 2013, Now On-line

The Fall 2013 issue of the Journal of Economic Perspectives is now available on-line. All articles from this issue back to the first issue in Summer 1987 are freely available on-line, courtesy of the American Economic Association. I\’ll post on some of the articles from this issue in the next week or two. But for now, I just want to let people know what\’s in the issue.  (Full disclosure: I\’ve been the Managing Editor of JEP since the first issue in 1987, so I am predisposed to believe that everything in the issue is well worth reading.) This issue starts with a six-paper symposium on \”The First 100 Years of the Federal Reserve,\” followed by a two-paper exchange on \”Economics and Moral Virtues,\” and then by several individual articles.

Symposium: The First 100 Years of the Federal Reserve

\”A Century of US Central Banking: Goals, Frameworks, Accountability,\” by Ben S. Bernanke

Several key episodes in the 100-year history of the Federal Reserve have been referred to in various contexts with the adjective \”Great\” attached to them: the Great Experiment of the Federal Reserve\’s founding, the Great Depression, the Great Inflation and subsequent disinflation, the Great Moderation, and the recent Great Recession. Here, I\’ll use this sequence of \”Great\” episodes to discuss the evolution over the past 100 years of three key aspects of Federal Reserve policymaking: the goals of policy, the policy framework, and accountability and communication. The changes over time in these three areas provide a useful perspective, I believe, on how the role and functioning of the Federal Reserve have changed since its founding in 1913, as well as some lessons for the present and for the future.
Full-Text Access | Supplementary Materials

\”Central Bank Design,\” by Ricardo Reis

Starting with a blank slate, how could one design the institutions of a central bank for the United States? This paper explores the question of how to design a central bank, drawing on the relevant economic literature and historical experiences while staying free from concerns about how the Fed got to be what it is today or the short-term political constraints it has faced at various times. The goal is to provide an opinionated overview that puts forward the trade-offs associated with different choices and identifies areas where there are clear messages about optimal central bank design.
Full-Text Access | Supplementary Materials

\”The Federal Reserve and Panic Prevention: The Roles of Financial Regulation and Lender of Last Resort,\” by Gary Gorton and Andrew Metrick

This paper surveys the role of the Federal Reserve within the financial regulatory system, with particular attention to the interaction of the Fed\’s role as both a supervisor and a lender-of-last-resort. The institutional design of the Federal Reserve System was aimed at preventing banking panics, primarily due to the permanent presence of the discount window. This new system was successful at preventing a panic in the early 1920s, after which the Fed began to discourage the use of the discount window and intentionally create \”stigma\” for window borrowing — policies that contributed to the panics of the Great Depression. The legislation of the New Deal era centralized Fed power in the Board of Governors, and over the next 75 years the Fed expanded its role as a supervisor of the largest banks. Nevertheless, prior to the recent crisis the Fed had large gaps in its authority as a supervisor and as lender of last resort, with the latter role weakened further by stigma. The Fed was unable to prevent the recent crisis, during which its lender of last resort function expanded significantly. As the Fed begins its second century, there are still great challenges to fulfilling its original intention of panic prevention.
Full-Text Access | Supplementary Materials

\”Shifts in US Federal Reserve Goals and Tactics for Monetary Policy: A Role for Penitence?\” by Julio J. Rotemberg

This paper considers some of the large changes in the Federal Reserve\’s approach to monetary policy. It shows that, in some important cases, critics who were successful in arguing that past Fed approaches were responsible for mistakes that caused harm succeeded in making the Fed averse to these approaches. This can explain why the Fed stopped basing monetary policy on the quality of new bank loans, why it stopped being willing to cause recessions to deal with inflation, and why it was temporarily unwilling to maintain stable interest rates in the period 1979-1982. It can also contribute to explaining why monetary policy was tight during the Great Depression. The paper shows that the evolution of policy was much more gradual and flexible after the Volcker disinflation, when the Fed was not generally deemed to have made an error.
Full-Text Access | Supplementary Materials

\”Does the Federal Reserve Care about the Rest of the World?\” by Barry Eichengreen

Many economists are accustomed to thinking about Federal Reserve policy in terms of the institution\’s \”dual mandate,\” which refers to price stability and high employment, and in which the exchange rate and other international variables matter only insofar as they influence inflation and the output gap — which is to say, not very much. In fact, this conventional view is heavily shaped by the distinctive and peculiar circumstances of the last three decades, when the influence of international considerations on Fed policy has been limited. In fact, the Federal Reserve paid significant attention to international considerations in its first two decades, followed by relative inattention to such factors in the two-plus decades that followed, then back to renewed attention to international aspects of monetary policy in the 1960s, before the recent period of benign neglect of the international dimension. I argue that in the next few decades, international aspects are likely to play a larger role in Federal Reserve policy making than at present.
Full-Text Access | Supplementary Materials

\”An Interview with Paul Volcker,\” by Martin Feldstein

Martin Feldstein interviewed Paul Volcker in Cambridge, Massachusetts, on July 10, 2013, as part of a conference at the National Bureau of Economic Research on \”The First 100 Years of the Federal Reserve: The Policy Record, Lessons Learned, and Prospects for the Future.\” Volcker was Chairman of the Board of Governors of the Federal Reserve System from 1979 through 1987. Before that, he served stints as President of the Federal Reserve Bank of New York from 1975 to 1979, as Deputy Undersecretary for International Affairs in the US Department of the Treasury from 1969 to 1974, as Deputy Undersecretary for Monetary Affairs in the Treasury from 1963 to 1965, and as an economist at the Federal Reserve Bank of New York from 1952 to 1957. He has led and served on a wide array of commissions, including chairing the President\’s Economic Recovery Advisory Board from its inception in 2009 through 2011.
Full-Text Access | Supplementary Materials

Symposium: Economics and Moral Virtues

\”Market Reasoning as Moral Reasoning: Why Economists Should Re-engage with Political Philosophy,\” by Michael J. Sandel

In my book What Money Can\’t Buy: The Moral Limits of Markets (2012), I try to show that market values and market reasoning increasingly reach into spheres of life previously governed by nonmarket norms. I argue that this tendency is troubling; putting a price on every human activity erodes certain moral and civic goods worth caring about. We therefore need a public debate about where markets serve the public good and where they don\’t belong. In this article, I would like to develop a related theme: When it comes to deciding whether this or that good should be allocated by the market or by nonmarket principles, economics is a poor guide. Deciding which social practices should be governed by market mechanisms requires a form of economic reasoning that is bound up with moral reasoning. But mainstream economic thinking currently asserts its independence from the contested terrain of moral and political philosophy. If economics is to help us decide where markets serve the public good and where they don\’t belong, it should relinquish the claim to be a value-neutral science and reconnect with its origins in moral and political philosophy.
Full-Text Access | Supplementary Materials

\”Reclaiming Virtue Ethics for Economics,\” by Luigino Bruni and Robert Sugden

Virtue ethics is an important strand of moral philosophy, and a significant body of philosophical work in virtue ethics is associated with a radical critique of the market economy and of economics. Expressed crudely, the charge sheet is this: The market depends on instrumental rationality and extrinsic motivation; market interactions therefore fail to respect the internal value of human practices and the intrinsic motivations of human actors; by using market exchange as its central model, economics normalizes extrinsic motivation, not only in markets but also in social life more generally; therefore economics is complicit in an assault on virtue and on human flourishing. We will argue that this critique is flawed, both as a description of how markets actually work and as a representation of how classical and neoclassical economists have understood the market. We show how the market and economics can be defended against the critique from virtue ethics, and crucially, this defense is constructed using the language and logic of virtue ethics. Using the methods of virtue ethics and with reference to the writings of some major economists, we propose an understanding of the purpose (telos) of markets as cooperation for mutual benefit, and identify traits that thereby count as virtues for market participants. We conclude that the market need not be seen as a virtue-free zone.
Full-Text Access | Supplementary Materials

Articles

\”Gifts of Mars: Warfare and Europe\’s Early Rise to Riches,\” Nico Voigtländer and Hans-Joachim Voth

Western Europe surged ahead of the rest of the world long before technological growth became rapid. Europe in 1500 already had incomes twice as high on a per capita basis as Africa, and one-third greater than most of Asia. In this essay, we explain how Europe\’s tumultuous politics and deadly penchant for warfare translated into a sustained advantage in per capita incomes. We argue that Europe\’s rise to riches was driven by the nature of its politics after 1350 — it was a highly fragmented continent characterized by constant warfare and major religious strife. No other continent in recorded history fought so frequently, for such long periods, killing such a high proportion of its population. When it comes to destroying human life, the atomic bomb and machine guns may be highly efficient, but nothing rivaled the impact of early modern Europe\’s armies spreading hunger and disease. War therefore helped Europe\’s precocious rise to riches because the survivors had more land per head available for cultivation. Our interpretation involves a feedback loop from higher incomes to more war and higher land-labor ratios, a loop set in motion by the Black Death in the middle of the 14th century.
Full-Text Access | Supplementary Materials

\”The Economics of Slums in the Developing World,\” by Benjamin Marx, Thomas Stoker and Tavneet Suri

The global expansion of urban slums poses questions for economic research as well as problems for policymakers. We provide evidence that the type of poverty observed in contemporary slums of the developing world is characteristic of that described in the literature on poverty traps. We document how human capital threshold effects, investment inertia, and a \”policy trap\” may prevent slum dwellers from seizing economic opportunities offered by geographic proximity to the city. We test the assumptions of another theory — that slums are a just transitory phenomenon characteristic of fastgrowing economies — by examining the relationship between economic growth, urban growth, and slum growth in the developing world, and whether standards of living of slum dwellers are improving over time, both within slums and across generations. Finally, we discuss why standard policy approaches have often failed to mitigate the expansion of slums in the developing world. Our aim is to inform public debate on the essential issues posed by slums in the developing world.
Full-Text Access | Supplementary Materials

\”Recommendations for Further Reading,\” by Timothy Taylor
Full-Text Access | Supplementary Materials

The Safety of Bioengineered Crops

When I first had to write about genetically modified crops, there was no evidence upon which to draw. It was 1986, and I was working as an editorial writer for the San Jose Mercury News. The first genetically modified plant, a variety of tobacco, had been created in a lab in 1983. A local company called Advanced Genetic Sciences was proposing the first field test of a genetically modified organism. Their plan was to spray some strawberries and potatoes with a bacteria often known as\” ice-minus,\” because it helped prevent frost damage to crops.

The step was quite controversial. For opponents, no buffer zone and no set of precautions could justify spraying this bacteria: indeed, the night before the trial was to take place, protesters broke into the fields and pulled up many of the plants. But I wrote some editorials supporting the trial, in large part because the ice-minus bacteria were already fairly widespread in nature. It was perfectly legal to culture the natural ice-minus bacteria and spray it; this trial just involved spraying genetically identical ice-minus bacteria that had been created in a laboratory. The trial worked fine: that is, the crops were more frost-resistant, and there were no observable negative effects to the environment.

By 1994, the first genetically modified plant was commercially sold when the FlavrSavrTM tomato hit the market. But the introduction of genetically engineered crops in the field is usually dated to 1996, when field crops that were genetically engineered to be pest-resistant and herbicide-tolerant were introduced. Now, after the technology has been in widespread use for 17 years, the studies are piling up. In a forthcoming article posted online at Critical Reviews in Biotechnology, Alessandro Nicolia, Alberto Manzo, Fabio Veronesi, and Daniele Rosellini provide \”An overview of the last 10 years of genetically engineered crop safety research.\” They built a comprehensive database of the research literature from 2002 through 2012, consisting of 1,783 articles on different aspects of these first-generation genetically engineered crops. Here is the bottom line of their survey, from the abstract:

\”The technology to produce genetically engineered (GE) plants is celebrating its 30th anniversary and one of the major achievements has been the development of GE crops. The safety of GE crops is crucial for their adoption and has been the object of intense research work often ignored in the public debate. We have reviewed the scientific literature on GE crop safety during the last 10 years, built a classified and manageable list of scientific papers, and analyzed the distribution and composition of the published literature. We selected original research papers, reviews, relevant opinions and reports addressing all the major issues that emergedin the debate on GE crops, trying to catch the scientific consensus that has matured since GE plants became widely cultivated worldwide. The scientific research conducted so far has not detected any significant hazards directly connected with the use of GE crops; however, the debate is still intense.\” 

More specifically, they look at how genetically engineered crops have interacted with biodiversity; at risks of \”gene flow\” to other crops, wild plants, or through the soil; at health effects for animals that eat feed from genetically engineered crops; and potential health effects for human consumers. As they discuss, some of the evidence raises warning signs that are worth more investigation,  and in cases certain genetically engineered crops are no longer grown, or not grown in certain areas, because of these concerns. But as they write in the conclusion: \”We have reviewed the scientific literature on GE [genetically engineered] crop safety for the last 10 years that catches the scientific
consensus matured since GE plants became widely cultivated worldwide, and we can conclude that the scientific research conducted so far has not detected any significant hazard directly connected with the use of GM [genetically modified] crops.\” In short, if you like to believe that you follow the scientific evidence, you should believe that the first generation of pest-resistant and herbicide-tolerate genetically engineered crops has been a great success, substantially increasing crop yields and reducing the need for heavy chemical use.

I support all sorts of rules and regulations and follow-up studies to make sure that genetically engineered crops continue to be safe for the environment and for consumers. After all, the first-generation genetically engineered field crops were all about pest resistance and herbicide-tolerance, and as new types of genetic engineering are proposed, they should be scrutinized. But for me, the purpose of these regulations is to create a clear pathway so that the technology can be more widely used in a safe way, not to create a set of paperwork hurdles to block the future use of the technology.

Farmers have been breeding plants and animals for desired characteristics for centuries. Genetic engineering holds the possibility of speeding up that process of agricultural innovation, so that agriculture can better meet a variety of human needs. Most obviously, genetically engineered crops are likely to be important as world population expands and world incomes continue to rise (so that meat consumption rises as well). In addition, remember that plants serve functions other than calorie consumption. Plant that were more effective at fixing carbon in place might be a useful tool in limiting the rise of carbon in the atmosphere. Genetically modified plants are one of the possible paths to making plant-based ethanol economically viable. Plants that can thrive with less water or fewer chemicals can be hugely helpful to the environment, and to the health of farmworkers around the world. The opportunity cost of slowing the progress of agricultural biotechnology is potentially very high.

Unavoidable Realities of Insurance: Health Care Edition

Insurance markets are unavoidably unpopular, because of a basic fact and an unpalatable implication.

Here\’s the basic fact about all insurance markets: What gets paid out must equal what gets paid in. Or to put it another way, what is paid in by the average person in premiums must be pretty much equal (with some minor differences noted below) to what is received by the average person in benefits.

And here\’s the unpalatable implication: Some people will receive much more from insurance payments than they pay in. They might buy life insurance and die young, or buy car insurance and suffer a severe accident, or buy health insurance and face a costly period of hospitalization. But in turn, because of the basic fact that the average person can\’t receive more more in benefits than the average person pays in premiums, there will inevitably be many people–probably a majority of insurance buyers–who will receive much less in insurance benefits than they pay in.

For example, I\’m a direct purchaser of insurance for my home, car , and (term) life, as well as an indirect purchaser of health and dental insurance through my employer. The best possible outcome for me is that I would pay premiums year after year, all my adult life, and never receive more than minimal insurance benefits. After all, receiving significant payments from an insurance company would mean that my family had experienced damage to our home, or a car accident, or sickness, injury, or death.

Any market where the good outcome experienced by a majority of buyers is to make payments all your life in exchange for little or no benefit is going to be continually unpopular.

A lot of energy goes into trying to ignore or deny either the basic fact about insurance markets or the unpalatable implication. The expansions of what health insurance policies must cover that are required by the Affordable Care and Patient Protection Act offer a vivid example. The rules expanding what a typical health insurance plan must cover mean that the average person will need to pay higher premiums, because the benefits being paid out of the health insurance system need to equal what is paid in. Moreover, some people are going to draw on these additional coverage provisions much more than others, so many of those who are unlikely to draw upon such policies will find an even bigger gap between what they are paying in insurance and the insurance benefits they personally receive. Indeed, many of these all-pay, little-benefit households–and many people have a pretty good idea whether they fall into this category–are being required under the new law to pay for others to receive a level of health insurance coverage that they had not previously chosen to receive for themselves.

There will always be a political dynamic to promise that the majority should receive more for their insurance, but that no one should need to pay more on their premiums. For example, the original Medicare legislation back in 1966 required that premiums paid by the elderly should cover 50% of the costs of Part B. But Medicare spending went up, premiums didn\’t, and in 1997 a law needed to be passed to assure that premiums paid by the elderly would cover 25% of the costs of Part B.

Another way to quarrel with the basic fact about insurance is to point out that private insurance companies spend money on administrative costs and on profits. In addition, insurance companies earn revenue from investing reserves. One can argue that insurance companies could be more efficient, or their profits could be more regulated, and that in these ways benefits might be increased somewhat without paying more in premiums. I\’m all for encouraging greater efficiency, and I think the government has been slow in pushing the private sector to coordinate on formats for electronic medical records and billing. But the National Association of Insurance Commissioners reports that in 2012, insurance companies spent 85.7% of the premiums they received, while 11.8% was paid in administrative expenses, and the other 2.7% was profit margin. In other words, the overwhelming amount of money paid into health insurance in premiums is indeed paid out to health care providers. The existence of private-sector insurance companies may bend the basic fact that what is paid out in insurance benefits must equal what is paid in, but it does not sever the connection.

Just to be clear, neither the fundamental fact about insurance nor the unpalatable implication goes away with a government-run or a single-payer insurance system. Whether it\’s Medicare or Medicaid or private sector health insurance or government-run exchanges, it still  holds true that benefits must be paid for. It\’s tempting, of course, to assert that the U.S. government could run a nonprofit health care system in a efficient and cost-effective manner that reduced administrative costs, but even if this assert is true, as noted a moment ago the additional revenue from greater administrative efficiency is a modest share of total health care spending. Also, given events in recent weeks, that assertion that the U.S. government could run an efficient and cost-effective health insurance system must be open to considerable dispute. There\’s a reason why, in a country as large and diverse as the United States, insurance regulation has typically been done at the state level rather than the federal level. Even with a government-run or single-payer system, it still holds true that because some people will experience very high costs, the typical person will pay into the health insurance system more than they ever get out–and should be perversely grateful to be lucky in this way.

I have for years favored having the government spend more to subsidize health insurance for those 50 million or so Americans who have lacked it. My preference has been to fund these subsidies by placing limits on the tax exemption given for employee compensation paid in the form of employer-paid health insurance. (For some estimates from the Congressional Budget Office of how such limits could raise $46-$101 billion per year in extra tax revenue when phased in 10 years from now, see Option 15 in Chapter Five of this recent CBO report.)

However, the Affordable Care and Patient Protection Act goes well beyond providing assistance to those without insurance. It has been promoted with set of promises that everyone in both the individual and employer-provided insurance market can have the same or more insurance coverage while everyone pays the same or lower private-sector insurance premiums–and while future increases in government health care spending are lower than than they otherwise would have been. In seeking to carry out these promises in defiance of the basic economics of insurance markets, the law will necessarily disrupt the health care arrangements for a sizable share of the 200 million or so Americans who already have private health insurance.

Thoughts about the Future of Print

My work life is mostly in the world of print. I run an academic economics journal. Yes, I give some lectures now and then, but I write this blog, and I have written an introductory economics textbook and a book for the general public on understanding economics. So I\’m always on the lookout for articles on the future of media in the digital age, and here are three that recently caught my eye.

Frank Rose writes on \”The Future of Media: First Print, Then Cable,\” in the Fourth Quarter 2013 issue of the Milken Institute Review. He sets the stage with a review of how

\”Contrary to popular belief, the 20th century did not end at the same time for everyone. For those in the music business, it sputtered to a halt with the close of 1999 – the year recorded-music sales in the United States reached their all-time high of $14 billion; by 2012, sales had dropped to half that. For newspapers, the end of the century arrived a year later, when U.S. ad revenues hit nearly $49 billion; now they’re down to a mere $22 billion. Hollywood was able to stave off the century’s demise until 2002, when box-office admissions in the United States and Canada peaked at 1.6 billion; last year they were down 15 percent from that (which is not bad, considering). Magazines, for their part, seemed to lead a charmed life – through 2006, anyway, when ad pages topped out at 253,000; since then, they have dropped by a third. As for television, the major broadcast networks have been on a slow, inexorable slide into the future for decades. In 1980, according to Nielsen, well over half the TV sets in America were tuned to ABC, NBC or CBS during prime time. By 2012, even with the addition of the Fox network, the portion tuned to a broadcast network was below one quarter.\”

Rose also describes what he sees as three key expectations that are emerging for media in the digital age. 

\”Over the past few years, as more and more people have become comfortable with smartphones, tablets and services that let you buy or stream stuff online, three key expectations have emerged. Any one of these would be challenging to a 20th-century media company. Taken together, they require a total rethinking of the enterprise. First, people are coming to expect an infinitude of choice. They want news and entertainment on their own terms. They see no reason they shouldn’t be able to watch movies or TV shows or listen to music whenever they want, wherever they happen to be, on whatever device they have with them – and with a minimum of advertising, thank you. …  Meanwhile, a second set of expectations has been developing. Humans have always
wanted to somehow inhabit the stories that move them – to be spellbound, entranced, transported to another realm. … Historically we’ve managed to plunge in with whatever technology is at hand, from books to film to TV. But digital raises the possibility of immersing ourselves even further, and in entirely new ways. … But there is a third, closely related expectation as well: more and more, audiences expect some kind of active involvement. No longer passive consumers, they now want to be participants – critics, providers and (through the act of sharing on social media) distributors of information.\”

But oddly enough, you may have noticed that Rose\’s list of media industries that have been overwhelmed by digital forces doesn\’t include plain old books. Evan Hughes offers an argument for why plain old books have held up so well in the digital era in \”Books Don\’t Want to Be Free,\” which appears in the October 21 issue of The New Republic.

\”If you’re in the business of selling journalism, moving images, or music, you have seen your work stripped of value by the digital revolution. Translate anything into ones and zeroes, and it gets easier to steal and harder to sell at a sustainable price. Yet people remain willing to fork over a decent sum for books, whether in print or in electronic form. “I can buy songs for 99 cents, I can read most newspapers for free, I can rent a $100 million movie tonight for $2.99,” Russ Grandinetti, Amazon’s vice president of Kindle content, told me in January. “Paying $9.99 for a best-selling book—paying $10 for bits?—is in many respects a very strong accomplishment for the business.”

Hughes emphasizes two main reasons why books have held up so well in the digital era. First, books are they are less vulnerable to disaggregation: you can pull apart and sell individual songs on an album of popular music much more easily than you can pull apart chapters of most books. Second, books are already so low-tech that they are in an odd way less affected by digital technology.

\”Part of the problem for journalism, music, and television is that they are vulnerable to disaggregation. Their products are made up of songs and articles and shows that have long been consumed in those individual units. Once the Internet made it possible to ignore the unwanted material, overall value slipped. Easy access to favorite singles opened those up to impulse buys—but also made purchasing the whole record feel almost indulgent, a splurge for audiophiles and diehard fans. Now the TV viewer wants “Breaking Bad” without bills from Comcast. The ability to score individual articles by the clicks and ad dollars they reap has exposed vital but embattled forms like international reporting and arts criticism to further pruning.

Hollywood has fallen victim not to disaggregation but to its opposite: Netflix and the like have bundled films into affordable smorgasbords, undermining the perceived value of each movie. (More recently, Pandora and Spotify have hit music from this second flank.) The dark side of digitization—piracy— is obviously another force at play. Sites like BitTorrent, following in Napster’s wake, have helped convince a generation that they shouldn’t have to pay for the goods. In publishing, meanwhile, the deal with the customer has always been dead simple, and the advent of digital has not changed it: You pay the asking price, and we give you the whole thing. It would make little sense to break novels or biographies into pieces, and they’re not dependent on the advertising that has kept journalism and television artificially inexpensive and that deceives the consumer into thinking the content is inexpensive to make …

Some print stalwarts find e-readers a step down, but the fundamental experience of immersing yourself in a text is not bound up in any particular medium or venue. Reading never depended too much on sensual verisimilitude, only a mental leap from the words to the ideas they represent. The book is so low-tech, it’s hard for technology to degrade it.

Finally, what a lot of digital media comes down to is the issue of how to get and keep the attention of possible consumers. Hal Varian emphasizes this point in \”The Economics of the Newspaper Business.\” a talk he gave at an Italian journalism awards ceremony in September. The talk is an absolutely elegant example of how to say something real and interesting in a very short time window. Varian argues: 

\”The fundamental challenge facing newspapers is to increase the time people spend on their content. … Subscribers to physical newspapers spend about 25 minutes a day reading them. The typical time spent on an online news site in the US and UK is about 2-4 minutes, roughly one-eighth as much. Interestingly, newspapers in the US make about one-eighth of their total ad revenue from online ads. If readers spent as much time reading the online content as the offline content, ad revenue from online content would be much closer to the offline revenue.\”

My job is to be Managing Editor of the Journal of Economic Perspectives, which is one of seven journals published by the American Economic Association. A couple of years ago, the AEA made a decision that all issues of the journal, from the first issue back in 1987 up to the most recent, are freely available online. Both individual articles and entire issues can be downloaded to readers like Kindle or Nook, too. Of course, making a publication free only works because the AEA has income from other sources, like dues from its members, subscriptions from libraries, and payments for the EconLit service it runs that indexes the articles in academic economics journals. This blog, like many blogs, is created from curiosity, obsessiveness, and attention-seeking, but is otherwise done without any specific payment.

Goldilocks Fiscal Policy: Just About Right

There\’s an old saying that the problem with trying to be middle-of-the-road is that you get hit by traffic going both directions. Nonetheless, standing in the middle of those who argue vehemently that the budget deficits should be much smaller (because the recession ended in June 2009 and the deficits didn\’t help much in stimulating the economy anyway) and those who argue that the deficits should be much larger (because unemployment remains stubbornly high at 7.3% and a bigger deficit would help to reduce that rate), I actually think the budget deficit is about where it should be.

The Monthly Budget Review that Congressional Budget Office published on November 7 has an overview of the budget picture through Fiscal Year 2013, which ended on September 30. In the figure, federal spending and revenue are shown in the way I prefer–as a percentage of GDP.

The story of fiscal policy since 1980 is encapsulated in these lines. For example, the Reagan tax cuts and spending increases that led to such large budget deficits in the 1980s are clear.

In the 1990s, there\’s a slow decline in federal spending (as a share of GDP) in the 1990s as defense spending fell and a growing economy reduced the need for income support programs. On the revenue side, Bill Clinton and the Democrats passed a tax increase in 1993 with no Republican votes at all, which together with the surge in tax revenues from the dot-com boom of the 1990s raised the federal tax take.

In the 2000s, tax revenue first falls with the end of the dot-com boom and the Bush tax cuts, and then rises with the construction and housing price bubble in mid-decade. Spending rises a bit in the early 2000s as a recession increases the need for government income support programs and the aftermath of 9/11 pushes up defense-related spending again.

When the Great Recession hits in late 2007, spending turns sharply up and tax revenue turns sharply down. In part, this change is built into government programs. When the economy slumps and income falls, tax revenues will drop automatically, too. Similarly, when the economy slumps more people become eligible for government support programs. These \”automatic stabilizers,\” as they are called, will increase the deficit, and then additional tax cuts or spending increases will raise it further. (Here\’s a post about a CBO report from last March laying out what happened with automatic stabilizers during the recession.)

Naturally, when the recession ends and economic growth resumes–even sluggish and halting growth–the automatic stabilizers reverse themselves and the temporary fiscal stimulus packages come to an end. Tax revenues rise; spending falls. In fiscal year 2013, government spending was 20.8% of GDP, which was slightly above the average over the last 40-year of 20.4%. Federal government tax revenues were 16.7% in 2013, which was below the 17.9% from back in 2007 before the Great Recession and also below the 17.4% average over the last 40 years.

The budget deficit for 2013 was 4.1% of GDP. This is considerably lower than the gargantuan deficits from fiscal years 2009 to 2012 (which were 9.8, 8.8, 8.4 and 6.8% of GDP, respectively). But after four years of gargantuan budget deficits, this lower budget deficit for 2013 hardly qualifies as austerity. As I have pointed out before, the total rise in federal debt in recent years was, when expressed relative to GDP at the time, about twice the size of what happened with the Reagan deficits of the mid-1980s, substantially larger than the federal debt increase incurred during the Great Depression of the 1930s, and about two-thirds the size of the debt increase incurred to fight World War II.

The 2013 government budget deficit of 4.1% of GDP is smaller than in the immediate past, but given the recession ended more than four years ago in June 2009, and unemployment has gradually worked down from 10% in October 2009 to 7.3% at present, it seems to me appropriate that the deficit should be declining, too. Moreover, by historical standards a budget deficit of 4.1% of GDP isn\’t tiny. If one goes back before the Great Recession, it\’s still the the biggest annual deficit since the 1991-92 recession year or the Reagan deficits of the mid-1980s.

The long-term projections for  higher budget deficits, driven by an aging population and continually rising health care costs, are troubling to me. But as for right now, the overall budget deficit seems just about right to me.  

Africa\’s Growth: Not Just Minerals

The lowest-income region of the world, sub-Saharan Africa, has been having a surge of economic growth during the last decade or so. The IMF reviews the current situation in one of its Regional Economic Outlook reports, \”Sub-Saharan Africa: Keeping the Pace,\” published in October.

The red line on the figure shows the rate of real GDP growth in the region sub-Saharan African since the late 1990s, measured on the left-hand axis. The green line, measured on the right-hand axis, shows the growth rate in sub-Saharan Africa compared with the 188 IMF member countries. Economic growth in the sub-Saharan region as a whole in the last decade has consistently been above the median country, and in recent years has been in the 60th to 70th percentile.

But this good news comes wrapped in some concerns. It\’s also been a time of high prices for oil and many minerals. Is Africa\’s economic growth really just about countries getting a better price for their minerals? Or is something deeper going on? The second chapter of the IMF report addresses this question head-on (as usual, footnotes and references to figure or tables are omitted for readability).

\”Sub-Saharan Africa has grown strongly since the mid-1990s. There is a common perception that this growth has been the result of relatively high commodity prices, particularly for natural resources such as oil and minerals, generating both higher commodity revenues and attracting substantial new investment. Although growth in some countries in the region is heavily dependent on the export of natural resources, many nonresource-intensive LICs [low-income countries] have also experienced rapid growth. In fact, 8 of the 12 fastest-growing economies in sub-Saharan Africa since 1995 were LICs considered nonresource-rich during this period, and as a group they have grown slightly faster than the oil exporters …\” 

The chapter then focuses on telling the story of six of these nonresource-intensive low-income countries. \”All sample countries—Burkina Faso, Ethiopia, Mozambique, Rwanda, Tanzania, and Uganda—have achieved strong and sustained growth since the mid-1990s, despite not having exploited natural resources on a large scale during this period. The countries were chosen on the basis of having experienced real output growth greater than 5 percent on average during the period 1995–2010, and real per capita GDP growth of more than 3 percent over the same period.\”

At the risk of oversimplifying the many details that differ across these six countries, I\’d say that two main themes run through the discussion of their success. First, they enacted economic policies that  made them attractive for investment, and capital arrived in the form of domestic investment, foreign direct investment, debt relief, and aid. Second, they saw a boost in their agricultural sector, which was a substantial share of the economy and a majority of the workforce and of poverty in all of these countries.

On the theme of increased investment levels, the IMF writes:

\”A common theme emerging from this analysis is that macroeconomic stabilization and decisive structural reforms were crucial in laying the foundations for sustained growth. Several of the countries emerged from armed conflict or had been characterized earlier by African socialism and state-led development strategies stemming from post-independence periods. Therefore, countries had to address major macroeconomic imbalances … Moreover, the six countries all saw higher aid flows, higher FDI [foreing direct investment], significant debt relief, and higher public capital expenditure than other nonresource-intensive LICs and fragile states. Although the nexus between investment and economic growth is not fully understood, this sustained
high investment is likely to have supported high growth in the sample countries.\”

Along with running more sensible macroeconomic policies and reducing the role of price controls, these countries also saw improved governance overall. \”The sample countries registered improvements in governance relative to the comparator group in four out of the five dimensions considered in the World Bank’s Worldwide Governance Indicators: control of corruption, government effectiveness, political stability and absence of violence, and regulatory quality.\”

These economic changes took effect in some important part through the agricultural sector. As economies develop, agriculture will typically play a smaller role in the economy, which as this graph illustrates has been happening for these six countries.

But the share of GDP that is in agriculture may understate the importance of agriculture in economic development, because and even larger share of low-income workers are in the agricultural sector. \”Most of the labor force in the sample countries remains concentrated in the agricultural sector, and more significant productivity increases will be crucial to promoting inclusive growth and lifting more people out of poverty. It is estimated that the bulk of the sample countries’ active labor force is engaged in the agricultural sector—about 80–81 percent in Mozambique and Burkina Faso, 71 percent in Uganda, and 65 percent in Tanzania … Moreover, most of the extremely poor rely on subsistence farming for their livelihoods.\”

A final intriguing point in the report is that these six countries are quite different from each other in various ways. Burkina Faso uses the CFA franc as its currency and has a large cotton sector. Ethiopia has seen growth in industries like tourism, air travel, and cut flowers. Mozambique has large projects underway to produce and transmit electricity and gas. Rwanda has seen growth in coffee and tourism. Tanzania is an example of broad-based reforms across all sectors, and Uganda has been diversifying its exports. As the report argues:

\”Initial conditions are not so important. The majority of the countries in the sample are landlocked, and many were just emerging from conflict situations in the early 1990s. Postconflict and other fragile states can move onto rapid growth paths once a clear vision is formulated and consistently implemented.\”

In other words, a number of countries in sub-Saharan Africa are reaching a point where sensible economic policy–no miracles needed!–can help them move to more rapid economic growth. For some earlier posts about Africa\’s growth, see:

Why U.S. Economic Growth is Getting Harder

The formula for economic growth is clear enough: more workers, improved human capital as measured in education and experience, high levels of capital investment, and ongoing adoption of new technologies, all mixed together in a market-oriented economic environment that provides incentives for work and innovation. The problem for the U.S. economy is that all of these factors are looking shaky for the decade or two ahead. Brink Lindsey lays out the issues in \”Why Growth Is Getting Harder,\” written as Policy Analysis #737 (October 8, 2013) for the Cato Institute.

For example, when it comes to the quantity of labor in the U.S. economy, it\’s become well-known that that the share of adults who have jobs or are looking for them–the \”labor force participation rate\”–has been dropping for most of the 21st century and dropping faster in the Great Recession and its aftermath. Lindsey offers an interesting graph of per capital hours worked. This measure takes into account changes in the share of adults working, as well as changes in the work week. Looking at this graph, and thinking about the ongoing retirement of the Baby Boom generation, it\’s hard to think that a surge in hours worked will be a main driver for future US economic growth.

When it comes to capital investment, it\’s been true for decades that the US is a low saving and low investment economy by the world standards. Maybe at some point in the future these trends will turn, so that Americans save more and the U.S. government borrows substantially less, both leading to more funds for domestic investment. But it\’s hard to look at this long-run pattern and think that a surge of saving and physical capital investment will be a main driver for long-run U.S. growth.

Along with physical capital, it\’s at least as important to invest in human capital, a broad term that include both education levels and skills gained through job experience. U.S. adults as a group tend lag behind the average of other high-income countries when it comes to literacy, numeracy, and problem solving. This graph shows that while the U.S. had a surge in educational attainment between about 1930 and 1970, the change since then has been much more modest. Meanwhile, the rest of the world isn\’t standing still in expanding educational attainment for their populations.  

So that leaves us with technology as a potential source of future economic growth. Measurements of productivity are quite volatile, depending on the business cycle, but the smoothed trend line shows that there is no reason to think that productivity growth is trending up. As just a little extra salt in the wound, Lindsey includes a line (measured on the left-hand axis) showing that in fact the U.S. workforce has experienced a rising share of workers in science, technology, engineering, and mathematics. Maybe this has helped to prevent the rate of productivity growth from falling further, but that\’s the best case one can make. For some perspectives on the controversy about whether a surge of US productivity growth is likely in the future, see here and here.

U.S. economy continues to have some enormous strengths and opportunities, of course. For example, the U.S. is uniquely positioned to be a crossroads of the evolving global economy, with the communications and transportation infrastructure, legal and financial industries, and personal and business connections around the world. The possibility that the U.S. could have less expensive domestically produced energy supplies could give the economy a genuine boost. Finding ways for typical jobs in factories, offices, schools, and hospitals to work more effectively with information technology holds the possibility of enormous gains. But at their core, all of these kinds of changes go back to the basic building blocks of workers with an increasing level of skill, working with a rising stock of physical capital investment, finding and adapting new technologies.

The bottom line here is not that the U.S. economy goes into  a tailspin, but rather that slow growth accumulates over time. Lindsey writes:

How important is such a [growth] slowdown? Thanks to the power of compound interest, relatively small differences in the growth rate add up to huge differences in living standards over time. Using the so-called “rule of 70,” you can figure out roughly how long it takes for output per person to double by dividing 70 by the growth rate. Thus, a 2 percent annual growth rate doubles incomes in 35 years, whereas with a 1.5 percent annual growth rate it takes 47 years for incomes to double. Consider the case of a 22-year-old college graduate, just starting in the workplace now. If the long-term average growth rate falls from 2 percent to 1.5 percent, the economy at the time our new college grad retires at age 65 will be almost 20 percent less wealthy than it would have been if the growth rate had remained on trend.

As I sometimes like to say, whatever your political goal, whether it is higher spending on social programs or a lower tax burden for citizens, it\’s easier to achieve that goal when the economy is growing more rapidly.