Hide table of contents

This post was crossposted from Oliver Kim's Global Developments blog, with the author's permission, by the Forum team. The author may not see comments.


A Review of Morten Jerven's Poor Numbers, A Decade On

Poor Numbers: How We Are Misled by African Development Statistics and What to Do About It, by Morten Jerven. 2013.

I think most users of GDP presume that, when they download an Excel spreadsheet from a slick-looking official website like the World Bank or the International Monetary Fund, what they get is something FDA-approved for human consumption—that, somewhere down the line, maybe in some other room down the hall, the real experts have done the shoe-leather work to make sure that the numbers are accurate and reliable.

Morten Jerven’s 2013 book, Poor Numbers, is about why that isn’t true.

The Late 1980s Tanzanian Growth Catastrophe

Perhaps the best illustration comes from Tanzania, where the socialist leader Julius Nyerere famously pursued a policy of ujamaa—“brotherhood” in Swahili—in which the newly independent state assumed control of large portions of the economy.

Initially, Nyerere’s ambitious project of socialist nation-building captured the global imagination. In some respects, it was a remarkable success—to this day, Tanzanians have a stronger sense of shared national identity than their neighbors. But it also had a dark side: a process of forced villagization, where Tanzania’s scattered rural population was coercively—sometimes violently—resettled into villages planned by the state.

By the 1980s, it was clear that ujamaa had fallen short in terms of delivering development. Under pressure from the IMF and the World Bank, Tanzania began to implement neoliberal reforms, loosening the state’s control of the economy.

According to the Penn World Tables, one of academic economists’ favorite datasets, between 1985 and 1986, GDP per capita fell from $1,717 to $1,130 in 2017 international dollars—a 33% collapse, larger than what the United States experienced from 1929-33 during the Great Depression:

The only problem? It likely didn’t happen.

The Late 1980s Tanzanian Growth Catastrophe is largely a statistical artifact—a product of the simple truth that GDP measures what the state can see.

As ujamaa collapsed, the Tanzanian state atrophied, and economic activity retreated into the informal sector, with farmers switching from export to subsistence crops. But the GDP statistics largely missed this shift. In 1997, the GDP methodology was finally revised, using 1992 prices as a base. With this revision, the Bureau of Statistics admitted that GDP had previously been underestimated by anywhere between 30 and 200(!) percent.[1]

This left the question of how to reconcile the new numbers with the old ones. For somewhat murky reasons, the Penn World Tables chose to stitch the two series together in such a way to create the Late 1980s Tanzanian Growth Collapse. Not only that, but to get the reconciliation to work, the Penn data also shows a miraculous recovery over the late 1980s, with per capita growth rates averaging over 5% a year from 1995 to 1999. (This recovery has erroneously been used to tout the success of Tanzania’s structural adjustment program.)

But no other estimate of Tanzanian GDP shows such dramatic volatility:

Jerven (2013), pg. 71. Figures may be slightly different from PWT above as they use an older vintage.

In the current Penn World Tables, Tanzania’s one-year GDP collapse is a large outlier—a three-standard deviation event. But it is by no means alone. In 2013, Nigeria revised its GDP estimates and, overnight, Nigerian GDP jumped by 89%, leapfrogging South Africa to become Africa’s largest economy. In 2010, a similar revision by Ghana caused its GDP to increase by 60%.[2]

In a sense, drastic revisions like the Late 1980s Tanzanian Growth Collapse or the Great 2010 Ghanaian Rebasing Jump are straightforward to deal with. If you know about them, you can drop them from your data. The scary thing is that decisions like these likely undergird most GDP numbers from poor countries. Your real concern should be the dogs that don’t bark—it should be the data fictions that lurk in plain sight.

“What Happens If I Disappear?”

Through a series of country case studies, Poor Numbers documents how African economic statistics are actually produced. Jerven’s examples, though not comprehensive, are collectively shocking. Hollowed out by years of state neglect, African statistical agencies are now often unable to conduct basic survey and sampling work. Jerven writes:

In 2010, I returned to Zambia and found that the national accounts now were prepared by one man alone… Until very recently he had had one colleague, but that man was removed from the National Accounts Division to work on the 2010 population census. To make matters worse, lack of personnel in the section for industrial statistics and public finances meant that the only statistician left in the National Accounts Division was responsible for these data as well. (pg. x)

Without the staff to collect and analyze survey data, statistical agencies are usually forced to improvise, guessing the size of the economy from population figures, which are themselves extrapolated from censuses that are decades-old.[3]

To name one prominent example, Nigeria (thought to be the world’s sixth-largest country) has not conducted a national census since 2006. With almost twenty years since the last data collection, the population figure of 230,842,743 that you can currently pull from Wikipedia is a quixotic monument to false precision.

Indeed, it would not be a great exaggeration to say that we have never known how large Nigeria is. The first census after independence, in 1962, was marred by regional infighting, because of its role in determining federal spending and assembly seats. The initial results, which reduced the population share of the politically dominant north, were rejected by the Northern People’s Congress. A redo in 1963 was also rejected as fraudulent—this time largely by southern politicians. The 1973 census was largely rejected as illegitimate; the 1980s and 1990s saw no censuses under military dictatorship; and the 2001 census produced an implausibly low number.[4] Even in the relatively successful 2006 attempt, census enumerators were attacked and killed, and the results were still disputed.

Many African states are failing at the basic task of knowing how many people live in their borders—let alone accurately measuring their economic activity. The vast, unobserved informal sector (which includes subsistence farming, and something like 60% of working people) is usually estimated just as a direct function of population.[5] Lacking direct harvest yields, estimates of agricultural output are often produced using FAO models based on planting-season rainfall data.[6] Even the minimal task of measuring the goods traveling across borders—in theory, the easiest thing for a sovereign state to accomplish—is occasionally beyond the reach of statistical agencies. Until 2008, landlocked Uganda only collected trade data on goods that eventually passed through the Kenyan port of Mombasa, ignoring the four other countries on its borders.[7]

In the absence of good underlying data, the prevailing approach for GDP in developing Africa can be summarized as:

Income estimates… derived by multiplying up per capita averages of doubtful accuracy by population estimates equally subject to error. (p. 39)

This is not the fault of the statisticians, who are doing the best they can with shoestring budgets. It is because of their work that we have estimates at all for what are some of the poorest countries in the world—and the most vulnerable. Earlier this year, Oxfam estimated that 84 out of Zambia’s 114 districts were hit by drought, leaving around a million farming households at risk of hunger. We are left with imprecise statements ("around a million”) and poor targeting of those in need, precisely because of our statistical ignorance.

I’m haunted by the words of the lone Zambian statistician, sitting in his empty office, who asks Jerven plaintively: “What happens if I disappear?”

A Marginal Proposal

Poor Numbers came out in 2013, attracting a wave of scholarly and policy attention (including by Bill Gates). Once you’ve heard its arguments, it’s virtually impossible to look at a GDP statistic the same way again.

But what actual progress has been made in the statistical capacity of nations?

Seemingly, not much. In late 2014, perhaps in response to Jerven’s book, the World Bank relaunched its website for its Statistical Capacity Indicator—a metric on a 0-100 scale which scores countries based on the strength of their “Methodology”, “Source Data”, and “Periodicity and Timeliness”. But even by this clunky internal metric, progress has been glacially slow: in 2004, the average score for African countries was 58.2; in 2019, it was just 61.4.

Moreover, over six years of an economics PhD, I have never heard of any economist using this statistic. Poor Numbers is well-cited and well-read (at least by Africa specialists), but its lessons about the fundamental unreliability of statistics have largely not been absorbed in how we actually do economics.[8] A separate data series with a clunky, hard-to-interpret scale is simply not going to cut it. The problem, I think, is that we economists need a tool that is internal to our statistical and quantitative frameworks.

And we actually have one: margins of error.

Strangely, economists pay a great deal of attention to statistical error when it comes to causal inference or survey data they collect themselves—but very little when it comes to national statistics. The presumption is the one with which I began this piece—and I shared, until fairly recently—that somewhere down the line, surely somebody has to have checked the work on something as important as GDP.

Here’s a simple bounding exercise. Remember Nigeria’s 89% GDP revision? If we assumed that Nigeria’s GDP estimate was exactly correct every other year since 1960, the implied margin of error of GDP would be around 20%.[9] This aligns roughly with the pioneering work of the statistician Derek Blades, who estimated in 1980 that the errors in aggregate GDP for Botswana, Kenya, Malawi, Nigeria, and Tanzania ranged from 20-35%, and the errors in GDP growth rates were at least 3%.[10] (The correlation of year-to-year GDP estimates helps bring down the margin of error when computing growth rates.)

I am writing this in November 2024, not long after the bulk of the news-reading population of the United States just got a crash course in statistical literacy. (There will be just enough time in four years for everyone to forget again.) But hopefully we’ve learned enough to realize that, with a 3% margin of error, empirical arguments about policies that might raise GDP growth by around 0.5% are—like Kamala Harris’s 2-point polling lead in Pennsylvania—largely academic. We can recommend good policies on the merits, but their growth effects are essentially undetectable with our crude instruments.

Putting this on here as a reminder about statistical humility. Yes, it hurts me to see too. Source.

Problems of statistical volatility of macro data extend, by the way, to the recent past of developed countries. A famous paper on the US by Christina Romer shows that much of the perceived postwar fall in unemployment volatility may simply be a statistical artifact—applying the old techniques to new data yielded far more volatile series. Developing countries and the developed countries of the past exist on a continuum of statistical precision.

The far-and-away first-best option, of course, would be for African countries to sustainably develop their indigenous statistical capacities. (Never forget that large-scale sample surveying was an invention of the Global South!) But a second-best approach would be for researchers and policymakers to begin developing better estimates of the statistical error of GDP. An excellent habit would be to mimic political polling, and to start printing margins of error alongside GDP growth estimates. Seeing that Nigeria posted GDP growth of 2.9% in 2023 feels substantially different from seeing that Nigeria posted GDP growth of 2.9 ± 5%. The former is a reified certainty; the latter, a statistical estimate.

The world is fiendishly complicated: fuzzy, gooey, hairy, spiky, oozing with contradiction. The job of statistical measurement is to squish all this unimaginable complexity down to dimensions that human beings can manage. This exercise is never value-free, nor one immune to error. I haven’t even gotten into all the thorny ideological complications of GDP; proposing margins of error is a purely internal statistical critique—a modest first step. Our goal should be a social scientific approach that doesn’t ignore this messiness, but is honest about how the data is made—one with a little less certainty and a little more epistemic humility—one that, when confronted with a graph of a particularly dubious GDP estimate, always remembers to draw out the missing confidence intervals with the mind’s eye.

  1. ^

    Jerven, pg. 70.

  2. ^

    Strictly speaking, these are rebasings, and earlier series were also revised. But the fact remains that, after a year, your estimate of the economy turned out to be 50% wrong.

  3. ^

    Jerven, pg. 56.

  1. ^

    The final estimate was 88 million; the World Bank rejected this estimate as too low and published 99.9 million instead.

  2. ^

    Jerven, pg. 14.

  3. ^

    Jerven, pg. 79.

  4. ^

    Jerven, pg. 51.

  5. ^

    An exception is the growing use of nightlight data.

  6. ^

    I’m assuming, from the position of sanity, that Nigeria’s economy did not actually move by 90% in a single year.

  7. ^

    Derek Blades, “What Do We Know about Levels and Growth of Output in Developing Countries? A Critical Analysis with Special Reference to Africa,” in Economic Growth and Resources: Proceedings of the Fifth World Congress, International Economic Association, Tokyo vol. 2, Trends and Factors, pgs. 60–77.

Show all footnotes
Comments3


Sorted by Click to highlight new comments since:

Wow thanks so much this is fantastic.

I'm not an economist but I resonate with this- guilty as charged. I usually assume national figures are kind of the gospel truth and now that I think about it I'm not sure why...

"Strangely, economists pay a great deal of attention to statistical error when it comes to causal inference or survey data they collect themselves—but very little when it comes to national statistics."

I've also been interested in economic growth as global health EA funding looks to increasingly be funneled into ways to improve this. Given that GDP data is so poor how on earth can we figure out what interventions might reliably improve growth in poor countries?

Executive summary: GDP statistics from developing countries, particularly in Africa, are far less reliable than commonly assumed due to severe data collection limitations and statistical capacity constraints, suggesting we should treat these numbers with much greater uncertainty and explicit error margins.

Key points:

  1. Statistical agencies in many African countries lack basic capacity to collect data, often relying on outdated censuses and crude estimations
  2. Major GDP revisions (e.g., Nigeria's 89% overnight increase in 2013) reveal the scale of potential measurement error
  3. Informal sectors (≈60% of working people) are typically estimated as a simple function of population, which itself is often poorly measured
  4. Statistical error margins for developing country GDP are estimated at 20-35%, with growth rate errors of at least 3%
  5. Recommendation: GDP statistics should include explicit margins of error (e.g., "2.9% ± 5%") to better reflect uncertainty

 

 

This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.

Curated and popular this week
LintzA
 ·  · 15m read
 · 
Cross-posted to Lesswrong Introduction Several developments over the past few months should cause you to re-evaluate what you are doing. These include: 1. Updates toward short timelines 2. The Trump presidency 3. The o1 (inference-time compute scaling) paradigm 4. Deepseek 5. Stargate/AI datacenter spending 6. Increased internal deployment 7. Absence of AI x-risk/safety considerations in mainstream AI discourse Taken together, these are enough to render many existing AI governance strategies obsolete (and probably some technical safety strategies too). There's a good chance we're entering crunch time and that should absolutely affect your theory of change and what you plan to work on. In this piece I try to give a quick summary of these developments and think through the broader implications these have for AI safety. At the end of the piece I give some quick initial thoughts on how these developments affect what safety-concerned folks should be prioritizing. These are early days and I expect many of my takes will shift, look forward to discussing in the comments!  Implications of recent developments Updates toward short timelines There’s general agreement that timelines are likely to be far shorter than most expected. Both Sam Altman and Dario Amodei have recently said they expect AGI within the next 3 years. Anecdotally, nearly everyone I know or have heard of who was expecting longer timelines has updated significantly toward short timelines (<5 years). E.g. Ajeya’s median estimate is that 99% of fully-remote jobs will be automatable in roughly 6-8 years, 5+ years earlier than her 2023 estimate. On a quick look, prediction markets seem to have shifted to short timelines (e.g. Metaculus[1] & Manifold appear to have roughly 2030 median timelines to AGI, though haven’t moved dramatically in recent months). We’ve consistently seen performance on benchmarks far exceed what most predicted. Most recently, Epoch was surprised to see OpenAI’s o3 model achi
Dr Kassim
 ·  · 4m read
 · 
Hey everyone, I’ve been going through the EA Introductory Program, and I have to admit some of these ideas make sense, but others leave me with more questions than answers. I’m trying to wrap my head around certain core EA principles, and the more I think about them, the more I wonder: Am I misunderstanding, or are there blind spots in EA’s approach? I’d really love to hear what others think. Maybe you can help me clarify some of my doubts. Or maybe you share the same reservations? Let’s talk. Cause Prioritization. Does It Ignore Political and Social Reality? EA focuses on doing the most good per dollar, which makes sense in theory. But does it hold up when you apply it to real world contexts especially in countries like Uganda? Take malaria prevention. It’s a top EA cause because it’s highly cost effective $5,000 can save a life through bed nets (GiveWell, 2023). But what happens when government corruption or instability disrupts these programs? The Global Fund scandal in Uganda saw $1.6 million in malaria aid mismanaged (Global Fund Audit Report, 2016). If money isn’t reaching the people it’s meant to help, is it really the best use of resources? And what about leadership changes? Policies shift unpredictably here. A national animal welfare initiative I supported lost momentum when political priorities changed. How does EA factor in these uncertainties when prioritizing causes? It feels like EA assumes a stable world where money always achieves the intended impact. But what if that’s not the world we live in? Long termism. A Luxury When the Present Is in Crisis? I get why long termists argue that future people matter. But should we really prioritize them over people suffering today? Long termism tells us that existential risks like AI could wipe out trillions of future lives. But in Uganda, we’re losing lives now—1,500+ die from rabies annually (WHO, 2021), and 41% of children suffer from stunting due to malnutrition (UNICEF, 2022). These are preventable d
Rory Fenton
 ·  · 6m read
 · 
Cross-posted from my blog. Contrary to my carefully crafted brand as a weak nerd, I go to a local CrossFit gym a few times a week. Every year, the gym raises funds for a scholarship for teens from lower-income families to attend their summer camp program. I don’t know how many Crossfit-interested low-income teens there are in my small town, but I’ll guess there are perhaps 2 of them who would benefit from the scholarship. After all, CrossFit is pretty niche, and the town is small. Helping youngsters get swole in the Pacific Northwest is not exactly as cost-effective as preventing malaria in Malawi. But I notice I feel drawn to supporting the scholarship anyway. Every time it pops in my head I think, “My money could fully solve this problem”. The camp only costs a few hundred dollars per kid and if there are just 2 kids who need support, I could give $500 and there would no longer be teenagers in my town who want to go to a CrossFit summer camp but can’t. Thanks to me, the hero, this problem would be entirely solved. 100%. That is not how most nonprofit work feels to me. You are only ever making small dents in important problems I want to work on big problems. Global poverty. Malaria. Everyone not suddenly dying. But if I’m honest, what I really want is to solve those problems. Me, personally, solve them. This is a continued source of frustration and sadness because I absolutely cannot solve those problems. Consider what else my $500 CrossFit scholarship might do: * I want to save lives, and USAID suddenly stops giving $7 billion a year to PEPFAR. So I give $500 to the Rapid Response Fund. My donation solves 0.000001% of the problem and I feel like I have failed. * I want to solve climate change, and getting to net zero will require stopping or removing emissions of 1,500 billion tons of carbon dioxide. I give $500 to a policy nonprofit that reduces emissions, in expectation, by 50 tons. My donation solves 0.000000003% of the problem and I feel like I have f