Michael Spagat is a professor of economics at Royal
Holloway, University of London. He has
written extensively upon survey work done in Iraq, specifically on the two Lancet
reports on estimated deaths in Iraq following the 2003 invasion, and two
other studies done on child mortality rates in the country during the 1990s
sanctions period. He’s also studied the Iraq Living Conditions Survey, the Iraq
Family Health Survey, and the 2013 PLOS Medicine Journal Survey on Iraqi
fatalities. Professor Spagat has found anomalies with almost all of these
papers that undermine their findings. Unfortunately, most of his work is only
known in academia. What follows is an interview with Prof. Spagat exploring his
critiques of these famous papers.
1. In October
2013 a new survey was released on estimated deaths in Iraq after 2003, which
was published in the PLOS Medicine Journal. One of the authors, Gilbert Burnham
was also a co-author of the two Lancet reports on Iraqi fatalities. The press reported that the new survey
believed 500,000 Iraqis had died as a result of the war since the U.S. invasion,
but its estimate for violent adult deaths was actually 132,000. The headline estimate was for what they
called “excess deaths” which they view as caused by the war, although many of
these are non-violent. What did you
think of that new paper?
The Hagopian
et al. report (PLOS) did two separate surveys simultaneously. One was a
sibling survey. The other was the more typical household survey. These are two
different methods to cut up the population into mutually exclusive groups that
exhaust the whole. So you either have a bunch of households that, hopefully,
don’t overlap or groups of individuals matched with their siblings. The
traditional household survey, which has pretty much monopolized the media
attention of the study, didn’t give an estimate for violent deaths. The central
violent-death estimate for the sibling survey, 132,000 non-elderly adults, is a
bit below the Iraq Body Count (IBC)
number for civilians plus combatants.
(IBC focuses on civilian deaths but they now also publish counts of
combatants killed. The IBC total for the Hagopian et al. period is around
160,000, including children and the elderly.)
To their credit, the PLOS authors do post
their data online, so you can do your own analysis. I’ve taken the
opportunity to investigate the household survey data together with a Royal Holloway
PhD student, Stijn Van Weezel. The data yield an estimate of around 200,000
violent war-related deaths, i.e., about 40,000 higher than the IBC number for
civilians plus combatants (160,000).
Hagopian et al. have stressed what they call “excess deaths”
rather than violent deaths. Excess deaths are meant to be both violent and
non-violent deaths that have been caused by the war. Unfortunately, the
excess-death concept is a pretty squishy one based on a poorly designed counter
factual exercise. The calculation hinges on constructing a death rate that
would have occurred, theoretically, if there had never been a war. This is
something we can never measure directly. We’d like to run history twice, once
as it actually happened with a war, and once without the war. We would then
measure death rates under both scenarios and call the difference between the
two the excess death rate caused by the war. Obviously, we can’t ever do this
exercise or even anything that resembles it.
There are, nevertheless, a couple of approaches that have
been tried with the hope of pinpointing the causal effect of a war on death
rates. The most common one is to measure both a pre-war death rate and a
during-war death rate and to then assume that the difference between the two is
caused by the war and nothing else.
Unfortunately, such a before-and-after exercise is
problematic. It pretty much boils down to saying that because “b” comes after
“a”, then “a” has caused “b.” This is a known logical fallacy. In the case of
Iraq a lot that can affect death rates has happened since March 2003. The start
of the war is the most obvious and dramatic factor but it is only one of these
things. In conflict situations there can be an event like a drought that
directly causes deaths but also exacerbates tensions, leading eventually to
war. If you attribute all increased deaths to just the war then you’re missing
the fact that there was also a drought that was also probably causing deaths. You
wind up exaggerating the number or deaths caused by the war.
In fact, the standard excess-deaths concept leads to an
interesting conundrum when combined with an interesting fact exposed in the
next-to-latest Human
Security Report; in most countries child mortality rates decline during
armed conflict (chapter 6). So if you believe the usual excess-death causality
story then you’re forced to conclude that many conflicts actually save the
lives of many children. Of course, the idea of wars savings lives is pretty
hard to swallow. A much more sensible understanding is that there are a variety
of factors that determine child deaths and that in many cases the factors that
save the lives of children are stronger than the negative effects that conflict
has on child mortality.
Anyway, Hagopian et al. didn’t bother much with the above
reflections but, rather, charged straight in and estimated 400,000 excess
deaths. However, they have quite a crazy confidence interval around this estimate
- 50,000 to 750,000. So even if you accept their notion of excess deaths at
face value you still have to say that this is not a very informative
estimate.
My student Stijn and I are taking a different tack in our
analysis. We say that if the war is causing non-violent death rates to increase
then you would expect non-violent deaths to increase more in the violent parts
of Iraq then they do in the non-violent parts of Iraq. To the contrary, we find
this just isn’t so. At least in our preliminary analysis, there seems to be
very little correlation between violence levels and changes in non-violent
death rates. This should make us wonder whether there is any reality behind the
excess deaths claims that have been based on this Iraq survey. In fact, we should
question the conventional excess-deaths idea in general.
Nevertheless, the authors and the media have stressed this
excess death estimate while obscuring the great uncertainty that surrounds it. Remember,
the estimate is 400,000, give or take 350,000. Yet somehow the authors were
able to talk
that up to 500,000 deaths and assert this number as a sort of minimum. Thus,
the uncertainty was expunged, and then there was inflation from 400,000, for
which there is some supporting data, to 500,000, which is more of a speculation
than a finding. Obviously 500,000 is a media friendly number - people like the
idea of half a million.
Although the household survey of Hagopian et al. tells us
little about excess non-violent deaths it does bring to bear some useful evidence
about violent deaths. The new study suggests that the full number of violent
deaths in the Iraq war is a bit higher than the IBC number (200,000 versus 160,000
civilians and combatants). Much other evidence points in this direction but
such an understanding has not been universal. Strangely, though, Hagopian et
al. seem to believe their findings are at odds with IBC, perhaps because they
are unclear in their minds about the distinction between excess deaths and
violent deaths.
However, the new survey completely flies in the face Burnham
et al. study (second Lancet). In fact, another problem with the
media campaign surrounding the PLOS report was that Gilbert Burnham tried
to claim their new study is consistent with the Burnham et al. one. It isn’t.
I have looked a little bit at just the time period that
was covered by the Burnham et al. 2006 study that had found 600,000 violent
deaths. The Hagopian et al. data will come in at around 100,000 deaths for that
same time period. So there is a factor-of-six discrepancy between the two. To
say these are consistent with each other is really farfetched.
Comparing the way the Hagopian et al. survey has been
presented, and the way the Roberts
et al. 2004 Lancet survey was presented is also interesting. In both
cases you have a central estimate of excess deaths with almost comical uncertainty
surrounding it. For Roberts et al. this was an estimate of 98,000 with a
confidence interval of 8,000 to 194,000. Then there is a public relations
campaign that erases the uncertainty, leaving behind just the central estimate
- 100,000 for Roberts et al. and 400,000 for Hagopian et al. Finally, the
central estimate
is promoted as a sort of minimum, with the “likely” number being even
higher than their central estimate. Actually, Hagopian et al. went one step
further, inflating up by another 100,000 before declaring a minimum of 500,000.
2. When you read
the PLOS report it seemed like they definitely recognized all the criticisms of
the 2006 Lancet paper, because it said they did all these steps to avoid
those problems. Then when they went to the media they said there was no problem
with the Lancet paper at all, and our new report backs it up. It seemed
like what they said to the press, and what they actually wrote were two
different things.
Right, I completely agree with that. Of course, if the
numbers had come out similar between the two survey then they would have, said
“look, Burnham et al. was criticized for all these reasons. We fixed all of
those things, but it didn’t make a difference, so the criticism was not
important.” In fact, what happened was that the Hagopian et al. report fixed
most of those things and then the numbers plummeted. Unfortunately, the authors
don’t yet seem willing to come to terms with this fact in the public dialogue.
3. Let’s turn to the two Lancet reports. One of
your main critiques of the 2006 Lancet report was what you called the “main
street bias.” Could you explain what that was and what you thought were the
major problems with that Lancet paper?
The main street bias critique is that the 06 Lancet
surveyed main thoroughfares where there would be higher likelihood of violence and
thus overestimate for deaths in Iraq (Defence
And Peace Economics)
That was among the critiques, and that was the first one I
made together with some other people. That was just from reading the
description of the sampling method Burnham et al. wrote in the paper. This
originally arose in a discussion that Neil Johnson and I were having with
Burnham mediated by a reporter at Science magazine. He got our input,
and then he forwarded that on to Burnham, and then Burnham made a response that
came back to us….and so on and so forth. We read carefully what the report
said, which was that the people who did the interviews selected a main street
at random from a list of main streets. Then interviewers selected a random
cross street to that main street, and did their interviews along that random cross
street. We argued that such places would tend to be more violent than average,
so sampling using that method would tend to overestimate.
At some point in that discussion (that eventually turned
into an
article for Science magazine) Burnham said they actually didn’t do
what was described in the Lancet paper.
He said there was a sentence that had been cut from the paper at the
demands of the editors to save space, although the paper was actually well
below the maximum length for a paper in the Lancet, and there was a lot
of text left in the final paper that was certainly more superfluous than
ensuring an accurate description of their sampling methodology. That last
sentence supposedly said that if there were streets that were not cross streets
to main streets those were included in the procedure as well. I thought it
really made no sense to have such a procedure. You take a main street at
random. Then you choose a cross street to the main street, but if there are
streets that aren’t cross streets to main streets, they also included those in
some unspecified way. Of course, pretty much wherever you go there are going to
be streets that aren’t cross streets to main streets, so why do you even bother
to select a main street and a cross street? Inevitably you’ll just find out
that there are other kinds of streets as well so you’ll then have to figure out
a way to include these too. And how can you operate without a well-defined procedure
for selecting streets? That was the moment I realized that something weird was
going on with this survey. It seemed that Burnham didn’t even know what his
field teams were doing. It also seemed like he was willing to change arguments
on the fly without knowing what he was talking about.
We informed the Lancet that we had been told that
the authors hadn’t followed their published sampling procedures so maybe there
should be a correction, but there was never any correction.
Neil Johnson and some other colleagues still wanted to
pursue the logic of what would be implied if the field teams actually followed
the procedures they claimed to have followed. We
worked this out in more detail and developed a plausible range of assumptions
that suggested that the impact of following main-street-biased procedures could
potentially be quite large. We suggested likely scenarios that could lead to
overestimations by even a factor of three. It seems that in practice Burnham et
al. overestimated by a factor of six or so. Perhaps main-street bias can
explain a good chunk of this overestimation. I don’t think it really has the
potential to explain all of it. At the end of the day I’m not confident that
main-street-bias explains much of anything given that we have a glaring
ambiguity about what actually happened on the ground in this survey. Burnham
says they didn’t actually do what they claimed to have done in the published
paper, but he has never specified a viable alternative. Where does that leave
us in the end?
4. Do you have
any other critiques of the 2006 Lancet?
There are many others. For example, there's a long sad
story having to do with the trend in that survey. If you go back on the Lancet
website there’s a
podcast that was put out right when the Burnham et al. study was published.
Gilbert Burnham was asked by an interviewer how he can be confident in these
phenomenally high numbers that are so far out of line with other sources. His
answer was that he is very confident, because although the numbers are
considerably higher than the Iraq Body Count numbers the trends match IBC’s
trends quite closely. So that was the confirmation - they got the same trends as
IBC.
Then there’s a graph in the paper (figure 4) where they
compare the trends from IBC and their own trends. I never understood that graph
until there were letters in the Lancet about it. One
of the authors was Jon Pederson who was the main person behind the Iraq Living
Conditions survey, and Josh
Dougherty of IBC also had a letter about this. There were many flaws with
the graph, but a crucial one was how they compared the trends. They have three
time periods, each of 13 months. Their own (Burnham et al.) figures are just
what you’d expect – one for the first 13 months, one for the second and one for
the third. But the IBC figures are
cumulative. So the first IBC figure covers a 13-month period just like the comparable
Burnham et al. figure. However, the second IBC figure covers 26 months and is
compared with a 13-month Burnham et al. figure. The third IBC figure covers 39
months and is compared with a 13-month Burnham et al. figure. In short, they
present a graph comparing cumulative figures with non-cumulative figures! And
do you know what? The IBC cumulative figures sky rocket up just like the
non-cumulative Burnham et al. figures. And that’s the confirmation that makes
them so confident in their outlying numbers. However, if you compare like with
like you see that the Burnham et al. numbers rise much faster than IBC’s, and
follow a different pattern.
There was never any follow up to that interview. If you
ever interview Gilbert Burnham you might want to ask him: “now that the basis
for your confidence in your numbers has been exposed as false will you now be
changing your position?”
5. There were
two others surveys, the Iraq Living Conditions Survey and the Iraq Family
Health Survey. They had radically
different findings than the Lancet surveys had. A lot of people compared
those, so what were the differences between those other surveys and the two Lancet
ones?
The Iraq Family Health
Survey (IFHS) covered the same time frame as the 2006 Burnham et al. study.
They published a central estimate of 150,000 violent deaths. That would compare
to the 600,000 in the Burnham et al., so those were apart by a factor of four. That
said, the people who did the IFHS really went into contortions to try to raise
their number up as high as possible, so the real distance is actually greater
than a factor of four.
The main estimate in the IFHS report was calculated in a
different way than is normal. If they had done the usual thing their estimate
would have come out around 100,000 or even 80,000. So they did two things to
push their number upward. One was to adjust for clusters that had been selected
in their randomization procedures, but where they had not been able to complete
their interviews because they considered those places too dangerous to enter at
the time the survey was done. So they applied an adjustment that had the effect
of raising their estimate from about 80,000 up to 100,000. That was not a crazy
thing to do although it was quite a dramatic adjustment. It had the implication
that the clusters in Baghdad where they hadn’t managed to interview were about
four times as violent as the ones where they did. That is a rather bold
assumption to make, but leave that aside.
Next the IFHS did an arbitrary fudge upward of an
additional 50%. They basically just declared without evidence that surveys tend
to under estimate violent deaths. So they raised their number from 100,000 to
150,000 with hardly an attempt at justification.
I would argue that in reality the IFHS found around 80,000
to 100,000 - take your pick.
Even if you accept the fudge up to 150,000 the IFHS is
still completely out of line with the Burnham et al. survey, and not just for
the overall number. For example, the Burnham et al. survey had a few
governorates with incredibly high numbers that aren’t at all supported by other
evidence. Burnham et al. also had a dramatic upward trend that isn’t matched by
the IFHS or IBC or any kind of other measurement that’s been taken there.
The Iraq Living
Conditions Survey really only covered slightly more than the first year of
the war. The first Lancet survey by Roberts et al. covered a bit more,
about the first 18 months, so they’re not exactly comparable. The best way to
think about the first Lancet survey is that it produced virtually no
information. They had an estimate of 98,000 excess deaths with a confidence
interval of 8,000 to 194,000. Right off the bat it’s just kind of useless
because estimates with that kind of uncertainty tell you nothing. They didn’t
actually calculate the confidence interval correctly either. If it is
calculated correctly it comes out even wider than what was published, although
in the end this probably doesn’t even matter.
The Iraq Living Conditions Survey didn’t estimate excess
deaths so it is a little bit hard to compare it with Roberts et al. However,
you can sort of bridge the gap because there was some data released on Roberts
et al., and you can use it to get rid of the deaths after the time period the
Iraq Living Conditions Survey was finished. Then you need to focus just on
violent deaths. Roberts et al. then has about 70% more violent deaths than the
Iraq Living Conditions did. They are not really compatible with one another,
but they’re not wildly out of line either. It’s the Burnham et al. survey that
is seriously at odds with everything else.
I prefer to focus more on violent deaths. Certainly if
you’re trying to compare all of the different sources you have to do this. In
some sense you can say that all the excess deaths estimates are kind of
compatible with one another because the confidence intervals are so wide that
the only reasonable conclusion is that we’ve hardly got any idea about excess
deaths, even if you accept that the whole notion of excess deaths as defined in
this paper makes sense.
6. You also had
problems with how they estimated their excess deaths. You had an article “The Iraq
Sanctions Myth” that was talking about a letter published in the Lancet by
Sarah Zaidi in 1995 that claimed that half a million children, died due to
sanctions. The other was “Sanctions
and Childhood Mortality in Iraq” by Mohamed Ali and Iqbal Shah that was
also in Lancet in 2000. Subsequent work extrapolated from this work and
found 400,000-500,000 excess child deaths in Iraq from 1990-1998. You said that
there were problems with their estimates, so a lot of the subsequent surveys
were using problematic surveys from before to figure out what the death rate
was in Iraq before the war to make their estimate for the excess deaths
afterward as well right?
Comparison of
survey estimates on child mortality in Iraq during and after the sanctions
period (Pacific Standard Magazine)
To answer your last question first, sanctions-era
estimates have not been carried forward to feed into excess death estimates
made during the war. All the estimates discussed earlier in the interview have
used their own surveys to estimate pre-war death rates.
I’m also very critical of the sanctions era estimates of
how many children were supposedly killed due to sanctions. These numbers were
first based on a survey done and later
retracted by Sarah Zaidi. She subcontracted her field work to some
government workers in Iraq and, on the basis of the data they gathered,
estimated half a million excess child deaths. This number was then cited by Leslie Stahl in her famous
interview with Madeleine Albright. Stahl actually won two awards for that
interview including an Emmy, but the basis for it turned out to be a survey
that was later retracted. The story is that some people found anomalies in the
survey. So Zaidi, to her credit, went to Baghdad herself and re-interviewed
many of the same households. She found that a lot of the deaths the Iraqi
surveyors had reported simply weren’t there. The data were just wrong so this
calculation falls even before you question that whole methodology of looking at
pre versus post as I do earlier in this interview.
However, the critique of the excess-deaths concept
certainly does apply to child deaths in Iraq in the 1990’s. It is not convincing
to assume that any differences between pre and post child death rates are due
entirely to sanctions. There was so much going on in Iraq besides just
sanctions. There was the first Gulf War, there were uprisings both in the south
and the north that were suppressed, etc. To the extent that there was an
increase in child death rates there could have been a lot of causes besides
just sanctions. However, in this case you can just leave that whole critique
aside, because the basic measurement was wrong.
Shortly after the Zaidi survey was retracted, UNICEF
did a new survey, again subcontracting the fieldwork to Iraqi government
officials. They found basically the same thing that Zaidi had found initially,
which should have raised red flags straight away. One person goes in and
conducts a survey that pretty clearly was manipulated. I don’t think this was
Zaidi’s fault and I’ve always praised her for correcting the record, which is
rare. However, if the corrected record is true then why is someone else finding
something that would completely contradict this newly corrected record? You
might also ask why, if we already saw Iraqi government workers manipulate one survey,
does UNICEF then create an opportunity for the same thing to happen again? In
this particular case we also need to consider that it was a central policy of
the Iraqi government to convince the outside world to drop sanctions against
it. One of the arguments they were using was that sanctions were hurting Iraqi
civilians, in particular Iraqi children. Why then give that government an
opportunity to do a UN-sponsored survey to reinforce their foreign policy
position? How confident can you be in these results?
So UNICEF got similar results to the ones that Zaidi had
just retracted. And those UNICEF results remained the conventional wisdom for
several years, going right up to the beginning of the 2003 war and beyond. It
was widely believed that sanctions were responsible for the deaths of hundreds
of thousands of Iraqi children, but the problem is that since then there have
been four further surveys that have all failed to find the massive and
sustained spike in the child mortality rate in the 1990’s that Zaidi had found
and lost and that the UNICEF survey had supposedly rediscovered. At this point
there’s so much evidence piled up against the UNICEF survey that I don’t think
a rational individual can believe any more in the sanctions-excess-child-death
story that we were sold before the war. You don’t have to even question the
excess death concept to grasp this point. All you have to do is look at what
all the surveys find. In order to get this massive number of excess deaths you
have to have a huge and sustained spike in the child death rate after the
sanctions come in, and this simply doesn’t happen in any of the surveys since
the UNICEF one from the late '90s.
7. I want to try
to address some of the arguments made by people who defend the two Lancet surveys. Some of the most common ones that I’ve heard
were that it was published in the Lancet that is a respected journal, it
was peer reviewed, and that they did it during a war so you’re never going to
get perfect work during that time. Given
all that people say that others shouldn’t be so critical of the two
surveys. What do you think of that kind
of defense?
First of all, saying that something has to be right or is
probably right because it has been peer reviewed is quite a weak defense. Peer
review is a good thing, and it is a strength of scientific journals that there
is that level of scrutiny, but if you look at the list of scientific claims
that have turned out to be wrong and that have been published in peer reviewed
journals….well…the list just goes on and on and on. Publishing in a peer
reviewed journal is no guarantee that something is right. Some of the people
who do the referee reports are more conscientious than others. In almost no
cases does refereeing ever include an element of replication. Often referees
don’t even know enough about literature cited to judge whether claims about the
current state of knowledge are accurate or otherwise. Mostly people just assume
what they’re being told by the authors of the paper is correct and valid. Peer
review is better than no peer review, but it hardly guarantees that something
is going to be correct. (Let’s not
forget the graph discussed earlier in this interview which survived the Lancet’s
peer review procedures.)
Journal peer review is just the beginning of a long peer
review process. Thinking that journal peer review is the end of this process is
a serious misunderstanding. Peer review is an ongoing thing. It is not something
that ends with publication. Everything in science is potentially up for grabs,
and people are always free to question. Anyone might come up with valid
criticisms.
If you look at Burnham et al. there have been a number of
peer reviewed articles that have critiqued it, and said it is wrong. So if you
think peer review has to always be correct then you’re immediately in a logical
conundrum because you’ve got peer reviewed articles saying opposite things. What
do you do now?
As for the Lancet, as a scientific journal over the
last decade or more it has had quite a spotty record. Much of what it has
published has turned out to be wrong. The Lancet is not considered one
of the more reliable scientific journals and it has a reputation for
sensationalism. You have to remember that at the end of the day the Lancet is
a profit making operation. It is chockablock full of advertising. Library
subscriptions are extremely expensive. It brings in millions of pounds of
revenue. Sensationalism sells, so by some metric Richard Horton has been a
successful journal editor, because he’s gotten a lot of media attention. It’s
good for subscriptions, good for advertising, but articles in the Lancet
still need to be scrutinized on a case-by-case basis, as is the case with any
other journal.
I’m happy to give people credit for doing difficult
research in war zones. And I’m happy to admire the courage of people who do
dangerous field work. But doing courageous field work doesn’t make your
findings correct and we shouldn’t accept false claims just because someone had
the guts to go out in the field and gather data. Science is a ruthless process.
We have to seek the truth. Courage is not an adequate rebuttal to being
wrong.
SOURCES
Boseley, Sarah, “UK scientists attack Lancet study
over death toll,” Guardian, 10/23/06
Burnham, Gilbert, Doocy, Shannon, Dzeng, Elizabeth, Lafta,
Riyadh, Roberts, Les, “The Human Cost of the War in Iraq, A Mortality Study,
2002-2006,” Bloomberg School of Public Health Johns Hopkins University, School
of Medicine Al Mustansiriya University, 9/26/06
Burnham, Gilbert, Lafta, Riyadh, Doocy, Shannon, Roberts,
Les, “Mortality after the 2003 invasion of Iraq: a cross-sectional cluster
sample survey,” The Lancet, 10/11/06
Giles, Jim, “Death toll in Iraq: survey team takes on its
critics,” Nature, 3/1/07
Johnson, Neil, Spagat, Michael, Gourley, Sean, Onnela,
Jukka-Pekka, and Reinert, Gesine, “Bias in Epidemiological Studies of Conflict
Mortality,” Journal of Peace Research, September 2008
Kaplan, Fred, “Number Crunching Taking another look at the
Lancet’s Iraq study,” Slate, 10/20/06
Onnela, J.-P., Johnson, N.F., Gourley, S., Reinert, G.,
and Spagat, M., “Sampling bias in systems with structural heterogeneity and
limited internal diffusion,” EPL, January 2009
Roberts, Les, Lafta, Riyahd, Garfield, Richard, Khudhairi,
Jamal, Burnham, Gilbert, “Mortality before and after the 2003 invasion of Iraq:
cluster sample survey,” The Lancet, 10/29/04
Spagat, Michael, “Ethical and Data-Integrity Problems In
The Second Lancet Survey of Mortality in Iraq,” Defense and Peace Economics,
February 2010
- “The Iraq Sanctions Myth,” Pacific Standard Magazine,
4/26/13
- “Mainstreaming an Outlier: The Quest to Corroborate the
Second Lancet Survey of Mortality in Iraq,” Department of Economics
Department, University of London, February 2009
11 comments:
Thanks for this Joel. An excellent interview with clear and well-articulated responses from Dr. Spagat. I’ve engaged here before on this ongoing refutation of the Lancet (and other) studies. I’m (again) left with the takeaway of … where do we go from here?
“Journal peer review is just the beginning of a long peer review process. Thinking that journal peer review is the end of this process is a serious misunderstanding. Peer review is an ongoing thing. It is not something that ends with publication.”
This is a critical point and one that cannot be emphasized enough. The methodological critiques put forward above should give pause to the certainty with which many express the ‘take-away’ numbers. The critique behind much of what Dr. Spagat outlines leads one to believe that not only were errors committed in study design and execution, as well as a subsequent failure of reviewers to find obvious errors; but, that the entire enterprise of the studies published was imbued with an agenda that made such a lack of rigour possible. Simply, the researchers of the studies critiqued were biased in their efforts. Assumably this was in an effort to find both the violent death totals as well as the ancillary ‘excess’ death totals to be egregious and therefore noteworthy.
The interlocked suffering of Iraqis - from the decades of authoritarian rule, the Iran-Iraq war, 1991 Gulf War, decade+ of sanctions, 2003 Iraq War, and subsequent Anglo-American occupation and redesign of Iraqi statehood - have seen undeniable misery and death visited on the peoples of Mesopotamia. I would not assume, but raise the prospect, that studies attributing large casualty and death totals to the abuses of the Ba’th regime and the hostilities of the Iran-Iraq war might also see revision.
This raises the question I have regarding politicization.
“… doing courageous field work doesn’t make your findings correct and we shouldn’t accept false claims just because someone had the guts to go out in the field and gather data. Science is a ruthless process. We have to seek the truth. Courage is not an adequate rebuttal to being wrong.”
You had me until the word “wrong” emerged. The flaws pointed out in the research undermine the truth claim(s) being made; they allow for a well-meaning person to respectfully disagree and not be persuaded that the death totals and especially the excess death totals (conceptually or in their empirical reality) are plausible - that the results of the research are “incorrect”.
The basis of the critique presented above by Dr. Spagat is one of scientific rigour - however, what is missing is the basis for that effort in a scholarly environment: an alternative theory/number/explanation. To be clear, this was not the venue for such output and Dr. Spagat is not the sole researcher responsible to provide such alternative(s) - however, one needs to emerge. Otherwise, this is solely an exercise in denial.
The central tenet of the sides engaged in this ‘debate’ is a disagreement over the casualty numbers discussed for three decades of Anglo-American political engagement with Iraq. Simply, it seems a debate over ‘who is blameworthy?’ and to what degree their policies are responsible for the humanitarian disaster represented by today’s Iraq.
Unless social scientists are willing to simply fob such declarations and outputs off onto historians of the distant future - alternatives should begin to accompany critique.
Thanks again for engaging in such a critical discussion. The Iraqi people deserve nothing less that the truth.
jmeasor,
If you're looking for someone to blame then you need to look beyond surveys on estimated deaths. That wasn't in the scope of their surveys really. There are dozens and dozens of books on Iraq that can be a better source for who was behind the violence and fighting
If you're looking for what is the best source for deaths Iraq Body Count is by far the most reliable and they are always working to expand the scope of their work.
Thanks Joel; I'm not looking for someone to blame - I'm suggesting that the framing of the 'Iraq question' in the English-speaking world is largely one of:
1. Iraq is a big mess and the mess was created by sanctions and the war(s), and
2. largely the Anglo-American policies that brought about sanctions and war(s) are to blame for the humanitarian disaster;
This fits within a "blame America" anti-Imperial guise generally and specifically is an effort to critique Anglo-American policy vis-a-vis first sanctions and then the 2003 invasion and occupation. The 'half-million Iraqi children dying due to sanctions' and the '500K-one million dying due to the impact(s) of the 2003 war' clearly add an exclamation point to such opposition to Anglo-American policies.
Having spoken publicly, in the media, and through my academic research I've been dealing with this framing for over fifteen years. For the sake of argument I'm accepting the above critique of the Lancet studies; moreover, I'm not insulting you or Dr. Spagat by insinuating that you are 'pro-war' or in support of the policies enunciated by the PMO in the UK or the past four U.S. administrations.
Rather, I'm asking, *if* the number of victims is dramatically lower than assumed by many in the public, where does that leave us? I'm assuming its not a simple binary of 'the war was good' as I don't think that is what you are suggesting. However, with the repeated critiques posted here regarding the higher death totals, the studies that brought them to the public's attention, and even the suggestion of mendacity on the part of their authors, I'm asking "where does this leave us?".
Dr. Spagat eloquently laid forth the scientific and methodological failings of the earlier studies. Are you aware - are you in favor? - of further studies to achieve a more plausible casualty number? Does the altered death total(s) alter our perceptions of Anglo-American policy? Adopting the scientific method as critique also suggests that it will be adopted wholly; meaning, that a new theory, study, or argument need replace the accepted wisdom.
I'm not attacking anyone - I'm simply suggesting that the 'meaning' of what occurred in Iraq will be a large (and heated!) debate over the coming decades.
Iraqis will come to their own conclusions, greatly impacted by what develops in their experiences' moving forward as much as the actions of the 1990s and post-2003 era.
It seems to my mind that having some accurate and agreed casualty totals and a nuanced view of the causation of the suffering is an important task of venues such as 'Musings on Iraq' - you have a global audience :)
All my best,
John
Hi John,
I think the debate about Iraq is largely over in the U.S. and England, but is still raging in Iraq.
I think if you look at Iraq Body Count you still get a sense of the huge loss of life that occurred after 03 even if it isn't as high as studies like the Lancet one.
The causes are multiple and not easy to break down. The U.S. invasion did cause the insurgency. This took many forms from nationalists to Islamists to foreign jihadists. The insurgency gained wide spread support amongst Sunnis because of the new sectarian identity that was being formed after the fall of Saddam. This was mainly based on the idea that the Shiite and Kurdish parties that took power were not really Iraqi but rather foreigners or controlled by foreigners. Sunnis came to see themselves as victims of this foreign plot which they believed wanted to destroy Iraq.
The foreign fighters and people like Zarqawi had similar views about Shiite being apostates and Kurds not being Arabs. Part of Zarqawi's plan was to start a sectarian war which would bring down the U.S. and the Iraqi government. Through his constant bombings of Shiites and the Samarra bombing in 06 he got the war, but he ended up dead and the Sunnis lost.
The U.S. is also to blame because of its lack of planning for post-war Iraq, its failure to develop a strategy to win the war until the Surge in 07, and then not really caring about Iraq after the withdrawal.
The Shiite and Kurdish parties also came into Iraq feeling like they were victims of Saddam and therefore had their own largely uncompromising agendas plus their own militias and sought to seize state power and shut out their rivals.
That mix of all those parties I would say are responsible for the continued violence in the country.
'... I’m also very critical of the sanctions era estimates of how many children were supposedly killed due to sanctions. These numbers were first based on a survey done and later retracted by Sarah Zaidi. She subcontracted her field work to some government workers in Iraq and, on the basis of the data they gathered, estimated half a million excess child deaths. This number was then cited by Leslie Stahl in her famous interview with Madeleine Albright. Stahl actually won two awards for that interview including an Emmy, but the basis for it turned out to be a survey that was later retracted. ...'
What Leslie Stahl got an Emmy and fame for was not estimating deaths or reporting an estimation, but for eliciting agreement from Madeleine Albright that such a level of mortality had occurred, followed by 'We think it is worth it.' In other words, for a moment, the face of a murderous imperialism was revealed. Many Americans seem to feel guilty about this sort of thing, assuage their guilt by shifting the blame onto their supposedly nefarious leaders, and reward those who manage the shift with fame and awards. The accuracy of the estimated deaths was unimportant; the confession of Albright, priceless.
I agree with Anarcissie. Albright believed the premise of the question and gave a shocking answer and that was revealing.
I'm not sure I agree that this was genius. It would be nice if Stahl would at least correct the record. People still watch that clip and it continues to spread a falsehood about sanctions in Iraq even while it reveals a truth about Madeleine Albright.
jmeasor, I think that making sense of anything has to begin with learning the truth. Some numbers may feel convenient but if they are wrong one needs to recognize this fact and move forward.
I didn't mention this in the interview but Tony Blair finds the sanctions-child-death numbers convenient and he cites them to justify the war, applying the same kind of excess death calculus that I criticize in the interview. Blair claims that the war has saved the lives of tens of thousand of Iraqi children because the child death rate is massively lower now than it was before the war (it is claimed). In fact, if the sanctions-era child death estimates were correct then Blair would have a point. But they are not correct and child mortality rates now are not, in fact, massively lower than they were under sanctions.
So the false claim that sanctions killed hundreds of thousands of Iraqi children whipsaws back and is used to justify the current war.
I also don't think one brings dignity to real individuals who have been killed by mixing them in with hundreds of thousands of fictitious people who have been estimated to have been killed using methods that can't be defended.
It's best to build meaning on top of truth. Certainly it's true that the invasion of Iraq has led to the deaths of a huge number of Iraqis. It's in the hundreds of thousands.
I stumbled across this blog post in 2017 while conducting historical research.
Some scholars conclude far fewer Jews died during the Holocaust than is often claimed. Does Mr. Spagat encourage further review of those materials or funding deeper study, since, as he states, "I don't think it brings dignity to real individuals who have been killed by mixing them in with hundreds of thousands of fictitious people who have been estimated to have been killed using methods that can't be defended" ?
Perhaps he would... but it would certainly not be my priority, because petty quibbling over such a massive crime against humanity is usually not driven by an honest concern for truth in history, but by a political and/or anti-Semitic agenda. And that is what I detect in Mr. Spagat's obsessive efforts to minimize Iraqi suffering here and elsewhere on the internet.
Mr Spagat claims inflated estimates of child deaths from the sanctions were used to justify the war on Iraq, as if to say research like his could potentially have spared the the Iraqis of being invaded on a false premise. That is disingenuous-- we all know that it was lies about WMD and Al Qaeda that were used to justify the war. The fact that Blair may have mentioned child mortality rates once or twice to bolster his case is irrelevant. Those child deaths were overwhelmingly cited by anti-war activists as a reason to end the sanctions and leave Iraq in peace.
Those who spent time living in Iraq in the 1990s, learned the language and got to know ordinary Iraqis, (yes I am one of those Arabists), would tell you the sanctions destroyed the Iraqi people. The word Assistant Secretary General Denis Halliday used after he quit the UN in protest was "genocide." The Iraqi people have suffered beyond measure from US policy and they continue to suffer up to this day. Mr. Spagat's herculean effort to downplay that is, at minimum, strange.
I think Spagat's work on both the PLOS and Lancet reports on fatalities in Iraq either under sanctions or during the war was to point out that the studies had so many problems that they were meaningless. If you want to report on the deaths going on in the country you need realistic estimates and these studies failed at that. You can be against the sanctions but you need realistic numbers to gauge their effects.
I see this as a key question going forward - similar to the debate(s) surrounding Vietnam - and (as above in 2014) I believe Spagat - and in supporting his research without critique or buffer vs the research published in The Lancet (and elsewhere) raises eyebrows because the academic debate did not debunk The Lancet methodology as definitively as you make it appear.
The methodology was sound, the data - as you rightly point out and highlight - was highly problematic. If we dismiss The Lancet study's methodology blithely we can also dismiss all casualty figures - which is why the WWII analogy by 'Berkley California'resonates.
This does not mean that the high figures of The Lancet should be dismissed entirely IMO. Perhaps one could see them as the highest end of the band for possible or plausible figures.
While the Iraq Body Count figures benefit from the vastly stronger data - it also can not be seen as definitive due to the overwhelming concerns with data collection.
The point here is to move beyond ad hominem and simplistic summaries and accept that a great many people were killed that needn't have.
Anything less allows those who make decisions that lead to warfare and humanitarian crises to avoid criticism.
I think the problem with the last two comments is the belief that because Spagat doesn't think the infant death survey was sound that he doesn't believe the sanctions were bad. Then if he doesn't believe that the Lancet study was believable that therefore he doesn't believe a lot of Iraqis died in the Iraq war. Neither of those surveys was sound, that's his point. If we want to talk about how bad the sanctions were and how costly the war with Iraq was we need to have good data and neither of those studies were that.
2 new articles back up Spagat on how Saddam's govt manipulated many of these studies on infant mortality rates during the sanctions period to turn public opinion.
https://www.healio.com/pediatrics/practice-management/news/online/%7Bc279fab6-2421-4d19-8db7-4a96d7365889%7D/iraqi-government-under-hussein-falsified-child-mortality-reports
https://www.washingtonpost.com/news/worldviews/wp/2017/08/04/saddam-hussein-said-sanctions-killed-500000-children-that-was-a-spectacular-lie/?hpid=hp_hp-more-top-stories_wv-hussein-1037am%3Ahomepage%2Fstory&utm_term=.751b41c6897b
Post a Comment