Abstract
We organized two evaluations of the paper: "Water Treatment and Child Mortality: A Meta-analysis and Cost-effectiveness Analysis"[1]. To read these evaluations, please see the links below. The first was a team effort by Hugh Sharma Waddington and Edoardo Masset, who recently published a related systematic review; this evaluation focuses mainly on the meta-analysis and the studies therein. The second (anonymous) evaluation focuses more on the cost-effectiveness analysis. Both evaluators agree that this paper tackles an important question and provides an advance on our prior understanding of the effects of water treatment on child mortality. Both also found areas where the paper could be improved or clarified.
Evaluations
1. Sharma Waddington and Masset
Anonymous Evaluator 2
Overall ratings
We asked evaluators to provide overall assessments as well as ratings for a range of specific criteria.
I. Overall assessment (see footnote)
II. Journal rank tier, normative rating (0-5): On a ‘scale of journals’, what ‘quality of journal’ should this be published in? (See ranking tiers discussed here.) Note: 0= lowest/none, 5= highest/best.
| Overall assessment (0-100) | Journal rank tier, normative rating (0-5) |
Sharma Waddington & Masset | 50 | 2.5 |
Anonymous evaluator 2 | 85 | 3.7 |
See “Metrics” below for a more detailed breakdown of the evaluators’ ratings across several categories. To see these ratings in the context of all Unjournal ratings, with some analysis, see our data presentation here. See here for the current full evaluator guidelines, including further explanation of the requested ratings.
Evaluation summaries
Sharma Waddington & Masset
We evaluated the conduct, analysis and interpretation of the meta-analysis findings. The paper concerns an important topic for the global burden of infectious disease, particularly in middle-income countries where most of the included studies were conducted. It presents useful information on childhood mortality following water treatment and protection interventions. However, we had major concerns about conduct and reporting, which did not follow systematic review standards of transparency, and we also had major concerns about the interpretation of the findings for policy and research.
Anonymous evaluator 2
1. I do not see any major technical issues with the paper.
2. The paper can be positioned more robustly i.e., the “need for it” needs to be better argued.
3. While the “effect” (mortality reduction) side of the paper has been done well, the “cost” part of the paper could be deepened. This would position it better and be supremely useful to policymakers.
Metrics
Ratings
See here for details on the categories below, and the guidance given to evaluators.
| Evaluator 1 Sharma Waddington & Masset | | | Evaluator 2 Anonymous | | |
---|
Rating category | Rating (0-100) | 90% CI (0-100)* | Comments | Rating (0-100) | 90% CI (0-100)* | Comments |
---|
Overall assessment | 50 | (40, 60) | | 85 | (75, 95) | |
---|
Advancing knowledge and practice | 80 | (70, 90) | | 65 | (45, 85) | |
---|
Methods: Justification, reasonableness, validity, robustness | 50 | (40, 60) | | 75 | (65, 85) | |
---|
Logic & communication | 50 | (30, 70) | | 95 | (90, 100) | |
---|
Open, collaborative, replicable | 35 | (20, 50) | | 90 | (80, 100) | |
---|
Real-world relevance | 20 | (0, 40) | | 50 | (25, 75) | |
---|
Relevance to global priorities | 45 | (30, 60) | | 90 | (85, 95) | |
---|
Journal ranking tiers
See here for more details on these tiers.
| Evaluator 1 Sharma Waddington & Masset | | Evaluator 2 Anonymous | | |
---|
Judgment | Ranking tier (0-5) | 90% CI | Ranking tier (0-5) | 90% CI | Comments |
---|
On a ‘scale of journals’, what ‘quality of journal’ should this be published in? | 2.5 | (2.0, 3.0) | 3.7 | (3.2, 4.2) | |
---|
What ‘quality journal’ do you expect this work will be published in? | 3.5 | (2.0, 5.0) | 3.7 | (3.2, 4.7) | |
---|
See here for more details on these tiers. | We summarize these as: 0.0: Marginally respectable/Little to no value 1.0: OK/Somewhat valuable 2.0: Marginal B-journal/Decent field journal 3.0: Top B-journal/Strong field journal 4.0: Marginal A-Journal/Top field journal 5.0: A-journal/Top journal
|
Evaluation manager’s discussion
The Unjournal solicited two evaluations of the paper “Water treatment and child mortality: a meta-analysis and cost-effectiveness analysis.” The paper was flagged to us as important by a staff member at a major philanthropic organization, and our screening indeed suggested it could be highly influential in moving money to water treatment.
The paper does two things. It first pools evidence from many studies and so allows the authors to have the statistical precision to examine the effect of water treatment on child mortality directly. This is an improvement over past work that attempted to back out such an effect indirectly, by looking at much more common indirect signals of mortality like diarrhea.
The second thing the paper does is use the meta-analytic results to do a cost-effectiveness exercise to see how cost-effective is water treatment at avoiding death (or increasing DALYs). The main takeaway is direct estimates of water treatment show much higher reductions in mortality than do indirect methods. In turn, this means that water treatment is estimated to be much more cost-effective at averting deaths than was previously thought.
I solicited two reviews of this paper. While all the evaluators looked at the entire paper, I asked one team of evaluators to focus more on the meta-analysis and studies therein and a second evaluator to focus more on the cost-effectiveness analysis. Both evaluators agree that the paper tackles an important question and provides an advance on our prior understanding of the effects of water treatment on child mortality. Both also found areas where the paper could be improved or clarified. Our first evaluators, who worked in a team, raised many questions and helpful criticisms. These include:
Questions about study inclusion and exclusion and transparency.
Issues of data quality and analysis transparency, such as what appear to be undisclosed deviations from the pre-registration plan and changes in the odds ratios of the underlying studies between drafts.
Questions for scale up and policy, such as the representativeness of the sample, questions around mechanisms, and concerns around the persistence of the behavioural change.
Our second evaluation focused more heavily on the CEA model and had two major suggestions. The first was around the framing of the study. The second focused on ways to make the CEA more useful for policymakers, including:
Disaggregating costs to allow individual policymakers to essentially make selections off of a menu to match the CEA to their situation more closely.
Sensitivity analysis of the CEA, making use of the estimates of uncertainty from the meta-analysis.
The authors have responded to both evaluations and updated the working paper in response. While not all of the changes to the paper have yet been implemented, my personal reading is that the evaluators noted a number of ways that the paper could be improved, it in turn was improved, and while the headline result remains largely the same, we have now learned some new things about:
The stability of the results of various changes in the data
Additional important points in the CEA that were not clear to me up front, such as the important role of compliance in affecting the results
Important limitations in most existing economic research that prevents meta-analyses from being conducted or greatly limits their usefulness. In particular I’m thinking of participant flow diagrams in studies.
Anyone with an interest in this topic should read the evaluations and response in full as my summary comments contain some of my own interpretation and both the evaluators and authors may disagree with parts. A major advantage of open review is that one can read the exchange directly and form their own opinion.
I extend my sincere thanks to the evaluators and the authors for what I view as a fulfilling and transparent cooperative effort to reach the best possible answer to an important question.
Unjournal process notes
Note on versions: This working paper had two public drafts. The evaluators were recruited after the first (NBER) draft was posted online but when we learned that the authors were close to publishing a second draft we froze the process and waited for the newer draft to be published. The evaluation is of the second (May 2024) draft, though reviewers may at times make points that compare across the drafts as they read both.
Why we chose this paper
This work is very relevant to charity recommendations on water interventions. From our team notes “Kremer’s work on WASH inform[ed] Evidence Action’s $64 million grant”. GiveWell was one of the funders of this research and they indicated a strong interest in having it evaluated. Furthermore, at least one senior academic and another senior applied researcher outside our team specifically suggested this.
For a ~nontechnical discussion, see GiveWell’s blog “Research Strategy: Water” (updated July 2024).
How we chose (and tasked) the evaluators
We selected evaluators with WASH research experience and we drew them from both the public health and development economics communities. As noted above “one team of evaluators was asked to focus more on the meta-analysis and studies therein and a second evaluator to focus more on the cost-effectiveness analysis.” The ‘team evaluation’ mode seemed particularly promising, and something we may pursue again in future evaluations.