Skip to main content

Main menu

  • Home
  • About
  • Who we are
  • News
  • Events
  • Publications
  • Search

Secondary Menu

  • Independent Science for Development CouncilISDC
    • Who we are
    • News
    • Events
    • Publications
    • Featured Projects
      • Inclusive Innovation
        • Agricultural Systems Special Issue
      • Proposal Reviews
        • 2025-30 Portfolio
        • Reform Advice
      • Foresight & Trade-Offs
        • Megatrends
      • QoR4D
      • Comparative Advantage
  • Standing Panel on Impact AssessmentSPIA
    • About
      • Who We Are
      • Our Mandate
      • Impact Assessment Focal Points
      • SPIA Affiliates Network
    • Our Work
      • Country Studies
        • Community of Practice
        • Bangladesh Study
        • Ethiopia Study
        • Uganda Study
        • Vietnam Study
      • Causal Impact Assessment
      • Use of Evidence
      • Cross-Cutting Areas
        • Capacity Strengthening
        • Methods and Measurement
        • Guidance to IDTs
    • Resources
      • Publications
      • Blog Series on Qualitative Methods for Impact Assessment
      • SPIA-emLab Agricultural Interventions Database
    • Activities
      • News
      • Events
      • Webinars
  • Evaluation
    • Who we are
    • News
    • Events
    • Publications
    • Evaluations
      • Science Group Evaluations
      • Platform Evaluations
        • CGIAR Genebank Platform Evaluation
        • CGIAR GENDER Platform Evaluation
        • CGIAR Excellence in Breeding Platform
        • CGIAR Platform for Big Data in Agriculture
    • Framework and Policy
      • Evaluation Method Notes Resource Hub
      • Management Engagement and Response Resource Hub
      • Evaluating Quality of Science for Sustainable Development
      • Evaluability Assessments – Enhancing Pathway to Impact
      • Evaluation Guidelines
  • Independent Science for Development CouncilISDC
  • Standing Panel on Impact AssessmentSPIA
  • Evaluation
Back to IAES Main Menu

Secondary Menu

  • About
    • Who We Are
    • Our Mandate
    • Impact Assessment Focal Points
    • SPIA Affiliates Network
  • Our Work
    • Country Studies
      • Community of Practice
      • Bangladesh Study
      • Ethiopia Study
      • Uganda Study
      • Vietnam Study
    • Causal Impact Assessment
    • Use of Evidence
    • Cross-Cutting Areas
      • Capacity Strengthening
      • Methods and Measurement
      • Guidance to IDTs
  • Resources
    • Publications
    • Blog Series on Qualitative Methods for Impact Assessment
    • SPIA-emLab Agricultural Interventions Database
  • Activities
    • News
    • Events
    • Webinars
Blog

Can Impacts Really Travel?

You are here

  • Home
  • Standing Panel on Impact AssessmentSPIA
  • News
  • Can Impacts Really Travel?

In 1991, researchers in Jamaica studied 129 undernourished children, providing some with weekly home play sessions led by community health aides. The play stimulation showed substantial effects on child development. Similar early stimulation programs were implemented at national scale in Bangladesh and Colombia, integrated into existing government nutrition and parenting programs. The programs still helped children, but the improvements in child development measures were much smaller: 0.17 and 0.16 standard deviations compared to 0.8 in Jamaica.

This pattern shows up constantly in development research. A program works brilliantly at small scale. It is expanded to reach millions more people. And then the impact changes: sometimes shrinking, sometimes shifting to different groups or outcomes, occasionally disappearing altogether.

This isn't a simple story about failure but instead reveals something deeper: that we are surprisingly bad at predicting what will happen when we take something that works in one place and try it somewhere else, or at a much larger scale. As a recent SPIA technical note investigates, this is what experts call the problem of scaling (expanding a successful pilot) and external validity (whether results hold in new contexts). Addressing these challenges means making predictions through extrapolation (applying results to new settings) or interpolation (filling gaps within familiar ones). However, both extrapolation and interpolation are a lot harder than they sound for a variety of reasons, as shown below.

When Good Programs Go Bad 

Consider this head-scratcher from Kenya: hire a teacher on a fixed-term contract through an international NGO, and students learn significantly more. Have the government hire teachers on the exact same contracts? Zero impact. Same contracts, different employer/implementer, very different outcomes. The culprit? Political opposition to contract reforms that changed on-ground implementation dynamics. 

The pattern repeats in other places. "Nudge" interventions, that social scientists adore, aim to subtly influence people’s behaviors by altering the way options are presented. These interventions appear to work four times better in controlled academic studies than in Nudge Units within governments or organizations which implement them at scale[1]. While programs that plug into existing government infrastructure generally tend to fare better, as these examples demonstrate, government involvement does not guarantee success. 

Why Success Doesn't Always Travel

The reasons for this are frustratingly human. When California tried to replicate Tennessee's successful small-class initiative statewide, they had to hire more teachers rapidly. Many were relatively inexperienced or uncertified. Consequently, smaller classes helped, but not as much as anticipated.

In the same vein, a pilot migration loans program in Bangladesh increased temporary migration by 25-40 percentage points but fell to 12 percent at scale. Why? The incentives for loan officers changed. In the small pilot, they had discretion to give loans to people who would benefit most from migrating. At scale, they started giving loans to people most likely to repay (implying more returns for them).

Sometimes the problem is even more peculiar. Researchers simulated what would happen if fertilizer subsidies, tested in small areas of Uganda, rolled out nationwide. They found that more than 80 percent of households would experience wildly different effects than the original experiment suggested and a third would see changes of 50 percent or more, in either direction. Why? Because when everyone changes their behavior at once, the economic dynamics are likely to shift.

These are scaling and external validity challenges in action: what worked in one context doesn’t automatically translate elsewhere, even with careful replication.

The CGIAR Reality Check

These issues matter enormously for CGIAR, which heavily invests in agricultural research and innovations meant to reach millions of farmers. In the past, when asked to project future benefits, optimistic projections were produced based on expert knowledge and literature, which regrettably could not substitute hard evidence for what actually happens when you scale. The uncomfortable truth is that we cannot simply multiply localized successes by the target populations to derive benefit projections. When interventions scale, impacts and incentives will shift and need to be accounted for. 

So what can be done? Two promising methods offer hope. One, AI and machine learning techniques can help identify populations most similar to those in successful pilots and offer workarounds for external validity constraints. Two, remote sensing data can help track climatic and weather conditions, yield variability, productivity, and intervention coverage, particularly in contexts where traditional data is scarce or expensive to collect. Despite their significant potential, these methods require heavy caution as data complexity and measurement errors remain persistent issues.

The Takeaway

In the face of these challenges, the main takeaway is to maintain honesty about uncertainty. Scaling something up or trying an intervention someplace new is akin to re-running an experiment. And experiments have uncertain outcomes. The question, therefore, is not whether to scale successful interventions, but rather being realistic about what to expect when we attempt to transport its impacts.

 

Read the Technical Note Here
 

[1] Although this may be influenced by various factors such as publication bias (i.e, academic studies publishing trials with large effects) or that government-run nudges operate in more complex settings.

Share on

Impact SPIA
Oct 15, 2025

Written by

  • Swetha Ramachandran

    Senior Officer, SPIA Use of Evidence
    • Use of Evidence Team
  • Astewale Bimr Melaku

    Post Doctoral Fellow, SPIA Causal Impact Assessment

Related News

Posted on
15 Oct 2025
by
  • SPIA Team

SPIA's Meeting Highlights in Cali: Strengthening Collaboration and Innovation

Posted on
01 Oct 2025
by
  • SPIA

SPIA Welcomes New Panel Member: Shilpa Aggarwal

Posted on
22 Sep 2025
by
  • SPIA

SPIA Welcomes New Research Officer in Data Systems for the SPIA Country Studies

More News

Related Publications

Technical Notes
Impact SPIA
Issued on 2025

Interpolating and Extrapolating Impact Estimates: Implications of Ex Post Impact Evaluation Evidence for Projecting Benefits Ex Ante

Technical Notes
Impact SPIA
Issued on 2025

Does An Innovation’s Reach Reveal Anything About Its Impact? Under The Right Conditions: Possibly

Technical Notes
Impact SPIA
Issued on 2025

Estimating the Reach of Biofortified Crops to Farm Households: The HarvestPlus Model vs. National Sampling and Genotyping in Uganda

More publications

CGIAR Independent Advisory and Evaluation Service (IAES)

Alliance of Bioversity International and CIAT
Via di San Domenico,1
00153 Rome, Italy
  • IAES@cgiar.org
  • (39-06) 61181

Follow Us

  • LinkedIn
  • Twitter
  • YouTube
JOIN OUR MAILING LIST
  • Terms and conditions
  • © CGIAR 2025

IAES provides operational support as the secretariat for the Independent Science for Development Council and the Standing Panel on Impact Assessment, and implements CGIAR’s multi-year, independent evaluation plan as approved by the CGIAR’s System Council.