<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	
	>
<channel>
	<title>Comments on: Randomized Controlled Trials: panacea or mirage?</title>
	<atom:link href="https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/feed/" rel="self" type="application/rss+xml" />
	<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/</link>
	<description>How active citizens and effective states can change the world</description>
	<lastBuildDate>Sat, 08 Jul 2017 06:58:57 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=4.2.15</generator>
	<item>
		<title>By: Stas Spector</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-19324</link>
		<dc:creator><![CDATA[Stas Spector]]></dc:creator>
		<pubDate>Thu, 27 Feb 2014 15:50:55 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-19324</guid>
		<description><![CDATA[Like any other system, RCTs are continually improving based on past experience and new ideal and technology. For example, our RCT consultancy http://researchtool.org  has built an extensive cloud system to minimise costs, improve access and streamline processes. It is healthy to have critics, but there is no doubt that there is no substitute for a well designed Randomised Controlled Trial.

One thing that could accelerate the evolution of RCTs is a better communication between researchers. Hopefully now it will be changing, as there is a new community forum dedicated to RCT http://rctrials.org]]></description>
		<content:encoded><![CDATA[<p>Like any other system, RCTs are continually improving based on past experience and new ideal and technology. For example, our RCT consultancy <a href="http://researchtool.org" rel="nofollow">http://researchtool.org</a>  has built an extensive cloud system to minimise costs, improve access and streamline processes. It is healthy to have critics, but there is no doubt that there is no substitute for a well designed Randomised Controlled Trial.</p>
<p>One thing that could accelerate the evolution of RCTs is a better communication between researchers. Hopefully now it will be changing, as there is a new community forum dedicated to RCT <a href="http://rctrials.org" rel="nofollow">http://rctrials.org</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Gricelda Schein</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1235</link>
		<dc:creator><![CDATA[Gricelda Schein]]></dc:creator>
		<pubDate>Tue, 27 Sep 2011 06:57:25 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1235</guid>
		<description><![CDATA[As I have written up, this is a enterprise tool. It’s not really a miracle piece of equipment. It could easily change into another expense unless you put the persistence into learning how to take advantage of it.]]></description>
		<content:encoded><![CDATA[<p>As I have written up, this is a enterprise tool. It’s not really a miracle piece of equipment. It could easily change into another expense unless you put the persistence into learning how to take advantage of it.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Hugh Waddington (3ie)</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1234</link>
		<dc:creator><![CDATA[Hugh Waddington (3ie)]]></dc:creator>
		<pubDate>Tue, 31 Aug 2010 18:41:54 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1234</guid>
		<description><![CDATA[There is a massive need for evidence on what works, when, why and for how much in the field of international development, as acknowledged in the Evaluation Gap Working Group report. The International Initiative for Impact Evaluation (3ie) was set up to help fill this gap, by promoting and funding impact evaluations of social and economic development interventions and summaries of the existing evidence. We read Duncan’s original post and the additional contributions with great interest, and provide a few comments below.

As stated in Duncan’s posting, the vast majority of social and economic development interventions are probably not randomisable. It’s been estimated that only 5% are, because of ethical reasons, since the control group population needs to be isolated from the intervention, at least during the study period, or for political reasons, such as the desire to reach certain populations or the neediest. RCTs may also not be feasible for certain types of programme, such as a national policy change. But there are other approaches to estimating impact rigorously which don’t require randomised assignment of an intervention to people or communities – so-called quasi-experiments.

Generalisability of findings is indeed incredibly important. What works in post-conflict Liberia may not be readily transferable to Rwanda, and even more so in Latin America. This is one argument why more impact evaluations in different contexts are needed. But, while the RCTs that were done in the US in the 1960s only provided a number for impact – what Melvin Mark calls “bare bone RCTs” – the emphasis in international development programme evaluation, especially with the push from 3ie and impact evaluation network NONIE, is for theory-based impact evaluation which aims to answer not just what works, or not, but why and in which contexts.

A good theory-based impact evaluation “opens the black box” by collecting information on the reasons for effectiveness, or lack of it, which can then be used to inform programme planning elsewhere. It builds on a theory of change and assesses information along the causal chain, as set out for example in a log-frame, to assess the reasons why interventions succeed or fail – what Chris Blattman has called Impact Evaluation 2.0. This information is just as likely to be opinions of participants and practitioners collected using beneficiary interviews and focus groups, as it is to be ‘hard’ numbers collected using surveys or administrative data. Use of mixed methods to triangulate data is a key tool in theory-based impact evaluation.

Despite earlier doubts, the range of interventions which are amenable to rigorous impact evaluation is large. For example, 3ie is funding RCTs of:
- Community driven development in Sierra Leone
- Monitoring patient compliance with tuberculosis treatment in Pakistan
- Expanding secondary education in Ghana
- Mexico&#039;s Payments for Ecosystem Services Program
- Scaling up male circumcision service provision in Malawi
- BRAC’s Graduation Model in Ghana
- Micro-entrepreneurship support in Chile
- Diffusion of health knowledge through social networks in Burkina Faso
- Property taxation in Pakistan

Innovations are being piloted so that impact evidence collected and disseminated is relevant to all stakeholders, not just researchers in ivory towers. For example, 3ie’s Policy Window enables policy makers in developing country governments and NGOs to request programmes that they want to be evaluated. Systematic reviews aim to provide an unbiased assessment and synthesis of all the available evidence on an intervention; this includes the unpublished studies which may remain hard to access because they do not provide evidence supporting positive effects. 3ie, DFID and AusAID will be launching a joint call for systematic review proposals in late September.

We sincerely look forward to seeing the growth of theory-based impact studies and systematic reviews, and their dissemination and use by policy makers and practitioners in international development.

Duncan: thanks Hugh, really interesting comment. For Chris Blattman on Impact Evaluation 2.0, go to http://chrisblattman.com/2008/02/16/impact-evaluation-2-0/]]></description>
		<content:encoded><![CDATA[<p>There is a massive need for evidence on what works, when, why and for how much in the field of international development, as acknowledged in the Evaluation Gap Working Group report. The International Initiative for Impact Evaluation (3ie) was set up to help fill this gap, by promoting and funding impact evaluations of social and economic development interventions and summaries of the existing evidence. We read Duncan’s original post and the additional contributions with great interest, and provide a few comments below.</p>
<p>As stated in Duncan’s posting, the vast majority of social and economic development interventions are probably not randomisable. It’s been estimated that only 5% are, because of ethical reasons, since the control group population needs to be isolated from the intervention, at least during the study period, or for political reasons, such as the desire to reach certain populations or the neediest. RCTs may also not be feasible for certain types of programme, such as a national policy change. But there are other approaches to estimating impact rigorously which don’t require randomised assignment of an intervention to people or communities – so-called quasi-experiments.</p>
<p>Generalisability of findings is indeed incredibly important. What works in post-conflict Liberia may not be readily transferable to Rwanda, and even more so in Latin America. This is one argument why more impact evaluations in different contexts are needed. But, while the RCTs that were done in the US in the 1960s only provided a number for impact – what Melvin Mark calls “bare bone RCTs” – the emphasis in international development programme evaluation, especially with the push from 3ie and impact evaluation network NONIE, is for theory-based impact evaluation which aims to answer not just what works, or not, but why and in which contexts.</p>
<p>A good theory-based impact evaluation “opens the black box” by collecting information on the reasons for effectiveness, or lack of it, which can then be used to inform programme planning elsewhere. It builds on a theory of change and assesses information along the causal chain, as set out for example in a log-frame, to assess the reasons why interventions succeed or fail – what Chris Blattman has called Impact Evaluation 2.0. This information is just as likely to be opinions of participants and practitioners collected using beneficiary interviews and focus groups, as it is to be ‘hard’ numbers collected using surveys or administrative data. Use of mixed methods to triangulate data is a key tool in theory-based impact evaluation.</p>
<p>Despite earlier doubts, the range of interventions which are amenable to rigorous impact evaluation is large. For example, 3ie is funding RCTs of:<br />
&#8211; Community driven development in Sierra Leone<br />
&#8211; Monitoring patient compliance with tuberculosis treatment in Pakistan<br />
&#8211; Expanding secondary education in Ghana<br />
&#8211; Mexico&#8217;s Payments for Ecosystem Services Program<br />
&#8211; Scaling up male circumcision service provision in Malawi<br />
&#8211; BRAC’s Graduation Model in Ghana<br />
&#8211; Micro-entrepreneurship support in Chile<br />
&#8211; Diffusion of health knowledge through social networks in Burkina Faso<br />
&#8211; Property taxation in Pakistan</p>
<p>Innovations are being piloted so that impact evidence collected and disseminated is relevant to all stakeholders, not just researchers in ivory towers. For example, 3ie’s Policy Window enables policy makers in developing country governments and NGOs to request programmes that they want to be evaluated. Systematic reviews aim to provide an unbiased assessment and synthesis of all the available evidence on an intervention; this includes the unpublished studies which may remain hard to access because they do not provide evidence supporting positive effects. 3ie, DFID and AusAID will be launching a joint call for systematic review proposals in late September.</p>
<p>We sincerely look forward to seeing the growth of theory-based impact studies and systematic reviews, and their dissemination and use by policy makers and practitioners in international development.</p>
<p>Duncan: thanks Hugh, really interesting comment. For Chris Blattman on Impact Evaluation 2.0, go to <a href="http://chrisblattman.com/2008/02/16/impact-evaluation-2-0/" rel="nofollow">http://chrisblattman.com/2008/02/16/impact-evaluation-2-0/</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Roving Bandit</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1233</link>
		<dc:creator><![CDATA[Roving Bandit]]></dc:creator>
		<pubDate>Tue, 22 Jun 2010 10:06:39 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1233</guid>
		<description><![CDATA[An ungated version is now available at http://karlan.yale.edu/]]></description>
		<content:encoded><![CDATA[<p>An ungated version is now available at <a href="http://karlan.yale.edu/" rel="nofollow">http://karlan.yale.edu/</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: David Carreon</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1232</link>
		<dc:creator><![CDATA[David Carreon]]></dc:creator>
		<pubDate>Tue, 01 Jun 2010 08:42:32 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1232</guid>
		<description><![CDATA[Great article and great discussion!

The RCT bonanza seems to me to be a reaction to the data-less (or at least data poor) development of the past and a desire to be associated with the success of medical science over the past decades (explicitly stated by Duflo). RCTs are the gold standard of medicine, but the assumption of external validity is much more reasonable. Human beings, more or less, have the same biochemistry; if a drug works in an RCT, it probably will work for everyone (yet personalized medicine is challenging even this assumption). Human behavior is much more variable than human biochemistry, and so I agree with Duncan and others who criticize development RCTs on these grounds. I think both for personalized medicine and for development, new techniques will need to emerge to monitor and characterize local and individual differences.

It’s important to consider not just the factual truth of a claim, but also its local validity. To an NGO, “low external validity” doesn’t mean that we shouldn’t give out bed nets or deworm kids. It means that those interventions are probably a good place to start, but progress should be monitored and the intervention changed if it’s not effective. I’m very glad that I can search the Literature and find a hundred studies on the cost effectiveness of every possible intervention under the sun. But that only gives me a ranked list of options, not a rigid prescription. Neither medicine nor development can ever be done completely by the book; I am no more in favor of cookbook medicine than I am of cookbook development.

Like good doctors, development workers need to monitor the progress and impact of their work. There have been some pretty exciting developments in technology that I’ve been trying to stitch together for my NGO, Nuru International. I’ve been able to set up a mobile data collection system for our seed project in Kenya that has been extremely helpful. We use cell phones and the web browser on the phone to collect data with Google Forms at very low cost; the total set up was about $5000 for 75 users (every member of our field staff), and has been costing around $1/user/month in data costs. For more info: bit.ly/cOYOyg .]]></description>
		<content:encoded><![CDATA[<p>Great article and great discussion!</p>
<p>The RCT bonanza seems to me to be a reaction to the data-less (or at least data poor) development of the past and a desire to be associated with the success of medical science over the past decades (explicitly stated by Duflo). RCTs are the gold standard of medicine, but the assumption of external validity is much more reasonable. Human beings, more or less, have the same biochemistry; if a drug works in an RCT, it probably will work for everyone (yet personalized medicine is challenging even this assumption). Human behavior is much more variable than human biochemistry, and so I agree with Duncan and others who criticize development RCTs on these grounds. I think both for personalized medicine and for development, new techniques will need to emerge to monitor and characterize local and individual differences.</p>
<p>It’s important to consider not just the factual truth of a claim, but also its local validity. To an NGO, “low external validity” doesn’t mean that we shouldn’t give out bed nets or deworm kids. It means that those interventions are probably a good place to start, but progress should be monitored and the intervention changed if it’s not effective. I’m very glad that I can search the Literature and find a hundred studies on the cost effectiveness of every possible intervention under the sun. But that only gives me a ranked list of options, not a rigid prescription. Neither medicine nor development can ever be done completely by the book; I am no more in favor of cookbook medicine than I am of cookbook development.</p>
<p>Like good doctors, development workers need to monitor the progress and impact of their work. There have been some pretty exciting developments in technology that I’ve been trying to stitch together for my NGO, Nuru International. I’ve been able to set up a mobile data collection system for our seed project in Kenya that has been extremely helpful. We use cell phones and the web browser on the phone to collect data with Google Forms at very low cost; the total set up was about $5000 for 75 users (every member of our field staff), and has been costing around $1/user/month in data costs. For more info: bit.ly/cOYOyg .</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Nic</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1231</link>
		<dc:creator><![CDATA[Nic]]></dc:creator>
		<pubDate>Tue, 25 May 2010 18:11:00 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1231</guid>
		<description><![CDATA[An interesting piece on the limitations/correct uses of RCTs can be found here:

http://www.princeton.edu/~deaton/downloads/Instruments_of_Development.pdf]]></description>
		<content:encoded><![CDATA[<p>An interesting piece on the limitations/correct uses of RCTs can be found here:</p>
<p><a href="http://www.princeton.edu/~deaton/downloads/Instruments_of_Development.pdf" rel="nofollow">http://www.princeton.edu/~deaton/downloads/Instruments_of_Development.pdf</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: James Stevenson</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1230</link>
		<dc:creator><![CDATA[James Stevenson]]></dc:creator>
		<pubDate>Fri, 14 May 2010 09:46:34 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1230</guid>
		<description><![CDATA[Duncan,

Good to see this post. Panacea or mirage? Neither, but a very useful development in a sector that is hardly the most transparent about success and failure.

I would second everything that Harry has posted about pitfalls, especially assumptions about external validity. Most methods, RCTs included, have low external validity - i.e. it is difficult to extrapolate research findings to other places and times. This is especially so for interventions where the causal mechanism plays out in a socially contingent, open system - contrast with a medical intervention where human bodies are more similar to each other than they are different.

RCTs have the highest internal validity of any method - they are excellent at isolating specific causal factors. The problem comes in conflating high internal validity with &quot;rigour&quot;. We usually want to be able to make inferences for how things will play out for populations, not samples, and so external validity is also important. I like Nancy Cartwright&#039;s writing on this topic (the philosopher, not the woman who does the voice for Bart Simpson...):

http://personal.lse.ac.uk/cartwrig/PapersOnEvidence/Are%20RCTs%20the%20gold%20standard.pdf

That said, I think one of the best things about the randomistas is their sense of humility and inventiveness. These are not ivory tower academics. I really appreciate the spirit in which they work, and the method allows us them to be very transparent in testing out new ideas. For example, Oxfam America are running some on a savings group intervention (I think) in Africa. This really helps them understand whether this kind of intervention does work in that context, and if they do it well, they&#039;ll be able to also understand why.

The reason RCTs get so much attention is because policy-makers / NGO directors (I hope so) can understand the methods and the results much more easily than a complex bit of econometrics, and so have a greater degree of trust in the findings.

See also Chris Blattmans&#039; excellent note on Impact Evaluation 2.0:
http://www.chrisblattman.com/documents/policy/2008.ImpactEvaluation2.DFID_talk.pdf

All power to the elbow of those with an interest in promoting more transparent and credible evaluation methodology in development agencies of all stripes -  it&#039;s a hard row to hoe.
Cheers,
James]]></description>
		<content:encoded><![CDATA[<p>Duncan,</p>
<p>Good to see this post. Panacea or mirage? Neither, but a very useful development in a sector that is hardly the most transparent about success and failure.</p>
<p>I would second everything that Harry has posted about pitfalls, especially assumptions about external validity. Most methods, RCTs included, have low external validity &#8211; i.e. it is difficult to extrapolate research findings to other places and times. This is especially so for interventions where the causal mechanism plays out in a socially contingent, open system &#8211; contrast with a medical intervention where human bodies are more similar to each other than they are different.</p>
<p>RCTs have the highest internal validity of any method &#8211; they are excellent at isolating specific causal factors. The problem comes in conflating high internal validity with &#8220;rigour&#8221;. We usually want to be able to make inferences for how things will play out for populations, not samples, and so external validity is also important. I like Nancy Cartwright&#8217;s writing on this topic (the philosopher, not the woman who does the voice for Bart Simpson&#8230;):</p>
<p><a href="http://personal.lse.ac.uk/cartwrig/PapersOnEvidence/Are%20RCTs%20the%20gold%20standard.pdf" rel="nofollow">http://personal.lse.ac.uk/cartwrig/PapersOnEvidence/Are%20RCTs%20the%20gold%20standard.pdf</a></p>
<p>That said, I think one of the best things about the randomistas is their sense of humility and inventiveness. These are not ivory tower academics. I really appreciate the spirit in which they work, and the method allows us them to be very transparent in testing out new ideas. For example, Oxfam America are running some on a savings group intervention (I think) in Africa. This really helps them understand whether this kind of intervention does work in that context, and if they do it well, they&#8217;ll be able to also understand why.</p>
<p>The reason RCTs get so much attention is because policy-makers / NGO directors (I hope so) can understand the methods and the results much more easily than a complex bit of econometrics, and so have a greater degree of trust in the findings.</p>
<p>See also Chris Blattmans&#8217; excellent note on Impact Evaluation 2.0:<br />
<a href="http://www.chrisblattman.com/documents/policy/2008.ImpactEvaluation2.DFID_talk.pdf" rel="nofollow">http://www.chrisblattman.com/documents/policy/2008.ImpactEvaluation2.DFID_talk.pdf</a></p>
<p>All power to the elbow of those with an interest in promoting more transparent and credible evaluation methodology in development agencies of all stripes &#8211;  it&#8217;s a hard row to hoe.<br />
Cheers,<br />
James</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Sceptical Secondo</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1229</link>
		<dc:creator><![CDATA[Sceptical Secondo]]></dc:creator>
		<pubDate>Wed, 12 May 2010 17:40:19 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1229</guid>
		<description><![CDATA[Not that RCTs are any less good or bad than so many other approaches but once I heard Duflo&#039;s speech ... sigh! I feel like, once again, do American academics know what distinguishes method from methodology?]]></description>
		<content:encoded><![CDATA[<p>Not that RCTs are any less good or bad than so many other approaches but once I heard Duflo&#8217;s speech &#8230; sigh! I feel like, once again, do American academics know what distinguishes method from methodology?</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: didier</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1228</link>
		<dc:creator><![CDATA[didier]]></dc:creator>
		<pubDate>Tue, 11 May 2010 19:18:07 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1228</guid>
		<description><![CDATA[I like the methodology but the presentaion made me wonder about some things.  The example of using food to incentivize mothers to come and get their kids vaccinated is a pretty old stragtegy dating back to the 1970&#039;s and has been applied on a widespread basis by USAID, WFP, UNICEF and almost every Ministry of Health in every developing country that I can remember with the coperation of large private development organizations such as CARE, CRS, World Vision etc.. for over three decades. Does the RCT in India confirm the wisdom of 30 years of maternal-child health interventions? If so, what made all these agencies so smart to pursue such programs without the benefit of an RCT during any of that time?  Can we say that there are may paths to heaven?]]></description>
		<content:encoded><![CDATA[<p>I like the methodology but the presentaion made me wonder about some things.  The example of using food to incentivize mothers to come and get their kids vaccinated is a pretty old stragtegy dating back to the 1970&#8217;s and has been applied on a widespread basis by USAID, WFP, UNICEF and almost every Ministry of Health in every developing country that I can remember with the coperation of large private development organizations such as CARE, CRS, World Vision etc.. for over three decades. Does the RCT in India confirm the wisdom of 30 years of maternal-child health interventions? If so, what made all these agencies so smart to pursue such programs without the benefit of an RCT during any of that time?  Can we say that there are may paths to heaven?</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Harry Jones</title>
		<link>https://oxfamblogs.org/fp2p/randomized-controlled-trials-panacea-or-mirage/#comment-1227</link>
		<dc:creator><![CDATA[Harry Jones]]></dc:creator>
		<pubDate>Tue, 11 May 2010 10:45:21 +0000</pubDate>
		<guid isPermaLink="false">http://www.oxfamblogs.org/fp2p/?p=2478#comment-1227</guid>
		<description><![CDATA[Dear Duncan et al

You&#039;ve made some good points - many of which are corroborated by last year&#039;s ODI study of the production and use of impact evaluations (see link at the bottom).

Aside from it being unwise to engage in methodological argument with people who are so hot on that sort of thing, randomistas should, to a certain extent, be left to carry out their work which is essentially a scientific enterprise, producing &#039;global public good&#039; knowledge. There are a few problems with this however:
- RCTs do tend to be carried out where they are methodologically convenient rather than where the new knowledge is really needed
- there is massive publication bias, by our count over 95% of published RCTs show &#039;positive&#039; impact, which severely limits the ability to really learn

The bigger problem is about the relationships between RCTs and policy. They were hugely fashionable in the US in the &#039;60s until it was clear that they couldn&#039;t deliver all that was hoped, and now it seems that development policy makers are in love with the idea of having clear numbers on the impact of all of their work. The problems:
- policy makers often assume a much higher level of external validity than is actually appropriate, and tend to ignore the careful caveats which come along with RCTs
-  As even people like Duflo would have to admit, the RCT model is only suitable for measuring impact in a subset, nay a minority of the kinds of intervention required for development. However they are being given a disproportionate amount of attention
- this is an issue of cost-effectiveness as you say, about where we choose to spend evaluation budgets and with what coverage
- but this is also about the danger of demanding ‘rigorous impact evaluation’ (which =RCTs for some) incentivising those programmes which fit easily into that model. These are generally output-driven programmes like distributing bed-nets, and things which will struggle are governance, capacity building, budget support, policy influencing.

I wrote up my worries about these issues in this ODI opinion piece: http://www.odi.org.uk/resources/download/2811.pdf
And here is the ODI working paper on the production and use of IEs http://www.odi.org.uk/resources/download/3177.pdf

Cheers

Harry]]></description>
		<content:encoded><![CDATA[<p>Dear Duncan et al</p>
<p>You&#8217;ve made some good points &#8211; many of which are corroborated by last year&#8217;s ODI study of the production and use of impact evaluations (see link at the bottom).</p>
<p>Aside from it being unwise to engage in methodological argument with people who are so hot on that sort of thing, randomistas should, to a certain extent, be left to carry out their work which is essentially a scientific enterprise, producing &#8216;global public good&#8217; knowledge. There are a few problems with this however:<br />
&#8211; RCTs do tend to be carried out where they are methodologically convenient rather than where the new knowledge is really needed<br />
&#8211; there is massive publication bias, by our count over 95% of published RCTs show &#8216;positive&#8217; impact, which severely limits the ability to really learn</p>
<p>The bigger problem is about the relationships between RCTs and policy. They were hugely fashionable in the US in the &#8217;60s until it was clear that they couldn&#8217;t deliver all that was hoped, and now it seems that development policy makers are in love with the idea of having clear numbers on the impact of all of their work. The problems:<br />
&#8211; policy makers often assume a much higher level of external validity than is actually appropriate, and tend to ignore the careful caveats which come along with RCTs<br />
&#8211;  As even people like Duflo would have to admit, the RCT model is only suitable for measuring impact in a subset, nay a minority of the kinds of intervention required for development. However they are being given a disproportionate amount of attention<br />
&#8211; this is an issue of cost-effectiveness as you say, about where we choose to spend evaluation budgets and with what coverage<br />
&#8211; but this is also about the danger of demanding ‘rigorous impact evaluation’ (which =RCTs for some) incentivising those programmes which fit easily into that model. These are generally output-driven programmes like distributing bed-nets, and things which will struggle are governance, capacity building, budget support, policy influencing.</p>
<p>I wrote up my worries about these issues in this ODI opinion piece: <a href="http://www.odi.org.uk/resources/download/2811.pdf" rel="nofollow">http://www.odi.org.uk/resources/download/2811.pdf</a><br />
And here is the ODI working paper on the production and use of IEs <a href="http://www.odi.org.uk/resources/download/3177.pdf" rel="nofollow">http://www.odi.org.uk/resources/download/3177.pdf</a></p>
<p>Cheers</p>
<p>Harry</p>
]]></content:encoded>
	</item>
</channel>
</rss>
