Search
158 results
Filter search resultsDo health improvement programmes fit with MRC guidance on evaluating complex interventions?
This article, authored by MacKenzie, O'Donnell, Halliday, E.ResourceWhat scientific idea is ready for retirement: Large randomized controlled trials
This comment, written by Dean Ornish and published on the Edge.org blog What scientific idea is ready for retirement, argues that larger studies do not always equate to more rigorous or definitive reResourceExcel charts: Pie charts
This webpage, written by Jorge Camoes for ExcelCharts, outlines the arguments for and against using pie charts and then provides detailed advice and some dos and don'ts for using them.ResourceIDEAS book launch: Randomized control trials in the field of development, a critical perspective [Webinar]
This webinar to launch the book Randomized Control Trials in the Field of Development: A Critical Perspective brings together five representatives of the book's editors and authors for a discussion around some of the keyResourceCanva
Canva is a very simple, free to use, online infographic creation platform. It has a drag and drop interface and a range of templates that you can adapt.ResourceContemporary thinking about causation in evaluation
This paper was produced following a discussion between Thomas Cook and Michael Scriven held at The Evaluation Center and Western Michigan University’s Interdisciplinary PhD in Evaluation program jointly hosted Evaluation Cafe´ event onResourceDo labor market policies have displacement effects? Evidence from a clustered randomized experiment
This resource reports the results from a randomized experiment intended to evaluation the direct and indirect (displacement) impacts of job placement assistance on the labor market outcomes of young, educated job seekers in France.ResourceWeek 47: Rumination #3: Fools' gold: the widely touted methodological "gold standard" is neither golden nor a standard
This week's post is an abbreviated version of a "rumination" from theBlogDigital dividends in natural resource management
The report sets out research findings on the "digital dividends" of various types of technology on natural resource management in low and middle-income countries.ResourceConditions to consider in the use of randomized experimental designs in evaluation
This paper, written by George Julnes, University of New Mexico, Melvin M. Mark, Penn State University, and Stephanie Shipman, U.S.ResourceInfographics
An infographic (short for 'information graphic') represents data visually so that the information is able to be quickly and easily understood.MethodRandomised control trials for the impact evaluation of development initiatives: a statistician's point of view
This paper from the Institutional Learning and Change (ILAC) Initiative provides a range of technical and practical reflections on the use of randomised control trials in impact evaluation.ResourceIntroduction to randomized control trials
This video lecture given by Dr Annette Brown for the Asian Development Bank (ADB) and the International Initiative for Impact Evaluation (3ie) describes how to create a valid counterfactual using randomizeResourceRandomized controlled trials (RCTs) video guide
This video guide, produced by UNICEF, summarises the key features of RCTs with a particular emphasis on their use in impact evaluation.ResourceRandomized controlled trials (RCTs)
This guide, written by Howard White, Shagun Sabarwal and Thomas de Hoop for UNICEF, looks at the use of Randomized Control Trials (RCTs) in Impact Evaluation.ResourceUNICEF webinar: Randomized controlled trials
What are the key features of an RCT? Are RCTs really the gold standard? What ethical and practical issues do I need to consider before deciding to do an RCT?Resource6: Sample size and power calculations
This presentation explores methods for identifying the right sample size for randomized evaluations so that results are defendable.ResourceMultiple lines and levels of evidence
Multiple lines and levels of evidence (MLLE) is a systematic approach to causal inference that involves bringing together different types of evidence (lines of evidence) and considering the strength of the evidence in terms of different indMethodJournals and logs
Journals and logs are forms of record-keeping tools that can be used to capture information about activities, results, conditions, or personal perspectives on how change occurred over a period of time.MethodIntegrity
Integrity refers to ensuring honesty, transparency, and adherence to ethical behaviour by all those involved in the evaluation process.MethodCultural competency
Cultural competency involves ensuring that evaluators have the skills, knowledge, and experience necessary to work respectfully and safely in cultural contexts different from their own.MethodFeasibility
Feasibility refers to ensuring that an evaluation can be realistically and effectively implemented, considering factors such as practicality, resource use, and responsiveness to the programme's context, including factors such as culture andMethodInclusion of diverse perspectives
Inclusion of diverse perspectives requires attention to ensure that marginalised people and communities are adequately engaged in the evaluation.MethodIndependence
Independence can include organisational independence, where an evaluator or evaluation team can independently set a work plan and finalise reports without undue interference, and behavioural independence, where evaluators can conduct and reMethodEvaluation accountability
Evaluation accountability relates to processes in place to ensure the evaluation is carried out transparently and to a high-quality standard.MethodTransferability
Transferability involves presenting findings in a way that they can be applied in other contexts or settings, considering the local culture and context to enhance the utility and reach of evaluation insights.MethodUtility
Utility standards are intended to increase the extent to which program stakeholders find evaluation processes and products valuable in meeting their needs.MethodProfessionalism
Professionalism within evaluation is largely understood in terms of high levels of competence and ethical practice.MethodPropriety
Propriety refers to ensuring that an evaluation will be conducted legally, ethically, and with due regard for the welfare of those involved in it and those affected by its results.MethodSystematic inquiry
Systematic inquiry involves thorough, methodical, contextually relevant and empirical inquiry into evaluation questions. Systematic inquiry is one of the guiding principles of the American Evaluation Association:MethodTransparency
Transparency refers to the evaluation processes and conclusions being able to be scrutinised.MethodEthical practice
Ethical practice in evaluation can be understood in terms of designing and conducting an evaluation to minimise any potential for harm and to maximise the value of the evaluation.Method