Search
152 results
Filter search resultsParticipatory video and the most significant change. A guide for facilitators
The toolkit is designed to support you in planning and carrying out evaluations using participatory video (PV) with the most significant change (MSC) technique, or PVMSC for short.ResourceAction and reflection: a guide for monitoring and evaluating participatory research
This paper from the International Development Research Centre (IDRC) was designed to support those involved in participatory research and development projects with monitoring and evaluation stratResourceWeek 50: Feedback loops – new buzzword, old practice?
Recently, I had the good fortune to start collaboration with The MasterCard Foundation, which is strongly committed to what it calls ‘listening deeply and elevating voices’.BlogParticipation not for you? Four reflections that might just change your mind
This month we start a series on participation in evaluation by Leslie Groves and Irene Guijt. This blog series aims to explore one simple question: How can we best open up evaluation processes to include those intended to benefit from a specificBlogPositioning participation on the power spectrum
In the second blog in the 4-part series about participation in evaluation, Irene Guijt and Leslie Groves focus on making power relationships and values in 'participatory' evaluation processes explicit to avoid tokenistic partBlogChoices about voices
In this third blog in the participation in evaluation series, Irene Guijt and Leslie Groves share frameworks to approach and make decisions about the level of stakeholder involvement during different evaluation stages.BlogStill Hesitating? Let's bust some myths around increasing stakeholder participation in evaluation
In the final blog in the 4-part series, Leslie Groves and Irene Guijt address some of the most common forms of resistance to increasing levels of participation in evaluation.BlogWhat does it mean to ‘un-box’ evaluation?
This guest blog by Jade Maloney is the first in a series about un-boxing evaluation – the theme of aes19 in Sydney, Australia.BlogL’évaluation en contexte de développement
Ce manuel est destiné aux personnes souhaitant s’initier à l’évaluation de programmes, en particulier en contexte de développement et de coopération internationale. À cet égard, tout en déroulant le fil d’une démarche évaluative classique, il présente…ResourceEvaluation contract checklist
This checklist is designed to help evaluators and clients identify key contractual issues that require agreement before an evaluation can commence.ResourceData party
A data party is a time-limited event of several hours where diverse stakeholders come together to collectively analyse data that have been collected.MethodA report on the mini-participatory learning and action (PLA) exercise in Zambia
This report is a concise summary of research evaluating the suitability and availability of reproductive health services for youth in Zambia using the Participatory Learning and Action (PLA) approach.ResourceMultiple lines and levels of evidence
Multiple lines and levels of evidence (MLLE) is a systematic approach to causal inference that involves bringing together different types of evidence (lines of evidence) and considering the strength of the evidence in terms of different indMethodJournals and logs
Journals and logs are forms of record-keeping tools that can be used to capture information about activities, results, conditions, or personal perspectives on how change occurred over a period of time.MethodIntegrity
Integrity refers to ensuring honesty, transparency, and adherence to ethical behaviour by all those involved in the evaluation process.MethodCultural competency
Cultural competency involves ensuring that evaluators have the skills, knowledge, and experience necessary to work respectfully and safely in cultural contexts different from their own.MethodFeasibility
Feasibility refers to ensuring that an evaluation can be realistically and effectively implemented, considering factors such as practicality, resource use, and responsiveness to the programme's context, including factors such as culture andMethodInclusion of diverse perspectives
Inclusion of diverse perspectives requires attention to ensure that marginalised people and communities are adequately engaged in the evaluation.MethodIndependence
Independence can include organisational independence, where an evaluator or evaluation team can independently set a work plan and finalise reports without undue interference, and behavioural independence, where evaluators can conduct and reMethodEvaluation accountability
Evaluation accountability relates to processes in place to ensure the evaluation is carried out transparently and to a high-quality standard.MethodTransferability
Transferability involves presenting findings in a way that they can be applied in other contexts or settings, considering the local culture and context to enhance the utility and reach of evaluation insights.MethodUtility
Utility standards are intended to increase the extent to which program stakeholders find evaluation processes and products valuable in meeting their needs.MethodProfessionalism
Professionalism within evaluation is largely understood in terms of high levels of competence and ethical practice.MethodPropriety
Propriety refers to ensuring that an evaluation will be conducted legally, ethically, and with due regard for the welfare of those involved in it and those affected by its results.MethodSystematic inquiry
Systematic inquiry involves thorough, methodical, contextually relevant and empirical inquiry into evaluation questions. Systematic inquiry is one of the guiding principles of the American Evaluation Association:MethodTransparency
Transparency refers to the evaluation processes and conclusions being able to be scrutinised.MethodEthical practice
Ethical practice in evaluation can be understood in terms of designing and conducting an evaluation to minimise any potential for harm and to maximise the value of the evaluation.MethodAccuracy
Accuracy refers to the correctness of the evidence and conclusions in an evaluation. It may have an implication of precision.MethodAccessibility
Accessibility of evaluation products includes consideration of the format and access options for reports, including plain language, inclusive print design, material in multiple languages, and material in alternative formats (such as online,MethodCompetence
Competence refers to ensuring that the evaluation team has or can draw on the skills, knowledge and experience needed to undertake the evaluation.MethodOutcome harvesting
Outcome Harvesting collects (“harvests”) evidence of what has changed (“outcomes”) and, working backwards, determines whether and how an intervention has contributed to these changes.Approach52 weeks of BetterEvaluation: Week 16: Identifying and documenting emergent outcomes of a global network
Global voluntary networks are complex beasts with dynamic and unpredictable actions and interactions. How can we evaluate the results of a network like this? Whose results are we even talking about?Blog