Part of our commitment to better evaluation is making sure that evaluation itself is evaluated better. Like any intervention, evaluations can be evaluated in different ways.
This week, Arnaldo Pellini (Senior Research Fellow, Overseas Development Institute and Lead for Learning at the Knowledge Sector Initiative, Indonesia) and Louise Shaxson (Senior Research Fellow, Overseas Development Institute) reflect on some of the challenges around monitoring, evaluating and learning (MEL) from adaptive programmes.
Last week, we started our focus on Adaptive Management with a blog post by Patricia Rogers that explored how monitoring and evaluation can support adaptive management. This week, we're continuing this series with a guest blog from Fred Carden and Arnaldo Pellini, in which they discuss what they learned about adaptive management in a major project on developing capacity for evidence-based policy.
Adaptive management is usually understood to refer to an iterative process of reviewing and making changes to programmes and projects throughout implementation. Commonly associated with environment and resource management, it's becoming more common in other areas of program management and development. Over the next few weeks, we'll be focusing on the increasing interest in how monitoring and evaluation can support adaptive management.
All too often conferences fail to make good use of the experience and knowledge of people attending, with most time spent presenting prepared material that could be better delivered other ways, and not enough time spent on discussions and active learning. With closing dates for two evaluation conferences fast approaching (the Australasian Evaluation Society and the American Evaluation Association), could you propose something more useful, that would demonstrate how much we know and care about communicating and using information?
The wonderful thing about BetterEvaluation is that it is, at its core, a platform to co-create and share knowledge about how to better conduct, use and manage evaluations.
Evaluation practitioners and managers, experts and partner organizations work together to create and learn from improved knowledge and practice in monitoring and evaluation. We support three interconnected areas of activity - capacity strengthening, M&E research and development, and the BetterEvaluation toolbox, which includes the Rainbow Framework and the BetterEvaluation resource library.
Last week we launched our newest theme page Sustained and Emerging Impacts Evaluation, authored by Jindra Cekan (Valuing Voices), Laurie Zivetz (Valuing Voices), and Patricia Rogers (BetterEvaluation/ANZSOG). The page argues for the need to go back and evaluate the impacts of a project or programme some time after the end of an intervention, and gives some advice on how to do this.
This blog post was originally posted on the Australia and New Zealand School of Government (ANZSOG) blog. It has been republished with ANZSOG's permission. You can find the BetterEvaluation resource page for the article discussed here.
Happy New Year! From everyone on the BetterEvaluation team we hope you’ve had a great start to your year. In our first newsletter for the year, we're sharing some highlights from 2016 and some of our plans for 2017... We'd love to hear yours at the end of the blog so please drop us a line or leave a comment down the bottom!
Happy Holidays! In celebration of a great year, we've collected a list of favourite resources of 2016 that are freely available online so that we could share them with you as a bit of an end of year gift. The resources aren't necessarily new - but they are things that have been discovered for the first time or rediscovered in 2016. We have a wide range of cool tools, guides, examples and other resources that help us do evaluation better and think about it differently.