Hello
I will probably regret this hasty response, especially given that the
main reason my family can eat is a tiny tiny portion of those millions
spent on evaluation. (I'd be surprised if it was much more than a
million in total each year for the museum sector though. Most contracts
for evaluation are between £1k and 10k so it may even be much less.)
I think there is much room for improvement in the way the sector
approaches evaluation. One place to start is to revamp the old MLA
Generic Learning Outcomes and Generic Social Outcomes into a much more
sophisticated-but-simple tool based on the Theory of Change which
integrates learning and social outcomes, and incorporates digital.
However, I don't think this is the main requirement. The reason why
evaluations make less impact than they should is due to the nature of
project funding, the closed endpoint and fixed outcomes.
The closed endpoint: What is the point of a summative evaluation if that
really is the end and there is nobody in post to enact any
recommendations for change? The main problem is that the summative
evaluation has to be done within a window of time that allows some
actions to be taken, yet too early to prove any kind of medium to long
term impact. I'm keen, as an evaluator, to act as a critical friend that
helps institutions increase the impact and longevity of their
initiatives from the start, throughout, and beyond the conclusion of the
initiating phase. Mostly the briefs don't allow me to act like this, and
this is usually because the evaluation report is required by the funder
to prove that the funds were spent as planned.
The fixed outcomes: Funders require projects to state in too fixed a way
what the outcomes will be. Evaluation processes can be designed to pick
up on unpredicted outcomes but they aren't seen as very important, or
there is no mechanism in the project for those unpredicted outcomes to
be built on.
I hope this is helpful testimony!
Best wishes
Bridget
On 06/06/2012 12:51, Mia wrote:
> There's an interesting post called 'Why evaluation doesn't measure up'
> on the Museums Association site
> http://www.museumsassociation.org/museums-journal/comment/01062012-why-evaluation-doesnt-measure-up
> or http://bit.ly/L9FlQz where they say:
>
> "No one seems to have done the sums, but UK museums probably spend
> millions on evaluation each year. Given that, it’s disappointing how
> little impact evaluation appears to have, even within the institution
> that commissioned it."
>
> and:
>
> "Summative evaluations are expected to achieve the impossible: to help
> museums learn from failure, while proving the project met all its
> objectives. Is it time to rethink how the sector approaches
> evaluation?"
>
> I'm curious to know what others think. Are they right? Or are they
> missing something?
>
> Cheers, Mia
>
> --------------------------------------------
> http://openobjects.org.uk/
> http://twitter.com/mia_out
>
> ****************************************************************
> website: http://museumscomputergroup.org.uk/
> Twitter: http://www.twitter.com/ukmcg
> Facebook: http://www.facebook.com/museumscomputergroup
> [un]subscribe: http://museumscomputergroup.org.uk/email-list/
> ****************************************************************
****************************************************************
website: http://museumscomputergroup.org.uk/
Twitter: http://www.twitter.com/ukmcg
Facebook: http://www.facebook.com/museumscomputergroup
[un]subscribe: http://museumscomputergroup.org.uk/email-list/
****************************************************************
|