How very reassuring, Gill.  I agree totally, BUT this is an anathema to the traditional [stereotypical] systematic reviewer, who has REPLACED reflection and theory-building with a technical lexicon of coding, categorising, checking, double-checking and above all COUNTING.   The link between theory-building systematic review and reflective practice is key, and one that has not previously been made explicit in many how-to-do-it guides.

 

Some of this is just the good old principles of qualitative research, where validity (and scholarship more generally) are defined in terms of reflexivity, making assumptions explicit, immersion in the data, critical awareness of the inevitability of error, etc etc rather than in terms of the 'view from nowhere' and some universal, context-free truth. 

 

However, I think we're getting somewhere beyond this.  The role of the systematic [theory-building] reviewer is not merely to reflect (and thereby build theory) but also to prompt others to reflect.  Dare I introduce a Bakhtinian perspective: there is no story without an audience, and the way the story is told is shaped and constrained by the relationship between the narrator and the audience.  When we produce a review, especially when we do so for particular policymakers in particular contexts – and especially when we are trying to pull out theories of change – we form a triangle: the reviewer + the review + the audience, all of which exist in a dynamic tension.  This tension is perhaps most pressing when we’re doing a ‘rapid review’ to a deadline for a painfully tiny budget!  But always the audience is present, even if only in a distant and assumed way.

 

It’s also true that the [realist] evaluator exists in a three-way tension: the evaluator, the policymakers/practitioners and the ‘case’.  The evaluator seeks to pull out the theories-in-use – and in order to do so, s/he reflects and asks the practitioners to reflect.  A good evaluator will tease out theories of action and facilitate the process by which the practitioners engage with and add to emerging interpretations.  So realist evaluation is inextricably bound up with reflection/facilitation.  As you say Gill, the realist reviewer both loses something (direct access to the reflections of the practitioners and the authors of the original studies) and gains something (a degree of critical distance).

 

Can anyone take this thread forward or challenge it.  My day job too calls now!

 

(does anyone else find it ironic that we’re doing our thinking outside our day jobs?)

 

trish

 

Trisha Greenhalgh

Professor of Primary Health Care and Director, Healthcare Innovation and Policy Unit

Centre for Primary Care and Public Health

Blizard Institute

Barts and The London School of Medicine and Dentistry

Yvonne Carter Building

Turner Street

London E1 2AT

t : 020 7882 7325 (PA) or 7326 (dir line)

f : 020 7882 2552

e: [log in to unmask]

 

 

http://www.icms.qmul.ac.uk/chs/staff/trishagreenhalgh.html

 

 

-----Original Message-----
From: Realist and Meta-narrative Evidence Synthesis: Evolving Standards [mailto:[log in to unmask]] On Behalf Of Gill Westhorp
Sent: 13 July 2011 00:33
To: [log in to unmask]
Subject: Re: How much should we impugn when authors don't make things explicit?

 

Hi all

Ah yes, I so agree.  I consider theory building to be a necessary part of science and therefore of good science.  I consider good theory building practice to include being as well informed as one can be, within all the practical constraints.  And I see realist (and probably meta-narrative but I haven't tried one of those) approaches as being purpose designed for the task... it's one of the reasons why I like it.

 

To expound:  I believe that there is no such thing as 'interpretation free' analysis  -  there are after all levels of interpretation and decision-making reflected in every step of every piece of prior research or evaluation that we synthesise, as well as every step of our review process itself.  I don't see a problem with taking that one step further and making 'next step interpretations' (ie theory building) - so long as we are explicit about the fact that that's what we're doing.  In fact, I almost consider it a duty.  Who else is in a better position to do it than those who have immersed themselves in the research and evidence?  So while I have sympathy for the desire to stay 'close to the literature', I see that as 'being as well informed as one can be' before proposing a theory that accounts for the findings.

 

As for practitioners not having theory - here I in fact disagree.  I think they do have theories - NB plural - albeit at the naieve level sometimes!  It's possible in realist evaluation to unpack those.  I tell stories in my training about experiences in doing so - it's often when practitioners find out that they're operating on different theories than their colleagues, even if they work together all the time.  This provides a wonderful opportunity to deepen reflective practice, and to assist them to access relevant MRT and evidence related to same.  The difficulty in realist synthesis is that we don't have the same direct access to the practitioners to find out what their theories are, and those who did the primary research/evaluation didn't always either find out or didn't record same. 

 

That's both a strength and a weakness for the synthesis analysts - a weakness because it reduces the clues about where to look for theory and because it means we can't check whether interventions were in fact built on 'the same' theory (now there's an issue for reflective thought: what would constitute adequate evidence that the theories were in fact 'the same'?). But it can be turned into a strength because our job as synthesists is not necessarily to test THEIR theories but to build and test theory ACROSS (insert here whichever version you're doing) ... across manifestations of an intervention; across interventions using similar theories; across interventions using different theories but similar mechanisms; or in MNR, across similar topics but from widely disparate theoretical bases...) 

 

Gotta go.  Day job calls.

Cheers

Gill