Greetings all, I searched the list archives, and could not find this question addressed (though there are several that discuss the formation and uses of PICO - as well as Andrew Booth's collection of wild-type PICO variants from 2009). A colleague of mine wishes to know: Has anyone developed and validated a rubric or other tool for evaluating the quality of a PICO question? The trick to this question, I think, is defining what "quality" is: output from a search? adequate representation of a learning need? or merely well-structured (which is somewhat recursive to the definition of the PICO question)? I imagine the outcomes for a output-based validation study would be the quality of the output of a database search (medline, embase, etc.), but the more I think about how one would assess this quality, things start spinning in my head. In the sense of learning need, I'm also somewhat of the mind that a "good" PICO question is best "validated" by the person asking the question themselves - does it really represent what they assess is their own learning need (or might that be adjudicated by something like cognitive task analysis)? Assessing the degree of "well-constructedness" of a PICO question - seems fairly straightforward and may not require "validation" as it's a relatively artificial construct to begin with. Does anyone have any pertinent references I can share with my colleague? Am I thinking about this the right way? Thanks in advance, John John Epling, MD, MSEd, FAAFP Associate Professor and Chair Department of Family Medicine Co-Director, Studying-Acting-Learning-Teaching Network (SALT-Net) Associate Professor, Public Health and Preventive Medicine SUNY-Upstate Medical University Syracuse, NY [log in to unmask] Clinical: http://www.upstate.edu/findadoc/eplingj Faculty: http://www.upstate.edu/faculty/eplingj |