Can someone send me a link explaining the GRADE system of ASSESSMENT of studies in detail please?
Rinku
--------------------------------------------
On Tue, 10/2/15, P.M.M. Bossuyt <[log in to unmask]> wrote:
Subject: Re: Genetic tests and Predictive validity
To: [log in to unmask]
Date: Tuesday, 10 February, 2015, 13:16
Thanks to you all
for an interesting discussion.
Sensitivity,
specificity, negative and positive predictive values are all
group-based statistics
So are relative
risks and risk differences, estimated in RCT of
interventions….
As elsewhere in EBM,
applying group-based statistics to the problems of an
individual patient requires additional steps and
assumptions, some of which are problematic.
Yes, sensitivity and
specificity can (also) vary with prevalence.
Leeflang MM, Bossuyt PM, Irwig L.
Diagnostic test
accuracy may vary with prevalence: implications for
evidence-based diagnosis.
J Clin Epidemiol.
2009 Jan;62(1):5-12. doi: 10.1016/j.jclinepi.2008.04.007.
There is definitely
“spin” in reporting test accuracy studies, as was noted
in some contributions. Sometimes the primary outcome measure
(as registered) changes
to negative predictive value in the final publication. You
can guess why…
Korevaar DA, Ochodo
EA, Bossuyt PM, Hooft L.
Publication and
reporting of test accuracy studies registered in
ClinicalTrials.gov.
Clin Chem. 2014
Apr;60(4):651-9. doi:
10.1373/clinchem.2013.218149.
Patrick
Bossuyt
AMC - University of
Amsterdam
From: Evidence based health (EBH)
[mailto:[log in to unmask]]
On Behalf Of Huw Llewelyn [hul2]
Sent: Monday, 9 February, 2015 23:17
To: [log in to unmask]
Subject: Re: Genetic tests and Predictive
validity
I agree that the
terminology for diagnosis is ambiguous and probably
confusing for those not immersed in its practical
application day in day out. .
A diagnostic test in its broad sense is any test that leads
to a diagnosis but also when deciding to treat (a form of
diagnostic refinement) and also monitoring the outcome.
A symptom or physical sign is the 'result' of the
'test' of listening or examining the patient.
Symptoms, signs and test results are all 'diagnostic
findings'. The use of 'diagnostic' in this sense
does not imply confirmatory (we say at times that findings
are 'diagnostic', i.e. 'pathognomonic'). It
is combinations of findings that usually confirm a
diagnosis.
I regard a screening test result as a form of presenting
complaint that also leads to a differential diagnosis. Both
bring to our attention patients with a higher probability of
a 'diagnosis of interest' in a big population. The
subsequent reasoning may lead
to changing the probabilities of the differential diagnoses
and hopefully confirming one of them by showing the presence
of a 'sufficient' diagnostic criterion. (It is at
this stagfe that 'over-diagnosis' happens - because
of faulty definitive diagnostic criteria.)
We then hope to show that the expected benefits from a
treatment (e.g. avoiding metastases) outweigh the expected
harms. (This can be modelled using Decision Analysis.) Some
findings are better at doing this than others. As far as I
can understand, it is this
final stage that Teresa's data was about.
I explain how to obtain evidence for the value of
'diagnostic' findings at these different stages of
the medical problem solving process in the final chapter of
Oxford Handbook of Clinical Diagnosis.
Huw
From: Brian Alper
MD <[log in to unmask]>
Sender:
"Evidence based health (EBH)" <[log in to unmask]>
Date: Mon, 9 Feb
2015 12:09:54 +0000
To: <[log in to unmask]>
ReplyTo: Brian
Alper MD <[log in to unmask]>
Subject: Re:
Genetic tests and Predictive validity
As Huw recently
shared evaluation of diagnostic/predictive tests can be
different depending on the purpose. Huw’s list
was:
1.
For population
screening
2.
For differential
diagnosis
3.
For diagnostic
confirmation
4.
For diagnostic
exclusion
5.
For predicting
outcomes (predicting future risk)
These concepts are
further complicated by imprecise use of
language. Many of us use “screening”
to mean testing for a diagnosis in people with no
symptoms.
In this context screening differs from diagnostic testing
not so much in the science/math/statistical approach but
often in the baseline risk (lower prevalence/baseline
risk/pretest probability in the screened population) and in
the values/preferences for
weighing benefits and harms – leading many to consider a
higher threshold for confidence in benefit (greater demand
for evidence for benefit) to recommend screening for an
asymptomatic person than to recommend a diagnostic test for
a symptomatic person.
But this does get
confused in general language because testing is often a
multi-stage process, so the terminology used could be a
“screening test” and a “confirmatory
test” and that language may get used for screening or
diagnosis in the earlier description of the
terms.
So there is a
substantial problem with the terminology when the terms
themselves are used in many different ways.
A diagnostic test is
a test used in symptomatic persons (to distinguish from a
screening test)
A diagnostic test is
a test which is able to confirm the diagnosis (as distinct
from earlier testing that increases or decreases our
suspicion for the diagnosis)
A diagnostic test is
any test that implies an increase or decrease in the
likelihood of the condition (and thus includes all the other
tests noted above by any
term)
A diagnostic test is
used to describe the result of the test rather than the test
itself. If we have certainty after testing then it was
a diagnostic test.
All of this makes
communication and education around diagnostic testing more
challenging.
Brian S. Alper, MD,
MSPH, FAAFP
Founder of DynaMed
Vice President of EBM Research and Development, Quality
& Standards
dynamed.ebscohost.com
From: Evidence based health (EBH) [mailto:[log in to unmask]]
On Behalf Of OWEN DEMPSEY
Sent: Monday, February 02, 2015 7:49 AM
To: [log in to unmask]
Subject: Re: Genetic tests and Predictive
validity
Dear All,
Brian, you said:
"But another
consideration is sometimes tests are used for
“diagnostic” purposes – Does the patient have or not
have a certain diagnosis? – an in these cases
sensitivity,
specificity, PPV*, NPV*, positive likelihood ratio, and
negative likelihood ratio (* with prevalence to put into
perspective) are clear."
What about the screening
situation, e.g. a breast cancer
screening mammography leads to a biopsy and a pathology
report: if the report is genuinely 'borderline' e.g.
the pathologist reports seeing some kind of atypia,
'indolent changes', in-situ changes etc. (changes
for which I think there is no evidence for any
net benefit of treatment; ref below) How much clarity is
there then? Is this a kind of 'no gold standard
situation'? So the so called diagnostic tests ROC
curve(s) becomes guesswork? Maybe this shouldn't be
called a diagnostic test?
Owen;
(Esserman LJ, Thompson
IM, Reid B. Overdiagnosis and overtreatment in cancer: an
opportunity for improvement. JAMA. 2013 Aug 28;
310(8):797-8.)
On 2 February 2015 at
10:08, Raatz Heike <[log in to unmask]>
wrote:
Dear
all,
Question is though:
is the genetic test you want to evaluate actually used as a
diagnostic test? From what I understand from the case
mentioned by Teresa she is interested not in whether a
genetic test accurately recognizes whether you have a
certain genotype but whether you will in the future develop
a certain phenotype.
So you are not trying to find out whether the patient
currently suffers from a condition but the risk of
developing a condition in the future. Now unless you have a
dominant gene that will always lead to the expression of a
certain phenotype (like Huntingtons)
you need to consider whether that genotype is not just one
of many factors that can lead to a certain condition. For
the examples mentioned like Mammaprint prognostic modelling
seems much more appropriate to me than diagnostic accuracy
though ultimately you
need RCTs to prove that they improve patient reported
outcomes and from what I saw last those don’t
exist.
Best
wishes, Heike
Heike Raatz, MD, MSc
Basel Institute for
Clinical Epidemiology and Biostatistics
Hebelstr.
10
4031
Basel
Tel.:
+41 61 265 31 07
Fax:
+41 61 265 31 09
E-Mail:
[log in to unmask]
From: Evidence based
health (EBH) [mailto:[log in to unmask]]
On Behalf Of Majid Artus
Sent: Monday, February 02, 2015 10:04 AM
To:
[log in to unmask]
Subject: Re: Genetic tests and Predictive
validity
Dear All,
Is it, or is it not,
correct that one should follow the classic teaching that
(loosely and notwithstanding the false partition here):
from
patient's perspective, sensitivity/specificity are what
is relevant; and from clinician's perspective, PPV/NPPV
are what is relevant?
Also, it is, isn't it, crucial to
consider the media take on outcome of research and how
careful researchers need to be in selecting the way they
present the outcome of their research? MRI (NMR) in
diagnosing autism is one example that springs to mind - the
high sensitivity was jumped on by the media presenting it as
a very accurate test missing the role of the varying
prevalence in certain settings.
I find this discussion trail hugely
thought provoking!
Best Regards
Majid
On 2 February 2015 at 00:19, Mark V Johnston
<[log in to unmask]>
wrote:
At the same time, don’t we need to
know whether the patient probably
has or does not have the condition of
interest? Yes, prevalence and other
factors affect PPV and NPV, but in my
opinion we need to move away from the
oversimplified notion that test interpretation depends
on a single factor.
From:
Evidence based health (EBH) [mailto:[log in to unmask]]
On Behalf Of Mayer, Dan
Sent: Friday, January 30, 2015 8:08 PM
To:
[log in to unmask]
Subject: Re: Genetic tests and Predictive
validity
Hi
Teresa,
You are
absolutely correct. This is why we should demand that
diagnostic studies ONLY present the results of Sensitivity,
Specificity and Likelihood ratios.
This issue
has been a serious problem for many years and it is about
time that more people spoke up about it. Also, journal
editors and peer reviewers should be up in arms against
the practice of reporting PPV and NPV.
Best
wishes
Dan
From: Evidence based
health (EBH) [[log in to unmask]]
on behalf of Benson, Teresa [[log in to unmask]]
Sent: Friday, January 30, 2015 11:59 AM
To:
[log in to unmask]
Subject: Genetic tests and Predictive
validity
I’ve just started reading the
literature on genetic tests, and noticing how many of them
tend to focus on predictive value—that is, if a certain
test accurately
predicts whether a patient will or won’t get a particular
phenotype (condition), the authors suggest the test should
be used. But if we’re deciding whether to order the
test in the first place, shouldn’t we be focused on
sensitivity and specificity instead, not PPV
and NPV? Predictive value is so heavily dependent on
disease prevalence. For example, if I want to get
tested for a disease with a 2% prevalence in people like me,
I could just flip a coin and regardless
of the outcome, my “Coin Flip Test” would show an NPV
of 98%! So what does NPV alone really tell me, if
I’m not also factoring out prevalence—which would be
easier done by simply looking at sensitivity and
specificity? Someone please tell me where my thinking
has gone awry!
For a
concrete example, look at MammaPrint, a test which reports
binary results. In addition to hazard ratios, study
authors often tout statistically significant
differences between the probabilities of recurrence-free
survival in the MammaPrint-High Risk vs. MammaPrint-Low Risk
groups (essentially the test’s predictive values).
In the RASTER study (N = 427), 97% of the patients with a
“Low Risk” test result did not
experience metastasis in the next 5 years. Sounds
great, right? But when you look at Sensitivity, you
see that of the 33 patients in the study who
did experience metastasis, only 23 of them were
classified as “High Risk” by MammaPrint, for a 70%
sensitivity. If patients and clinicians are looking
for a test to inform their decision about adjuvant
chemotherapy for early stage breast cancer, wouldn’t
the fact that the test missed 10 out of 33 cases be more
important than the 97% NPV, an artifact of the extremely low
5-year prevalence of metastasis in this cohort (only 33 out
of 427, or 0.7%)?
Drukker et al. A prospective evaluation
of a breast cancer prognosis signature in the observational
RASTER study. Int J Cancer 2013. 133(4):929-36.
http://www.ncbi.nlm.nih.gov/pubmed/23371464
Retel et al. Prospective
cost-effectiveness analysis of genomic profiling in breast
cancer. Eur J Cancer 2013. 49:3773-9.
http://www.ncbi.nlm.nih.gov/pubmed/23992641
(Provides actual true/false positive/negative
results)
Thanks so much!
Teresa Benson, MA, LP
Clinical Lead, Evidence-Based
Medicine
McKesson Health Solutions
18211 Yorkshire
Ave
Prior Lake, MN 55372
[log in to unmask]
Phone:
1-952-226-4033
Confidentiality Notice:
This e-mail message, including any attachments, is for the
sole use of
the intended recipient(s) and may contain confidential and
privileged information. Any unauthorized review, use,
disclosure or distribution is prohibited. If you are not the
intended recipient, please contact the sender by reply
e-mail and destroy all copies
of the original message.
-----------------------------------------
CONFIDENTIALITY NOTICE: This email and any attachments may
contain confidential information that is protected by law
and is for the sole
use of the individuals or entities to which it is
addressed. If you are not the intended recipient, please
notify the sender by replying to this email and destroying
all copies of the communication and attachments. Further
use, disclosure, copying, distribution
of, or reliance upon the contents of this email and
attachments is strictly prohibited. To contact Albany
Medical Center, or for a copy of our privacy practices,
please visit us on the Internet at
www.amc.edu.
--
Please note my new email address:
[log in to unmask]
Dr Majid Artus PhD
NIHR Clinical Lecturer in General Practice
Arthritis Research UK Primary Care
Centre
Research Institute for
Primary Care & Health Sciences
Keele
University
Staffordshire, ST5 5BG
Tel: 01782 734826
Fax: 01782 733911
http://www.keele.ac.uk/pchs/
Please
consider the environment before printing this email.
This email and its attachments are intended for the above
named only and may be confidential. If it has come to you in
error you should not copy or show it to anyone; nor should
you take any action based on it, other than to notify the
sender of the error by
replying to the sender. Keele University staff and students
are required to abide by the University's conditions of
use when sending email. Keele University email is hosted by
a cloud provider and may be stored outside of the
UK.
--
Owen Dempsey MBBS MSc
MRCGP RCGP cert II
07760 164420
GPwsi Substance Misuse
Locala and Kirklees Lifeline Addiction Service
AMC Disclaimer :
http://www.amc.nl/disclaimer
|