Title: Making Sense of Multiple Studies
1Making Sense of Multiple Studies
- Causal Analysis of Multiple Studies
2MULTIPLE TESTING
- Whenever we use multiple hypothesis tests to
examine the same basic biological or medical
question, or to search for relationships among
many variables ("fishing expedition"), the
results must be interpreted with great caution
3WHY?
- Because the chance of finding some statistically
significant relationship increases with the
number of tests we make, even when nothing
whatsoever is really going on biologically
4Cont.
- This chance is sometimes called the "class-wise"
error rate when applied to a specific class (or
group) of tests, and the "experiment-wise" error
rate when applied to all the hypothesis tests
used in a particular study. The experiment-wise
error rate may be much, much, much, much higher
than the "nominal" error rate a of any single one
of the tests.
5(No Transcript)
6EXAMPLE CASE-CONTROL TRAWLING.
- You run a case-control study to try to find out
the cause of some disease. Not knowing the cause,
you ask the patients and controls about each of
100 exposures that might have been involved. If
none of these exposures really had anything to do
with generating the disease, how many would you
expect to look suspicious (i.E. Statistically
significant exposure odds-ratio)?
7EXAMPLE OVERDOSING ON HYPOTHESIS TESTS.
- To evaluate the effectiveness of a new
pharmacologic agent and to find the optimal dose,
you randomly assign patients to placebo or 100,
200, 300 or 400 mg of the drug daily. At the end
of the study, you separately test whether each of
the doses was followed by a better results than
the placebo, using a5. If the drug doesn't work
at all, what's the chance that some dose will
appear to work?
8EXAMPLE MULTIPHASIC HEALTH SCREENING.
- Routinely, as part of the office physical you
give to every new patient and periodically to
continuing patients, you draw a blood sample and
send it to the lab for a smac-20. The report
labels every result outside the middle 95 of
values of healthy people as "abnormal." How
likely is a perfectly healthy patient to produce
a perfectly clean lab report?
9EXAMPLE MULTIPLE OUTCOMES.
- In a comparative trial of two surgical
procedures, you define 10 outcome variables to
describe the results of these procedures. Some
describe basic critical results, e.G. Life or
death and major surgical complications, while
others describe several aspects of quality of
life after surgery. If the average results of the
two procedures were identical on all scores,
what's the chance you'd see a spurious difference
anyway?
10EXAMPLE MULTIPLE LOOKS AT CLINICAL TRIALS
- In comparing two therapies over time, you
compare the patient's current status via an
hypothesis test every three months for ten years.
What's the chance you'll see no difference over
all that time?
11CRITERIA OF CAUSALITY
- Strength of association
- Consistency
- Temporality
- Dose response
- Reversibility
- Biologic plausibility
- Specificity
- Analogy
12STRENGTH OF ASSOCIATION
- The larger the relative risk, the stronger the
evidence for causality however, a small relative
risk (lt2) does not rule out cause and effect. - CONSISTENCYRepeatedly observed in different
studies, different populations, different times
and places, different study designs
13TEMPORALITY
- Cause precedes effect problem outcome (at a
sub-clinical level) may cause change in exposure
e.g. 1. Heart disease and exercise 2. Peptic
disease and spicy food 3. Lung cancer and recent
smoking cessation
14DOSE-RESPONSE
- Larger exposures associated with higher rates of
disease. Problems 1. Confounder can explain
dose-response relationship 2. Threshold
phenomenon higher level exposure (beyond a
threshold level) does not increase outcome rate
any further.
15REVERSIBILITY
- Reversing the exposure is associated with lower
rates of disease. (E.G. Smoking cessation). Can a
confounder explain this one? - BIOLOGIC PLAUSIBILITY Can a biologic mechanism
explain how the exposure causes the effect? But
beware biologic plausibility is always absent
for new discoveries
16SPECIFICITY
- One cause leads to one effect
- ANALOGY Cause and effect relationship already
established for a similar exposure-disease
combination - THE BETTER THE STUDY DESIGN THE STRONGER THE
EVIDENCE
17HIERARCHY OF RESEARCH DESIGN IN ESTABLISHING CAUSE
- 1. CASE REPORT
- 2. CASE SERIES
- 3. CASE-CONTROL
- 4. COHORT STUDY
- 5. CLINICAL TRIAL
18METANALYSIS
- CRITERION-BASED
- POOLING
- CRITERION-BASED
- EXAMPLE BCG
19(No Transcript)
20(No Transcript)
21POOLING
- Combining the patients in all trials and
reanalyzing the data, as if they had come from a
single large but stratified study
22EXAMPLES
- Quinidine for atrial fibrillation
- IV streptokinase for acute MI
23(No Transcript)
24(No Transcript)
25Pooling cont.
- Criteria for the study, design
- criteria for patients.
- both included.
- Used mostly for randomized studies.
- Each individual study may have little power, but
the combined analysis has a much higher power
(smaller confidence interval).
26Metanalysis Cont.
- Sometimes the individual studies had little power
because they were addressing a different (more
common) end point. - used to plan sample size for future studies.
27PROBLEMS WITH METAANALYSIS
- Remember metaanalysis deals only with "chance" by
increasing power and decreasing confidence
interval widths. It does not correct for bias and
confounding. - Patients, treatment, outcome evaluation may
differ across studies. What then?
28Publication bias
- One potential problem with metaanalysis is
publication bias That studies with positive
findings are more likely to have been published,
and hence be included in the metaanalysis than
negative studies.
29(No Transcript)
30Funnel Plot
31WHY READ THE JOURNALS
- Other sources of clinical guidelines
- Expert opinion
- Textbooks
- National bodies
- Editorials
32Embargo
- Why?
- How is it done?
- The WHI study
33SPAF Trial
- Stroke Prevention in Atrial Fibrillation
- 627 patients were randomized to warfarin,
aspirin, or placebo - Warfarin was superior.
34Average Clinicians Conclusion
- Every patient with atrial fibrillation who
doesnt have any of the exclusion criteria should
receive warfarin. - You would reach that conclusion if you read the
articles abstract and conclusion.
35My Conclusion
- Not every patient with atrial fibrillation should
be started on warfarin. - My reason?
- I read the whole article.
- What did I find?
36They Started with 18,376 Patients!
- Thats not bad by itself.
- Most were excluded for appropriate criteria
- BUT among the excluded
- 717 refused
- Another 1,084 their doctors refused
- Another 2,262 no reason was recorded
- Another 239 patient or doctor refused
anticoagulation. (So they were randomized into
the aspirin versus placebo trial.)
37Compare These Numbers
- 717 refused
- Another 1,084 their doctors refused
- Another 2,262 no reason was recorded
- Another 239 patient or doctor refused
anticoagulation. (So they were randomized into
the aspirin versus placebo trial.) - To 627 randomized (210 ended on warfarin)
38(No Transcript)
39Why not point out that at 12 mcg/L PSA IS 100
specific
- Misses early disease.
- Subjects an occasional patient to endless
invasive testing. - You are responsible for all of those.
- When proven wrong gives competitors an automatic
publication at your expense.
40HOW TO READ THE JOURNALS
- Read the conclusion
- If valid, would this be useful or interesting
(common disease, new finding)?
41Types of articles
- 1. THERAPY Randomization
- 2. DIAGNOSIS Independent blind comparison with
a gold standard - 3. CAUSE
- 4. COURSE/PROGNOSIS
- Inception
- Adequate follow up
42Clinical trial, Look for
- 1. Design randomization
- 2. Hypothesis stated 3. Clinically
significant end
points 4.
Sample size estimation 5. Subject selection,
description of subjects 6. Exclusions why? How
many
43Look for cont.
- 7. Blinding, outcome criteria 8.
Intervention, specific 9. Interim review,
guidelines 10. Drop outs, intention to treat
11. All clinically relevant outcomes 12. Few
comparisons (data derived hypotheses)
44Any study that
- Randomizes
- Patients similar to the ones I see
- And looks at all clinically relevant outcomes
45Clinical Trials Jargon
- Consecutive patients (versus a random sample)
- Baseline characteristics of patients (to see if
randomization worked) - Number of subjects and average duration of
follow-up (versus patient years) - Interim analysis, problems
- Cumulative incidence (versus incidence density)
46Jargon cont
- Relative risk (hopefully lt1)is rate of outcome
in a drug group rate of outcome in a placebo
group - Relative risk reduction (similar to attributable
risk ) (But here it is 1-RR) - Absolute difference in risk (similar to AR, very
important, sometimes not reported - Relative risk reduction versus absolute
difference in risk - Number needed to treat
47Jargon cont.
- Subgroup analyses (looking for effect modifiers)
- Problems with subgroup analyses (multiplicity,
data-derived hypotheses) other methods of data
torturing - "Intention to treat" versus efficacy (hazards of
non-randomized comparisons)
48DIAGNOSIS
- SENSITIVITY Spectrum
- SPECIFICITY Similar conditions
- WHO GETS THE GOLD STANDARD?
- Is it dependent on the test?
- Is it blinded?
49BAD ARTICLES
- Negative with no power
- Test with no evaluation of negatives
50Review of Bias and Confounding
51BIASES CASE-CONTROL STUDIES
- 1. SELECTION BIAS CASES Could the way you
selected your cases be related to exposure?
Diagnosis referral response availability
survival
52Biases cont.
- Controls if this person had the disease, would
she have been a case? - 2. Measurement (observation, information) bias
subjects recall, lying interviewer
53BIASES COHORT STUDIES
- 1. Selection bias randomization could the
way you assigned the subjects (or they assigned
themselves) (to exposed or unexposed) be related
to outcome? Volunteer non participation
compliance
542. MEASUREMENT BIAS
- Migration loss to follow up
misclassification random differential
55CONFOUNDING
- A confounder is 1. Associated with the exposure
2. Statistically associated with outcome (risk
factor) independent of the exposure 3. Not
necessarily a cause of the outcome 4. Not a
result of the exposure
56CONTROL OF BIAS AND COUNFOUNDING
- In the design
- 1. Randomization 2. Restriction
-disadvantages 3. Matching -as a concept
-matched pairs -disadvantages
-overmatching
57In the design cont.
- Additional controls
- Include assessment of known risk factor
- Include data on all risk factors
- Reliability check
58IN THE ANALYSIS
-
- 1. Stratification 2. Adjustment 3.
Multivariate analysis