520 likes | 655 Views
Evaluating Medical Literature. Dr. Shounak Das Aug. 26, 2008. Problem-Solving Using the Medical Literature. Searching for the answer – 4 types of clinical questions: Therapy determine the effect of a treatment Harm ascertain the effect of exposure to a harmful agent Diagnosis
E N D
Evaluating Medical Literature Dr. Shounak Das Aug. 26, 2008
Problem-Solving Using the Medical Literature • Searching for the answer – 4 types of clinical questions: • Therapy • determine the effect of a treatment • Harm • ascertain the effect of exposure to a harmful agent • Diagnosis • establish the power of an intervention to distinguish those with and without the target condition • Prognosis • estimating the future course of a patient’s disease
Problem-Solving Using the Medical Literature • Sources of evidence: • Prefiltered • authors have already accumulated best of published +/- unpublished evidence (i.e. a systematic review) • examples = Best Evidence, Cochrane Library, UpToDate, and Clinical Evidence • Unfiltered • MEDLINE = U.S. National Library of Medicine database (contains primary studies and reviews) • world wide web
Evaluation of Primary Studies Assessing Therapies • When using the medical literature to answer a clinical question, approach the study using three discrete steps* : • Are the results of the study valid? • What are the results? • How can I apply these results to patient care? *these 3 steps are applicable to primary studies assessing harm, diagnosis, and prognosis as well as therapies; they are also applicable for the evaluation of systematic reviews
1. Are the Results of the Study Valid? • Assessing validity requires answering 2 further questions: • Did experimental and control groups begin the study with a similar prognosis? • Did experimental and control groups retain a similar prognosis after the study started?
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • To determine whether or not the experimental and control groups began the study with similar prognoses, we ask 4 questions: • Were patients randomized? • Was randomization concealed (blinded or masked)? • Were patients analyzed in the groups to which they were randomized? • Were patients in the treatment and control groups similar with respect to known prognostic factors?
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • Were patients randomized? • randomization is important so that the experimental and control groups are matched for both known and unknown prognostic factors
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • There are many examples of RCT’s contradicting the findings of observational studies, i.e.: Clinical Question Observational Study Findings RCT Findings
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • Was randomization concealed (blinded or masked)? • Some years ago, a group of Australian investigators undertook a randomized trial of open vs laparoscopic appendectomy. • At night, the attending surgeon's presence was required for the laparoscopic procedure but not the open one. • When an eligible patient appeared, the residents checked the attending staff and the lineup for the operating room and, depending on the personality of the attending surgeon and the length of the lineup, held the translucent envelopes containing orders up to the light. As soon as they found one that dictated an open procedure, they opened that envelope. • If patients who presented at night were sicker than those who presented during the day, the residents' behavior would bias the results against the open procedure.
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • Were patients analyzed in the groups to which they were randomized? • This principle of attributing all patients to the group to which they were randomized results in an intention-to-treat analysis, which is analysis of outcomes based on the treatment arm to which patients were randomized, rather than which treatment they actually received. This strategy preserves the value of randomization: prognostic factors that we know about--and those we do not know about -- will be, on average, equally distributed in the two groups; and the effect we see will result simply from the treatment assigned.
Let us assume that the treatment is ineffective and that the true underlying event rate in both treatment and control patients is 20%. If the 20 nonadherent patients are sicker and their event rate (60%) is higher, the nonadherent patients will suffer 12 of the 20 events destined to occur in the treated patients. If one compares only the adherent patients (with an event rate of 8/80, or 10%) with the control group (event rate 20/100, or 20%), one will mistakenly conclude that treatment cuts the event rate in half. 1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis?
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? 20 events 12 events 8 events 8/80 = 10% vs. 20/100 = 20% 50% RRR in treatment arm 20/100 = 20% vs. 20/100 = 20% 0% RRR in treatment arm 20 events
1a. Did Experimental and Control Groups Begin the Study with a Similar Prognosis? • Were patients in the treatment and control groups similar with respect to known prognostic factors? • Clinicians should look for a display of prognostic features of the treatment and control patients at the study's commencement -- the baseline or entry prognostic features. Although we never will know whether similarity exists for the unknown prognostic factors, we are reassured when the known prognostic factors are well balanced.
1. Are the Results of the Study Valid? • Assessing validity requires answering 2 further questions: • Did experimental and control groups begin the study with a similar prognosis? • Did experimental and control groups retain a similar prognosis after the study started?
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started? • To determine whether or not the experimental and control groups retain a similar prognosis, we ask 4 questions: • Were patients aware of group allocations? • Were clinicians aware of group allocations? • Were outcome assessors aware of group allocations? • Was follow-up complete?
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started? • Were patients aware of group allocations? • Patients who take a treatment that they believe is efficacious may feel and perform better than those who do not, even if the treatment has no biologic action – the placebo effect. • The best way to avoid the placebo effect skewing the results is to ensure that patients are unaware of whether they are receiving the experimental treatment.
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started? • Were clinicians aware of group allocations? • If randomization succeeds, treatment and control groups in a study begin with a very similar prognosis. However, randomization provides no guarantee that the two groups will remain prognostically balanced. Differences in patient care other than the intervention (cointerventions) under study can bias the results. • Effective blinding eliminates the possibility of either conscious or unconscious differential administration of effective (co)interventions to treatment and control groups.
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started? • Were outcome assessors aware of group allocations? • Unblinded study personnel who are measuring or recording outcomes such as physiologic tests, clinical status, or quality of life may provide different interpretations of marginal findings or may offer differential encouragement during performance tests, either one of which can distort results.
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started? • Was follow-up complete? • The greater the number of patients who are lost to follow-up, the more a study's validity is potentially compromised. The reason is that patients who are lost often have different prognoses from those who are retained. The situation is completely analogous to the reason for the necessity for an intention-to-treat analysis.
1b. Did Experimental and Control Groups Retain a Similar Prognosis After the Study Started?
1. Are the Results of the Study Valid? • The final assessment of validity is never a "yes" or "no" decision. Rather, think of validity as a continuum ranging from strong studies that are very likely to yield an accurate estimate of the treatment effect to weak studies that are very likely to yield a biased estimate of effect.
Evaluation of Primary Studies Assessing Therapies • When using the medical literature to answer a clinical question, approach the study using three discrete steps : • Are the results of the study valid? • What are the results? • How can I apply these results to patient care?
2. What are the Results? • Once validity of a study is assessed, the next step is to analyze the results. This involves looking at: • How large was the treatment effect? • How precise was the estimate of the treatment effect?
2a. How Large Was the Treatment Effect? • Most frequently, randomized clinical trials carefully monitor how often patients experience some adverse event or outcome. Examples of these dichotomous outcomes ("yes" or "no" outcomes -- ones that either happen or do not happen) include cancer recurrence, myocardial infarction, and death. Patients either do or do not suffer an event, and the article reports the proportion of patients who develop such events.
2a. How Large Was the Treatment Effect? • Even if the outcome is not one of these dichotomous variables, investigators sometimes elect to present the results as if this were the case. For example, in a study of the use of forced expiratory volume in 1 second (FEV1) in the assessment of the efficacy of oral corticosteroids in patients with chronic stable airflow limitation, investigators defined an event as an improvement in FEV1 over baseline of more than 20%. • The investigators' choice of the magnitude of change required to designate an improvement as "important" can affect the apparent effectiveness of the treatment
When clinicians consider the results of clinical trials, they are interested in the association between a treatment and an outcome. Consider three different treatments that reduce mortality administered to three different populations. Although all three treatments reduce the risk of dying by a third, this piece of information is not adequate to fully capture the impact of treatment. Expressing the strength of the association as a relative risk (RR), a relative risk reduction (RRR), an absolute risk reduction (ARR) or risk difference (RD), an odds ratio (OR), or a number needed to treat (NNT) or number needed to harm (NNH) conveys a variety of different information. 2a. How Large Was the Treatment Effect?
2a. How Large Was the Treatment Effect? Risk RR RRR 28% 63% 37% 45% 1. In this study, the risk of death in the ligation arm is 18/64 = 28% 2. The risk of death in the sclerotherapy arm is 29/65 = 45% 3. The relative risk (RR) of death in the ligation arm is 28%/45% = 63% 4. The relative risk reduction (RRR) of death in the ligation arm is (45% - 28%)/45% = 17%/45% = 37% • the RRR tells us the proportion of baseline risk that is removed by the therapy • using nontechnical language, we would say that ligation decreases the relative risk of death by 37% compared to sclerotherapy
2a. How Large Was the Treatment Effect? Risk ARR NNT 28% 17% 6 45% (28%) • The absolute risk reduction (ARR) = 45% – 28% = 17% • the ARR tells us what proportion of patients are spared the adverse outcome if they receive the experimental therapy, rather than the control therapy The number needed to treat (NNT) = 1/ARR = 6 • the NNT is the number of patients one would need to treat to prevent an adverse event
2a. How Large Was the Treatment Effect? Risk Odds OR 28% 0.39 0.49 45% 0.81 (28%) The odds of death in the ligation arm is 18/46 = 0.39 The odds of death in the sclerotherapy arm is 29/36 = 0.81 • The odds ratio (OR) = 0.39 ÷ 0.81 = 0.49 • the (OR) represents the proportion of patients with the target event divided by • the proportion without the target event • in most instances in medical investigation, odds and risks are approximately equal
2a. How Large Was the Treatment Effect? • Relative Risk and Odds Ratio vs Absolute Risk Reduction: Why the Fuss? • distinguishing between OR and RR will seldom have major importance • we must pay much more attention to distinguishing between the OR and RR vs the ARR (or its reciprocal the NNT)
2a. How Large Was the Treatment Effect? • Forrow and colleagues demonstrated that clinicians were less inclined to treat patients after presentation of trial results as the absolute change in the outcome compared with the relative change in the outcome. • The pharmaceutical industry's awareness of this phenomenon may be responsible for their propensity to present physicians with treatment-associated relative risk reductions. (ARR) (10%) (0.5%)
2. What are the Results? • Once validity of a study is assessed, the next step is to analyze the results. This involves looking at: • How large was the treatment effect? • How precise was the estimate of the treatment effect?
2b. How Precise Was the Estimate of the Treatment Effect? • Realistically, the true risk reduction can never be known. The best we have is the estimate provided by rigorous controlled trials, and the best estimate of the true treatment effect is that observed in the trial. This estimate is called a point estimate, a single value calculated from observations of the sample that is used to estimate a population value or parameter.
2b. How Precise Was the Estimate of the Treatment Effect? • Investigators often tell us the neighborhood within which the true effect likely lies by calculating confidence intervals. • We usually (though arbitrarily) use the 95% confidence interval. • You can consider the 95% confidence interval as defining the range that includes the true relative risk reduction 95% of the time.
2b. How Precise Was the Estimate of the Treatment Effect? This study enrolled 1000 patients • In this hypothetical situation, 2 different-sized studies with the same RRR of 25% have widely varying confidence intervals. • It is evident that the larger the sample size, the narrower the confidence interval. (RRR = 0.25 95% confidence intervals 0.09 – 0.41) (RRR = 0.25 95% confidence intervals -0.38 – 0.59) This study enrolled 100 patients
2b. How Precise Was the Estimate of the Treatment Effect? • When is the sample size big enough? • In a positive study -- a study in which the authors conclude that the treatment is effective -- if the lowest RRR that is consistent with the study results is still important (that is, it is large enough for you to recommend the treatment to the patient), then the investigators have enrolled sufficient patients. • In a negative study -- in which the authors have concluded that the experimental treatment is no better than control therapy -- if the RRR at the upper boundary would, if true, be clinically important, the study has failed to exclude an important treatment effect (i.e. has not enrolled sufficient patients).
Evaluation of Primary Studies Assessing Therapies • When using the medical literature to answer a clinical question, approach the study using three discrete steps : • Are the results of the study valid? • What are the results? • How can I apply these results to patient care?
3. How Can I Apply These Results to Patient Care? • Before embarking on therapy based on a study for a particular patient, ask 3 questions: • Were the study patients similar to the patient in my practice? • Were all clinically important outcomes considered? • Are the likely treatment benefits worth the potential harm and costs?
3a. Were the study patients similar to the patient in my practice? • If the patient had qualified for enrollment in the study -- that is, if she had met all inclusion criteria and had violated none of the exclusion criteria -- you can apply the results with considerable confidence. Even here, however, there is a limitation. Treatments are not uniformly effective in every individual. Conventional randomized trials estimate average treatment effects. Applying these average effects means that the clinician will likely be exposing some patients to the cost and toxicity of the treatment without benefit. • A better approach than rigidly applying the study's inclusion and exclusion criteria is to ask whether there is some compelling reason why the results should not be applied to the patient.
3b. Were all clinically important outcomes considered (and were the outcomes that were measured clinically significant)? • Treatments are indicated when they provide important benefits. Demonstrating that a bronchodilator produces small increments in forced expired volume in patients with chronic airflow limitation does not provide a sufficient reason for administering this drug. What is required is evidence that the treatment improves outcomes that are important to patients, such as reducing shortness of breath during the activities required for daily living. We can consider forced expired volume as a substitute or surrogate outcome. Investigators choose to substitute these variables for those that patients would consider important, usually because they would have had to enroll many more patients and follow them for far longer periods of time.
3b. Were all clinically important outcomes considered? • Even when investigators report favorable effects of treatment on one clinically important outcome, you must consider whether there may be deleterious effects on other outcomes. For example a cancer chemotherapeutic agent may lengthen life but decreases its quality.
3c. Are the likely treatment benefits worth the potential harm and costs? • If you can apply the study's results to a patient, and its outcomes are important, the next question concerns whether the probable treatment benefits are worth the effort that you and the patient must put into the enterprise. • the NNT can help clinicians judge the degree of benefit and the degree of harm patients can expect from therapy.
3c. Are the likely treatment benefits worth the potential harm and costs? • As a result of taking aspirin, patients with hypertension without known coronary artery disease can expect a reduction of approximately 15% in their relative risk of cardiovascular related events. • For an otherwise low-risk woman with hypertension and a baseline risk of cardiovascular related event of between 2.5% and 5%, this translates into an NNT of approximately 200 during a 5-year period. • For every 161 patients treated with aspirin, one would experience a major hemorrhage (NNH = 161). • Thus, in 1000 of these patients, aspirin would be responsible for preventing five cardiovascular events, but it would also be responsible for causing approximately six serious bleeding episodes.
3c. Are the likely treatment benefits worth the potential harm and costs? • Trading off benefit and risk requires an accurate assessment of medication adverse effects. Although RCTs are the correct vehicle for reporting commonly occurring side effects, reports regularly neglect to include these outcomes. Clinicians must often look to other sources of information – often characterized by weaker methodology – to obtain an estimate of the adverse effects of therapy. • The preferences or values that determine the correct choice when weighing benefit and risk are those of the individual patient.
Evaluation of Primary Studies Assessing Therapies SUMMARY • Are the results valid? • Randomized? • Blinded? • Intention to treat analysis? • Follow-up complete? • What are the results? • How large was the treatment effect (ARR/NNT)? • How precise was the estimate (confidence intervals)? • How can I apply these results to patient care? • How is my patient different from the study patient? • Was the outcome measured clinically significant (surrogate outcomes), and were all clinically significant outcomes considered? • What is the risk of treating vs. the risk of not treating (benefit vs. harm)?
User’s Guides to the Medical Literature • The “User’s Guides to the Medical Literature” is a valuable resource. Please consult this for analyzing studies on harm, diagnosis, prognosis, and systematic reviews. The principles are the same as presented today, but the steps are different. • There is a direct link to the “User’s Guides” from the JAMA home page – www.jama.ama-assn.org