370 likes | 699 Views
E N D
1. Critical appraisal:Systematic Reviews andClinical Practice Guidelines forDrug Therapy Nancy J. Lee, PharmD, BCPS
Research fellow, Drug Effectiveness Review Project
Oregon Evidence-based Practice Center
Oregon Health and Science University
To receive 1.25 AMA PRA Category 1 Credits you must review this section and answer CME questions at the end.
Release date: January 2009 Expiration date: January 2012
Hello and welcome to the module entitled: Critical appraisal of systematic reviews and clinical practice guidelines. My name is Nancy Lee and I am a research fellow with the Drug Effectiveness Review Project at the Oregon Evidence-based Practice Center.
To receive continuing medical education credit you must review this section and answer the CME questions at the end of this module with a passing score of 75%.
Keep in mind, the contents of this module build upon concepts already reviewed in a prior module titled: Critical Appraisal of Randomized Controlled trials. Review of this previous module is recommended but not required before proceeding.
Next slideHello and welcome to the module entitled: Critical appraisal of systematic reviews and clinical practice guidelines. My name is Nancy Lee and I am a research fellow with the Drug Effectiveness Review Project at the Oregon Evidence-based Practice Center.
To receive continuing medical education credit you must review this section and answer the CME questions at the end of this module with a passing score of 75%.
Keep in mind, the contents of this module build upon concepts already reviewed in a prior module titled: Critical Appraisal of Randomized Controlled trials. Review of this previous module is recommended but not required before proceeding.
Next slide
2. The attachments tab in the upper right had corner contains documents that supplement this presentation. Handouts of slides and a glossary of terms can be found under this tab and are available to print out for your use. The URL to online resources are also located here as well.
Next slide The attachments tab in the upper right had corner contains documents that supplement this presentation. Handouts of slides and a glossary of terms can be found under this tab and are available to print out for your use. The URL to online resources are also located here as well.
Next slide
3. This work was made possible by a grant from the state Attorney General Consumer and Prescriber Education programwhich is funded by the multi-state settlement of consumer fraud claims regarding the marketing of the prescription drug Neurontin.
Next slideThis work was made possible by a grant from the state Attorney General Consumer and Prescriber Education programwhich is funded by the multi-state settlement of consumer fraud claims regarding the marketing of the prescription drug Neurontin.
Next slide
4. This program has been planned and implemented in accordance with the Essential Areas and Policies of the Accreditation Council for CME jointly sponsored by the University of Texas Southwestern Medical Center and the Federation of State Medical Boards Research and Education Foundation.
Next slideThis program has been planned and implemented in accordance with the Essential Areas and Policies of the Accreditation Council for CME jointly sponsored by the University of Texas Southwestern Medical Center and the Federation of State Medical Boards Research and Education Foundation.
Next slide
5. CME information The course director is Barbara Schneidman, MD, MPH with the Federation of State Medical Boards Research and Education Foundation, and the
The program directors are as follows:
David Pass, MD
Director of Health Resources Commission at the Oregon Office for Health Policy and Research
Dean Haxby, PharmD
Associate professor of pharmacy practice at the Oregon State University College of Pharmacy
Daniel Hartung, PharmD, MPH
Assistant professor of pharmacy practice also at the Oregon State University College of Pharmacy
This educational activity is intended for persons who are involved in committees dealing with medication use policies and for health care professionals who are involved with medication prescribing.
The educational objectives are described below but will be reviewed in detail in the following slides.
Next slide
The course director is Barbara Schneidman, MD, MPH with the Federation of State Medical Boards Research and Education Foundation, and the
The program directors are as follows:
David Pass, MD
Director of Health Resources Commission at the Oregon Office for Health Policy and Research
Dean Haxby, PharmD
Associate professor of pharmacy practice at the Oregon State University College of Pharmacy
Daniel Hartung, PharmD, MPH
Assistant professor of pharmacy practice also at the Oregon State University College of Pharmacy
This educational activity is intended for persons who are involved in committees dealing with medication use policies and for health care professionals who are involved with medication prescribing.
The educational objectives are described below but will be reviewed in detail in the following slides.
Next slide
6. CME policies The continuing education sponsors require us to report the policies for this program which are described below.
Next slide.The continuing education sponsors require us to report the policies for this program which are described below.
Next slide.
7. Each speaker for this program has completed and signed a conflict of interest statement. The faculty members relationships to any commercial entities are listed on this slide.
Next slideEach speaker for this program has completed and signed a conflict of interest statement. The faculty members relationships to any commercial entities are listed on this slide.
Next slide
8. Learning objectives I. Systematic reviews
Recognize benefits and limitations
Assess quality of systematic reviews
Identify the differences between systematic reviews, narrative reviews, and meta-analyses
Recognize components of forest and funnel plots used in systematic reviews with meta-analyses
II. Guidelines
Identify strengths and weaknesses
Assess and recognize quality components
Review grading of the strength of evidence used in guidelines
For this module we will be reviewing systematic reviews and clinical practice guidelines. By the end of both sessions, our objectives are for you to be able to recognize the benefits and limitations of systematic reviews, be able to assess quality of systematic reviews, appreciated differences among systematic reviews, narrative reviews, and meta-analyses, and recognize components of forest and funnel plots that may be used in systematic reviews with meta-analyses. Similarly, for guidelines, our objectives will be able for you to appreciate strengths and weaknesses of guidelines, be able to assess and recognize quality factors, and review grading of the strength of evidence used in guideline development.
I would also like to mention that our focus for this module is regarding systematic reviews and guidelines of drug therapy and not diagnostic tests or on prognosisalthough key concepts on quality assessment can be used for these different areas.
Next slide.
For this module we will be reviewing systematic reviews and clinical practice guidelines. By the end of both sessions, our objectives are for you to be able to recognize the benefits and limitations of systematic reviews, be able to assess quality of systematic reviews, appreciated differences among systematic reviews, narrative reviews, and meta-analyses, and recognize components of forest and funnel plots that may be used in systematic reviews with meta-analyses. Similarly, for guidelines, our objectives will be able for you to appreciate strengths and weaknesses of guidelines, be able to assess and recognize quality factors, and review grading of the strength of evidence used in guideline development.
I would also like to mention that our focus for this module is regarding systematic reviews and guidelines of drug therapy and not diagnostic tests or on prognosisalthough key concepts on quality assessment can be used for these different areas.
Next slide.
9. I. Systematic Reviews:Outline Why, When, What?
Benefits and limitations
Steps in conducting Systematic Reviews
Scientific process
Quality assessment of Systematic Reviews
Tools and checklists
Before we begin, we will first address the why, when, and whats? Of systematic reviews and also review the benefits and limitations.
Next slide
Before we begin, we will first address the why, when, and whats? Of systematic reviews and also review the benefits and limitations.
Next slide
10. Why are systematic reviews needed? Too much information
Not enough time
More than 2 million articles published yearly from more than 200 biomedical journals
Results can often be contradicted by subsequent trials
Taken together, a clearer picture can emerge
Minimize biases
Increase statistical power
Improve generalizability
Improve allocation of resources for other needed trials = minimize funding of unnecessary trials So, why are systematic reviews needed?
Over the years weve been taught hat RCT design is the gold standardwhich is more often, than not, a true statement when evaluating drug therapy
.But quite often we are faced with too much information and not enough time (animate) to get through all the trials that are published every year.
It is estimated (animate) that more than 2 million articles are published annually from more than 200 different biomedical journals and keeping up with such literature can be overwhelming. In addition, not all trials show similar findings and results of isolated trials (animate) are frequently contradicted by subsequent studies which can lead to confusion. In fact, a study by Ioannidis (2005) and colleagues, the authors found that about 1/3 of the studies evaluated were either contradicted or reported stronger effects than subseq studiessuggesting that relying on a single high profile trials can have the potential to misrepresent the true effect and can therefore be harmful in decision making.
Well-conducted systematic reviews can help synthesize information from various trials and when appropriate, observational studies can also be includedwhen these publications are taken together, (animate) a clearer picture can emerge. There is the potential for minimizing biases observed among trials, potential for increasing statistical power esp when only smaller trials exist, potential for improving generalizability, and potential for better allocation of resources by minimizing the funding of unnecessary or duplicate trials.
Next slide
Ioannidis JP (2005) Contradicted and initially stronger effects in highly cited
clinical research. JAMA 294:21828So, why are systematic reviews needed?
Over the years weve been taught hat RCT design is the gold standardwhich is more often, than not, a true statement when evaluating drug therapy
.But quite often we are faced with too much information and not enough time (animate) to get through all the trials that are published every year.
It is estimated (animate) that more than 2 million articles are published annually from more than 200 different biomedical journals and keeping up with such literature can be overwhelming. In addition, not all trials show similar findings and results of isolated trials (animate) are frequently contradicted by subsequent studies which can lead to confusion. In fact, a study by Ioannidis (2005) and colleagues, the authors found that about 1/3 of the studies evaluated were either contradicted or reported stronger effects than subseq studiessuggesting that relying on a single high profile trials can have the potential to misrepresent the true effect and can therefore be harmful in decision making.
Well-conducted systematic reviews can help synthesize information from various trials and when appropriate, observational studies can also be includedwhen these publications are taken together, (animate) a clearer picture can emerge. There is the potential for minimizing biases observed among trials, potential for increasing statistical power esp when only smaller trials exist, potential for improving generalizability, and potential for better allocation of resources by minimizing the funding of unnecessary or duplicate trials.
Next slide
Ioannidis JP (2005) Contradicted and initially stronger effects in highly cited
clinical research. JAMA 294:21828
11. Fergusson D, et al. Clin Trials 2005; 2:218-32 One example of where results from systematic reviews could have had beneficial impact on resource allocation was in a study by Fergusson and colleagues titled: randomized controlled trials of aprotinin in cardiac surgery: could clinical equipoise have stopped the bleeding? In the context of this particular study, clinical equipoise was defined as whether or not there was awareness of the existence of medical evidence that could indicate whether randomization of individuals to competing drug therapies or placebo was justified.
Briefly, aprotinin is a serine protease inhibitor used to limit perioperative bleeding and therefore reduce the need for allogeneic or donor red blood cell transfusions. Randomized trials studying aprotinin' s effects on the proportion of patients receiving at least 1 transfusion in cardiac surgery have been published since 1987. And since this time, 2 systematic reviews one in 1992 and one in 1997 were conducted. Both reviews showed that aprotinin was more effective than the control group in reducing the need for transfusions. Despite these results from the systematic reviews, 19 more RCTs were funded and conducted. Begging the question, did trialists review all the literature before conducting their study?
Next slide
One example of where results from systematic reviews could have had beneficial impact on resource allocation was in a study by Fergusson and colleagues titled: randomized controlled trials of aprotinin in cardiac surgery: could clinical equipoise have stopped the bleeding? In the context of this particular study, clinical equipoise was defined as whether or not there was awareness of the existence of medical evidence that could indicate whether randomization of individuals to competing drug therapies or placebo was justified.
Briefly, aprotinin is a serine protease inhibitor used to limit perioperative bleeding and therefore reduce the need for allogeneic or donor red blood cell transfusions. Randomized trials studying aprotinin' s effects on the proportion of patients receiving at least 1 transfusion in cardiac surgery have been published since 1987. And since this time, 2 systematic reviews one in 1992 and one in 1997 were conducted. Both reviews showed that aprotinin was more effective than the control group in reducing the need for transfusions. Despite these results from the systematic reviews, 19 more RCTs were funded and conducted. Begging the question, did trialists review all the literature before conducting their study?
Next slide
12. As you may have guessed, trialists did not systematically review the medical literature before submitting their proposal for their own trials.
Fergusson and colleagues identified a total of 69 trials (N=8000 patients) on aprotinin and conducted a cumulative meta-analysis which was ordered by study trial date. This quantitative analysis produces an updated measure of the aprotinin' s effect on bleeding by pooling or combining trials after each study is completed. This forest plot depicts the cumulative analysis.
As you can see, the very 1st trial (1987) showed a benefit with aprotinin (albeit with a large confidence interval). By around 1992, and after trial #12, the cumulative effect estimate stabilized around OR of 0.25-0.35 which indicates that fewer patients in the aprotinin group received at least 1 blood transfusion than the control group. Throughout the meta-analysis the upper limit of the confidence interval never crossed 0.65.
Even after the largest trial (which enrolled 1784 patients) was published, only 16% of subsequent trials referenced this particular study. Overall, about 20% of trials conducted and published actually referenced results from previously published studies on aprotinin in cardiac surgery. The results of this review not only make us wonder about the necessity of all these subsequent trials but also make us wonder if all these trials really made a clinical impact?
This is just 1 example of the importance of conducting high quality systematic reviews in influencing practice. There are more examples like this that have demonstrated the benefit of systematic reviews in answering and confirming medical practices that help save lives.
Next slideAs you may have guessed, trialists did not systematically review the medical literature before submitting their proposal for their own trials.
Fergusson and colleagues identified a total of 69 trials (N=8000 patients) on aprotinin and conducted a cumulative meta-analysis which was ordered by study trial date. This quantitative analysis produces an updated measure of the aprotinin' s effect on bleeding by pooling or combining trials after each study is completed. This forest plot depicts the cumulative analysis.
As you can see, the very 1st trial (1987) showed a benefit with aprotinin (albeit with a large confidence interval). By around 1992, and after trial #12, the cumulative effect estimate stabilized around OR of 0.25-0.35 which indicates that fewer patients in the aprotinin group received at least 1 blood transfusion than the control group. Throughout the meta-analysis the upper limit of the confidence interval never crossed 0.65.
Even after the largest trial (which enrolled 1784 patients) was published, only 16% of subsequent trials referenced this particular study. Overall, about 20% of trials conducted and published actually referenced results from previously published studies on aprotinin in cardiac surgery. The results of this review not only make us wonder about the necessity of all these subsequent trials but also make us wonder if all these trials really made a clinical impact?
This is just 1 example of the importance of conducting high quality systematic reviews in influencing practice. There are more examples like this that have demonstrated the benefit of systematic reviews in answering and confirming medical practices that help save lives.
Next slide
13. When are systematic reviews needed? When an important question needs to be addressed
Gaps in the literature or conflicting results
When there is uncertainty regarding an intervention
Uncertainty may lie in:
Population, Intervention, Outcomes
When several primary studies exist
Lack of strong evidence So, weve just reviewed reasons why systematic reviews are needed and also discussed some of their potential benefits. But, of course the next question is: when are they needed?
Systematic reviews are not needed for every clinical scenario. Systematic reviews are needed: when an important question needs to be addressedsuch as in areas where there are gaps or when there are conflicting results. Another area is when there is uncertainty regarding an intervention particularly within a population, dose, or type of outcome. And last, systematic reviews are needed when several smaller primary studies exist where there is a lack of strong evidence.
Pause
.
Next slide
Lancetrequiring SR before trial publication (Young and Horton 2005)So, weve just reviewed reasons why systematic reviews are needed and also discussed some of their potential benefits. But, of course the next question is: when are they needed?
Systematic reviews are not needed for every clinical scenario. Systematic reviews are needed: when an important question needs to be addressedsuch as in areas where there are gaps or when there are conflicting results. Another area is when there is uncertainty regarding an intervention particularly within a population, dose, or type of outcome. And last, systematic reviews are needed when several smaller primary studies exist where there is a lack of strong evidence.
Pause
.
Next slide
Lancetrequiring SR before trial publication (Young and Horton 2005)
14. Limitations of systematic reviews Only as good as what is available and what is included
Issue of publication bias
Restricted to published results
Quality of individual trials
Garbage In, Garbage Out
Good quality systematic reviews typically do not address all the issues relevant for decision making
Evidence outside the scope of the review may be relevant and needed for decision making
Cost and implementation implications may not always be addressed Although systematic reviews are helpful in synthesizing large bodies of evidence and in answering or confirming therapies in clinical practice, systematic reviews of course, have their limitations. One of which is that: systematic reviews are only as good as what is available and what is included. With regard to publication biasif negative trials are not published and therefore not available for synthesis, the results may be biased towards positive published studies. Publication bias is a significant issue that is difficult to controland we will further discuss this issue later in this module.
The second issue deals with quality of included trials. If the quality of the included studies is poor then final product is also poor regardless of how well the systematic review was conductedthis is the same as the garbage in, garbage out concept.
Another limitation is that good quality reviews cannot include all the issues relevant or necessary for decision-making for clinicians and policy makers. Evidence outside the scope of the review may be needed and often times cost information specific for each different decision-making scenario may not be addressed and may need to be considered separately.
Next slideAlthough systematic reviews are helpful in synthesizing large bodies of evidence and in answering or confirming therapies in clinical practice, systematic reviews of course, have their limitations. One of which is that: systematic reviews are only as good as what is available and what is included. With regard to publication biasif negative trials are not published and therefore not available for synthesis, the results may be biased towards positive published studies. Publication bias is a significant issue that is difficult to controland we will further discuss this issue later in this module.
The second issue deals with quality of included trials. If the quality of the included studies is poor then final product is also poor regardless of how well the systematic review was conductedthis is the same as the garbage in, garbage out concept.
Another limitation is that good quality reviews cannot include all the issues relevant or necessary for decision-making for clinicians and policy makers. Evidence outside the scope of the review may be needed and often times cost information specific for each different decision-making scenario may not be addressed and may need to be considered separately.
Next slide
15. Limitations of systematic reviews Unrealistic expectations
What if results conflict with a good quality large landmark trial?
About 10-23% of large trials disagreed with meta-analyses*
May not always include the most up to date studies
When was the last literature search conducted?
Estimate: 3-5 years**
Does not make decisions for the user
These are not guidelines
The reader uses their own judgment
*Ioannidis, et al. JAMA 1998; 279:1089-93. **Shojoania, et al. Ann Intern Med 2007; 147:224-33. Other limitations result from readers unrealistic expectations that results from systematic reviews must always be rightwhat if results from systematic reviews conflict with good quality large landmark trials? It is estimated that about 10-20% large clinical trials differ from MAnot to say that one study design is better than the other but that discrepancies can occur.
Another important limitation of systematic reviews is that systematic reviews may not always include the most up to date studies. Therefore, it is always important to ask, when was the literature last searched? Depending on the topic area, it is estimated that the information contained in systematic reviews become out of date within 3-5 years. And so, it is wise to check the date of the last search and date of publication.
And finally, systematic reviews do not make decisions for usersthese are not guidelines. The user must use their judgment based on the evidence to make decisions.
Next slideOther limitations result from readers unrealistic expectations that results from systematic reviews must always be rightwhat if results from systematic reviews conflict with good quality large landmark trials? It is estimated that about 10-20% large clinical trials differ from MAnot to say that one study design is better than the other but that discrepancies can occur.
Another important limitation of systematic reviews is that systematic reviews may not always include the most up to date studies. Therefore, it is always important to ask, when was the literature last searched? Depending on the topic area, it is estimated that the information contained in systematic reviews become out of date within 3-5 years. And so, it is wise to check the date of the last search and date of publication.
And finally, systematic reviews do not make decisions for usersthese are not guidelines. The user must use their judgment based on the evidence to make decisions.
Next slide
16. What it is and isnt Adapted from Cook DJ, et al. Ann Intern Med 1997; 126:376-80. Before going any further, I would like to take some time to highlight some of the differences between systematic reviews and narrative/or traditional reviews.
First, systematic reviews are different from narrative reviews (also referred to as traditional reviews). Generally speaking, systematic reviews refers to the entire process of collecting, reviewing, and presenting all evidence. Unlike narrative reviews, systematic reviews tend to have a more focused clinical question than narrative reviews which are often broad in scope and can cover information from pharmacokinetics, pharmacodynamics, and pathophysiology in addition to summarizing trial results.
Although more narrative reviews are searching bibliographic databases such as Medline or Embaseprespecified and comprehensive search methods are typically not performed.
Study eligibility are often not specified in narrative reviews as well and if they are reported, the criteria for study inclusion or exclusion may not be uniformly appliedwhich can introduce several biases such as study selection bias. You may have come across very biased narrative reviews that seemed to have picked certain studies that supported the authors preformed conclusions and omitted other studies that did not support their conclusions.
The key difference between narrative and systematic reviews, however, is in the appraisal of the literature (shaded highlight)more specifically, the quality assessment of the included trials. Authors of narrative reviews may appraise the evidence well but more often than not, authors of narrative reviews do not consistently assess methodologic rigor of included trials or the internal validity.
Next slide
Adapted from Cook DJ, Ann Intern Med 1997; 126:376-380Before going any further, I would like to take some time to highlight some of the differences between systematic reviews and narrative/or traditional reviews.
First, systematic reviews are different from narrative reviews (also referred to as traditional reviews). Generally speaking, systematic reviews refers to the entire process of collecting, reviewing, and presenting all evidence. Unlike narrative reviews, systematic reviews tend to have a more focused clinical question than narrative reviews which are often broad in scope and can cover information from pharmacokinetics, pharmacodynamics, and pathophysiology in addition to summarizing trial results.
Although more narrative reviews are searching bibliographic databases such as Medline or Embaseprespecified and comprehensive search methods are typically not performed.
Study eligibility are often not specified in narrative reviews as well and if they are reported, the criteria for study inclusion or exclusion may not be uniformly appliedwhich can introduce several biases such as study selection bias. You may have come across very biased narrative reviews that seemed to have picked certain studies that supported the authors preformed conclusions and omitted other studies that did not support their conclusions.
The key difference between narrative and systematic reviews, however, is in the appraisal of the literature (shaded highlight)more specifically, the quality assessment of the included trials. Authors of narrative reviews may appraise the evidence well but more often than not, authors of narrative reviews do not consistently assess methodologic rigor of included trials or the internal validity.
Next slide
Adapted from Cook DJ, Ann Intern Med 1997; 126:376-380
17. The advantage of using carefully done, systematic reviews becomes clear when we observe how often mistakes are made when research is reviewed non-systematically, whether by experts or others. The costs of mistaken conclusions based on non-systematic reviews can be high. I believe Andrew Oxman, Director of the Health Services Research Unit at the National Institute of public Health in Oslo Norway best summarized the differences between narrative or traditional reviews with well-conducted systematic reviews by stating, the advantage of using carefully done systematic reviews becomes clear when we observe how often mistakes are made when research is reviewed non-systematically, whether by experts or others. The costs of mistaken conclusions based on non-systematic reviews can be high.
Next slideI believe Andrew Oxman, Director of the Health Services Research Unit at the National Institute of public Health in Oslo Norway best summarized the differences between narrative or traditional reviews with well-conducted systematic reviews by stating, the advantage of using carefully done systematic reviews becomes clear when we observe how often mistakes are made when research is reviewed non-systematically, whether by experts or others. The costs of mistaken conclusions based on non-systematic reviews can be high.
Next slide
18. I. Systematic Reviews:Outline Why, When, What?
Benefits and limitations
Steps in conducting Systematic Reviews
Scientific process
Quality assessment of Systematic Reviews
Tools and checklists
Conducting a good-or high quality systematic review is based on a scientific process similar to concepts observed in good trial designand requires us to take a closer look at the steps involved in conducting a systematic review before going over how to assess its quality.
Next slide
Conducting a good-or high quality systematic review is based on a scientific process similar to concepts observed in good trial designand requires us to take a closer look at the steps involved in conducting a systematic review before going over how to assess its quality.
Next slide
19. Systematic Reviews: A scientific process Figure 1. Copyright Š1997 BMJ Publishing Group Ltd. from Greenhalgh T. BMJ 1997;315:672-5. Performing systematic reviews is a scientific process and involves multiple steps as shown in this figure. We will go over each step in detail in the following slides but just like a well-conducted clinical trial, well done systematic reviews also try and minimize potential sources of bias by prespecifying the key questions, inclusion/exclusion criteria, performing dual review of the selection of studies, data abstraction, and quality assessmentin addition to evaluating appropriateness of quantitative methods.
Next slide.
Greenhalgh T, BMJ 1997; 315:672-75
Performing systematic reviews is a scientific process and involves multiple steps as shown in this figure. We will go over each step in detail in the following slides but just like a well-conducted clinical trial, well done systematic reviews also try and minimize potential sources of bias by prespecifying the key questions, inclusion/exclusion criteria, performing dual review of the selection of studies, data abstraction, and quality assessmentin addition to evaluating appropriateness of quantitative methods.
Next slide.
Greenhalgh T, BMJ 1997; 315:672-75
20. Developed a priori
Most important
Relevant and sensible to practitioners and patients?
Typically not changed during the review process
What are we asking?
Efficacy
Effectiveness
Well-defined?
PICOS
Any exclusions?
Language restrictions or type of study design Whats the purpose and question? Before formulating a question, a clear purpose for conducting a systematic review should be specifiedfor instance: what will this review address? Will this fill any gaps? Or will this add new information?
Formulating a relevant, clear and focused question is probably the most important step in the process and is developed a prioriThe key question or questions of the systematic review sets the tone or scope of the report and the question should be relevant and sensible to clinicians and patients. The key question is typically not changed in the middle of a systematic review unless there is a significant and important reason for doing soreasons for changing a question should be explicitly reported in the review for transparency.
In the case of comparative drug effectiveness reviews, there are generally, 2 types of questions that are addressed:1) can it work or 2) how well it works over time? The first question addresses efficacy of a treatment while the 2nd question deals more with the effectiveness of treatment. In both types of questions, harms of the drug therapy should also be assessed
Good questions that are well-defined also include information about pop, intervention, outcome, in some instances study setting. It is also a good idea to specify if there are any exclusions such as any language restrictions or any type of study design restrictions.
Next slideBefore formulating a question, a clear purpose for conducting a systematic review should be specifiedfor instance: what will this review address? Will this fill any gaps? Or will this add new information?
Formulating a relevant, clear and focused question is probably the most important step in the process and is developed a prioriThe key question or questions of the systematic review sets the tone or scope of the report and the question should be relevant and sensible to clinicians and patients. The key question is typically not changed in the middle of a systematic review unless there is a significant and important reason for doing soreasons for changing a question should be explicitly reported in the review for transparency.
In the case of comparative drug effectiveness reviews, there are generally, 2 types of questions that are addressed:1) can it work or 2) how well it works over time? The first question addresses efficacy of a treatment while the 2nd question deals more with the effectiveness of treatment. In both types of questions, harms of the drug therapy should also be assessed
Good questions that are well-defined also include information about pop, intervention, outcome, in some instances study setting. It is also a good idea to specify if there are any exclusions such as any language restrictions or any type of study design restrictions.
Next slide
21. What was the study eligibility? Determines what studies get included in a systematic review
Formed a priori
Applied uniformly by at least 2 reviewers (dual review)
Study inclusion and exclusion criteria should relate to the areas defined by PICO(S)
Population
Intervention
Comparator
Outcome
Setting/study design The next step is determining the study eligibilityor determining what studies get included in a systematic review. Study eligibility criteria is also often written a priori to minimize study selection bias and is applied uniformly by at least 2 reviewersagain, to minimize bias. The study eligibility is written in relation to the areas defined by PICOS.
Next slide.
The next step is determining the study eligibilityor determining what studies get included in a systematic review. Study eligibility criteria is also often written a priori to minimize study selection bias and is applied uniformly by at least 2 reviewersagain, to minimize bias. The study eligibility is written in relation to the areas defined by PICOS.
Next slide.
22. Study eligibility What are the consequences of being too inclusive or exclusive?
Too inclusive
Scope is too large
Lose focus of question
Main point may be lost
May be difficult to interpret
Too exclusive
Scope is too narrow
Potential to exclude important trials
May end up not having enough evidence
If unaware, could lead to biased conclusions Although determining study eligibility seems easy enough, there are consequences to criteria that are too inclusive or exclusive.
If the study eligibility criteria are too broad or inclusivethis could impact the scope of the review and as a consequence muddle the focus of the key question resulting in a loss in the main point of the reviewwhich can make the review difficult to interpret and use.
If the study eligibility criteria are too stringent or exclusivethis could significantly narrow the scope of the review and exclude potentially important studies. Sometimes being too exclusive can also lead to biased conclusions or potentially create a situation the reviewer has to make assumptions regarding missing information.
Next slideAlthough determining study eligibility seems easy enough, there are consequences to criteria that are too inclusive or exclusive.
If the study eligibility criteria are too broad or inclusivethis could impact the scope of the review and as a consequence muddle the focus of the key question resulting in a loss in the main point of the reviewwhich can make the review difficult to interpret and use.
If the study eligibility criteria are too stringent or exclusivethis could significantly narrow the scope of the review and exclude potentially important studies. Sometimes being too exclusive can also lead to biased conclusions or potentially create a situation the reviewer has to make assumptions regarding missing information.
Next slide
23. Example: Study eligibility Here is an example of an eligibility criteria that was adapted from a systematic review that assessed the efficacy, effectiveness, and harms of sitagliptin (a new DPP-4 inhibitor) compared with placebo and compared with other antihyperglycemic agents in patients with type 2 diabetes.
As you can see the population includes: adults and children; the intervention spells out which treatments will be included. The outcomes of interest are listedgenerally, the outcomes should be health outcomes or patient-oriented outcomes rather than intermediate or surrogate markers of health outcomes.
The study design can be specified here as wellfor instance, for efficacy and effectivenessRCTs and existing good quality systematic reviews can be usedfor harmsRCT, systematic reviews, and large comparative cohort observational studies were allowed.
A minimum duration was specified as >12 weeks based on the minimum time period for a change in A1c or glycemic control to be observed.
And finally, if any studies were to be excludedthis should be specified here. Arbitrarily excluding relevant studies or certain studies because the results conflict with a preformed conclusion is a serious flaw!! Reasons for study exclusion (again) should be explicitly reported and should be reasonable.
Next slideHere is an example of an eligibility criteria that was adapted from a systematic review that assessed the efficacy, effectiveness, and harms of sitagliptin (a new DPP-4 inhibitor) compared with placebo and compared with other antihyperglycemic agents in patients with type 2 diabetes.
As you can see the population includes: adults and children; the intervention spells out which treatments will be included. The outcomes of interest are listedgenerally, the outcomes should be health outcomes or patient-oriented outcomes rather than intermediate or surrogate markers of health outcomes.
The study design can be specified here as wellfor instance, for efficacy and effectivenessRCTs and existing good quality systematic reviews can be usedfor harmsRCT, systematic reviews, and large comparative cohort observational studies were allowed.
A minimum duration was specified as >12 weeks based on the minimum time period for a change in A1c or glycemic control to be observed.
And finally, if any studies were to be excludedthis should be specified here. Arbitrarily excluding relevant studies or certain studies because the results conflict with a preformed conclusion is a serious flaw!! Reasons for study exclusion (again) should be explicitly reported and should be reasonable.
Next slide
24. Finding all relevant studies:Search strategy Medical librarian important
Key search terms should at the very least be reported
Were any significant studies missing?
If yes, why? Once study eligibility has been laid outthe next step is to search for the evidence by creating a net of search terms to use to cast out into various bibliographic databases.
Development of a comprehensive search strategy is another important step in systematic reviews and involving a medical librarian in this step is highly recommended. Remember, the development of comprehensive searches minimizes the potential for missing potentially relevant trials.
At a minimum key search terms should be reported in systematic reviews and an assessment of any missing studies should be routinely performed by both the authors and readers of the systematic review.
Next slide.Once study eligibility has been laid outthe next step is to search for the evidence by creating a net of search terms to use to cast out into various bibliographic databases.
Development of a comprehensive search strategy is another important step in systematic reviews and involving a medical librarian in this step is highly recommended. Remember, the development of comprehensive searches minimizes the potential for missing potentially relevant trials.
At a minimum key search terms should be reported in systematic reviews and an assessment of any missing studies should be routinely performed by both the authors and readers of the systematic review.
Next slide.
25. Example: Search strategy This is just a little snap shot of 1 part of a comprehensive search strategy.
The first line reports which database was searched and the time period it was searchedso if you notice that a large trial was missed you may want to check the dates to see if this is a possible reason. As you can see, Medline was searched in this snap shot from 1950 to 2007 and then was updated to 2008.
The next lines with numbers is the actual searching strategy with terms usedthe first search term is the CAS number which is a unique numerical identifier for the exenatide compound before the drug was even named.
The next search terms are the various names for exenatide or terms used to describe exenatide.
Line 8 here shows that the search was limited to human subjects and publications in English
Next slideThis is just a little snap shot of 1 part of a comprehensive search strategy.
The first line reports which database was searched and the time period it was searchedso if you notice that a large trial was missed you may want to check the dates to see if this is a possible reason. As you can see, Medline was searched in this snap shot from 1950 to 2007 and then was updated to 2008.
The next lines with numbers is the actual searching strategy with terms usedthe first search term is the CAS number which is a unique numerical identifier for the exenatide compound before the drug was even named.
The next search terms are the various names for exenatide or terms used to describe exenatide.
Line 8 here shows that the search was limited to human subjects and publications in English
Next slide
26. Finding all relevant studies: Sources Electronic databases
MEDLINE (Ovid/PubMed)
Cochrane Library
EMBASE
PsychINFO
CINAHL
Hand searching
Reference lists of trials and/or reviews
Journals
Sources for unpublished information
FDA website
Clinical Trials.gov
Registeries
Industry dossiers
Part of accomplishing a comprehensive search involves searching multiple sources of information . This entails searching more than 1 bibliographic database. In fact, searching multiple databases can still miss about 15% of relevant trials and thus hand searching and searching websites and dossiers is necessary to find as many potentially relevant studies as possible.
Next slide.Part of accomplishing a comprehensive search involves searching multiple sources of information . This entails searching more than 1 bibliographic database. In fact, searching multiple databases can still miss about 15% of relevant trials and thus hand searching and searching websites and dossiers is necessary to find as many potentially relevant studies as possible.
Next slide.
27. Selection of studies Review titles and abstracts from initial search
Review of full text articles
Uniform application of study eligibility criteria
Dual review for each step
Disagreements resolved by consensus Once the search for the evidence is completed, studies need to be selected. Typically, systematic reviewers receive long lists of titles and abstracts. These citations are reviewed by at least 2 reviewers. Then full text articles are retrieved for another round of study selectionthis step is also performed in dual review.
During the study selection process, study eligibility criteria should be uniformly applied and disagreements are usually resolved by consensus or with input from a third party.
Next slideOnce the search for the evidence is completed, studies need to be selected. Typically, systematic reviewers receive long lists of titles and abstracts. These citations are reviewed by at least 2 reviewers. Then full text articles are retrieved for another round of study selectionthis step is also performed in dual review.
During the study selection process, study eligibility criteria should be uniformly applied and disagreements are usually resolved by consensus or with input from a third party.
Next slide
28. Issue of publication bias Adapted from Cochrane Open Learning. Module 15. Publication bias 2002. As previously mentioned, another aspect of accomplishing a comprehensive search strategy and minimizing bias in the study selection process is the issue of publication bias. Publication bias is probably one of the larger challenges to reviewers because it is hard to manage and controland most systematic reviews will be subject to some publication bias. Publication bias refers to the tendency for more positive studies to be published in biomedical journals than negative studies.
In general, positive trials are likely to be published more rapidly, are more likely published in English, and are likely to be published more than once. All of these issues may exaggerate treatment effects observed in systematic reviews and meta-anlayses.
The failure to publish negative results by peer reviewers, journal editors, investigators, and pharma may knowingly or unknowingly influence the results of systematic reviews and meta-analyses toward the positive trials.
Next slide
_______________________
As previously mentioned, another aspect of accomplishing a comprehensive search strategy and minimizing bias in the study selection process is the issue of publication bias. Publication bias is probably one of the larger challenges to reviewers because it is hard to manage and controland most systematic reviews will be subject to some publication bias. Publication bias refers to the tendency for more positive studies to be published in biomedical journals than negative studies.
In general, positive trials are likely to be published more rapidly, are more likely published in English, and are likely to be published more than once. All of these issues may exaggerate treatment effects observed in systematic reviews and meta-anlayses.
The failure to publish negative results by peer reviewers, journal editors, investigators, and pharma may knowingly or unknowingly influence the results of systematic reviews and meta-analyses toward the positive trials.
Next slide
_______________________
29. Scargle. J of Scientific Explor 2000; 14(1):91-106. The issue of publication bias is not new and has been around for about 2 decades and as Rosenthal put it
.---- (type 1 errors or false positive findings)
Next slideThe issue of publication bias is not new and has been around for about 2 decades and as Rosenthal put it
.---- (type 1 errors or false positive findings)
Next slide
30. Investigating for presence of publication bias Visually check for asymmetry in funnel plots
NOT a tool to diagnose bias
Potential sources of asymmetry
True heterogeneity
Data irregularities
Chance
Other statistical methods
Ask a biostatistician
Egger, et al. BMJ 1997; 315:629-34. Figure 1 from Peters, et al. JAMA 2006; 295:676-80. A method that is commonly used to assess publication bias in systematic reviews that include meta-analyses is thru a funnel plotwhich plots each trials result against the sample sizeallowing for a visual check for asymmetry. I include this slide not for you to be able to understand every detail of funnel plots
but for you to be aware of this method and their limitations esp if it is reported in a systematic review with meta-analysis.
Ideally, the funnel plot should look balancedor as a symmetric inverted funnel as shown in this figure. Unfortunately, funnel plotting is not the most reliable method to check for or diagnose publication bias. There are other potential sources that can cause asymmetry in funnel plots. For instance, true clinical heterogeneity among the included trials in a review can skew the plot as well as data irregularities and by chance. There are other statistical methods to further assess the asymmetry but these are typically just further exercises that may not help answer the question.
So, the bottomline with funnel plots is that these need to be interpreted with caution.
Next slideA method that is commonly used to assess publication bias in systematic reviews that include meta-analyses is thru a funnel plotwhich plots each trials result against the sample sizeallowing for a visual check for asymmetry. I include this slide not for you to be able to understand every detail of funnel plots
but for you to be aware of this method and their limitations esp if it is reported in a systematic review with meta-analysis.
Ideally, the funnel plot should look balancedor as a symmetric inverted funnel as shown in this figure. Unfortunately, funnel plotting is not the most reliable method to check for or diagnose publication bias. There are other potential sources that can cause asymmetry in funnel plots. For instance, true clinical heterogeneity among the included trials in a review can skew the plot as well as data irregularities and by chance. There are other statistical methods to further assess the asymmetry but these are typically just further exercises that may not help answer the question.
So, the bottomline with funnel plots is that these need to be interpreted with caution.
Next slide
31. Ways to minimize publication bias in the review process Identify duplicate publications
Contact study authors or manufacturer
Often difficult to obtain information
Time intensive
Check sources for grey literature
FDA review documents
Clinical trial registries
Databases
Check for any language restrictions
Rising, et al. PLoS Med 5(11):e217. Not all is lost with publication bias however, there are a few ways to minimize its potential:
The first is for the author and reader to critically evaluate and check whether there are duplicate publications of the same study. Double counting results in meta-analyses can further skew the treatment effect and therefore it is essential that duplicate publications be accounted for.
Next, authors of systematic reviews can contact lead investigators or the manufacturer for unpublished trial data. Often, this is a very time intensive process and in the end information may not be provided; however, it may be worth the time in some cases.
Thirdly, authors of reviews can check sources for grey literature such as FDA review documents, clinical registeries, or databases for unpublished information. And finally, in some instances, authors may need to consider whether language restrictions may have an impact on publication bias..is this a topic area where I suspect there will be more trials published in non-english journals? If this is the case, it might be necessary to invest the funds to hire someone to translate non-english studies.
Next slideNot all is lost with publication bias however, there are a few ways to minimize its potential:
The first is for the author and reader to critically evaluate and check whether there are duplicate publications of the same study. Double counting results in meta-analyses can further skew the treatment effect and therefore it is essential that duplicate publications be accounted for.
Next, authors of systematic reviews can contact lead investigators or the manufacturer for unpublished trial data. Often, this is a very time intensive process and in the end information may not be provided; however, it may be worth the time in some cases.
Thirdly, authors of reviews can check sources for grey literature such as FDA review documents, clinical registeries, or databases for unpublished information. And finally, in some instances, authors may need to consider whether language restrictions may have an impact on publication bias..is this a topic area where I suspect there will be more trials published in non-english journals? If this is the case, it might be necessary to invest the funds to hire someone to translate non-english studies.
Next slide
32. Quality assessment of included studies >25 different tools
Jadad scale, Risk of Bias tool, DERP method (for trials)
Other scales or checklists (for observational studies)
How were poor-or low quality trials handled in the review?
Were these excluded?
Sensitivity analyses? Now that weve touched on comprehensive search methods, study selection, and the issue of publication bias
each article should be assessed for its quality or internal validity
.was quality assessment conducted and reported?
As mentioned in a previous module, there are more than 25 different quality assessment tools and checklists. These tools evaluate the different components of internal validity which include: randomization, allocation concealment, blinding, statistical analysis, and attrition. Evaluating each component for potential risk of bias is important because inclusion of poorer quality trials can inflate overall treatment effects observed in systematic reviews. These concepts are discussed in more detail in the module for randomized controlled trials.
In addition to quality assessment of individual trials, it is important to evaluate
how were poor or lower quality trials were handled in the systematic review. Generally, authors of systematic reviews should report what they did with poor quality studies. Were these excluded? Or were sensitivity analyses that included and excluded poor quality studies conducted and explored?
Next slideNow that weve touched on comprehensive search methods, study selection, and the issue of publication bias
each article should be assessed for its quality or internal validity
.was quality assessment conducted and reported?
As mentioned in a previous module, there are more than 25 different quality assessment tools and checklists. These tools evaluate the different components of internal validity which include: randomization, allocation concealment, blinding, statistical analysis, and attrition. Evaluating each component for potential risk of bias is important because inclusion of poorer quality trials can inflate overall treatment effects observed in systematic reviews. These concepts are discussed in more detail in the module for randomized controlled trials.
In addition to quality assessment of individual trials, it is important to evaluate
how were poor or lower quality trials were handled in the systematic review. Generally, authors of systematic reviews should report what they did with poor quality studies. Were these excluded? Or were sensitivity analyses that included and excluded poor quality studies conducted and explored?
Next slide
33. Example Bjelakovic, et al. Lancet 2004; 364:1219-28. This table was included in a systematic review of antioxidant supplements for the prevention of GI cancers. This nice feature of this table is that it reports the overall quality of each included study and also reports the results for each component of internal validity that the authors assessed, making it easy for the reader to evaluate what trials were included and help assess the overall quality of the review and implications of the final results.
Next slide.This table was included in a systematic review of antioxidant supplements for the prevention of GI cancers. This nice feature of this table is that it reports the overall quality of each included study and also reports the results for each component of internal validity that the authors assessed, making it easy for the reader to evaluate what trials were included and help assess the overall quality of the review and implications of the final results.
Next slide.
34. Data abstraction Dual abstraction and review
Types of data abstracted:
Study design
Setting
Population characteristics (age, sex, ethnicity)
Inclusion/exclusion criteria
Interventions
Comparisons
Number screened, eligible, enrolled
Number withdrawn
Method of outcome ascertainment
Results
Adverse events At the same time studies are being evaluated for quality, data abstraction of study characteristics and the results can occur. Again, well-conducted systematic reviews will tend to use at least 2 reviewers to data abstract or at 1 person to abstract while another checks the work.
Types of data abstracted are listed on this slide: this list may vary depending on the topic area. Authors of systematic reviews may pilot their abstraction tool using a few studies, however, changing the data abstraction components frequently during the review is not recommended. Frequently changing what data to abstract may introduce bias into the review.
Next slideAt the same time studies are being evaluated for quality, data abstraction of study characteristics and the results can occur. Again, well-conducted systematic reviews will tend to use at least 2 reviewers to data abstract or at 1 person to abstract while another checks the work.
Types of data abstracted are listed on this slide: this list may vary depending on the topic area. Authors of systematic reviews may pilot their abstraction tool using a few studies, however, changing the data abstraction components frequently during the review is not recommended. Frequently changing what data to abstract may introduce bias into the review.
Next slide
35. Data synthesis Two methods: qualitative and quantitative
Qualitative
Discussion of results (synthesis)
in relation to each other
in relation to study quality
Not a reporting of results from each study
Adapted from Cochrane Collaboration open learning materials for reviewers 2002-2003. After data has been abstracted
the information must be synthesized. There are 2 general methods of synthesisqualitative and quantitative.
Systematic reviews are qualitative in nature and authors of reviews should at a minimum discuss the results of studies in relation to each other and in relation to the quality of the study. This is different from just reporting results from each study which narrative reviews tend to do
systematic reviews should try and synthesize and summarize the major findings.
Next slide
After data has been abstracted
the information must be synthesized. There are 2 general methods of synthesisqualitative and quantitative.
Systematic reviews are qualitative in nature and authors of reviews should at a minimum discuss the results of studies in relation to each other and in relation to the quality of the study. This is different from just reporting results from each study which narrative reviews tend to do
systematic reviews should try and synthesize and summarize the major findings.
Next slide
36. Data synthesis Quantitative or meta-analyses
Statistical method for combining results from >1 study
Advantage: provides an estimate of treatment effect
Disadvantage: misleading estimate if used inappropriately
Misuse of terminology
Systematic review and Meta-analysis = NOT the same
Adapted from Cochrane Collaboration open learning materials for reviewers 2002-2003. The second method of data synthesis is: Quantitative analyses also referred to as meta-analyses. Meta-analysis can be used in addition to qualitative synthesis when combining or pooling results from more than 1 study.
The advantage of meta-analysis is that it provides a numerical estimate of the treatment effect (in the form of RR, OR, NNT, or mean difference). The disadvantage of this type of analysis is that the numerical estimate can be misleading if trials are combined or pooled inappropriately.
Another complication with meta-analysis, is with the terminology. Oftentimes, people misuse the terms systematic review and meta-analysis and use these interchangeably. The Venn diagram shows that you can have systematic reviews with or without meta-analysesand you can have meta-analyses that are not a part of systematic reviews at all. Someone could easily find a few studies that support their preformed conclusions and calculate a result with no attempt to be systematic. These types of meta-analyses are very biased and should not be confused or referred to as being systematic in nature. Remember, meta-analysis is a statistical method for combining results!
Next slideThe second method of data synthesis is: Quantitative analyses also referred to as meta-analyses. Meta-analysis can be used in addition to qualitative synthesis when combining or pooling results from more than 1 study.
The advantage of meta-analysis is that it provides a numerical estimate of the treatment effect (in the form of RR, OR, NNT, or mean difference). The disadvantage of this type of analysis is that the numerical estimate can be misleading if trials are combined or pooled inappropriately.
Another complication with meta-analysis, is with the terminology. Oftentimes, people misuse the terms systematic review and meta-analysis and use these interchangeably. The Venn diagram shows that you can have systematic reviews with or without meta-analysesand you can have meta-analyses that are not a part of systematic reviews at all. Someone could easily find a few studies that support their preformed conclusions and calculate a result with no attempt to be systematic. These types of meta-analyses are very biased and should not be confused or referred to as being systematic in nature. Remember, meta-analysis is a statistical method for combining results!
Next slide
37. Meta-analysis The review should provide enough information about the included studies for you to judge whether combining results was appropriate.
Two types of heterogeneity
Clinical heterogeneity
Does it make clinical sense to combine these studies?
Statistical heterogeneity
Are there inconsistencies in the results?
Calculation of Q-or I-squared statistic
Common sources of heterogeneity
Clinical diversity between studies, conflicts of interest, and differences in study quality Adapted from Cochrane Collaboration open learning materials for reviewers 2002-2003. The decision to combine or not combine studies in meta-analyses involves an element of judgment and authors of systematic reviews and meta-analyses should discuss their reasons for combining or not combining results across studies. Authors of systematic review and meta-analysis should also provide the reader with enough information to judge whether or not combining results was reasonable.
There are 2 types of heterogeneity and the assessment of these sources of heterogeneity such as clinical and statistical heterogeneity are very important concepts to consider when determining appropriate combining of studies.
Clinical heterogeneity assesses whether the studies were reasonably similar with regard to population, intervention, and outcomeand also whether the combining of studies makes clinical sense at all. Statistical heterogeneity evaluates whether there are inconsistencies in the results among or between the included studies. The q-or I-squared statistic is usually used to determine statistical heterogeneity.
Common sources of heterogeneity include clinical diversity between studies such as patient characteristics, settings, interventions, or outcomes, conflicts of interest, and differences in study quality.
Next slideThe decision to combine or not combine studies in meta-analyses involves an element of judgment and authors of systematic reviews and meta-analyses should discuss their reasons for combining or not combining results across studies. Authors of systematic review and meta-analysis should also provide the reader with enough information to judge whether or not combining results was reasonable.
There are 2 types of heterogeneity and the assessment of these sources of heterogeneity such as clinical and statistical heterogeneity are very important concepts to consider when determining appropriate combining of studies.
Clinical heterogeneity assesses whether the studies were reasonably similar with regard to population, intervention, and outcomeand also whether the combining of studies makes clinical sense at all. Statistical heterogeneity evaluates whether there are inconsistencies in the results among or between the included studies. The q-or I-squared statistic is usually used to determine statistical heterogeneity.
Common sources of heterogeneity include clinical diversity between studies such as patient characteristics, settings, interventions, or outcomes, conflicts of interest, and differences in study quality.
Next slide
38. Example: Clinical heterogeneity? Here is an example of clinical heterogeneity from a meta-analysis by Nissen and colleagues from the NEJM. Part of the introduction reads:-----
Based on this introduction, we anticipate that studies included in this meta-analysis include patients with type 2 diabetes who used rosiglitazone. However, when looking at the actual included populations, you can see that other populations such as those with chronic psoriasis, Alzheimer's disease, or impaired glucose were included.
Inclusion of these very different populations with those with diabetes adds significant heterogeneity and could bias or lead to confusing or meaningless results.
Next slide.Here is an example of clinical heterogeneity from a meta-analysis by Nissen and colleagues from the NEJM. Part of the introduction reads:-----
Based on this introduction, we anticipate that studies included in this meta-analysis include patients with type 2 diabetes who used rosiglitazone. However, when looking at the actual included populations, you can see that other populations such as those with chronic psoriasis, Alzheimer's disease, or impaired glucose were included.
Inclusion of these very different populations with those with diabetes adds significant heterogeneity and could bias or lead to confusing or meaningless results.
Next slide.
39. How to read a Forest plot Forest plot adapted from Bjelakovic, et al. Lancet 2004; 364:1219-28. Another method used to help assess clinical heterogeneity (more specifically for meta-analyses) is to visually inspect a forest plot.
Each of the horizontal lines with boxes are the included trials in a meta-analysis.
The center vertical line represents the line of no effectand for this review this line is at 1 for RR. This line also divides the 2 treatment groups on the right favors--- and on the left, favoring ---.
For each trial, the square box indicates the point estimate or the result. The size of the square is proportional to the weight or size of the studythe larger the trialthe more precise the estimate.
The length of the horizontal line represents the confidence interval for the point estimate.
And finally the diamonds represent the pooled or combined results across the trials. The width of the diamond represents the confidence interval while the vertical tips indicate the point estimate.
Next slide
Another method used to help assess clinical heterogeneity (more specifically for meta-analyses) is to visually inspect a forest plot.
Each of the horizontal lines with boxes are the included trials in a meta-analysis.
The center vertical line represents the line of no effectand for this review this line is at 1 for RR. This line also divides the 2 treatment groups on the right favors--- and on the left, favoring ---.
For each trial, the square box indicates the point estimate or the result. The size of the square is proportional to the weight or size of the studythe larger the trialthe more precise the estimate.
The length of the horizontal line represents the confidence interval for the point estimate.
And finally the diamonds represent the pooled or combined results across the trials. The width of the diamond represents the confidence interval while the vertical tips indicate the point estimate.
Next slide
40. Two common methods
Fixed effects model
Assumes homogeneity
Random effects model
Assumes heterogeneity
Use both methods and select 1 to present
Should briefly discuss why a certain method was selected
What statistical method was used for the meta-analysis? Once, youve deemed that it is reasonable to combine studies, the next step is to determine which statistical method was used to combine the results? For this module, we will not be reviewing in detail the differences in these methods. But instead, we will just define these terms and what they assume so you are at least aware of these methods.
The 2 commonly reported statistical methods for combining results across studies are: fixed and random effects model.
A fixed effect model is based on a mathematical assumption that every study is evaluating a common treatment effect. This model assumes that results within studies differ by chance aloneand therefore assumes homogeneity.
Random effects model or Der Simonian-Laird method is an alternative approach that does not assume that a common or fixed treatment effect exists. This model assume that the true treatment effect in the studies may be different from each other. And therefore, assumes heterogeneity.
Typically, both models are conducted and 1 method is selected. The authors of the analysis should briefly discuss why a certain method was reported.
Next slideOnce, youve deemed that it is reasonable to combine studies, the next step is to determine which statistical method was used to combine the results? For this module, we will not be reviewing in detail the differences in these methods. But instead, we will just define these terms and what they assume so you are at least aware of these methods.
The 2 commonly reported statistical methods for combining results across studies are: fixed and random effects model.
A fixed effect model is based on a mathematical assumption that every study is evaluating a common treatment effect. This model assumes that results within studies differ by chance aloneand therefore assumes homogeneity.
Random effects model or Der Simonian-Laird method is an alternative approach that does not assume that a common or fixed treatment effect exists. This model assume that the true treatment effect in the studies may be different from each other. And therefore, assumes heterogeneity.
Typically, both models are conducted and 1 method is selected. The authors of the analysis should briefly discuss why a certain method was reported.
Next slide
41. Invalid methods of synthesis Picking and choosing
Pick what you like, ignore what you dont like
Searching for proof
Data dredging or data mining
Vote counting
Counting the number of studies with positive and negative results without considering study quality
We just reviewed 2 methods of data synthesis: qualitative and quantitaive methods. However, it is also common to observe invalid methods for data synthesis. Often these invalid methods of synthesis are seen in poor quality systematic reviews and narrative reviews. The methods can include: picking and choosing which evidence to report on while ignoring or minimizing what you dont like;
The next is
searching for proof which is similar to data dredging or data miningthat is, digging long and hard enough to find something that is of statistical significance
And finally vote counting which is counting the number of studies with positive and negative results without the consideration of study quality.
Next slideWe just reviewed 2 methods of data synthesis: qualitative and quantitaive methods. However, it is also common to observe invalid methods for data synthesis. Often these invalid methods of synthesis are seen in poor quality systematic reviews and narrative reviews. The methods can include: picking and choosing which evidence to report on while ignoring or minimizing what you dont like;
The next is
searching for proof which is similar to data dredging or data miningthat is, digging long and hard enough to find something that is of statistical significance
And finally vote counting which is counting the number of studies with positive and negative results without the consideration of study quality.
Next slide
42. Bridging the resultsto the conclusion Do conclusions reflect the uncertainty in the evidence?
Are gaps identified and recommendations for future research provided?
Now that weve finished going through each of the steps involved in the systematic review process and methods for data synthesis and analysis, the next step is to summarize the findings. For most clinicians and decision makers, the conclusion section is probably the most sought and read section of studies and systematic reviews. Therefore, it is crucial for authors and readers of these reviews to evaluate whether the conclusions were clear and accurately reflective of the results.
Do the conclusions reflect any uncertainty in the evidence that was found?
Are there still gaps in the evidence that need to be addressed in future research?
Next slideNow that weve finished going through each of the steps involved in the systematic review process and methods for data synthesis and analysis, the next step is to summarize the findings. For most clinicians and decision makers, the conclusion section is probably the most sought and read section of studies and systematic reviews. Therefore, it is crucial for authors and readers of these reviews to evaluate whether the conclusions were clear and accurately reflective of the results.
Do the conclusions reflect any uncertainty in the evidence that was found?
Are there still gaps in the evidence that need to be addressed in future research?
Next slide
43. I. Systematic Reviews:Outline Why, When, What?
Benefits and limitations
Steps in conducting Systematic Reviews
Scientific process
Quality assessment of Systematic Reviews
Tools and checklists
So, we just completed reviewing the steps involved in conducting systematic reviews. Evaluating whether each step of the systematic review process was followed with as little as bias as possible serves as the foundation for assessing its quality. Since weve already reviewed much of the process in detail and discussed some quality aspectswe will just briefly review a few tools and resources that can help you in assessing systematic review quality.
Next slideSo, we just completed reviewing the steps involved in conducting systematic reviews. Evaluating whether each step of the systematic review process was followed with as little as bias as possible serves as the foundation for assessing its quality. Since weve already reviewed much of the process in detail and discussed some quality aspectswe will just briefly review a few tools and resources that can help you in assessing systematic review quality.
Next slide
44. Key questions to ask when assessing quality of systematic reviews Is there a clear, focused, clinically relevant question?
Were study eligibility criteria reported and rationale provided (if needed)?
Was the search for relevant studies detailed and exhaustive?
Were included trials assessed for quality and were the assessments reproducible?
How was data synthesized and was this appropriate?
Are the conclusion statements clear and reflect the results from the evidence that was reviewed?
To assess the quality of systematic reviews, a few questions need to be addressed. These include the questions, we just reviewed which are:
Is there a clear, focused and clinically relevant question
How were the studies included in the review selected? Were reasons reported as to why certain studies were not included?
How systematic was the search for studies? Was the strategy fairly exhaustive?
Were the individual studies assess for their quality and are the assessments reproducible?
How was the data from the studies synthesized?
And finally, do the concluding statements make sense? Does it reflect the results from the evidence that was reviewed?
Next slideTo assess the quality of systematic reviews, a few questions need to be addressed. These include the questions, we just reviewed which are:
Is there a clear, focused and clinically relevant question
How were the studies included in the review selected? Were reasons reported as to why certain studies were not included?
How systematic was the search for studies? Was the strategy fairly exhaustive?
Were the individual studies assess for their quality and are the assessments reproducible?
How was the data from the studies synthesized?
And finally, do the concluding statements make sense? Does it reflect the results from the evidence that was reviewed?
Next slide
45. Tools and lists for assessing systematic review quality >10 different scales and checklists
Oxman and Guyatt
Sacks, et al
DERP method
As with assessing internal validity of randomized controlled trials, there are multiple different tools and checklists available to help guide you in evaluating quality.
The commonly referenced tools include: oxman and guyatt tool, sacks, et al tool, and the DERP method or tool for assessing systematic review quality.
It is important to note that there is no 1 tool that is considered the gold standard. There may be some quality assessment tools or checklists that have been validated by researchers but for the most part well thought out tools evaluate the key concepts in assessing systematic review quality. Also, remember that similar to assessing individual study quality or internal validity, this requires some judgment and has an element of subjectivitytherefore, it is prudent to also have dual review for evaluating quality assessment of systematic reviews as well.
Next slideAs with assessing internal validity of randomized controlled trials, there are multiple different tools and checklists available to help guide you in evaluating quality.
The commonly referenced tools include: oxman and guyatt tool, sacks, et al tool, and the DERP method or tool for assessing systematic review quality.
It is important to note that there is no 1 tool that is considered the gold standard. There may be some quality assessment tools or checklists that have been validated by researchers but for the most part well thought out tools evaluate the key concepts in assessing systematic review quality. Also, remember that similar to assessing individual study quality or internal validity, this requires some judgment and has an element of subjectivitytherefore, it is prudent to also have dual review for evaluating quality assessment of systematic reviews as well.
Next slide
46. Oxman and Guyatt Shea B, et al. Eval Health Prof 2002; 25(1):116-29. This is the Oxman and Guyatt tool. Each question is answered and a total score is given to rate the overall quality.
As you can see on this slide, each question evaluates whether the steps in the review process were conducted in a manner that was uniform and that would minimize potential biases.
Next slideThis is the Oxman and Guyatt tool. Each question is answered and a total score is given to rate the overall quality.
As you can see on this slide, each question evaluates whether the steps in the review process were conducted in a manner that was uniform and that would minimize potential biases.
Next slide
47. Using Oxman and Guyatt method From DERP report. http://www.ohsu.edu/drugeffectiveness Here is an example of the oxman and guyatt method in action. Dual review was conducted and this is the final product. The top of the form lists the abridged version of the questions that were on the previous slide. Each question was answered and a final score out of 7 is given.
As you can see the systematic review that was evaluated reported that searches were conducted in medline from 1966 to may 2007. The sources that were search were reportedand as you can see more than 5 different sources of literature were searchedsuggesting a comprehensive search. Study eligibility criteria also referred to as inclusion criteria was documentedand the authors of this specific systematic review also reported criteria for exclusion.
Next, dual review was performed in the selection of studies and each included trial was evaluated for its internal validity. The systematic review provided a table with their assessments.
Next slide Here is an example of the oxman and guyatt method in action. Dual review was conducted and this is the final product. The top of the form lists the abridged version of the questions that were on the previous slide. Each question was answered and a final score out of 7 is given.
As you can see the systematic review that was evaluated reported that searches were conducted in medline from 1966 to may 2007. The sources that were search were reportedand as you can see more than 5 different sources of literature were searchedsuggesting a comprehensive search. Study eligibility criteria also referred to as inclusion criteria was documentedand the authors of this specific systematic review also reported criteria for exclusion.
Next, dual review was performed in the selection of studies and each included trial was evaluated for its internal validity. The systematic review provided a table with their assessments.
Next slide
48. Using Oxman and Guyatt method(continued) Questions 6 and 7 were evaluatedand were found to be adequately fulfilled although there were a few areas of uncertainty relating to how poor quality studies were handled in the systematic review. As you may have inferred by the mention of random effects model and I-squared testthis systematic review also conducted a meta-analysis. And based on the provided study information in the systematic review and justifications provided by the authors, it was deemed that the findings were combined appropriately. And finally, the summary or conclusions reflected the results without over stepping the available evidence. An overall score of 6 out of 7 was given.
Next slideQuestions 6 and 7 were evaluatedand were found to be adequately fulfilled although there were a few areas of uncertainty relating to how poor quality studies were handled in the systematic review. As you may have inferred by the mention of random effects model and I-squared testthis systematic review also conducted a meta-analysis. And based on the provided study information in the systematic review and justifications provided by the authors, it was deemed that the findings were combined appropriately. And finally, the summary or conclusions reflected the results without over stepping the available evidence. An overall score of 6 out of 7 was given.
Next slide
49. Sacks, et al 1. Prospective design
a. Protocol
b. Literature search
c. Lists of trials analyzed
d. Log of rejected trials
e. Treatment assignment
f. Ranges of patients
g. Ranges of treatment
h. Ranges of diagnosis
2. Combinability
a. Criteria
b. Measurement
3. Control of bias
a. Selection bias
b. Data-extraction bias
c. Interobserver agreement
d. Source of support
4. Statistical analysis
a. Statistical methods
b. Statistical errors
c. Confidence intervals
d. Subgroup analysis
5. Sensitivity analysis
a. Quality assessment
b. Varying methods
c. Publication bias
6. Application of results
a. Caveats
b. Economic impact
7. Language
This is the Sacks and colleagues tool. Compared with the Oxman and Guyatt tool, this tool is more detailed for each area such as the design of the review, assessment of combinability of trials, control of bias, statistical analyses, sensitivity analyses, application of results, and language.
This tool also asks the reader to assess whether publication bias was evaluated by the authors of the systematic review.
Next slide.This is the Sacks and colleagues tool. Compared with the Oxman and Guyatt tool, this tool is more detailed for each area such as the design of the review, assessment of combinability of trials, control of bias, statistical analyses, sensitivity analyses, application of results, and language.
This tool also asks the reader to assess whether publication bias was evaluated by the authors of the systematic review.
Next slide.
50. DERP method 50 From DERP report. http://www.ohsu.edu/drugeffectiveness This is another example of a quality assessment tool for systematic reviews used by the Drug Effectiveness Review Project. This tool is similar to the 2 previous tools but unlike the oxman and guyatt tool which uses a scoring systemthis is based on a more qualitative approachof rating the systematic review as good, fair, or poor. The rating of good requires an almost perfect fulfillment of each criteria. A poor rating is either based on a serious fatal flaw or a combination of flaws
while the fair rating includes the rest. Each quality rating lies on a spectrum of good, fair, and poor and again, dual review is performed.
Feel free to take a closer look at each component.
Next slideThis is another example of a quality assessment tool for systematic reviews used by the Drug Effectiveness Review Project. This tool is similar to the 2 previous tools but unlike the oxman and guyatt tool which uses a scoring systemthis is based on a more qualitative approachof rating the systematic review as good, fair, or poor. The rating of good requires an almost perfect fulfillment of each criteria. A poor rating is either based on a serious fatal flaw or a combination of flaws
while the fair rating includes the rest. Each quality rating lies on a spectrum of good, fair, and poor and again, dual review is performed.
Feel free to take a closer look at each component.
Next slide
51. Database of Abstracts of Reviews of Effects Another great resource for readers of systematic reviews who desire a quick quality assessment of systematic reviews is The database of abstracts of reviews and effects or DARE . DARE is a service performed by the centre for reviews and dissemination. Its focus is on providing readers/clinicians/policy makers with brief, clear, and concise summaries of systematic reviews on health and social interventionsespecially highlighting strengths and weaknesses. Each month thousands of citations are screened to identify potential systematic reviews. These are independently assessed by two researchers for inclusion on the DARE website.
To ensure quality of the assessment of systematic reviews, abstracts are written and independently checked by health service researchers with in-depth knowledge and experience of systematic review methods. A copy of the abstract is sent to the original authors of the systematic review for any additional information. Authors are invited to reply with corrections to factual errors, further information and other relevant research. Where applicable, this information is added to the records. The DARE database is updated monthlybut due to the shear volume of reviewsthere appears to be a time lag.
Next slideAnother great resource for readers of systematic reviews who desire a quick quality assessment of systematic reviews is The database of abstracts of reviews and effects or DARE . DARE is a service performed by the centre for reviews and dissemination. Its focus is on providing readers/clinicians/policy makers with brief, clear, and concise summaries of systematic reviews on health and social interventionsespecially highlighting strengths and weaknesses. Each month thousands of citations are screened to identify potential systematic reviews. These are independently assessed by two researchers for inclusion on the DARE website.
To ensure quality of the assessment of systematic reviews, abstracts are written and independently checked by health service researchers with in-depth knowledge and experience of systematic review methods. A copy of the abstract is sent to the original authors of the systematic review for any additional information. Authors are invited to reply with corrections to factual errors, further information and other relevant research. Where applicable, this information is added to the records. The DARE database is updated monthlybut due to the shear volume of reviewsthere appears to be a time lag.
Next slide
52. Here is an example of an abstract that evaluated a systematic review on estrogen therapy for the treatment of hotflashes in postmenopausal women.
Next slideHere is an example of an abstract that evaluated a systematic review on estrogen therapy for the treatment of hotflashes in postmenopausal women.
Next slide
53. Here is the CRD summary/appraisal of the review with concluding remarks.
Next slideHere is the CRD summary/appraisal of the review with concluding remarks.
Next slide
54. Summary: Systematic Reviews Advantages and disadvantages
Can minimize biases that exist in individual studies
May not answer all questions of interest
Systematic reviews and meta-analyses are not synonymous
Meta-analysis is a statistical method of combining studies
Each step of the process should be questioned
Comprehensive search of evidence
Quality assessment of individual trials
Appropriate method of synthesis
So, this concludes the segment on appraising systematic reviews.
Weve discussed some advantages and disadvantages of this type of evidence. One advantage of systematic reviews is that these can help minimize potential biases observed in individual studies when taken together. One disadvantage is that systematic reviews have their limitations and may not answer all questions of interest.
Hopefully you understand that systematic reviews and meta-analyses are not the same. Meta-Analysis is a statistical method used for combining study results
And finally I hope you understand that a well-done high quality systematic review follows a rigorous scientific process and that each step of the process should be questioned. In particular, you should evaluate whether a comprehensive search of the evidence was performed, quality assessment of individual trials was achieved, and whether appropriate methods of synthesis were used.
Next slide.So, this concludes the segment on appraising systematic reviews.
Weve discussed some advantages and disadvantages of this type of evidence. One advantage of systematic reviews is that these can help minimize potential biases observed in individual studies when taken together. One disadvantage is that systematic reviews have their limitations and may not answer all questions of interest.
Hopefully you understand that systematic reviews and meta-analyses are not the same. Meta-Analysis is a statistical method used for combining study results
And finally I hope you understand that a well-done high quality systematic review follows a rigorous scientific process and that each step of the process should be questioned. In particular, you should evaluate whether a comprehensive search of the evidence was performed, quality assessment of individual trials was achieved, and whether appropriate methods of synthesis were used.
Next slide.
55. Appraisal of guidelines
Welcome to the critical appraisal of clinical practice guidelines!
Next slide
Welcome to the critical appraisal of clinical practice guidelines!
Next slide
56. II. Guidelines: Outline What is the purpose and what are the potential benefits and limitations?
Why do we need to critically assess guidelines?
Quality assessment of guidelines
Tools to help evaluate guidelines In this section, we will review the purpose of clinical practice guidelines, their benefits and limitations. We will also review a few reasons why critical evaluation is warranted, review of the key concepts in assessing quality of guidelines, and finally end with a review of a few tools that you can use to assess the quality of clinical guidelines.
Next slideIn this section, we will review the purpose of clinical practice guidelines, their benefits and limitations. We will also review a few reasons why critical evaluation is warranted, review of the key concepts in assessing quality of guidelines, and finally end with a review of a few tools that you can use to assess the quality of clinical guidelines.
Next slide
57. Guidelines: steps beyond a review Incorporates the judgments and values involved in making recommendations
Addresses larger spectrum of issues relevant for clinical decision making
Purpose:
Provide clinical practice recommendations
Improve quality of care and outcomes
Seek to influence change in clinical practice
Reduce inappropriate variation in practice
Shed light on gaps in the evidence
As mentioned before, systematic reviews are not guidelinesand guidelines are not systematic reviews even though this body of evidence can be used in the process. Development of clinical practice guidelines requires a few more steps beyond just a systematic review of the literature. It incorporates judgments and values of clinicians, patients, and other stakeholders by addressing broader clinically important issues.
The most obvious purpose of creating a guideline is to provide clinical practice recommendations resulting in the improvement of care and outcomes. Developing guidelines also try to influence a change in the current practice possibly by reducing inappropriate variations in practice. Because large bodies of evidence are considered in guideline development, gaps in the evidence may be revealed which could generate scientific research and progress.
Next slideAs mentioned before, systematic reviews are not guidelinesand guidelines are not systematic reviews even though this body of evidence can be used in the process. Development of clinical practice guidelines requires a few more steps beyond just a systematic review of the literature. It incorporates judgments and values of clinicians, patients, and other stakeholders by addressing broader clinically important issues.
The most obvious purpose of creating a guideline is to provide clinical practice recommendations resulting in the improvement of care and outcomes. Developing guidelines also try to influence a change in the current practice possibly by reducing inappropriate variations in practice. Because large bodies of evidence are considered in guideline development, gaps in the evidence may be revealed which could generate scientific research and progress.
Next slide
58. Guidelines are not intended to
Provide a black and white answer for complex situations
Substitute for clinical insight and judgment
Be a legal resource in malpractice cases
Prompt providers to withdraw availability or coverage of therapies
Hinder or discourage scientific progress Woolf S, et al. BMJ 1999; 318:527-30. Guidelines however, are not intended to provide clinicians or decision makers with a black and white or cookbook answer for complex medical situations. Nor are they meant to replace clinical insight and judgment. Guidelines are meant provide a clinicians with as much evidence as available to make decisions. Guidelines are also not mean to be used as a legal or binding document in malpractice cases. Often, guidelines can be misconstrued by both policy makers and clinicians as the only right way to practice medicine. In the same light, guidelines are not intended for providers or health groups to withdraw availability or coverage of therapiesagain, guidelines are guides. And finally, recommendations from guidelines are not meant to hinder or discourage scientific progress or research.
Next slide
Guidelines however, are not intended to provide clinicians or decision makers with a black and white or cookbook answer for complex medical situations. Nor are they meant to replace clinical insight and judgment. Guidelines are meant provide a clinicians with as much evidence as available to make decisions. Guidelines are also not mean to be used as a legal or binding document in malpractice cases. Often, guidelines can be misconstrued by both policy makers and clinicians as the only right way to practice medicine. In the same light, guidelines are not intended for providers or health groups to withdraw availability or coverage of therapiesagain, guidelines are guides. And finally, recommendations from guidelines are not meant to hinder or discourage scientific progress or research.
Next slide
59. Why is it necessary tocritically assess clinical practice guidelines? There are > 2,500 published guidelines
Multiple guidelines with differing recommendations
Not all guidelines are of good/high quality
Consensus-based
Evidence-based (systematic methods, transparent)
Many stakeholders who are invested in the influence of their guidelines
Government organizations and healthcare systems
Professional societies
Pharmaceutical industry So, why is it necessary to critically assess clinical practice guidelines? Isnt it enough that weve critically looked at trials and systematic reviews?
Evaluating clinical practice guidelines with scrutiny is needed because of the wide variability in the quality of how guidelines are produced, used, and abused.
There are more than 2500 guidelines most of which are consensus-based. These are more often than not heavily influenced by expert opinion rather than based on a systematic and transparent approach of using evidence and are not as time and resource dependent. Evidence-based or evidence-informed clinical practice guidelines on the other hand are limited in number compared with consensus-based methods because of its time and resource intensive approachthese types of guidelines are definitely preferred and are less likely to be biased and abused relative to consensus guidelines.
It is imperative that clinical guidelines be scrutinized because of their significant impact on clinical practice. Many stakeholders are invested and recommendations in guidelines affect patients, physicians, policies, professional societies, and industry.
Next slideSo, why is it necessary to critically assess clinical practice guidelines? Isnt it enough that weve critically looked at trials and systematic reviews?
Evaluating clinical practice guidelines with scrutiny is needed because of the wide variability in the quality of how guidelines are produced, used, and abused.
There are more than 2500 guidelines most of which are consensus-based. These are more often than not heavily influenced by expert opinion rather than based on a systematic and transparent approach of using evidence and are not as time and resource dependent. Evidence-based or evidence-informed clinical practice guidelines on the other hand are limited in number compared with consensus-based methods because of its time and resource intensive approachthese types of guidelines are definitely preferred and are less likely to be biased and abused relative to consensus guidelines.
It is imperative that clinical guidelines be scrutinized because of their significant impact on clinical practice. Many stakeholders are invested and recommendations in guidelines affect patients, physicians, policies, professional societies, and industry.
Next slide
60. A glance at guidelines from 1988-1998 3 items assessed
Description of professionals involved
Search undertaken
Explicit grading of evidence for recommendation
Grilli, et al. Lancet 2000; 355:103-6. Another important reason for the need to critically evaluate clinical practice guidelines is because of the increasing rate of clinical practice guidelines that are published by professional societies who have been taking active roles in the development process. The large growth in guidelines without critical evaluation of the process could undermine the credibility and benefits of having guidelines.
Grilli and colleagues wanted to empirically evaluate 3 aspects that affect quality of guidelines put forth by specialty groups over a 10 year period between 1988 and 1998.
431 guidelines were retrieved and evaluated on 3 aspects: 1) Type of professionals involved in guideline groups, 2) Sources of information used to retrieve evidence, and 3) whether there was explicit grading or quality assessment of the included evidence in guidelines.
As you can see, a large proportion of guidelines were from cardiology, oncology, and neurology specialty groups. More than 1/3rd of the guidelines were regarding treatment recommendations. And the largest surge of the guidelines were published between 1994-1998 over the 10 year span.
Next slideAnother important reason for the need to critically evaluate clinical practice guidelines is because of the increasing rate of clinical practice guidelines that are published by professional societies who have been taking active roles in the development process. The large growth in guidelines without critical evaluation of the process could undermine the credibility and benefits of having guidelines.
Grilli and colleagues wanted to empirically evaluate 3 aspects that affect quality of guidelines put forth by specialty groups over a 10 year period between 1988 and 1998.
431 guidelines were retrieved and evaluated on 3 aspects: 1) Type of professionals involved in guideline groups, 2) Sources of information used to retrieve evidence, and 3) whether there was explicit grading or quality assessment of the included evidence in guidelines.
As you can see, a large proportion of guidelines were from cardiology, oncology, and neurology specialty groups. More than 1/3rd of the guidelines were regarding treatment recommendations. And the largest surge of the guidelines were published between 1994-1998 over the 10 year span.
Next slide
61. Results from Grilli, et al Grilli and colleagues found that 67% of the guidelines from 1988-1998 did not describe the type of professionals involved in guideline groups,. Only 28% of guidelines included at least 1 professional or representative outside of the specialty. Astonishingly, 87% of the guidelines did not specify whether systematic search for evidence was undertaken. And less than 20% of guidelines provided information describing how the strength of the body of evidence was graded. Only 5% of guidelines fulfilled all 3 criteria compared with 54% who did not meet any of the 3 criteriathis example provides compelling evidence for the need to critically evaluate and scrutinize the quality of guidelines.
Next slideGrilli and colleagues found that 67% of the guidelines from 1988-1998 did not describe the type of professionals involved in guideline groups,. Only 28% of guidelines included at least 1 professional or representative outside of the specialty. Astonishingly, 87% of the guidelines did not specify whether systematic search for evidence was undertaken. And less than 20% of guidelines provided information describing how the strength of the body of evidence was graded. Only 5% of guidelines fulfilled all 3 criteria compared with 54% who did not meet any of the 3 criteriathis example provides compelling evidence for the need to critically evaluate and scrutinize the quality of guidelines.
Next slide
62. Assessing quality Who were involved in the decision making process?
Were all relevant perspectives considered?
To what extent were the funders of the guideline involved in the process?
Conflicts of interests declared for each participant?
Were all important practice options and clinically relevant outcomes considered?
What was excluded and was rationale provided?
How were the relative values of the outcomes weighed in terms of importance?
When assessing guideline quality, a few key questions need to be assessed. The first is, Who were involved in decision making and were all relevant perspectives included and considered during the guideline development process?
Well-conducted and thoughtful guidelines try and involve as many disciplines as possible. When a single specialty group is formed, inherently the group will be biased in favor of performing procedures or treatments in which they have a vested interest. Individual biases may be better balanced in multidisciplinary groups which may produce more valid guidelines
Next to what extent were the funders of the guideline involved in the process? And were conflicts of interests declared for each of the participants? Reporting of such information opens up the process to become more transparent to any potential sources of bias especially if there appears to be an imbalance in maybe one group being heavily represented relative to others.
It is also important to assess whether all important practice options and clinically relevant outcomes were considered. If certain outcomes were not considered, members of the guideline committee should provide rationale as to why certain items were excluded.
Once a list of relevant outcomes has been decided upon, it is worth asking how were the outcomes weighed in terms of importance? Is there a reason why this outcome is considered to be more important than another? Is the outcome preference influenced by the funders, one specialty society, or a reflection of a multidisciplinary group? This information is nice to have but may not always be available to readers of clinical practice guidelines.
Next slideWhen assessing guideline quality, a few key questions need to be assessed. The first is, Who were involved in decision making and were all relevant perspectives included and considered during the guideline development process?
Well-conducted and thoughtful guidelines try and involve as many disciplines as possible. When a single specialty group is formed, inherently the group will be biased in favor of performing procedures or treatments in which they have a vested interest. Individual biases may be better balanced in multidisciplinary groups which may produce more valid guidelines
Next to what extent were the funders of the guideline involved in the process? And were conflicts of interests declared for each of the participants? Reporting of such information opens up the process to become more transparent to any potential sources of bias especially if there appears to be an imbalance in maybe one group being heavily represented relative to others.
It is also important to assess whether all important practice options and clinically relevant outcomes were considered. If certain outcomes were not considered, members of the guideline committee should provide rationale as to why certain items were excluded.
Once a list of relevant outcomes has been decided upon, it is worth asking how were the outcomes weighed in terms of importance? Is there a reason why this outcome is considered to be more important than another? Is the outcome preference influenced by the funders, one specialty society, or a reflection of a multidisciplinary group? This information is nice to have but may not always be available to readers of clinical practice guidelines.
Next slide
63. Consensus guidelines: Assessment, diagnosis, and treatment of diabetic peripheral neuropathic pain. Mayo Clinic Proceedings 2006; 81(4):S1-36. This is an example of a front page of a consensus guideline for the treatment of diabetic neuropathy. 11 pain specialists were invited and involved in the development of this guideline over just a 2 day period in New Orleans. Looking at the list of participants, there were no primary care physicians involved, no internal medicine physicians, or geriatriciansfor whom was the target audience.
Financial and commercial conflicts of interest were reported for 10 out of 11 members. All of whom had received honoraria or research funding from at least 1 of the manufacturers of medications listed in the first-tier or recommended category. 8 members had some funding from Eli Lilly who provided an educational grant for the production of this guideline.
Next slide.This is an example of a front page of a consensus guideline for the treatment of diabetic neuropathy. 11 pain specialists were invited and involved in the development of this guideline over just a 2 day period in New Orleans. Looking at the list of participants, there were no primary care physicians involved, no internal medicine physicians, or geriatriciansfor whom was the target audience.
Financial and commercial conflicts of interest were reported for 10 out of 11 members. All of whom had received honoraria or research funding from at least 1 of the manufacturers of medications listed in the first-tier or recommended category. 8 members had some funding from Eli Lilly who provided an educational grant for the production of this guideline.
Next slide.
64. Assessing quality How was evidence retrieved?
Was it comprehensive?
Was there explicit description of how evidence was used?
Systematic reviews?
Was there an approach to the hierarchy of evidence?
Was quality of the evidence assessed and reported?
How was the body of evidence graded? The next series of questions that should be asked revolves around the evidence-base and the quality of the evidence used in guidelines which is probably the crux to determining more credible guidelines.
First, how was evidence retrieved? Was the process of retrieving the evidence explicit, transparent, and comprehensive? Was the process for using different levels evidence (such as expert opinion) or systematic reviews? Was there a specific approach to using the hierarchy of evidence?
Next, how was the quality of the evidence assessed in the guideline development process? Was this information reported? Based on the quality of the evidence how the body of evidence graded so that clinical recommendations can be made? Or was this based more on expert opinion and experience?
Next slideThe next series of questions that should be asked revolves around the evidence-base and the quality of the evidence used in guidelines which is probably the crux to determining more credible guidelines.
First, how was evidence retrieved? Was the process of retrieving the evidence explicit, transparent, and comprehensive? Was the process for using different levels evidence (such as expert opinion) or systematic reviews? Was there a specific approach to using the hierarchy of evidence?
Next, how was the quality of the evidence assessed in the guideline development process? Was this information reported? Based on the quality of the evidence how the body of evidence graded so that clinical recommendations can be made? Or was this based more on expert opinion and experience?
Next slide
65. Chou, et al. Ann Intern Med 2007; 147:505-14. Here is a snap shot of the methods section of a guideline for low back pain sponsored by the American Pain Society and American College of Physicians. This guideline specified who determined the expert panel involved in the process, reported the methods used in retrieving evidence, how evidence was used based on the hierarchy of evidence, what was excluded from the review, and so on
Next slideHere is a snap shot of the methods section of a guideline for low back pain sponsored by the American Pain Society and American College of Physicians. This guideline specified who determined the expert panel involved in the process, reported the methods used in retrieving evidence, how evidence was used based on the hierarchy of evidence, what was excluded from the review, and so on
Next slide
66. -ing the strength of the evidence and recommendations reported in guidelines
To provide a systematic and explicit approach to making judgments involved in a guideline process that can be used by all guideline developers
One method ofassessing the body of evidence One recent new method in assessing the body of evidence for guideline development is the GRADE method.
The GRADE method was developed by a multidisciplinary group of clinicians and methodologists who were interested in improving the old and variable methods of assessing the body of literature in guidelines into something more transparent, systematic, and replicable.
Next slideOne recent new method in assessing the body of evidence for guideline development is the GRADE method.
The GRADE method was developed by a multidisciplinary group of clinicians and methodologists who were interested in improving the old and variable methods of assessing the body of literature in guidelines into something more transparent, systematic, and replicable.
Next slide
67. The approach considers: Strength of the body of evidence
Study design
Risk of bias or limitations
Consistency of results
Precision
Directness of evidence
Strength of recommendation:
Strong vs. Weak Briefly, the GRADE approach considers a minimum of 4 domainshere we list 5: study design, risk of bias across trials, consistency of results across multiple trials, precision, and directness of the evidence from multiple trials.
An overall GRADE of the body of evidence is categorized as high, moderate, low, or very low quality. Grading the body of evidence in this manner helps with determining whether strong or weak recommendations will be made.
Next slide
Briefly, the GRADE approach considers a minimum of 4 domainshere we list 5: study design, risk of bias across trials, consistency of results across multiple trials, precision, and directness of the evidence from multiple trials.
An overall GRADE of the body of evidence is categorized as high, moderate, low, or very low quality. Grading the body of evidence in this manner helps with determining whether strong or weak recommendations will be made.
Next slide
68. Example: GRADE table This is just an example of a GRADE table with all the factors that were considered for determining the effectiveness of pneumococcal vaccine for treating invasive pneumococcal disease.
http://www.who.int/immunization/pp_pneumo_grade_tables/en/index.html
This is just an example of a GRADE table with all the factors that were considered for determining the effectiveness of pneumococcal vaccine for treating invasive pneumococcal disease.
http://www.who.int/immunization/pp_pneumo_grade_tables/en/index.html
69. Finally, This is the AGREE instrument or checklist for critically assessing clinical practice guidelines. This instrument was developed by researchers and policy makers who wanted to improve the quality of clinical practice guidelines by establishing a minimum framework for the process.
This tool evaluates and questions each step of the guideline development process which include: scope and purpose, stakeholder involvement, rigor of development, clarity of recommendations, applicability, and editorial independence.
The key concepts however continue to revolve around who was involved, how was the evidence retrieved and assessed, and how was the body of the evidence considered.
Next slide
Finally, This is the AGREE instrument or checklist for critically assessing clinical practice guidelines. This instrument was developed by researchers and policy makers who wanted to improve the quality of clinical practice guidelines by establishing a minimum framework for the process.
This tool evaluates and questions each step of the guideline development process which include: scope and purpose, stakeholder involvement, rigor of development, clarity of recommendations, applicability, and editorial independence.
The key concepts however continue to revolve around who was involved, how was the evidence retrieved and assessed, and how was the body of the evidence considered.
Next slide
70. Summary: Guidelines Incorporates values and judgments
Can improve care by reducing variation in practice
Not meant to provide black and white answers for complex problems
Not all guidelines are the same
Consensus-based approach
Evidence-informed approach
Important to question each step of the process
Who was involved?
How was evidence retrieved, synthesized, and graded?
How were recommendation decisions made?
In summary, unlike systematic reviews clinical practice guideline incorporate values and judgments of those involved in the development process. Clinical practice guidelines can be helpful in situations where there is wide variability in practice in light of good evidence with the caveat that guidelines are not meant to provide black and white answers for all situations or complex problems.
It is important to remember that not all guideline are created equally
there are significant differences and implications to consensus-based approaches as well as evidence-based or evidence-informed approaches.
And it is always important to question each step of the guideline development process: who was involved? How was evidence retrieved, synthesized, and graded? And how were the recommendations made?
This concludes this module.
Next slide.In summary, unlike systematic reviews clinical practice guideline incorporate values and judgments of those involved in the development process. Clinical practice guidelines can be helpful in situations where there is wide variability in practice in light of good evidence with the caveat that guidelines are not meant to provide black and white answers for all situations or complex problems.
It is important to remember that not all guideline are created equally
there are significant differences and implications to consensus-based approaches as well as evidence-based or evidence-informed approaches.
And it is always important to question each step of the guideline development process: who was involved? How was evidence retrieved, synthesized, and graded? And how were the recommendations made?
This concludes this module.
Next slide.
71. Acknowledgements Attorney General Consumer and Prescriber Education Program
Members of the technical advisory committee of this grant
Office for Oregon Health Policy and Research
The University of Texas Southwestern Medical Center
The Federation of State Medical Boards Research and Education Foundation
I would like to acknowledge the attorney general consumer and prescriber program, the members of the technical advisory committee for review of the module, the office for Oregon health policy and research, the univ of Texas southwestern med center, and the federation of state medical boards research and education foundation.
Next slideI would like to acknowledge the attorney general consumer and prescriber program, the members of the technical advisory committee for review of the module, the office for Oregon health policy and research, the univ of Texas southwestern med center, and the federation of state medical boards research and education foundation.
Next slide
72. CME instructions Please complete the survey, CME questions, and program evaluation after this slide
Dont forget to click the finish button at the end of the CME questions
You should be directly linked to a CME form which you will need to fill out and fax, email, or mail in order to receive credit hours
Please complete the survey, CME questions, and program evaluation before exiting the session. After you have completed the CME questions, dont forget to click he finish buttonthis will directly link you to a CME form which you will need to fill out and fax, email, or mail in order to receive your credit hours.Please complete the survey, CME questions, and program evaluation before exiting the session. After you have completed the CME questions, dont forget to click he finish buttonthis will directly link you to a CME form which you will need to fill out and fax, email, or mail in order to receive your credit hours.
73. systematic_quiz