1.2k likes | 4.95k Views
Critiquing Quantitative Research. Cynthia L. Russell, PhD, APRN, BC. There is no conflict of interest in the creation of this education program. Learner Objectives. 1. Describe 2 reasons for evaluating quantitative research. 2. Identify 5 sections of quantitative research reports.
E N D
Critiquing Quantitative Research Cynthia L. Russell, PhD, APRN, BC There is no conflict of interest in the creation of this education program.
Learner Objectives 1. Describe 2 reasons for evaluating quantitative research. 2. Identify 5 sections of quantitative research reports. 3. Discuss the 5 components of the methods section of a quantitative research report.
Reasons for Evaluating Quantitative Research Reports • To determine if the findings can be used in practice • (Caelli, Ray & Mill, 2003)
Reasons for Evaluating Quantitative Research Reports • To determine the current state of science on a particular topic to direct your next research study
Components of a Quantitative Research Report Quantitative research report components: • -- Introduction/Research problem • -- Review of literature • -- Methods • -- Data analysis • -- Results • -- Discussion • (Caelli, Ray & Mill, 2003)
Introduction • Describes the background of the problem • Prevalence and/or incidence of the problem • Impact of the problem on patient outcomes • Importance of the problem to health care providers and nurses, in particular • Builds a case for the need for the study • Develops the Research Problem
Research Problem • Broadly defines what needs to be studied • Flows directly from the introduction • Concludes the Introduction section • (Caelli, Ray & Mill, 2003)
Review of the Literature Summary of pertinent literature Organized and clear Evaluate quality of literature Strengths and weaknesses Organized from broad to specific Identify gaps in literature Need for the study should be clear Discuss theoretical/conceptual framework
Methods Describes the steps used by the researcher to conduct the study Design Sampling plan Instrumentation Procedures Protection of human subjects
Design Describes the plan for carrying out the study Non-experimental designs Types Descriptive Correlational Experimental designs Types Randomized control trials Quasi-experimental designs
Sampling Plan Describes the sample for the study Who will participate How they will be identified Characteristics of the target population (the population to which the findings are generalized) Sample procedure Sample size
Instruments Describes the instruments used to gather the data for the study Researcher should identify which instruments measured which study concepts May involve Interviews Questionnaires Scales Observation Biophysiological measures
Instrument Reliability and Validity • Reliability is the instrument’s ability to consistently measure the concept • Validity is the instrument’s ability to measure the concept that is supposed to measure
Protection of Human Participants • Report should state that Institutional Review Board Approval was obtained • In the Procedure section the report should state that informed consent was obtained
Informed Consent Why Am I Being Approached? Why Is This Study Being Done? How Many People Will Take Part In The Study? What Is Involved in the Study? How Long Will I Be in the Study?
Informed Consent What Are the Risks of the Study? Are There Benefits to Taking Part in the Study? What Other Options Are There? What about Confidentiality? What Are the Costs?
Informed Consent • Will I be Paid for Participating in the Study? • What if I am Injured? • What Are My Rights as a Participant? • Whom Do I Call if I Have Questions or Problems? • Subject’s signature • Signature of Study Representative • If patient is incompetent to give consent, additional steps are taken
Procedure • Recipe for the research process • Sufficient detail should be provided • Should be written clearly and flow logically • All steps of the procedure should be clearly stated
Data Analysis and Results • Can be intimidating • Confirm that the researcher has presented results that clearly answer the proposed research question(s) • Researchers frequently organize this section by research question to facilitate readability
Data Analysis and Results Review the research design Descriptive design should use descriptive statistics such as: Mean, mode, and median (which measure how the data tend to be similar or grouped together) and variance, Standard deviation, Range (which measure how the data tend to be spread out) Correlational design should use correlational statistics such as a correlation coefficient (identifies the strength and the direction of the relationship between two variables. Examples are: Person r Spearman rho.
Discussion • Should clearly flow from the data • Should place the study’s findings in context with what is already known • May be linked back to theoretical or conceptual framework • Author interpretations should be clearly identified as such • Reviewer must determine if the author’s interpretations are justified • Should present study limitations • Should present implications for practice, as appropriate • Should present directions for future research
Study Limitations Common study limitations include: Small sample size Sample limited to one area or one group of people (this decreases generalizability to other areas and other people) Weak designs used such as correlational or quasi-experimental
Checklist for Research Report Evaluation • Research Problem • Is the problem clearly stated? • Is the problem significant? • Review of the Literature • Is the literature summarized? • Is the literature critically evaluated? • Are gaps and inconsistencies in the literature described? • Is the literature current and the review complete? • If presented, is the theoretical or conceptual framework clearly described including concepts and relationships? • Does the problem clearly link to and flow from the theoretical or conceptual framework? • Design • Is the design clearly stated?
Checklist for Research Report Evaluation • Sample • Is the sample clearly identified? • Is it clear how the sample will be obtained? • Is the relationship between the sample and the target population clearly delineated? • Is the rationale for the sample size provided? • Instrumentation • Is it clear which instruments will measure which concepts? • Is the rationale for instrument selection acceptable? • Is the reliability for each instrument described and acceptable? • Is the validity for each instrument described and acceptable? • Procedure • Are sufficient details provided in the procedure? • Is the procedure written clearly? • Does the procedure flow logically? • Are all steps of the procedure clearly stated?
Checklist for Research Report Evaluation Human Participants Protection • Has the researcher provided sufficient protection of human participants? • Data Analysis and Results • Is the data analysis section well organized? • Is the statistical method used for analysis appropriate for the research question(s) and/or hypothesis and level of measurement? • Are tables and graphs clearly labeled? • Do the tables and graphs complement the text?
Checklist for Research Report Evaluation Discussion • Does the discussion clearly flow from the data? • Does the discussion place the study’s findings in context with what is already known? • If a theoretical or conceptual framework is presented, are the nature of the findings discussed in the context of the framework? • If the author presents interpretations of the findings are these clearly distinguished as such? • Are justifications offered for the author’s conclusions? • Are study limitations provided? • Are implications for practice and future research delineated?
Summary • You may see yourself as a novice reviewer and not trust your interpretations. • This is a common concern. • Read, review, and discuss research reports with others to increase your skills.
References • Alderman, S. (1998). Critiquing research for use in clinical nursing practice: A CD-ROM review. Nurse Educator, 23(2), 8. • Beck, A. T., Steer, R. A., & Garbin, M. G. (1988). Psychometric properties of the Beck Depression Inventory: Twenty-five years of evaluation. Clinical Psychology Review, 8(1), 77-100. • Beck, A. T., Ward, C. H., Mendelson, M., Mock, J., & Erbaugh, J. (1961). An inventory for measuring depression. Archives of General Psychiatry, 4, 561-571.
References • Beck, C. T. (1990). The research critique: general criteria for evaluating a research report. JOGNN, 19(1), 18-22. • Beyea, S. C. (1998). Critiquing research for use in clinical nursing practice (CD-ROM). Computers in Nursing, 16(1), 16-17. • Brink, P. J., & Wood, M. J. (2001). Basic Steps in Planning Nursing Research (5th ed.). Boston: Jones and Bartlett Publishers. • Haughey, B.P. (1994). Evaluating quantitative research designs: part 1. Critical Care Nurse, 14(5): 100-2.
References • Haughey, B. P. (1994). Evaluating quantitative research designs: part 2. Critical Care Nurse, 14(6): 69-72. • Holcomb, Z. C. (2002). Interpreting Basic Statistics (3rd ed.). Los Angeles, CA: Pyrczak Publishing. • Pieper, B. (1993). Basics of critiquing a research article. Journal of ET Nursing, 20, 245-250. • Polit, D., & Beck, C. T. (2004). Nursing Research: Principles and Methods (7th ed.). Philadelphia, PA: Lippincott, Williams, & Wilkins.
References • Polit, D., & Beck, C. T. (2008). Nursing Research: Principles and Methods (8th ed.). Philadelphia, PA: Lippincott, Williams, & Wilkins. • Russell, C. L., Brown, K. (2002). The effects of information and support on individuals awaiting transplant. Progress in Transplantation, 12(3), 201-207. • Russell, C. L. (2005). Evaluating quantitative research reports. Nephrology Nursing Journal, 32(1), 61-64.
References • Ryan-Wengar, N. M. (1992). Guidelines for critique of a research report. Heart & Lung, 21(4), 394-401. • Soeken, K. L. (1985). Critiquing research: steps for complete evaluation on an article. AORN Journal, 41(5), 882-893. • Summers, S. (1991). Defining components of the research process needed to conduct and critique studies. Journal of Post Anesthesia Nursing, 6(1), 50-55.