E N D
Chapter 7 Flashcards
overall plan that describes all of the elements of a research or evaluation study, and ideally the plan allows the researcher or evaluator to reach valid conclusions (e.g., questions or hypotheses to be addressed, number and types of participants to be included, number and types of variables to be studied, collection and analysis of data) Evaluation/research design
family of research and evaluation designs characterized by the systematic repeated measurement of a client’s outcome(s) at regular, frequent, predesignated intervals under different conditions (baselineandintervention), and the evaluation of outcomes over time and under different conditions in order to monitor client progress, identify intervention effects, and more generally, learn when, why, how, and the extent to which client change occurs. Also known as single-subject designs, single-system designs, N = 1 designs, or sometimes time series or interrupted time series designs Single-case design
period of time during which an outcome is measured repeatedly in the absence of an intervention in order to (1) describe the naturally occurring pattern of outcome data (e.g., level, trend, variability) and (2) determine the effect of an intervention on that outcome. Typically symbolized by the letter A Baseline phase
period of time during which an intervention is implemented while an outcome is measured repeatedly Intervention phase
period of time after an intervention has ended during which outcome data are collected to determine the extent to which a client’s progress has been maintained. Also known as a maintenance phase Follow-up phase
a variable (e.g.,intervention) that produces an effector is responsible for events or results (e.g., outcome) Cause
change in one variable (e.g., outcome) that occurred at least in part as the result of another variable (e.g., intervention) Effect
measure of the strength of the relationship between variables (e.g., effect of an intervention on an outcome, as quantified by any one of a number of different statistics). Effect size
conclusion based on evidence and reasoning that one variable (e.g., intervention) causes another (e.g., outcome) Causal inference
accuracy of conclusions based on evidence and reasoning about the presence, direction, and strength of relationships between variables (e.g., outcome is different during baseline than intervention). The ability to establish that one variable (e.g., intervention) is related to another (e.g., outcome) is a requirement for inferring that one variable caused another Statistical conclusion validity
accuracy of conclusions based on evidence and reasoning about causal relationships between variables (e.g., extent to which an intervention, as opposed to other factors, caused a change in an outcome). Internal validity
uncertainty about which of several events or processes caused an outcome Causal ambiguity
variable that is associated with the independent variable inadvertently influences the outcome, and consequently makes it difficult to determine the effect of the intervention on the outcome (e.g., an unknown event that occurs during intervention but not baseline and causes the pattern of outcome data to change from baseline to intervention). Also known as a confound or confounding variable, and such a result is said to be confounded Extraneous variable
plausible reasons for a relationship between an intervention and an outcome, other than that the intervention caused the outcome. Also known as alternative hypotheses. The ability to rule out alternative hypotheses is a requirement for inferring that one variable (intervention) caused another (outcome) Alternative explanations
reasons why it might be partly or completely wrong (i.e., invalid) to conclude that one variable (e.g., an intervention) caused another (e.g., an outcome). See also Ambiguous temporal precedence, History effect, Instrumentation effect, Maturation effect, Regression effect, and Testing effect Threats to internal validity
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by an external event that occurs at the same time as the intervention (e.g., a student who has trouble completing his homework for a new teacher improves his performance as he becomes accustomed to her and her expectations, and the improvement is misinterpreted as being due to the rewards he earns working with a social worker). History effect
potential threat to internal validity in which an apparent change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by a change in how the outcome is measured (e.g., an older man’s weight stabilizes when he begins weighing at his physician’s office rather than at home). Instrumentation effect
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by naturally occurring changes in clients over time (e.g., a toddler’s tantrums diminish not in response to an intervention, but due to maturing out of the terrible twos, or a child outgrows enuresis). Maturation effect
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by repeated measurement of the outcome (e.g., a pretest about health behaviors may sensitize a client to the need to make changes in diet and exercise). See also Fatigue effect and Practice effect. Testing effect
deterioration in an outcome caused by fatigue associated with repeated measurement of the outcome (e.g., a mother who is self-recording each instance of time-out with her preschooler reduces those time-outs to avoid recording the behavior) Fatigue effect
improvement in an outcome caused by repeated measurement of the outcome (e.g., taking multiple practice exams may improve a student’s score on the Graduate Record Exam simply because he or she becomes familiar with the format of the exam) Practice effect
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by the tendency of an individual with unusually high or low scores on a measure to subsequently have scores closer to the mean (e.g., clients who are depressed frequently seek help when they have hit bottom, and their scores are likely to improve somewhat in the following weeks, even without intervention). Also known as regression toward the mean Regression effect
result due to the order in which different interventions are administered (e.g., a couple may be more successful in an intervention designed to increase their pleasant time together each day if they first complete an intervention designed to increase their reflective listening). Also known as a sequence effect Order effect
single-casedesign(arguably) consisting of aninterventionphase(B) during which the outcome is measured repeatedly B-only design
single-casedesign(arguably, since there isn’t repeated measurement during baseline) consisting of one pre-interventionoutcome measurement followed by an interventionphase (B) during which the outcome is measured repeatedly B+ design
two-phase single-case design consisting of a pre-intervention baseline phase (A) followed by an intervention phase (B) A-B design
three-phase single-case design consisting of a pre-intervention baseline phase (A1); an intervention phase (B); and a second baseline phase (A2) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern A-B-A design
three-phase single-case design beginning with the intervention phase (B1), followed by the withdrawal of the intervention (A) to determine if the outcome changes in the absence of the intervention, and reintroduction of the intervention (B2) to see whether the initial intervention effects are replicated B-A-B design
four-phase single-case design consisting of a pre-intervention baseline phase (A1); an intervention phase (B1); a second baseline phase (A2) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern; and a reintroduction of the intervention (B2) to see whether the initial intervention effects are replicated. Also known as a reversal or withdrawal design A-B-A-B design
single-casedesignthat begins with a baseline during which the sameproblemis measured for a single client in two or more settings at the same time. Baseline is followed by the application of theinterventionin one setting while baseline conditions remain in effect for other settings, then the intervention is applied sequentially across the remaining settings to see whetherinterventioneffectsare replicated across different settings Multiple baseline across settings design
single-casedesign that begins with a baseline during which the same problem is measured for two or more clients at the same time in a particular setting. Baseline is followed by the application of the intervention to one client while baseline conditions remain in effect for other clients, then the intervention is applied sequentially to remaining clients to see whether interventioneffects are replicated across different clients Multiple baseline across subjects (clients) design
single-casedesign that begins with a baseline during which two or more problems are measured at the same time for a single client in a particular setting. Baseline is followed by the application of the intervention to one problem with baseline conditions remaining in effect for other problems, then the intervention is applied sequentially to the remaining problems to see whether interventioneffects are replicated across different problems Multiple baseline across behaviors (problems) design
three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase (C) in which a new intervention is introduced in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-C design
three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase in which a new intervention (C) is added to the first intervention in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-BC design
accuracy of conclusions based on evidence and reasoning about how well a causal relationship applies across or beyond the people, settings, treatment variables, and measurement variables that were studied (e.g., extent to which a causal relationship between an intervention and outcome is the same with different people) External validity