440 likes | 742 Views
Chapter 13 The Behavioral/ Social Learning Approach: Theory and Application. Behaviorism. John B. Watson (1878-1958). Watson was a fighter and a builder. In college, he was unsociable and uninterested in his studies.
E N D
Chapter 13The Behavioral/ Social Learning Approach: Theory and Application
John B. Watson (1878-1958) • Watson was a fighter and a builder. • In college, he was unsociable and uninterested in his studies. • He switched from philosophy to psychology at the University of Chicago, but preferred to study animals rather than humans. • After joining the faculty at Johns Hopkins University in 1908, he developed and promoted his ideas about behaviorism. • Behaviorism swept the field following the publication of his paper “Psychology as the Behaviorist Views It” in 1913. • Watson’s flourishing academic career ended in disgrace in 1920, in the wake of his affair with his research associate, Rosalie Rayner.
Watson’s view of behaviorism • Watson argued that if psychology were to be a science, psychologists must stop examining mental states and study overt, observable behavior instead. • Emotions, thoughts, experiences, values, reasoning, insight, and the unconscious would therefore be off-limits to behaviorists unless they could be defined in terms of observable behaviors. For example, Watson regarded thinking as a variant of verbal behavior that he called “subvocal speech.” • According to Watson, personality could be described as “the end product of our habit systems.” • Watson argued that he could take “a dozen healthy infants, well-formed” and condition them “to become any type of specialist” that he chose.
B.F. Skinner (1904-1990) • Skinner passed on joining his father’s law firm and studied English at Hamilton College with the intent of becoming a writer. • After producing nothing of consequence in the two years following his graduation, he went to Harvard to study psychology. • He became the new standard bearer for a view of behaviorism that he called radical behaviorism. • In the 1940s, he published Walden Two, a novel about a utopian community based on conditioning principles. • To the end, he remaining an adamant believer in the power of the environment, and conceded little to those who emphasized genetic determinants of behavior.
Skinner’s “radical behaviorism” • Skinner did not deny the existence of thoughts and inner experiences, but he argued that we often misattribute our actions to mental states when they should be attributed to our conditioning instead. • His position maintained that we often don’t know the reason for many of our behaviors, although we may think we do. • He therefore argued that our perception that we are free to act as we choose is, to a large extent, an illusion. Instead, we act in response to environmental contingencies. • Among Skinner’s many contributions to the study of conditioning are the so-called “Skinner box” and the discovery of partial reinforcement schedules.
Ivan Pavlov (1849-1936) • Ivan Pavlov was born in Ryazan, Russia. He began his higher education as a seminary student, but dropped out and enrolled at the University of Petersburg to study the natural sciences. He received his doctorate in 1879. • In the 1890s, Pavlov was investigating the digestive process in dogs by externalizing a salivary gland so he could collect, measure, and analyze the saliva produced in response to food under different conditions. • He noticed that the dogs tended to salivate before food was actually delivered to their mouths. He realized that this was more interesting than the chemistry of saliva, and changed the focus of his research, carrying out a long series of experiments in which he manipulated the stimuli occurring before the presentation of food. • He thereby established the basic laws for the establishment and extinction of what he called "conditional reflexes" — i.e., reflex responses, like salivation, that only occurred conditional upon specific previous experiences of the animal.
Classical (Pavlovian) conditioning (aka “signal learning”) • Classical conditioning begins with an existing stimulus-response (S-R) association. • Pavlov, a Russian physiologist who was studying the process of digestion in dogs, was astute enough to recognize the importance of the fact that the dogs learned to salivate even before they were given the food. • Understanding that there was an existing S-R association between the food (S) and the dog’s salivation, he quickly perceived that there might also be a learned or “conditioned” association between cues associated with feeding (S) and the dog’s salivation (R). • Using the sound of either a bell or a tuning fork as his conditioned stimuli, he found that he could indeed “condition” the response of salivation to the sound of a bell or a tuning fork.
Pavlov’s famous demonstration of classically conditioned salivation in a dog
Classical conditioning paradigm Unconditionedstimulus (UCS) Unconditioned response (UCR) Conditioned stimulus (CS) Conditioned response (CR)
Classical (Pavlovian) conditioning (aka “signal learning”) • Once the new S-R association is established, it can be used to condition yet another S-R association in a process called second-order conditioning. • For example, once the dog is reliably salivating to the sound of the bell, the bell can be paired with a green light and soon the dog will salivate whenever the green light comes on. • Both first-order and second-order classical conditioning are subject to extinction. • Other limitations of classical conditioning involve the length of the delay between the presentation of the CS and the presentation of the UCS (in general, one second is optimal) and the fact that some associations simply cannot be established through classical conditioning.
Operant (instrumental) conditioning (aka “consequence learning”) • Operant conditioning concerns the effect certain kinds of consequences have on the frequency of behavior. • A consequence that increases the frequency of a behavior is called a reinforcement. • A consequence that decreases the frequency of a behavior is called a punishment. • Whether a consequence is reinforcing or punishing varies according to the person and the situation. • There are two basic strategies for increasing the frequency of a behavior: positive reinforcement and negative reinforcement. • There are two basic strategies for decreasing the frequency of a behavior: extinction and punishment.
Edward L. Thorndike (1874-1949) • Edward L. Thorndike graduated from Wesleyan University in 1895, and received his Ph.D. from Columbia University in 1898 • He was appointed as an instructor in genetic psychology at Teachers College, Columbia, in 1899, he served there until 1940. • He devised methods to measure children’s intelligence and their ability to learn. He also conducted studies in animal psychology and the psychology of learning. His law of effect addressed the phenomenon we now call reinforcement. • Thorndike’s books include Educational Psychology (1903), Mental and Social Measurements (1904), Animal Intelligence (1911), A Teacher’s Word Book (1921), Your City (1939), and Human Nature and the Social Order (1940).
Operant (instrumental) conditioning (aka “consequence learning”) • Operant conditioning concerns the effect certain kinds of consequences have on the frequency of behavior. • A consequence that increases the frequency of a behavior is called a reinforcement. • A consequence that decreases the frequency of a behavior is called a punishment. • Whether a consequence is reinforcing or punishing varies according to the person and the situation. • There are two basic strategies for increasing the frequency of a behavior: positive reinforcement and negative reinforcement. • There are two basic strategies for decreasing the frequency of a behavior: extinction and punishment.
Problems with the use of punishment • Punishment does not teach what behavior is appropriate. It only teaches what behavior is inappropriate. • To be effective, punishment must be delivered immediately and consistently. • Punishment can have the negative side effect of inhibiting not only the undesirable behavior but also desirable behavior that is associated with it. • Punishment can result in the person who is punished coming to fear the person who administers the punishment. • Punishment may also serve as a behavior that is later modeled by the person being punished. • Punishment can create strong negative emotions that can interfere with learning the desired response. • For all of these reasons, punishment should be used sparingly and only when other operant conditioning procedures either cannot be used or will not work.
Other important operant conditioning concepts • Shaping: reinforcing successive approximations of the desired behavior until the complete response is well established • Generalization: displaying the response to stimulus situations that resemble the one in which the original response was acquired • Discrimination: selectively reinforcing the response to help ensure that it will only occur in the presence of the original stimulus and not ones that might resemble it
Julian B. Rotter (1916- ) • Julian Rotter first learned about psychology in the Avenue J Library in Brooklyn, where he spent much of his childhood and adolescence. • Believing that he couldn’t earn a living as a psychologist, he majored in chemistry at Brooklyn College. • While still in college, he discovered that Alfred Adler was teaching at the Long Island School of Medicine and began associating with Adler and his colleagues. • He then became a psychology major at the University of Iowa and received a Ph.D. in clinical psychology at the University of Indiana. • After serving as a psychologist in the Army during World War II, he later taught at the Ohio State University and at the University of Connecticut.
Key concepts in Rotter’s Social Learning Theory: perceptions, expectancies, and values • Behavior potential (BP): the likelihood of a given behavior occurring in a particular situation • Expectancy: the perceived likelihood that a given behavior will result in a particular outcome • Generalized expectancies: beliefs about how often our actions typically lead to reinforcements and punishments • Locus of control: generalized perceptions about the degree to which one’s outcomes are determined by internal versus external factors • Reinforcement value: the degree to which we prefer one reinforcer over another
Rotter’s basic formula for predicting behavior Behavior potential (BP) = Expectancy (E) X Reinforcement Value (RV) If either the expectancy or the reinforcement value is zero, then the behavior potential will be zero.
Examples of calculating the behavior potentials in an insult situation
Albert Bandura (1925- ) • Albert Bandura was raised in Alberta, Canada and received his bachelor’s degree at the University of British Columbia. • He completed his Ph.D. at the University of Iowa, where he was influenced by the prominent learning theorist Kenneth Spence. • After a year of clinical internship in Wichita, he accepted a position at Stanford University. • His career has been spent building bridges between traditional learning theory, cognitive personality theory, and clinical psychology. • Over the years, Bandura has received many professional honors and awards, and he continues to be active in his 80s.
Bandura’s reciprocal determinism model Behavior External factors (Rewards, punishments) Internal factors (Beliefs, thoughts, expectations)
Features of learning and cognition that are (relatively) unique to humans • The use of symbols and other cognitive representational structures to “re-create” the outside world within our own minds • The resulting abilities to imagine alternative courses of action and conduct mental stimulations to “project” what their outcomes are likely to be • The capacity for self-regulation through the application of self-reward and self-punishment, even in the face of strong external rewards and punishment • The capacity for vicarious or observational learning
Observational learning • The distinction between learning and performance is important: not every behavior is learned gets performed • Much of our learning occurs vicariously, through our observation of other people’s actions, and the consequences of those actions. • We are more likely to imitate a behavior we have seen other people display if the outcome of their behavior was a reward, rather than a punishment. • In a study by Bandura (1965), nursery school children were more likely to model the aggressive behaviors of an adult who they observed in a film segment if they saw the model get rewarded, rather than punished, for his aggressive behavior.
Mean number of aggressive responses performed (Bandura, 1965)
Application: behavior modification • Classical conditioning applications • Systematic desensitization • Aversion therapy • Operant conditioning applications • Changing behavior by changing contingencies • Reward • Punishment • Extinction through nonreinforcement • Token economy • Biofeedback
Diagram of Little Albert’s classical conditioning Loud noise (Unconditioned stimulus) Fear responses (Unconditioned response) White rat (Conditioned stimulus) Fear response (Conditioned response)
Application: behavior modification • Classical conditioning applications • Systematic desensitization • Aversion therapy • Operant conditioning applications • Changing behavior by changing contingencies • Reward • Punishment • Extinction through nonreinforcement • Token economy • Biofeedback
Application: self-efficacy therapy • The difference between an outcome expectancy and an efficacy expectancy. • Four sources of efficacy expectancies: • Enactive mastery experiences • Vicarious experiences • Verbal persuasion (“coaching”) • Physiological and affective states • Guided mastery as a step-by-step approach to achieve enactive mastery experiences • Problems that have been addressed through the application of self-efficacy beliefs include traumatic stress disorder, test anxiety, phobias, and bereavement.
Behavioral observation methods • Direct observation • Direct observation in the field, in the laboratory, or in the clinic • Analogue behavioral observation (e.g., staging a dance for clients who are being treated for shyness) • Role-play • The importance of reliable observations • Self-monitoring • The importance of consistent and objective self-observation • The therapeutic value of self-observation • Observation by others (parents, teachers, nurses, etc.)
Strengths and criticisms of the behavioral / social learning approach • Strengths • The approach has a solid foundation in empirical research with humans and with infrahuman species. • It led to the development of useful therapeutic procedures involving behavior modification. These intervention procedures, which are relatively quick, inexpensive, and easily administered, assess baseline levels of behavior and establish objective criteria for behavior change. • The social-cognitive extensions of the behavioral approach have greatly expanded the range of phenomena that can be addressed. • Criticisms • The approach gives inadequate attention to the role of heredity. • Not all responses can be successfully conditioned (e.g., fear of food). • Rewarding intrinsically-motivated behaviors can sometimes reduce their frequency of occurrence. • Reducing the problems of therapy patients to observable behaviors may fail to address the underlying problem in certain kinds of disorders.