Chapter 7: Learning
LOQ: How do we define learning, and what are some basic forms of learning?
We adapt to our environments by learning
learn to expect and prepare for significant events such as food or pain
classical conditioning
learn to repeat acts that bring rewards and avoid acts that bring unwanted result
operant conditioning
learn new behaviors by observing events and people, and through language and things we have neither experienced nor observed
cognitive learning
Learned associations also feed our habitual behaviors
Habits form when we repeat behaviors in a given context
As behavior becomes linked with the context, our next experience of that context will evoke our habitual response
Takes around 66 days of doing a task to make it habitual
Complex animals can learn to associate their own behavior with its outcomes
Linking two events closely together causes animals to exhibit associative learning
The animals have learned something important to its survival: anticipating the immediate future
This process of learning associations is called conditioning
There are two main forms:
Classical conditioning, we learn to associate two stimuli and thus to anticipate events. (A stimulus is any event or situation that evokes a response). We associate stimuli that we do not control, and we respond automatically (exhibiting respondent behavior)
Operant conditioning, we learn to associate a response (our behavior) and its consequence. Thus we (and other animals) learn to repeat acts followed by good results and avoid acts followed by bad results. These associations produce operant behaviors (which operate on the environment to produce a consequence)
Cognitive learning is another form of learning
we acquire mental information that guides our behavior
We learn from others’ experiences through observational learning (one form of cognitive learning)
Learning: the process of acquiring through experience new and relatively enduring information or behaviors.
Associative Learning: learning that certain events occur together. The events may be two stimuli (as in classical conditioning) or a response and its consequence (as in operant conditioning).
Stimulus: any event or situation that evokes a response.
Respondent Behavior: behavior that occurs as an automatic response to some stimulus.
Operant Behavior: behavior that operates on the environment, producing consequences
Cognitive Learning: the acquisition of mental information, whether by observing events, by watching others or through language.
LOQ: What was behaviorism’s view of learning?
Ivan Pavlov conducted famous experiments in the twentieth-century
Explored a phenomenon we call classical conditioning
His work became a foundation for many of John B. Watson’s ideas
Watson wanted to discard reference to inner thoughts, feelings, and motives
He said that the science of psychology should instead study how organisms respond to stimuli in their environments
psychology should be an objective science based on observable behavior
Called this study behaviorism
Behaviorism influenced North American psychology during the first half of the twentieth century
Pavlov and Watson shared both disdain for “mentalistic” concepts (ex. consciousness) and believed that the basic laws of learning were the same for all animals no matter the species
Few researchers today agree that psychology should ignore mental processes
Most researchers agree that classical conditioning is a basic form of learning that all organisms use to adapt to their environment
Behaviorism: the view that psychology (1) should be an objective science that (2) studies behavior without reference to mental processes. Most research psychologists today agree with (1) but not with (2).
Pavlov’s Experiments
LOQ: Who was Pavlov, and what are the basic components of classical conditioning?
Palvov started an experiment from seeing that seeing food, the food dish, or the person serving the food, caused a dog to instinatiously salivate
Pavlov considered these “psychic secretions” an annoyance
realized they pointed to a simple but fundamental form of learning
Pavlov and his assistants tried to think what the dog was thinking and feeling that it caused it to drool
Experimented by isolating the dog in a small room secured to a harness and attached to a device that transported the dogs slaiva into a measuring device
Presented the food in various stages to the dog
First, they slide a food bowl in
Then blew meat powder in the dog’s mouth at a specific moment
They then paired a variety of neutral stimuli (NS) events that the dog could see or hear but that weren’t associated with food
The dog did not learn to salivate in response to food in its mouth. Rather, food in the mouth automatically, unconditionally, triggers a dog’s salivary reflex
Pavlov called this drooling and the unconditioned response (UR)
Called the food and unconditioned stimulus (US)
Salvation in response to the tone is learned and is called conditioned response (CR)
The stimulus that was used to be neutral (the tone in this case) is the conditioned stimulus (CS)
After his demonstration of associative learning he explored five major conditioning processes with his associates:
Acquisition
Extinction
spontaneous recovery
generalization
discrimination
Neutral Stimulus (NS): In classical conditioning, a stimulus that elicits no
response before conditioning.
Unconditioned Response (UR): in classical conditioning, an unlearned, naturally occurring response (such as salivation) to an unconditioned stimulus (US) (such as food in the mouth).
Unconditioned Stimulus (US): in classical conditioning, a stimulus that unconditionally—naturally and automatically—triggers an unconditioned response (UR).
Conditioned Response (CR): in classical conditioning, a learned response to a previously neutral (but now conditioned) stimulus (CS).
Conditioned Stimulus (CS): in classical conditioning, an originally neutral stimulus that, after association with an unconditioned stimulus (US), comes to trigger a conditioned response (CR)
LOQ: In classical conditioning, what are the processes of acquisition, extinction, spontaneous recovery, generalization, and discrimination?
Pavolv tried to understand the acqusition, intial learning, by asking “How much time should elapse between presenting the NS (the tone, the light, the touch) and the US (the food)?”
This turned out to not to be super long- usually around 0.5 seconds
Conditioning usually won’t occur when the NS follows the US
classical conditioning is biologically adaptive
helps humans and other animals prepare for good or bad events
A new NS can become a new CS without a US through higher-order conditioning
Only needs to become associated with a previously conditioned stimulus
Higher-order conditioning tends to be weaker than first-order conditioning
also called second-order conditioning
Higher-Order Conditioning: a procedure in which the conditioned stimulus in one conditioning experience is paired with a new neutral stimulus, creating a second (often weaker) conditioned stimulus. For example, an animal that has learned that a tone predicts food might then learn that a light predicts the tone and begin responding to the light alone. (Also called second-order conditioning.)
Extinction and Spontaneous Recovery
If the tone was continued to play and there was no food brought to the dogs, then the dogs start to drool less and less and it continues
This reaction is known as extinction
But if the time between the tones sounding, the dogs would begin slaivating to the tone
This is called spontaneous recovery
Extinction: the diminishing of a conditioned response; occurs in classical conditioning when an unconditioned stimulus (US) does not follow a conditioned stimulus (CS); occurs in operant conditioning when a response is no longer reinforced.
Spontaneous Recovery: the reappearance, after a pause, of an extinguished conditioned response.
Pavlov and his students noticed that if a dog was conditioned to one tone, they would also somewhat respond to a new and different tone
If a dog was conditioned to salivate when rubbed would also drool a bit when scratched
This response is called generalization (or stimulus generalization)
Generalization can be adaptive
Gerealized fear can also linger
Stimuli similar to naturally disgusting objects will, by association, also evoke some disgust
Generalization: the tendency, once a response has been conditioned, for stimuli similar to the conditioned stimulus to elicit similar responses. (In operant conditioning, generalization occurs when responses learned in one situation occur in other, similar situations.)
Pavlov’s dogs also learned to respond to the sound of a particular tone and not to other tones
They learned the ability to distinguish between a conditioned stimulus (which predicts the US) and other, irrelevant stimuli (discrimination)
Slightly different stimuli can be followed by vastly different consequences
Discrimination: in classical conditioning, the learned ability to distinguish between a conditioned stimulus and similar stimuli that do not signal an unconditioned stimulus. (In operant conditioning, the ability to distinguish responses that are reinforced from similar responses that are not reinforced.)
LOQ: Why does Pavlov’s work remain so important?
Most psychologists now agree that classical conditioning is a basic form of learning
We now see that Pavlov’s ideas were incomplete with new knowledge and research
Pavolv’s work has remained important
He found that many other responses to many other stimuli can be classically conditioned in many other organisms
This showed that classical conditioning is one way that virtually all organisms learn to adapt to their environment
showed us how a process such as learning can be studied objectively
His meathods involved virtually no subjective judgments or guesses about what went on in a dog’s mind
LOQ: What have been some applications of Pavlov’s work to human health and well-being? How did Watson apply Pavlov’s principles to learned fears?
Pavlov’s principles can influence human health and well-being in three ways:
Drug cravings: Former drug users often feel a craving when they are again in the drug-using context—with people or in places they associate with previous highs. Thus, drug counselors advise their clients to steer clear of people and settings that may trigger these cravings
Food cravings: Classical conditioning makes dieting difficult. We readily associate sugary substances with an enjoyable sweet sensation. Researchers have conditioned healthy volunteers to experience cravings after only one instance of eating a sweet food
Eating one cookie can create hunger for another. People who struggle with their weight often have eaten unhealthy foods thousands of times, leaving them with strongly conditioned responses to eat the very foods that will keep them in poor health
Immune responses: Classical conditioning even works on the body’s disease-fighting immune system. When a particular taste accompanies a drug that influences immune responses, the taste by itself may come to produce an immune response
LOQ: What is operant conditioning?
Classical conditioning and operant conditioning are both forms of associative learning but are both very different
Classical Conditioning: forms associations between stimuli (a CS and the US it signals). It also involves respondent behavior—automatic responses to a stimulus (such as salivating in response to meat powder and later in response to a tone)
Operant Conditioning: organisms associate their own actions with consequences. Actions followed by reinforcers increase; those followed by punishments often decrease. Behavior that operates on the environment to produce rewarding or punishing stimuli is called operant behavior.
Operant Conditioning: a type of learning in which a behavior becomes more likely to recur if followed by a reinforcer or less likely to recur if followed by a punisher.
LOQ: Who was Skinner, and how is operant behavior reinforced and shaped?
B. F. Skinner (1904–1990) was a college English major and aspiring writer
enrolled as a graduate student in psychology
Eventually became modern behaviorism’s most influential and controversial figure
His work elaborated on t Edward L. Thorndike’s “law of effect”
Developed behavioral technology that revealed principles of behavior control
Skinner’s pioneering studies
designed an operant chamber
Also known as a Skinner box
The box has a bar (a lever) that an animal presses—or a key (a disc) the animal pecks
This releases a reward of food or water.
also has a device that records these responses
This creates a stage on which rats and other animals act out Skinner’s concept of reinforcement
Law of Effect: Thorndike’s principle that behaviors followed by favorable consequences become more likely, and that behaviors followed by unfavorable consequences become less likely
Operant Chamber: in operant conditioning research, a chamber (also known as a Skinner box) containing a bar or key that an animal can manipulate to obtain a food or water reinforcer; attached devices record the animal’s rate of bar pressing or key pecking.
Reinforcement: in operant conditioning, any event that strengthens the behavior it follows.
Skinner would tease rats to press a bar through shaping
Did this by gradually guiding the rat’s actions toward the desired behavior
You start by watching how the animal naturally behaves, so that you could build on its existing behavior
might give the rat a bit of food each time it approaches the bar
Then once the rat is approaching regularly, you would give the food only when it moves close to the bar, then closer still
Finally you would require it to touch the bar to get food
This rewards successive approximations by reinforcing responses towards your disired outcome
Shaping can also help us understand what nonverbal organisms perceive
Skinner noted that we continually reinforce and shape others’ everyday behaviors, even though we may not mean to do so
Shaping: an operant conditioning procedure in which reinforcers guide behavior toward closer and closer approximations of the desired behavior
LOQ: How do positive and negative reinforcement differ, and what are the basic types of reinforcers?
Positive reinforcement strengthens the response by presenting a typically pleasurable stimulus immediately after a response
Negative reinforcement strengthens a response by reducing or removing something negative
negative reinforcement is not punishment
negative reinforcement removes a punishing (aversive) even
Positive Reinforcement: increasing behaviors by presenting positive reinforcers. A positive reinforcer is any stimulus that, when presented after a response, strengthens the response. negative reinforcement
Negative Reinforcement: increasing behaviors by stopping or reducing negative stimuli. A negative reinforcer is any stimulus that, when removed after a response, strengthens the response. (Note: Negative reinforcement is not punishment.)
Primary reinforcers are unlearned
Ex. getting food when hungry
Conditioned reinforcers are learned through association with primary reinforcers
Also known as secondary reinforcers
Ex. a rat in a Skinner box learns that a light reliably signals a food delivery, the rat will work to turn on the light
Primary Reinforcer: an innately reinforcing stimulus, such as one that satisfies a biological need.
Conditioned Reinforcer: a stimulus that gains its reinforcing power through its association with a primary reinforcer; also known as a secondary reinforcer.
Immediate and Delayed Reinforcers
Humans respond to delayed reinforcers
Ex. paycheck at the end of the week, good grade at the end of a term, or a trophy after a sports season
We must learn to delay satisfaction to function effectively
Learning to control our impulses in order to achieve more valued rewards is a big step toward maturity
can later protect us from committing an impulsive crime
Children who delay gratification have tended to become socially competent and high-achieving adults
LOQ: How do different reinforcement schedules affect behavior?
Reinforcement schedules vary
Continous reinforcement learning occours rapidly
Makes it the best choice for mastering a behavior
But extinction also occurs rapidly
When reinforcement stops, the behavior soon stops (is extinguished)
Real-life e rarely provides continuous reinforcement.
Ex. salespeople do not make a sale with every pitch but they are persistent
This persistence is called partial (intermittent) reinforcement schedules
Responses are sometimes reinforced but sometimes not
Learning is slower to appear
Resistance to extinction is greater than continuous reinforcement
Fixed-ratio schedules reinforce behavior after a set number of responses.
Ex. a coffee shop gives you a free cup of coffee after 10 drinks are purchased
This causes continued animals to only pause briefly after a reinforcer before returning to a high rate of responding
Variable-ratio schedules provide reinforcers after a seemingly unpredictable number of responses
unpredictable reinforcement is what makes gambling and fly fishing so hard to extinguish even when they don’t produce the desired results
Because reinforcers increase as the number of responses increases, variable-ratio schedules produce high rates of responding
Fixed-interval schedules reinforce the first response after a fixed time period
Animals on this type of schedule tend to respond more frequently as the anticipated time for reward draws near
produces a choppy stop-start pattern rather than a steady rate of response
Ex. checking the cake more often as it cooks longer
Variable-interval schedules reinforce the first response after varying time intervals
unpredictable times, a food pellet rewarded Skinner’s pigeons for persistence in pecking a key
Ex. a longed-for message that finally rewards persistence in checking our phone
variable-interval schedules tend to produce slow, steady responding
there is no knowing when the waiting will be over
Overall, response rates are higher when reinforcement is linked to the number of responses (a ratio schedule) rather than to time (an interval schedule)
responding is more consistent when reinforcement is unpredictable (a variable schedule) than when it is predictable (a fixed schedule)
Reinforcement Schedule: a pattern that defines how often the desired response
will be reinforced.
Continuous Reinforcement: schedule reinforcing the desired response every
time it occurs.
Partial (Intermittent) Reinforcement Schedule Reinforcing: a response only
part of the time; results in the slower acquisition of a response but much greater
resistance to extinction than does continuous reinforcement.
Fixed-Ratio Schedule: in operant conditioning, a reinforcement schedule that reinforces a response only after a specified number of responses
Variable-Ratio Schedule: in operant conditioning, a reinforcement schedule that reinforces a response after an unpredictable number of responses.
Fixed-Interval Schedule: in operant conditioning, a reinforcement schedule that reinforces a response only after a specified time has elapsed.
Variable-Interval Schedule: in operant conditioning, a reinforcement schedule
that reinforces a response at unpredictable time intervals.
LOQ: How does punishment differ from negative reinforcement, and how does punishment affect behavior?
Reinforcement increases a behavior; punishment does the opposite
negative reinforcement increases the frequency of a preceding behavior
Withdraws something negative
a punisher is any consequence that decreases the frequency of a preceding behavior
Ex. a child gets burned after touching a hot stove
Criminal behavior is influenced more by swift and sure punishers than by the threat of severe sentences
Most of this behavior is impulsive
Many psychologists note four major drawbacks of physical punishment:
Punished behavior is suppressed, not forgotten. This temporary state may (negatively) reinforce parents’ punishing behavior.
child swears, the parent swats, the parent hears no more swearing and feels the punishment successfully stopped the behavior
Punishment teaches discrimination among situations.
In operant conditioning, discrimination occurs when an organism learns that certain responses, but not others, will be reinforced. Did the punishment effectively end the child’s swearing? Or did the child simply learn that while it’s not okay to swear around the house, it’s okay elsewhere?
Punishment can teach fear
A punished child may associate fear not only with the undesirable behavior but also with the person who delivered the punishment or where it occurred
may learn to fear a punishing teacher and try to avoid school, or may become more anxious
Physical punishment may increase aggression by modeling violence as a way to cope with problems.
Many aggressive delinquents and abusive parents come from abusive families
Studies find that spanked children are at increased risk for aggression
There is a debate to if physical punishmed chuldren may be more aggressive
It follows the same logic of people who have undergone psychotherapy are more likely to suffer depression since they had preexisting problems that triggered the treatments
Punishment: an event that tends to decrease the behavior that it follows.
LOQ: Why did Skinner’s ideas provoke controversy, and how might his operant conditioning principles be applied at school, in sports, at work, in parenting, and for self-improvement?
Skinner said argued that brain science isn’t needed for psychological science
Said “a science of behavior is independent of neurology”
urged people to use operant conditioning principles to influence others’ behavior at school, work, and home
argued that we should use rewards to evoke more desirable behavior
His critics objected and said that dehumanizing people by neglecting their personal freedom and by seeking to control their actions
Skinner’s response was that External consequences already haphazardly control people’s behavior
At School
Skinner said “Good instruction demands two things, students must be toldimmediately whether what they do is right or wrong and, when right, they must be directed to the step to be taken next.”
Many of his ideals over education are now possible
Students can now learn at their own pace with some programs
In Sports
Reinforcing small successes and then gradually increasing the challenge is the key to shaping behavior in athletic performance
Compared with children taught by conventional methods, those trained by this behavioral method have shown faster skill improvement
At Work
Organizations have invited employees to share the risks and rewards of company ownership
Others focus on reinforcing a job well done
Rewards are most likely to increase productivity if the desired performance is both well defined and achievable
Reward specific, achievable behaviors, not vaguely defined “merit”
Operant conditioning reminds us that reinforcement should be immediate.
In Parenting
parents can learn from operant conditioning practices
Parent-training researchers explain that, “Get ready for bed” and then caving into protests or defiance, parents reinforce such whining and arguing
Notice people doing something right and affirm them for it.
This disrupts the cycle
Give children attention and other reinforcers when they are behaving well. Target a specific behavior, reward it, and watch it increase
To Change Your Own Behaviors
State a realistic goal in measurable terms and announce it.
Decide how, when, and where you will work toward your goal.
Monitor how often you engage in your desired behavior
Reinforce the desired behavior.
Reduce the rewards gradually
LOQ: How does operant conditioning differ from classical conditioning?
Our biology and cognitive process influcene both classical and operant conditioning.
LOQ: How do biological constraints affect classical and operant conditioning?
Biological Limits on Classical Conditioning
Early behaviorists realized, an animal’s capacity for conditioning is limited by biological constraints
Ex. individual species’ predispositions prepare it to learn the associations that enhance its survival
This is a phenomenon called preparedness
Humans are also biologically prepared to learn some associations other than others
Ex. you develop aversions to that taste of oysters after becoming ill from eating them rather than the sight associated with oysters such as restaurants
Arthur Schopenhaue said that important ideas are first ridiculed, then attacked, and finally taken for granted
Our preparedness to associate a CS with a US that follows predictably and immediately is adaptive
Causes often do immediately precede effects
Ex. When chemotherapy triggers nausea and vomiting more than an hour following treatment, cancer patients may over time develop classically conditioned nausea (and sometimes anxiety) to the sights, sounds, and smells associated with the clinic
In normal circumstances, such revulsion to sickening stimuli would be adaptive.
Preparedness: a biological predisposition to learn associations, such as between taste and nausea, that have survival value.
Biological Limits on Operant Conditioning
Nature also constrains each species’ capacity for operant conditioning
We most easily learn and retain behaviors that reflect our biological predispositions.
This principle of this is: “Biological constraints predispose organisms to learn associations that are naturally adaptive.”
Instinctive Drift: the tendency of learned behavior to gradually revert to biologically predisposed patterns.
LOQ: How do cognitive processes affect classical and operant conditioning?
Cognition and Classical Conditioning
Early behaviorists believed that rats’ and dogs’ learned behaviors could be reduced to mindless mechanisms, so there was no need to consider cognition
Robert Rescorla and Allan Wagner showed that an animal can learn the predictability of an event.
If a shock always is preceded by a tone, and then may also be preceded by a light that accompanies the tone
Although the light is always followed by the shock, it adds no new information; the tone is a better predictor
more predictable the association, the stronger the conditioned response
the animal learns an expectancy, an awareness of how likely it is that the US will occur.
Classical conditioning with humans is not just CS-US association
Cognition and Operant Conditioning
Skinner believed that thoughts and emotions were behaviors that follow the same laws as other behaviors
growing belief that cognitive processes have a necessary place in the science of psychology and even in our understanding of conditioning
viewed “cognitive science
Animals on a fixed-interval reinforcement schedule respond more and more frequently as the time approaches when a response will produce a reinforcer
strict behaviorist would object to talk of “expectations,” the animals behave as if they expected that repeating the response would soon produce the reward
There is more to learning than associating a response with a consequence; there is also cognition
Promising people a reward for a task they already enjoy can backfire.
Excessive rewards can destroy intrinsic motivation
the desire to perform a behavior effectively and for its own sake
Cognitive Map: a mental representation of the layout of one’s environment. For example, after exploring a maze, rats act as if they have learned a cognitive map of it.
Latent Learning: learning that occurs but is not apparent until there is an incentive to demonstrate it.
Intrinsic Motivation: a desire to perform a behavior effectively for its own sake.
Extrinsic Motivation: a desire to perform a behavior to receive promised rewards or avoid threatened punishment.
LOQ: How does observational learning differ from associative learning? How may observational learning be enabled by neural mirroring?
Cognition supports observational learning
Learning without direct experience
Ex. watching and intermittent
This process is called modeling
Observational Learning: learning by observing others.
Modeling: the process of observing and imitating a specific behavior.
Vicarious reinforcement/ punishment
we learn to anticipate a behavior’s consequences
Mirrors and Imitation in the Brain
In humans, imitation is pervasive
catchphrases, fashions, ceremonies, foods, traditions, morals, and fads all spread by one person copying another
Even children and infants are natural imitators
children’s brains enable their empathy and their ability to infer another’s mental state, an ability known as theory of mind
Humans have brains that support empathy and imitation
Researchers can use fMRI scans to see brain activity associated with performing and with observing action
Our brain’s response to observing others makes emotions contagious
The brain simulates and vicariously experiences what we observe
Through reenactments of these we grasp others’ states of mind, postures, faces, voices, and writing styles,
By observing we unconsciously synchronize our own to theirs—which helps us feel what they are feeling
Imitation helps us gain friends
We mimic people we like
We yawn when other people yawn, we laugh when other people laugh
Pain imagined by an empathic romantic partner has triggered some of the same brain activity experienced by the loved one actually having the pain
we mentally simulatet the experiences described
Mirror Neurons: frontal lobe neurons that some scientists believe fire when we perform certain actions or observe another doing so. The brain’s mirroring of another’s action may enable imitation and empathy.
LOQ: What is the impact of prosocial modeling and of antisocial modeling?
Prosocial Effects
People’s modeling of prosocial (positive, helpful) behaviors can have prosocial effect
businesses use behavior modeling to help new employees learn communications, sales, and customer service skills
People who show nonviolent and helpful behavior can also prompt similar behavior
Parents are also powerful models
observational learning of morality begins early
Socially responsive toddlers who imitate their parents usually become preschoolers with a strong internalized conscience
Models are most effective when their actions and words are consistent
To encourage children to read, read to them and surround them with books and people who read.
Prosocial Behavior: positive, constructive, helpful behavior. The opposite of antisocial behavior.
Antisocial Behavior
Observational learning may also have antisocial effects
helps us know why abusive parents might have aggressive children
Critics note that such aggressiveness could be genetic
The lessons we learn as children are not easily replaced as adults, and they are sometimes visited on future generations.
Intellectual history is often made by people who risk going to extremes in pushing ideas to their limits. Researchers such as Ivan Pavlov, John Watson, B. F. Skinner, and Albert Bandura defined the issues and impressed on us the importance of learning. They illustrated the impact that can result from single-minded devotion to a few well-defined problems.
LOQ: How do we define learning, and what are some basic forms of learning?
We adapt to our environments by learning
learn to expect and prepare for significant events such as food or pain
classical conditioning
learn to repeat acts that bring rewards and avoid acts that bring unwanted result
operant conditioning
learn new behaviors by observing events and people, and through language and things we have neither experienced nor observed
cognitive learning
Learned associations also feed our habitual behaviors
Habits form when we repeat behaviors in a given context
As behavior becomes linked with the context, our next experience of that context will evoke our habitual response
Takes around 66 days of doing a task to make it habitual
Complex animals can learn to associate their own behavior with its outcomes
Linking two events closely together causes animals to exhibit associative learning
The animals have learned something important to its survival: anticipating the immediate future
This process of learning associations is called conditioning
There are two main forms:
Classical conditioning, we learn to associate two stimuli and thus to anticipate events. (A stimulus is any event or situation that evokes a response). We associate stimuli that we do not control, and we respond automatically (exhibiting respondent behavior)
Operant conditioning, we learn to associate a response (our behavior) and its consequence. Thus we (and other animals) learn to repeat acts followed by good results and avoid acts followed by bad results. These associations produce operant behaviors (which operate on the environment to produce a consequence)
Cognitive learning is another form of learning
we acquire mental information that guides our behavior
We learn from others’ experiences through observational learning (one form of cognitive learning)
Learning: the process of acquiring through experience new and relatively enduring information or behaviors.
Associative Learning: learning that certain events occur together. The events may be two stimuli (as in classical conditioning) or a response and its consequence (as in operant conditioning).
Stimulus: any event or situation that evokes a response.
Respondent Behavior: behavior that occurs as an automatic response to some stimulus.
Operant Behavior: behavior that operates on the environment, producing consequences
Cognitive Learning: the acquisition of mental information, whether by observing events, by watching others or through language.
LOQ: What was behaviorism’s view of learning?
Ivan Pavlov conducted famous experiments in the twentieth-century
Explored a phenomenon we call classical conditioning
His work became a foundation for many of John B. Watson’s ideas
Watson wanted to discard reference to inner thoughts, feelings, and motives
He said that the science of psychology should instead study how organisms respond to stimuli in their environments
psychology should be an objective science based on observable behavior
Called this study behaviorism
Behaviorism influenced North American psychology during the first half of the twentieth century
Pavlov and Watson shared both disdain for “mentalistic” concepts (ex. consciousness) and believed that the basic laws of learning were the same for all animals no matter the species
Few researchers today agree that psychology should ignore mental processes
Most researchers agree that classical conditioning is a basic form of learning that all organisms use to adapt to their environment
Behaviorism: the view that psychology (1) should be an objective science that (2) studies behavior without reference to mental processes. Most research psychologists today agree with (1) but not with (2).
Pavlov’s Experiments
LOQ: Who was Pavlov, and what are the basic components of classical conditioning?
Palvov started an experiment from seeing that seeing food, the food dish, or the person serving the food, caused a dog to instinatiously salivate
Pavlov considered these “psychic secretions” an annoyance
realized they pointed to a simple but fundamental form of learning
Pavlov and his assistants tried to think what the dog was thinking and feeling that it caused it to drool
Experimented by isolating the dog in a small room secured to a harness and attached to a device that transported the dogs slaiva into a measuring device
Presented the food in various stages to the dog
First, they slide a food bowl in
Then blew meat powder in the dog’s mouth at a specific moment
They then paired a variety of neutral stimuli (NS) events that the dog could see or hear but that weren’t associated with food
The dog did not learn to salivate in response to food in its mouth. Rather, food in the mouth automatically, unconditionally, triggers a dog’s salivary reflex
Pavlov called this drooling and the unconditioned response (UR)
Called the food and unconditioned stimulus (US)
Salvation in response to the tone is learned and is called conditioned response (CR)
The stimulus that was used to be neutral (the tone in this case) is the conditioned stimulus (CS)
After his demonstration of associative learning he explored five major conditioning processes with his associates:
Acquisition
Extinction
spontaneous recovery
generalization
discrimination
Neutral Stimulus (NS): In classical conditioning, a stimulus that elicits no
response before conditioning.
Unconditioned Response (UR): in classical conditioning, an unlearned, naturally occurring response (such as salivation) to an unconditioned stimulus (US) (such as food in the mouth).
Unconditioned Stimulus (US): in classical conditioning, a stimulus that unconditionally—naturally and automatically—triggers an unconditioned response (UR).
Conditioned Response (CR): in classical conditioning, a learned response to a previously neutral (but now conditioned) stimulus (CS).
Conditioned Stimulus (CS): in classical conditioning, an originally neutral stimulus that, after association with an unconditioned stimulus (US), comes to trigger a conditioned response (CR)
LOQ: In classical conditioning, what are the processes of acquisition, extinction, spontaneous recovery, generalization, and discrimination?
Pavolv tried to understand the acqusition, intial learning, by asking “How much time should elapse between presenting the NS (the tone, the light, the touch) and the US (the food)?”
This turned out to not to be super long- usually around 0.5 seconds
Conditioning usually won’t occur when the NS follows the US
classical conditioning is biologically adaptive
helps humans and other animals prepare for good or bad events
A new NS can become a new CS without a US through higher-order conditioning
Only needs to become associated with a previously conditioned stimulus
Higher-order conditioning tends to be weaker than first-order conditioning
also called second-order conditioning
Higher-Order Conditioning: a procedure in which the conditioned stimulus in one conditioning experience is paired with a new neutral stimulus, creating a second (often weaker) conditioned stimulus. For example, an animal that has learned that a tone predicts food might then learn that a light predicts the tone and begin responding to the light alone. (Also called second-order conditioning.)
Extinction and Spontaneous Recovery
If the tone was continued to play and there was no food brought to the dogs, then the dogs start to drool less and less and it continues
This reaction is known as extinction
But if the time between the tones sounding, the dogs would begin slaivating to the tone
This is called spontaneous recovery
Extinction: the diminishing of a conditioned response; occurs in classical conditioning when an unconditioned stimulus (US) does not follow a conditioned stimulus (CS); occurs in operant conditioning when a response is no longer reinforced.
Spontaneous Recovery: the reappearance, after a pause, of an extinguished conditioned response.
Pavlov and his students noticed that if a dog was conditioned to one tone, they would also somewhat respond to a new and different tone
If a dog was conditioned to salivate when rubbed would also drool a bit when scratched
This response is called generalization (or stimulus generalization)
Generalization can be adaptive
Gerealized fear can also linger
Stimuli similar to naturally disgusting objects will, by association, also evoke some disgust
Generalization: the tendency, once a response has been conditioned, for stimuli similar to the conditioned stimulus to elicit similar responses. (In operant conditioning, generalization occurs when responses learned in one situation occur in other, similar situations.)
Pavlov’s dogs also learned to respond to the sound of a particular tone and not to other tones
They learned the ability to distinguish between a conditioned stimulus (which predicts the US) and other, irrelevant stimuli (discrimination)
Slightly different stimuli can be followed by vastly different consequences
Discrimination: in classical conditioning, the learned ability to distinguish between a conditioned stimulus and similar stimuli that do not signal an unconditioned stimulus. (In operant conditioning, the ability to distinguish responses that are reinforced from similar responses that are not reinforced.)
LOQ: Why does Pavlov’s work remain so important?
Most psychologists now agree that classical conditioning is a basic form of learning
We now see that Pavlov’s ideas were incomplete with new knowledge and research
Pavolv’s work has remained important
He found that many other responses to many other stimuli can be classically conditioned in many other organisms
This showed that classical conditioning is one way that virtually all organisms learn to adapt to their environment
showed us how a process such as learning can be studied objectively
His meathods involved virtually no subjective judgments or guesses about what went on in a dog’s mind
LOQ: What have been some applications of Pavlov’s work to human health and well-being? How did Watson apply Pavlov’s principles to learned fears?
Pavlov’s principles can influence human health and well-being in three ways:
Drug cravings: Former drug users often feel a craving when they are again in the drug-using context—with people or in places they associate with previous highs. Thus, drug counselors advise their clients to steer clear of people and settings that may trigger these cravings
Food cravings: Classical conditioning makes dieting difficult. We readily associate sugary substances with an enjoyable sweet sensation. Researchers have conditioned healthy volunteers to experience cravings after only one instance of eating a sweet food
Eating one cookie can create hunger for another. People who struggle with their weight often have eaten unhealthy foods thousands of times, leaving them with strongly conditioned responses to eat the very foods that will keep them in poor health
Immune responses: Classical conditioning even works on the body’s disease-fighting immune system. When a particular taste accompanies a drug that influences immune responses, the taste by itself may come to produce an immune response
LOQ: What is operant conditioning?
Classical conditioning and operant conditioning are both forms of associative learning but are both very different
Classical Conditioning: forms associations between stimuli (a CS and the US it signals). It also involves respondent behavior—automatic responses to a stimulus (such as salivating in response to meat powder and later in response to a tone)
Operant Conditioning: organisms associate their own actions with consequences. Actions followed by reinforcers increase; those followed by punishments often decrease. Behavior that operates on the environment to produce rewarding or punishing stimuli is called operant behavior.
Operant Conditioning: a type of learning in which a behavior becomes more likely to recur if followed by a reinforcer or less likely to recur if followed by a punisher.
LOQ: Who was Skinner, and how is operant behavior reinforced and shaped?
B. F. Skinner (1904–1990) was a college English major and aspiring writer
enrolled as a graduate student in psychology
Eventually became modern behaviorism’s most influential and controversial figure
His work elaborated on t Edward L. Thorndike’s “law of effect”
Developed behavioral technology that revealed principles of behavior control
Skinner’s pioneering studies
designed an operant chamber
Also known as a Skinner box
The box has a bar (a lever) that an animal presses—or a key (a disc) the animal pecks
This releases a reward of food or water.
also has a device that records these responses
This creates a stage on which rats and other animals act out Skinner’s concept of reinforcement
Law of Effect: Thorndike’s principle that behaviors followed by favorable consequences become more likely, and that behaviors followed by unfavorable consequences become less likely
Operant Chamber: in operant conditioning research, a chamber (also known as a Skinner box) containing a bar or key that an animal can manipulate to obtain a food or water reinforcer; attached devices record the animal’s rate of bar pressing or key pecking.
Reinforcement: in operant conditioning, any event that strengthens the behavior it follows.
Skinner would tease rats to press a bar through shaping
Did this by gradually guiding the rat’s actions toward the desired behavior
You start by watching how the animal naturally behaves, so that you could build on its existing behavior
might give the rat a bit of food each time it approaches the bar
Then once the rat is approaching regularly, you would give the food only when it moves close to the bar, then closer still
Finally you would require it to touch the bar to get food
This rewards successive approximations by reinforcing responses towards your disired outcome
Shaping can also help us understand what nonverbal organisms perceive
Skinner noted that we continually reinforce and shape others’ everyday behaviors, even though we may not mean to do so
Shaping: an operant conditioning procedure in which reinforcers guide behavior toward closer and closer approximations of the desired behavior
LOQ: How do positive and negative reinforcement differ, and what are the basic types of reinforcers?
Positive reinforcement strengthens the response by presenting a typically pleasurable stimulus immediately after a response
Negative reinforcement strengthens a response by reducing or removing something negative
negative reinforcement is not punishment
negative reinforcement removes a punishing (aversive) even
Positive Reinforcement: increasing behaviors by presenting positive reinforcers. A positive reinforcer is any stimulus that, when presented after a response, strengthens the response. negative reinforcement
Negative Reinforcement: increasing behaviors by stopping or reducing negative stimuli. A negative reinforcer is any stimulus that, when removed after a response, strengthens the response. (Note: Negative reinforcement is not punishment.)
Primary reinforcers are unlearned
Ex. getting food when hungry
Conditioned reinforcers are learned through association with primary reinforcers
Also known as secondary reinforcers
Ex. a rat in a Skinner box learns that a light reliably signals a food delivery, the rat will work to turn on the light
Primary Reinforcer: an innately reinforcing stimulus, such as one that satisfies a biological need.
Conditioned Reinforcer: a stimulus that gains its reinforcing power through its association with a primary reinforcer; also known as a secondary reinforcer.
Immediate and Delayed Reinforcers
Humans respond to delayed reinforcers
Ex. paycheck at the end of the week, good grade at the end of a term, or a trophy after a sports season
We must learn to delay satisfaction to function effectively
Learning to control our impulses in order to achieve more valued rewards is a big step toward maturity
can later protect us from committing an impulsive crime
Children who delay gratification have tended to become socially competent and high-achieving adults
LOQ: How do different reinforcement schedules affect behavior?
Reinforcement schedules vary
Continous reinforcement learning occours rapidly
Makes it the best choice for mastering a behavior
But extinction also occurs rapidly
When reinforcement stops, the behavior soon stops (is extinguished)
Real-life e rarely provides continuous reinforcement.
Ex. salespeople do not make a sale with every pitch but they are persistent
This persistence is called partial (intermittent) reinforcement schedules
Responses are sometimes reinforced but sometimes not
Learning is slower to appear
Resistance to extinction is greater than continuous reinforcement
Fixed-ratio schedules reinforce behavior after a set number of responses.
Ex. a coffee shop gives you a free cup of coffee after 10 drinks are purchased
This causes continued animals to only pause briefly after a reinforcer before returning to a high rate of responding
Variable-ratio schedules provide reinforcers after a seemingly unpredictable number of responses
unpredictable reinforcement is what makes gambling and fly fishing so hard to extinguish even when they don’t produce the desired results
Because reinforcers increase as the number of responses increases, variable-ratio schedules produce high rates of responding
Fixed-interval schedules reinforce the first response after a fixed time period
Animals on this type of schedule tend to respond more frequently as the anticipated time for reward draws near
produces a choppy stop-start pattern rather than a steady rate of response
Ex. checking the cake more often as it cooks longer
Variable-interval schedules reinforce the first response after varying time intervals
unpredictable times, a food pellet rewarded Skinner’s pigeons for persistence in pecking a key
Ex. a longed-for message that finally rewards persistence in checking our phone
variable-interval schedules tend to produce slow, steady responding
there is no knowing when the waiting will be over
Overall, response rates are higher when reinforcement is linked to the number of responses (a ratio schedule) rather than to time (an interval schedule)
responding is more consistent when reinforcement is unpredictable (a variable schedule) than when it is predictable (a fixed schedule)
Reinforcement Schedule: a pattern that defines how often the desired response
will be reinforced.
Continuous Reinforcement: schedule reinforcing the desired response every
time it occurs.
Partial (Intermittent) Reinforcement Schedule Reinforcing: a response only
part of the time; results in the slower acquisition of a response but much greater
resistance to extinction than does continuous reinforcement.
Fixed-Ratio Schedule: in operant conditioning, a reinforcement schedule that reinforces a response only after a specified number of responses
Variable-Ratio Schedule: in operant conditioning, a reinforcement schedule that reinforces a response after an unpredictable number of responses.
Fixed-Interval Schedule: in operant conditioning, a reinforcement schedule that reinforces a response only after a specified time has elapsed.
Variable-Interval Schedule: in operant conditioning, a reinforcement schedule
that reinforces a response at unpredictable time intervals.
LOQ: How does punishment differ from negative reinforcement, and how does punishment affect behavior?
Reinforcement increases a behavior; punishment does the opposite
negative reinforcement increases the frequency of a preceding behavior
Withdraws something negative
a punisher is any consequence that decreases the frequency of a preceding behavior
Ex. a child gets burned after touching a hot stove
Criminal behavior is influenced more by swift and sure punishers than by the threat of severe sentences
Most of this behavior is impulsive
Many psychologists note four major drawbacks of physical punishment:
Punished behavior is suppressed, not forgotten. This temporary state may (negatively) reinforce parents’ punishing behavior.
child swears, the parent swats, the parent hears no more swearing and feels the punishment successfully stopped the behavior
Punishment teaches discrimination among situations.
In operant conditioning, discrimination occurs when an organism learns that certain responses, but not others, will be reinforced. Did the punishment effectively end the child’s swearing? Or did the child simply learn that while it’s not okay to swear around the house, it’s okay elsewhere?
Punishment can teach fear
A punished child may associate fear not only with the undesirable behavior but also with the person who delivered the punishment or where it occurred
may learn to fear a punishing teacher and try to avoid school, or may become more anxious
Physical punishment may increase aggression by modeling violence as a way to cope with problems.
Many aggressive delinquents and abusive parents come from abusive families
Studies find that spanked children are at increased risk for aggression
There is a debate to if physical punishmed chuldren may be more aggressive
It follows the same logic of people who have undergone psychotherapy are more likely to suffer depression since they had preexisting problems that triggered the treatments
Punishment: an event that tends to decrease the behavior that it follows.
LOQ: Why did Skinner’s ideas provoke controversy, and how might his operant conditioning principles be applied at school, in sports, at work, in parenting, and for self-improvement?
Skinner said argued that brain science isn’t needed for psychological science
Said “a science of behavior is independent of neurology”
urged people to use operant conditioning principles to influence others’ behavior at school, work, and home
argued that we should use rewards to evoke more desirable behavior
His critics objected and said that dehumanizing people by neglecting their personal freedom and by seeking to control their actions
Skinner’s response was that External consequences already haphazardly control people’s behavior
At School
Skinner said “Good instruction demands two things, students must be toldimmediately whether what they do is right or wrong and, when right, they must be directed to the step to be taken next.”
Many of his ideals over education are now possible
Students can now learn at their own pace with some programs
In Sports
Reinforcing small successes and then gradually increasing the challenge is the key to shaping behavior in athletic performance
Compared with children taught by conventional methods, those trained by this behavioral method have shown faster skill improvement
At Work
Organizations have invited employees to share the risks and rewards of company ownership
Others focus on reinforcing a job well done
Rewards are most likely to increase productivity if the desired performance is both well defined and achievable
Reward specific, achievable behaviors, not vaguely defined “merit”
Operant conditioning reminds us that reinforcement should be immediate.
In Parenting
parents can learn from operant conditioning practices
Parent-training researchers explain that, “Get ready for bed” and then caving into protests or defiance, parents reinforce such whining and arguing
Notice people doing something right and affirm them for it.
This disrupts the cycle
Give children attention and other reinforcers when they are behaving well. Target a specific behavior, reward it, and watch it increase
To Change Your Own Behaviors
State a realistic goal in measurable terms and announce it.
Decide how, when, and where you will work toward your goal.
Monitor how often you engage in your desired behavior
Reinforce the desired behavior.
Reduce the rewards gradually
LOQ: How does operant conditioning differ from classical conditioning?
Our biology and cognitive process influcene both classical and operant conditioning.
LOQ: How do biological constraints affect classical and operant conditioning?
Biological Limits on Classical Conditioning
Early behaviorists realized, an animal’s capacity for conditioning is limited by biological constraints
Ex. individual species’ predispositions prepare it to learn the associations that enhance its survival
This is a phenomenon called preparedness
Humans are also biologically prepared to learn some associations other than others
Ex. you develop aversions to that taste of oysters after becoming ill from eating them rather than the sight associated with oysters such as restaurants
Arthur Schopenhaue said that important ideas are first ridiculed, then attacked, and finally taken for granted
Our preparedness to associate a CS with a US that follows predictably and immediately is adaptive
Causes often do immediately precede effects
Ex. When chemotherapy triggers nausea and vomiting more than an hour following treatment, cancer patients may over time develop classically conditioned nausea (and sometimes anxiety) to the sights, sounds, and smells associated with the clinic
In normal circumstances, such revulsion to sickening stimuli would be adaptive.
Preparedness: a biological predisposition to learn associations, such as between taste and nausea, that have survival value.
Biological Limits on Operant Conditioning
Nature also constrains each species’ capacity for operant conditioning
We most easily learn and retain behaviors that reflect our biological predispositions.
This principle of this is: “Biological constraints predispose organisms to learn associations that are naturally adaptive.”
Instinctive Drift: the tendency of learned behavior to gradually revert to biologically predisposed patterns.
LOQ: How do cognitive processes affect classical and operant conditioning?
Cognition and Classical Conditioning
Early behaviorists believed that rats’ and dogs’ learned behaviors could be reduced to mindless mechanisms, so there was no need to consider cognition
Robert Rescorla and Allan Wagner showed that an animal can learn the predictability of an event.
If a shock always is preceded by a tone, and then may also be preceded by a light that accompanies the tone
Although the light is always followed by the shock, it adds no new information; the tone is a better predictor
more predictable the association, the stronger the conditioned response
the animal learns an expectancy, an awareness of how likely it is that the US will occur.
Classical conditioning with humans is not just CS-US association
Cognition and Operant Conditioning
Skinner believed that thoughts and emotions were behaviors that follow the same laws as other behaviors
growing belief that cognitive processes have a necessary place in the science of psychology and even in our understanding of conditioning
viewed “cognitive science
Animals on a fixed-interval reinforcement schedule respond more and more frequently as the time approaches when a response will produce a reinforcer
strict behaviorist would object to talk of “expectations,” the animals behave as if they expected that repeating the response would soon produce the reward
There is more to learning than associating a response with a consequence; there is also cognition
Promising people a reward for a task they already enjoy can backfire.
Excessive rewards can destroy intrinsic motivation
the desire to perform a behavior effectively and for its own sake
Cognitive Map: a mental representation of the layout of one’s environment. For example, after exploring a maze, rats act as if they have learned a cognitive map of it.
Latent Learning: learning that occurs but is not apparent until there is an incentive to demonstrate it.
Intrinsic Motivation: a desire to perform a behavior effectively for its own sake.
Extrinsic Motivation: a desire to perform a behavior to receive promised rewards or avoid threatened punishment.
LOQ: How does observational learning differ from associative learning? How may observational learning be enabled by neural mirroring?
Cognition supports observational learning
Learning without direct experience
Ex. watching and intermittent
This process is called modeling
Observational Learning: learning by observing others.
Modeling: the process of observing and imitating a specific behavior.
Vicarious reinforcement/ punishment
we learn to anticipate a behavior’s consequences
Mirrors and Imitation in the Brain
In humans, imitation is pervasive
catchphrases, fashions, ceremonies, foods, traditions, morals, and fads all spread by one person copying another
Even children and infants are natural imitators
children’s brains enable their empathy and their ability to infer another’s mental state, an ability known as theory of mind
Humans have brains that support empathy and imitation
Researchers can use fMRI scans to see brain activity associated with performing and with observing action
Our brain’s response to observing others makes emotions contagious
The brain simulates and vicariously experiences what we observe
Through reenactments of these we grasp others’ states of mind, postures, faces, voices, and writing styles,
By observing we unconsciously synchronize our own to theirs—which helps us feel what they are feeling
Imitation helps us gain friends
We mimic people we like
We yawn when other people yawn, we laugh when other people laugh
Pain imagined by an empathic romantic partner has triggered some of the same brain activity experienced by the loved one actually having the pain
we mentally simulatet the experiences described
Mirror Neurons: frontal lobe neurons that some scientists believe fire when we perform certain actions or observe another doing so. The brain’s mirroring of another’s action may enable imitation and empathy.
LOQ: What is the impact of prosocial modeling and of antisocial modeling?
Prosocial Effects
People’s modeling of prosocial (positive, helpful) behaviors can have prosocial effect
businesses use behavior modeling to help new employees learn communications, sales, and customer service skills
People who show nonviolent and helpful behavior can also prompt similar behavior
Parents are also powerful models
observational learning of morality begins early
Socially responsive toddlers who imitate their parents usually become preschoolers with a strong internalized conscience
Models are most effective when their actions and words are consistent
To encourage children to read, read to them and surround them with books and people who read.
Prosocial Behavior: positive, constructive, helpful behavior. The opposite of antisocial behavior.
Antisocial Behavior
Observational learning may also have antisocial effects
helps us know why abusive parents might have aggressive children
Critics note that such aggressiveness could be genetic
The lessons we learn as children are not easily replaced as adults, and they are sometimes visited on future generations.
Intellectual history is often made by people who risk going to extremes in pushing ideas to their limits. Researchers such as Ivan Pavlov, John Watson, B. F. Skinner, and Albert Bandura defined the issues and impressed on us the importance of learning. They illustrated the impact that can result from single-minded devotion to a few well-defined problems.