AP Psychology Unit 3 Learning: Conditioning, Cognition, Motivation, and Emotion
Classical Conditioning (Pavlov)
What classical conditioning is
Classical conditioning is a type of learning where you come to associate two stimuli, so that you begin to respond to a previously neutral stimulus as if it were meaningful. The key idea is that your nervous system learns that “this predicts that.”
Ivan Pavlov discovered classical conditioning while studying digestion. He noticed that dogs began salivating before food arrived—sometimes just from cues that reliably came before feeding (like footsteps or the food bowl). That observation matters because it shows learning can happen through simple associations, without deliberate decision-making.
Classical conditioning helps explain many everyday reactions that feel automatic: a spike of anxiety when you hear a dentist’s drill, nausea after a food that once made you sick, or excitement when you hear a particular notification sound.
The parts: unconditioned vs. conditioned
Classical conditioning uses a specific vocabulary because the exam often tests whether you can label the pieces correctly.
- Unconditioned stimulus (US): something that naturally, automatically triggers a response (food, a puff of air to the eye, pain).
- Unconditioned response (UR): the unlearned, automatic response to the US (salivation, blinking, fear).
- Neutral stimulus (NS): a stimulus that initially does not trigger the target response (a bell before learning).
- Conditioned stimulus (CS): the previously neutral stimulus after it has been associated with the US (the bell after repeated pairings with food).
- Conditioned response (CR): the learned response to the CS (salivation to the bell).
A common misconception is that the CS “causes” the US. It doesn’t. The CS becomes a predictor of the US, and your body prepares.
How classical conditioning works (step-by-step)
- Before learning: US → UR. The NS does not elicit the UR.
- During learning (acquisition): You repeatedly pair NS + US → UR.
- After learning: CS (formerly NS) → CR.
Acquisition refers to the initial phase when the association is being formed. In general, conditioning is strongest when the CS reliably predicts the US (and usually when the CS comes shortly before the US). If the CS is random or comes after the US, learning is weaker because it’s not a good predictor.
Key processes: generalization, discrimination, extinction, spontaneous recovery
Once a conditioned response exists, several predictable patterns show up:
- Stimulus generalization: you respond similarly to stimuli that resemble the CS. If you were bitten by a large black dog, you might feel nervous around similar-looking dogs.
- Discrimination: you learn to respond differently to different stimuli. You might learn that only one specific dog (or only large dogs) predicts danger.
Generalization is useful because it protects you without needing to relearn from scratch every time. Discrimination is useful because overgeneralizing can make life unmanageable (e.g., being afraid of all dogs).
- Extinction: the conditioned response weakens when the CS is repeatedly presented without the US. If the bell rings many times with no food, salivation decreases.
- Spontaneous recovery: after extinction and a rest period, the CR can reappear briefly when the CS is presented again.
A frequent student error is saying extinction “erases” learning. In classical conditioning, extinction is better thought of as new learning (“CS no longer predicts US”), which is why spontaneous recovery can happen.
Higher-order conditioning
Higher-order conditioning (also called second-order conditioning) happens when a well-learned CS is paired with a new neutral stimulus so that the new stimulus becomes another CS. For example, if a tone (CS) predicts food and then a light is repeatedly paired with the tone, the light may eventually trigger salivation too.
This matters because it explains how chains of associations can form in real life—your reaction can spread from the original trigger to related cues.
Classical conditioning in real life: emotions, phobias, and taste aversion
Classical conditioning is especially powerful for emotional learning.
- Phobias can form when a neutral object becomes associated with fear or pain. Even one strong pairing can be enough.
- Conditioned taste aversion is a special form where you associate a particular taste with nausea. This is notable because it can occur after a single pairing and even when there’s a long delay between eating and getting sick.
Taste aversion highlights an important limit on “anything can be associated with anything.” Humans and other animals are biologically prepared to form some associations more easily than others.
Biological constraints and cognition in classical conditioning
Classical conditioning is not purely mechanical.
- Biological constraints: Evolution shapes what associations are easy to learn (taste with nausea; certain cues with fear).
- Cognitive influences: Your expectations matter. If you learn that the CS no longer predicts the US, your conditioned response changes. In many situations, you are learning an implicit “rule” about prediction, not just forming a reflex.
Example: identifying US, UR, CS, CR (the AP-style skill)
Suppose a student gets sick after eating a particular cafeteria burrito. Later, the smell of that burrito makes them nauseated.
- US: the illness/nausea-inducing event (getting sick)
- UR: nausea in response to getting sick
- CS: the smell/taste of the burrito (once associated)
- CR: nausea in response to the burrito smell
Notice how the burrito smell started as neutral (or even pleasant) but became a predictor.
Exam Focus
- Typical question patterns:
- Labeling US/UR/CS/CR in short scenarios (often with emotion or nausea).
- Predicting what happens after extinction, or distinguishing extinction vs. spontaneous recovery.
- Applying generalization vs. discrimination to real-life examples.
- Common mistakes:
- Mixing up UR and CR (UR is to the US; CR is to the CS).
- Calling extinction “forgetting” or “erasing” rather than new learning.
- Choosing a CS that is not actually a predictor (a cue that occurs after the US is usually not an effective CS).
Operant Conditioning (Skinner)
What operant conditioning is
Operant conditioning is learning based on consequences. You learn which behaviors “work” because they are followed by desirable outcomes (or avoid undesirable ones). Where classical conditioning is about associating stimuli, operant conditioning is about associating behavior with outcomes.
B. F. Skinner expanded on earlier work by Edward Thorndike, who proposed the law of effect: behaviors followed by satisfying outcomes become more likely, and behaviors followed by unpleasant outcomes become less likely. Skinner’s big contribution was studying behavior systematically using controlled environments (like the “Skinner box”) and focusing on how reinforcement schedules shape behavior over time.
Operant conditioning matters because much of human learning—studying, practicing sports, social behaviors, habits—depends on consequences. It also provides tools for behavior change in classrooms, parenting, therapy, and workplaces.
Core idea: reinforcement increases behavior, punishment decreases behavior
In operant conditioning, consequences come in two broad categories:
- Reinforcement: increases the likelihood of a behavior.
- Punishment: decreases the likelihood of a behavior.
Each can be positive (adding something) or negative (removing something). “Positive” and “negative” here do not mean “good” and “bad”—they mean add vs. remove.
| Consequence | What happens | Effect on behavior | Example |
|---|---|---|---|
| Positive reinforcement | Add a desirable stimulus | Behavior increases | Give a dog a treat for sitting |
| Negative reinforcement | Remove an aversive stimulus | Behavior increases | Buckle seatbelt to stop beeping |
| Positive punishment | Add an aversive stimulus | Behavior decreases | Add extra chores for breaking curfew |
| Negative punishment | Remove a desirable stimulus | Behavior decreases | Take away phone for missing homework |
A very common misconception is confusing negative reinforcement with punishment. Negative reinforcement increases behavior; punishment decreases behavior.
How operant conditioning works: shaping and successive approximations
Many behaviors aren’t learned in one step. Shaping is reinforcing closer and closer versions of the desired behavior—called successive approximations.
For example, if you want to train a rat to press a lever, you might reinforce:
1) moving toward the lever, then 2) touching it, then 3) pressing it.
Shaping matters because it explains how complex behaviors can be built without “explaining” them verbally. Humans shape each other all the time through attention, praise, and feedback.
Primary vs. conditioned reinforcers
Not all reinforcers are equal.
- Primary reinforcers satisfy biological needs (food, water, warmth).
- Conditioned reinforcers gain value through association with primary reinforcers (money, grades, tokens).
A token economy is a system where people earn tokens (conditioned reinforcers) that can later be traded for rewards. Token economies are used in some classrooms and treatment settings because they make reinforcement immediate and consistent.
Reinforcement schedules: why timing patterns matter
Reinforcement doesn’t just depend on what consequence you get—it depends on when you get it.
- Continuous reinforcement: every correct response is reinforced. This is great for learning a new behavior quickly, but the behavior may extinguish faster when reinforcement stops.
- Partial (intermittent) reinforcement: only some responses are reinforced. This usually leads to slower acquisition but greater resistance to extinction.
Two key dimensions:
- Ratio schedules reinforce based on number of responses.
- Interval schedules reinforce based on time.
And each can be: - Fixed (predictable)
- Variable (unpredictable)
| Schedule | Reinforcement rule | Real-world feel | Example |
|---|---|---|---|
| Fixed ratio (FR) | After a set number of responses | High response rate with pauses | Paid per 10 items assembled |
| Variable ratio (VR) | After an unpredictable number of responses | Very high, steady responding | Slot machines |
| Fixed interval (FI) | First response after a set time | “Scallop” pattern (slows then speeds up) | Checking oven near end of bake time |
| Variable interval (VI) | First response after varying time | Steady, moderate responding | Email notifications arriving unpredictably |
Many students memorize these but struggle to apply them. A helpful approach is to ask two questions:
1) Is reinforcement based on responses (ratio) or time (interval)?
2) Is it predictable (fixed) or unpredictable (variable)?
Punishment: why it’s tricky
Punishment can suppress behavior, but it has drawbacks:
- It may teach fear or avoidance (you learn to avoid the punisher, not necessarily change the behavior).
- It doesn’t automatically teach a replacement behavior.
- It can increase aggression or anxiety.
- If it’s inconsistent, the behavior may persist.
In real settings, reinforcing a desired alternative behavior (and making expectations clear) is often more effective long-term than relying on punishment alone.
Cognition and biology in operant conditioning
Operant conditioning is influenced by thinking, not just consequences:
- People can develop expectancies about whether a behavior will pay off.
- Intrinsic motivation (doing something because it’s satisfying) can sometimes be reduced by excessive external rewards, especially if the rewards feel controlling. This is often discussed as the overjustification effect.
Biology also constrains operant learning. Some responses are more “natural” for a species, and attempts to reinforce unnatural behaviors may be less effective.
Example: reinforcement vs. punishment (and the negative reinforcement trap)
Scenario: A student has a headache. They take pain medication, and the headache goes away. Later, they are more likely to take the medication when a headache starts.
- Behavior: taking medication
- Consequence: removal of pain
- Type: negative reinforcement (behavior increases because an aversive state is removed)
This is not punishment because the behavior increased.
Exam Focus
- Typical question patterns:
- Identifying whether a consequence is positive/negative reinforcement/punishment in brief scenarios.
- Matching examples to reinforcement schedules, especially VR vs. VI.
- Applying shaping to how a complex behavior is trained.
- Common mistakes:
- Treating “negative” as “bad” instead of “removal.”
- Calling a consequence “reinforcement” just because it feels like a reward—if behavior decreases, it’s not reinforcement.
- Mixing up interval vs. ratio (time vs. number of responses).
Observational and Cognitive Learning
Why learning is more than conditioning
Classical and operant conditioning are powerful, but they don’t explain everything. Humans can learn by watching others, thinking about outcomes, and forming mental representations of the world. AP Psychology commonly groups these ideas under observational learning and cognitive learning.
These approaches matter because they explain rapid learning without direct reinforcement, and they better match many real-world situations—like learning social rules, avoiding risks, or mastering skills by demonstration.
Observational learning: learning by watching
Observational learning occurs when you learn new behaviors or information by watching others (models). Albert Bandura’s work emphasized that learning can happen even without performing the behavior during learning and without receiving direct reinforcement.
Bandura highlighted several processes that help explain when observational learning is likely:
- You have to pay attention to the model.
- You must remember what happened.
- You need to be able to reproduce the behavior.
- You must have motivation to imitate (often influenced by whether the model was rewarded or punished).
A key mechanism is vicarious reinforcement: seeing someone else get rewarded makes you more likely to copy them; seeing punishment makes you less likely.
Bandura’s famous Bobo doll studies showed that children who observed adults behaving aggressively were more likely to imitate aggression—especially when the adult’s behavior seemed unpunished or rewarded.
Observational learning also supports prosocial behavior. If children observe helping and cooperation being valued, those behaviors can spread too.
Self-efficacy: the belief that you can do it
Bandura also introduced self-efficacy, your belief in your ability to succeed at a task. Self-efficacy matters because it affects persistence: if you believe you can improve, you are more likely to keep trying, seek strategies, and tolerate setbacks.
A subtle misconception is to equate self-efficacy with self-esteem. Self-efficacy is task-specific (“I can learn to solve these problems”), while self-esteem is a broader evaluation of self-worth.
Cognitive learning: latent learning, insight, and mental maps
Cognitive learning focuses on internal mental processes.
- Latent learning is learning that occurs without obvious reinforcement and isn’t immediately demonstrated. Edward Tolman argued that rats in mazes could learn the layout without rewards; the learning became visible when motivation (like food) was introduced.
- Cognitive maps are mental representations of spatial layouts. You use them when you can take a shortcut even if you’ve never taken that exact route.
- Insight learning involves a sudden realization of a solution, rather than trial-and-error. Wolfgang Köhler’s work with chimpanzees is often used to illustrate insight (e.g., suddenly figuring out how to reach bananas with tools).
These concepts matter because they show that behavior can change due to changes in understanding, not only changes in reinforcement history.
Learned helplessness: when experience teaches “nothing works”
Learned helplessness occurs when repeated exposure to uncontrollable negative events leads an organism to stop trying to escape—even when escape becomes possible. Martin Seligman’s research helped establish this concept.
In humans, learned helplessness is relevant to motivation, academic persistence, and some patterns seen in depression. The learning component is crucial: people may come to expect that effort won’t matter.
A common mistake is to treat learned helplessness as laziness. The core is a learned expectation of lack of control.
The bridge to conditioning: cognition inside “simple” learning
Even in classical and operant conditioning, cognitive factors can matter:
- In classical conditioning, you often learn a prediction relationship.
- In operant conditioning, you often learn an expectancy that a behavior leads to a consequence.
So rather than seeing “conditioning vs. cognition” as opposites, it’s better to see them as levels of explanation: conditioning describes patterns of learning; cognitive ideas explain what information the learner may be representing internally.
Example: vicarious reinforcement
A middle schooler watches a classmate tell a joke and receive lots of laughs and attention. The next day, the student tells similar jokes more often.
This can be explained without direct reinforcement: the student learned through observational learning, with vicarious reinforcement providing motivation.
Exam Focus
- Typical question patterns:
- Identifying observational learning vs. operant conditioning in social scenarios.
- Applying Bandura terms like modeling, vicarious reinforcement, and self-efficacy.
- Distinguishing latent learning and insight from trial-and-error learning.
- Common mistakes:
- Assuming observational learning requires being rewarded personally (it doesn’t).
- Confusing self-efficacy with general confidence or self-esteem.
- Treating latent learning as “not real learning” because it isn’t immediately shown.
Motivation and Emotion
Why include motivation and emotion in learning
Motivation and emotion are tightly connected to learning because they influence what you pay attention to, what you repeat, what you avoid, and how strongly experiences are remembered. (In AP Psychology, motivation and emotion are often emphasized as their own major topic area, but they directly support your understanding of learning processes.)
If you think of learning as “what changes behavior and knowledge over time,” motivation explains why you engage in the behaviors that create practice and reinforcement, and emotion helps explain why some associations (especially fear) form quickly and persist.
Motivation: needs, drives, and incentives
Motivation refers to the processes that initiate, direct, and sustain behavior.
One classic approach is drive-reduction theory, which proposes that physiological needs create an aroused state (a drive) that motivates you to reduce the need and return to homeostasis (internal balance). Hunger and thirst fit this idea well.
However, not all motivated behavior is about reducing a biological drive. Incentives—external rewards or goals—can pull behavior even when you are not in a deprived state (e.g., eating dessert when not hungry). Curiosity and thrill-seeking can even increase arousal rather than reduce it.
A practical way to think about it is that behavior often reflects a mix:
- internal states (needs)
- learned expectations (what has paid off before)
- external incentives (what is available)
Hunger: biology plus learning
Hunger regulation involves multiple signals rather than a single “hunger center.” The hypothalamus plays a central role in appetite and satiety, integrating signals about energy balance.
Your body uses chemical signals that influence hunger and fullness:
- Ghrelin (from the stomach) is associated with hunger.
- Leptin (from fat cells) is associated with satiety and long-term energy regulation.
- Insulin (from the pancreas) is involved in regulating blood glucose and energy use.
Learning and environment strongly shape eating:
- Cues like time of day, smells, and social context can trigger eating through classical conditioning.
- Operant processes can reinforce snacking (e.g., eating reduces boredom or stress, which can function like negative reinforcement).
A common misconception is that hunger is purely willpower or purely biology. In reality, it’s an interaction: biology provides signals, but learning and environment determine many patterns.
Achievement motivation and intrinsic vs. extrinsic motivation
Achievement motivation involves the desire to meet goals and standards. Your persistence is affected by:
- the value you place on the goal
- your expectancy of success (closely related to self-efficacy)
- feedback and reinforcement history
Intrinsic motivation means doing an activity because it is inherently satisfying. Extrinsic motivation means doing it for external rewards or to avoid punishment.
Extrinsic rewards can help initiate behavior, especially for tasks you wouldn’t otherwise do. But if rewards feel controlling, they can reduce intrinsic motivation (the overjustification effect). This links directly back to operant conditioning: reinforcement works, but the meaning of the reinforcement to the learner matters.
Emotion: what it is and why theories differ
Emotion involves a coordinated response including:
- subjective experience (what you feel)
- physiological arousal (body responses)
- expressive behaviors (facial expression, posture)
- cognitive appraisal (interpretation)
Different emotion theories disagree about the order and importance of these components.
Major theories of emotion (AP favorites)
| Theory | Core idea | What comes first? | Quick example |
|---|---|---|---|
| James-Lange | Emotion follows physiological arousal | Arousal → emotion label | “My heart races, so I feel fear.” |
| Cannon-Bard | Arousal and emotion happen simultaneously | Arousal + emotion together | “I feel fear and my body reacts at the same time.” |
| Schachter-Singer (Two-factor) | Emotion requires arousal plus a cognitive label | Arousal + interpretation → emotion | “My heart races; in this context I call it excitement.” |
| Lazarus (cognitive appraisal) | Appraisal (interpretation) can define emotion | Appraisal is central | “I evaluate it as dangerous, so I feel fear.” |
Two-factor theory is especially testable because it emphasizes that the same arousal can be experienced as different emotions depending on the situation.
A common misunderstanding is treating these theories as mutually exclusive “either/or” truths. In practice, different emotions and situations may fit different patterns, and modern views often integrate physiology, appraisal, and brain pathways.
Fear, conditioning, and the brain
Fear learning is a strong example of how emotion and learning interact. Through classical conditioning, neutral cues can become fear triggers. This matters for understanding anxiety and phobias.
AP Psychology also commonly emphasizes that some emotional responses can happen quickly, sometimes before conscious analysis—supporting the idea that there are fast, automatic pathways for certain emotional reactions.
Emotion and learning in everyday life
Emotion influences learning in several ways:
- It guides attention: emotionally significant events grab your focus.
- It affects memory: emotionally arousing experiences are often remembered more strongly.
- It shapes avoidance and approach behaviors: fear encourages avoidance; positive emotions can reinforce approach and repetition.
You can see the conditioning connection clearly:
- If studying in one environment consistently leads to stress, that environment can become a conditioned trigger for anxiety.
- If practicing a skill leads to pride or social approval, those feelings can become reinforcers that sustain effort.
Example: two-factor theory in action
Imagine you’re walking quickly and your heart is pounding. If you realize you’re late to a test, you might label the arousal as anxiety. If you realize you’re rushing to meet friends for a surprise party, you might label it as excitement. The physiological arousal is similar, but your interpretation changes the emotion.
Exam Focus
- Typical question patterns:
- Matching scenarios to emotion theories (especially James-Lange vs. Cannon-Bard vs. two-factor).
- Explaining motivated behaviors using drives, incentives, and intrinsic/extrinsic motivation.
- Connecting emotion to learning via conditioning (fear learning, avoidance behaviors).
- Common mistakes:
- Mixing up James-Lange and Cannon-Bard (sequence vs. simultaneous).
- Treating extrinsic rewards as always harmful to intrinsic motivation (they can help; the effect depends on context and how controlling the reward feels).
- Explaining hunger as only biology or only choice, ignoring cue-driven and learned components.