Operant Conditioning - Complete Interactive Lesson
Part 1: Thorndike & Skinner
๐ง Operant Conditioning
Part 1 of 7 โ Thorndike, Skinner & the Foundations
Operant conditioning is learning through consequences โ behaviors that are followed by favorable outcomes are repeated, while behaviors followed by unfavorable outcomes are suppressed. Unlike classical conditioning (which involves involuntary reflexes), operant conditioning involves voluntary behaviors.
Core Definitions
| Term | Definition |
|---|---|
| Thorndike's Law of Effect | Behaviors followed by satisfying consequences are more likely to be repeated; behaviors followed by unpleasant consequences are less likely |
| Operant conditioning | A type of learning where behavior is strengthened or weakened by its consequences (reinforcement or punishment) |
| Skinner box | An operant conditioning chamber Skinner designed to study how animals learn through consequences (lever pressing โ food pellet) |
| Respondent vs. operant behavior | Respondent = involuntary/reflexive (classical conditioning); Operant = voluntary/chosen (operant conditioning) |
Real-World Example
Think about studying for a test. If you study hard and get an A (positive consequence), you're more likely to study hard again. If you skip studying and fail (negative consequence), you're less likely to skip again. You're learning through the consequences of your voluntary behavior โ that's operant conditioning.
Why This Matters for AP Psychology
Operant conditioning is one of the most heavily tested topics on the AP exam. You'll need to classify scenarios as reinforcement or punishment, identify schedules, and compare operant with classical conditioning. This part builds the foundation that everything else rests on.
Concept Check ๐ฏ
Deep Dive: From Thorndike to Skinner
Thorndike's Puzzle Box (1898) Edward Thorndike placed cats inside wooden crates with a latch mechanism. The cat had to figure out how to escape to reach food outside. At first, the cat scratched randomly โ but over trials, escape time decreased dramatically. The cat "stamped in" the successful behavior because it was followed by a satisfying consequence (food + freedom).
Skinner's Operant Chamber (1930sโ1950s) B.F. Skinner took Thorndike's principle and made it systematic. His "Skinner box" contained a lever (for rats) or a disk (for pigeons) that, when pressed, delivered food. Skinner could precisely control:
- When reinforcement was delivered
- How often it was delivered
- What type of consequence followed
Comparing the Pioneers
| Feature | Thorndike | Skinner |
|---|---|---|
| Era | Late 1800s | Mid 1900s |
| Apparatus | Puzzle box (cats) | Operant chamber (rats, pigeons) |
| Key concept | Law of Effect | Reinforcement schedules |
| Focus | Which behaviors get "stamped in" | Precisely controlling consequences |
| Legacy | Foundation of behaviorism |
Applied Recall โ๏ธ
1) What is the name of Thorndike's principle that behaviors followed by satisfying consequences are repeated?
2) What type of behavior does operant conditioning involve โ voluntary or involuntary?
3) What apparatus did Skinner use to study operant conditioning in animals?
Type the exact term.
Match the Concepts ๐
Common Misconceptions and Exam Strategy
Misconceptions to Avoid
- Operant conditioning is NOT the same as classical conditioning โ operant involves voluntary behaviors and consequences; classical involves involuntary reflexes and stimulus associations.
- Skinner did NOT invent operant conditioning โ Thorndike established the Law of Effect first. Skinner systematized and expanded the study.
- "Operant" does NOT mean "operation" or "surgery" โ it comes from "operate," meaning the organism operates on its environment to produce consequences.
- Negative reinforcement is NOT punishment โ "negative" means removing something, and reinforcement always INCREASES behavior. You'll explore this crucial distinction in Part 2.
AP Strategy Moves
- When an AP question describes a scenario, first ask: "Is this voluntary behavior (operant) or an involuntary reflex (classical)?"
- Know both Thorndike AND Skinner โ the exam tests which pioneer contributed what.
- The four consequences grid (positive/negative ร reinforcement/punishment) is the single most important framework in this unit. Master it.
- If a question mentions a "Skinner box" or "operant chamber," you're in operant conditioning territory.
Applied Scenarios ๐ฏ
Part 2: Reinforcement Types
Reinforcement Types
Part 2 of 7 โ Positive & Negative Reinforcement
Reinforcement is any consequence that increases the likelihood of a behavior being repeated. There are two types, and the key to understanding them is knowing what "positive" and "negative" mean in psychology:
- Positive = adding/presenting something
- Negative = removing/taking away something
Core Definitions
| Term | Definition | Example |
|---|---|---|
| Positive reinforcement (+R) | Adding a pleasant stimulus to increase behavior | Giving a dog a treat for sitting |
| Negative reinforcement (-R) | Removing an aversive stimulus to increase behavior | Taking aspirin removes a headache, so you take aspirin again |
| Primary reinforcer | Naturally satisfying โ no learning needed | Food, water, warmth, relief from pain |
| Secondary (conditioned) reinforcer | Learned through association with primary reinforcers | Money, grades, praise, tokens |
Real-World Example
Your car makes an annoying beeping sound until you buckle your seatbelt. When you buckle up, the beeping stops (removing an aversive stimulus). You're more likely to buckle up quickly next time. This is โ the removal of something unpleasant increases the behavior.
Part 3: Punishment
Punishment
Part 3 of 7 โ Positive & Negative Punishment
Punishment is any consequence that decreases the likelihood of a behavior being repeated. Just like reinforcement, punishment comes in two forms:
- Positive punishment = adding an aversive stimulus โ behavior decreases
- Negative punishment = removing a pleasant stimulus โ behavior decreases
Core Definitions
| Term | Definition | Example |
|---|---|---|
| Positive punishment (+P) | Adding an unpleasant stimulus after a behavior to decrease it | A speeding ticket (adding a fine) reduces speeding |
| Negative punishment (-P) | Removing a pleasant stimulus after a behavior to decrease it | Losing phone privileges reduces rule-breaking |
| Punishment limitations | Punishment suppresses behavior temporarily but doesn't teach what TO do; can cause fear, aggression, and avoidance | |
| Reinforcement vs. punishment | Reinforcement increases behavior; punishment decreases behavior |
Real-World Example
A teenager stays out past curfew. Their parents take away their car keys for a week (removing a pleasant stimulus). The teenager is less likely to break curfew again. This is negative punishment โ something desirable was removed to decrease the behavior.
Part 4: Schedules of Reinforcement
Schedules of Reinforcement
Part 4 of 7 โ When and How Often to Reinforce
So far, you know WHAT consequences do (reinforce or punish). Now the question is: WHEN should reinforcement be delivered? The schedule of reinforcement determines how quickly behavior is learned, how steadily it's performed, and how resistant it is to extinction.
Core Definitions
| Term | Definition | Example |
|---|---|---|
| Continuous reinforcement | Reinforce EVERY correct response | Vending machine gives candy every time you insert money |
| Partial (intermittent) reinforcement | Reinforce only SOME correct responses | Slot machine pays out unpredictably |
| Fixed-ratio (FR) | Reinforce after a SET NUMBER of responses | Earn a free coffee after every 10 purchases |
| Variable-ratio (VR) | Reinforce after an UNPREDICTABLE number of responses | Slot machines, fishing โ you never know which try will pay off |
| Fixed-interval (FI) | Reinforce the first response after a SET TIME period | Checking for a paycheck every two weeks |
| Variable-interval (VI) | Reinforce the first response after an UNPREDICTABLE time period |
Part 5: Shaping
Shaping & Applications
Part 5 of 7 โ Teaching Complex Behaviors
How do you train an animal to do something it would never do naturally, like a pigeon playing ping-pong? You can't wait for the full behavior to appear and then reinforce it. Instead, you use shaping โ reinforcing successive approximations toward the target behavior.
Core Definitions
| Term | Definition | Example |
|---|---|---|
| Shaping | Reinforcing successive approximations (closer and closer attempts) toward a desired behavior | Teaching a dog to roll over by first rewarding lying down, then turning, then full roll |
| Successive approximations | Each step that gets progressively closer to the target behavior | Turn head โ lean โ lie down โ roll partly โ full roll |
| Chaining | Linking a sequence of individually shaped behaviors into a complex chain | A rat pressing a lever, then pulling a chain, then climbing stairs |
| Token economy | A system where secondary reinforcers (tokens) are earned and exchanged for primary reinforcers | Classroom behavior chart โ earn stars, exchange for prizes |
| Applied Behavior Analysis (ABA) | A therapeutic application of operant conditioning, commonly used for autism spectrum disorder | Breaking social skills into steps and reinforcing each one |
Part 6: Problem-Solving Workshop
Problem-Solving Workshop
Part 6 of 7 โ Classifying Scenarios & Comparing Conditioning Types
The AP exam will give you real-world scenarios and ask you to classify them. This part gives you a systematic framework for tackling these questions confidently.
The Classification Framework
Step 1: Is this operant or classical conditioning?
- Does it involve a voluntary behavior and its consequence? โ Operant
- Does it involve an involuntary response paired with a stimulus? โ Classical
Step 2: If operant โ did behavior increase or decrease?
- Behavior INCREASED โ Reinforcement
- Behavior DECREASED โ Punishment
Step 3: Was something added or removed?
- Something ADDED โ Positive
- Something REMOVED โ Negative
Quick Reference: The Complete Classification
| Scenario cue | Classification |
|---|---|
| Pleasant stimulus added โ behavior increases | Positive reinforcement |
| Aversive stimulus removed โ behavior increases | Negative reinforcement |
| Aversive stimulus added โ behavior decreases | Positive punishment |
Part 7: AP Review
Synthesis & AP Review
Part 7 of 7 โ Putting It All Together
This final part integrates everything you've learned about operant conditioning and connects it to broader learning concepts that appear on the AP exam.
The Big Picture: Three Types of Learning
| Type | Mechanism | Behavior type | Key researchers |
|---|---|---|---|
| Classical conditioning | Association between stimuli (CS + UCS) | Involuntary/reflexive | Pavlov, Watson |
| Operant conditioning | Consequences shape behavior (R & P) | Voluntary/chosen | Thorndike, Skinner |
| Observational learning | Watching and imitating models | Can be either | Bandura |
Beyond Strict Behaviorism: Cognitive Influences
Strict behaviorists like Skinner argued that only observable behavior matters. But research showed that cognition plays a role even in conditioning:
- Cognitive maps (Tolman): Rats navigated mazes using mental representations of the layout โ not just stimulus-response connections. They had internal "maps" of the maze.
- Latent learning (Tolman): Rats that explored a maze without reinforcement learned the layout but didn't show it until a reward was introduced. Learning occurred WITHOUT observable behavior change โ contradicting strict behaviorism.