Operant Conditioning

Help Questions

AP Psychology › Operant Conditioning

Questions 1 - 10
1

A lab rat presses a lever; food is delivered only after 60 seconds pass. What schedule is used?

Variable-interval schedule: the first press after varying time intervals is reinforced, producing steady responding with little scalloping.

Fixed-interval schedule: the first lever press after a fixed time (60 seconds) is reinforced, often producing scalloped response patterns.

Variable-ratio schedule: food follows an unpredictable number of lever presses, yielding very high, steady pressing without timing effects.

Fixed-ratio schedule: food follows every set number of presses, so the rat must press a certain count regardless of time.

Explanation

This describes a fixed-interval (FI-60) schedule, where the first response after a fixed time period (60 seconds) is reinforced. This schedule typically produces a "scalloped" response pattern - the rat presses very little right after receiving food, then gradually increases pressing as the 60-second mark approaches. The rat learns to time the interval and doesn't waste energy pressing when food is unavailable. This differs from fixed-ratio schedules based on response count, and from variable schedules where timing is unpredictable. Fixed-interval schedules are common in real life, like checking for mail delivery that arrives at a consistent time.

2

A slot machine pays out after an unpredictable number of pulls. What schedule maintains gambling behavior?

Fixed-ratio schedule: payout after a predictable, set number of pulls, typically producing brief pauses after each reinforcement.

Fixed-interval schedule: payout for the first pull after a fixed time period, producing a scalloped increase near payoff time.

Variable-interval schedule: payout for the first pull after varying time intervals, producing steady, moderate response rates.

Variable-ratio schedule: payout after an unpredictable number of pulls, producing high and resistant-to-extinction responding.

Explanation

Slot machines operate on a variable-ratio (VR) schedule, where reinforcement (payout) occurs after an unpredictable, varying number of responses (lever pulls). This schedule produces the highest, most persistent response rates and is highly resistant to extinction, explaining why gambling can become so addictive. The gambler never knows which pull will pay off, maintaining hope and continued responding. Unlike fixed schedules where patterns are predictable, or interval schedules based on time, variable-ratio schedules keep people engaged because 'the next one might be the winner.' This powerful operant conditioning principle differs fundamentally from classical conditioning's reflexive responses, as gambling involves voluntary, persistent behavior maintained by intermittent reinforcement.

3

A child stops swearing after parents wash the child’s mouth with soap. Which operant process is this?

Negative reinforcement: removing soap after swearing to increase swearing through relief.

Variable-interval schedule: soap is applied after varying time intervals, not contingent on swearing.

Positive punishment: adding an aversive stimulus (soap taste) after swearing to decrease swearing.

Negative punishment: removing a desirable stimulus after swearing by adding soap to the mouth.

Explanation

Washing the child's mouth with soap represents positive punishment because it adds an aversive stimulus (unpleasant soap taste) following swearing to decrease this behavior. The soap creates an unpleasant experience contingent on swearing, which reduces future swearing behavior. Positive punishment involves presenting aversive consequences after undesired behaviors. While this particular method is controversial and not recommended by modern parenting experts, it technically fits the definition of positive punishment because something unpleasant is added to decrease behavior frequency.

4

A teacher gives a quiz every Friday; studying increases most on Thursday night. Which schedule explains this?

Negative punishment: removing a desired activity after not studying, which decreases not-studying behavior.

Fixed-ratio schedule: reinforcement after a set number of study sessions, producing rapid responding with pauses.

Fixed-interval schedule: reinforcement is tied to a predictable time, producing scalloped increases near the quiz.

Variable-ratio schedule: reinforcement after unpredictable study amounts, producing consistently high studying.

Explanation

This scenario illustrates a fixed-interval schedule where reinforcement (good quiz performance) becomes available after a predictable time period (every Friday). Fixed-interval schedules characteristically produce a "scalloped" response pattern where behavior is low immediately after reinforcement and gradually increases as the next reinforcement opportunity approaches. Students typically study less on Monday and Tuesday but intensify their studying on Wednesday and Thursday night, knowing the quiz is coming Friday. This creates the distinctive increase in behavior near the end of the interval.

5

A rat gets a food pellet after every 10 lever presses. Which reinforcement schedule is used?

Fixed-interval schedule: reinforcement becomes available after a fixed time period, so responding increases as the next interval ends.

Fixed-ratio schedule: reinforcement is delivered after a set number of responses, such as every 10 lever presses, often producing brief pauses.

Variable-ratio schedule: reinforcement occurs after an unpredictable number of lever presses, producing very high responding with minimal pauses.

Negative reinforcement schedule: an aversive stimulus is removed after a fixed time period, which increases lever pressing regardless of response count.

Explanation

This describes a fixed-ratio schedule of reinforcement, where the rat receives a food pellet after completing a specific, unchanging number of responses (10 lever presses). This schedule typically produces high rates of responding with brief pauses after each reinforcement, known as post-reinforcement pauses. Unlike variable-ratio schedules where the number varies unpredictably, fixed-ratio schedules create predictable patterns of behavior. This differs from interval schedules, which are based on time passage rather than response count, and from continuous reinforcement where every response is reinforced. Fixed-ratio schedules are common in piece-rate work situations and demonstrate how operant conditioning principles can maintain consistent behavioral output. The predictability allows organisms to anticipate reinforcement, unlike classical conditioning where stimulus-response connections are more automatic.

6

A child’s tantrums stop after parents consistently ignore them and provide no attention. What occurred?

Extinction: a previously reinforced behavior decreases when reinforcement (attention) is withheld consistently, so tantrums become less frequent over time.

Variable-ratio schedule: ignoring tantrums after an unpredictable number of outbursts, which typically produces extremely persistent tantrum behavior.

Negative punishment: removing an aversive stimulus after tantrums, decreasing tantrums because the child experiences relief when outbursts occur.

Positive reinforcement: adding attention after tantrums, increasing tantrums because the child learns that outbursts reliably produce social rewards.

Explanation

This demonstrates extinction in operant conditioning, where a previously reinforced behavior (tantrums) decreases when reinforcement (parental attention) is consistently withheld. The child's tantrums were likely maintained by the attention they produced, even if that attention was negative, making attention the reinforcer. By removing this reinforcement, the behavior loses its function and gradually diminishes, though extinction bursts (temporary increases) may occur initially. This differs from punishment, which would involve adding aversive consequences or removing positive ones, rather than simply withholding the maintaining reinforcer. Unlike classical conditioning extinction where a conditioned stimulus loses its power, operant extinction involves learned behaviors losing their reinforcement. The process requires consistency, as intermittent attention would create a variable-ratio schedule that actually strengthens the tantrum behavior.

7

A student buckles a seatbelt to stop the car’s loud beeping, increasing buckling. What is this?

Variable-interval schedule: buckling is reinforced after unpredictable time intervals, producing steady responding regardless of how often buckling occurs.

Negative punishment: removing a desirable stimulus (music) after buckling, which decreases seatbelt use because privileges are taken away.

Negative reinforcement: removing an aversive stimulus (beeping) after buckling, which increases seatbelt use in the future by strengthening the response.

Positive reinforcement: adding an aversive stimulus (beeping) after buckling, which increases seatbelt use because the noise signals safety.

Explanation

This exemplifies negative reinforcement, where an aversive stimulus (the loud beeping) is removed when the desired behavior (buckling the seatbelt) occurs. The key distinction is that something unpleasant is taken away rather than something pleasant being added, which would be positive reinforcement. The behavior of buckling increases because it successfully terminates the annoying sound, demonstrating how negative reinforcement strengthens behaviors through the removal of aversive stimuli. This differs from punishment, which aims to decrease behaviors, and from reinforcement schedules, which describe timing patterns rather than the reinforcement type. Unlike classical conditioning where responses are automatic, this operant conditioning involves a voluntary action (buckling) that produces a consequence (beeping stops). The effectiveness of negative reinforcement explains why many safety devices use warning sounds that cease when proper action is taken.

8

A pigeon receives food for the first key peck after exactly 60 seconds. Which schedule is this?

Negative reinforcement: food is removed after 60 seconds to increase pecking, so the pigeon pecks to end the loss of food.

Fixed-ratio schedule: food is delivered after a set number of pecks, such as every 60 pecks, producing high responding with brief pauses.

Fixed-interval schedule: the first response after a fixed time (60 seconds) is reinforced, often producing a scalloped response pattern.

Variable-interval schedule: the first peck after varying time intervals is reinforced, producing steady responding without predictable acceleration.

Explanation

This represents a fixed-interval schedule, where reinforcement is available for the first response after a fixed time period (60 seconds) has elapsed. This schedule typically produces a scalloped response pattern, with low responding immediately after reinforcement and accelerating responding as the next interval approaches. Unlike fixed-ratio schedules based on response count, fixed-interval schedules depend solely on time passage, regardless of how many pecks occur during the interval. This differs from variable-interval schedules where the time requirement changes unpredictably, producing steadier responding. The pigeon learns to time its responses, demonstrating how operant conditioning can shape temporal patterns of behavior. This contrasts with classical conditioning, where timing involves anticipating stimuli rather than optimizing response patterns for maximum reinforcement.

9

A dog sits when commanded, and its owner gives a treat, increasing future sitting. What occurred?

Positive reinforcement: adding a desirable stimulus after sitting to increase sitting frequency; the treat strengthens the sit response over time.

Positive punishment: adding an aversive consequence after sitting to decrease sitting frequency; the behavior should drop because it produces something unpleasant.

Fixed-interval schedule: providing a treat only after a set amount of time has passed since the last treat, regardless of sitting frequency.

Negative reinforcement: removing an unpleasant stimulus after sitting to increase sitting frequency in the future; behavior becomes more likely when discomfort ends.

Explanation

This scenario demonstrates positive reinforcement, a fundamental operant conditioning principle where a pleasant stimulus (the treat) is added following a desired behavior (sitting on command). The dog's sitting behavior increases in frequency because it results in something rewarding, distinguishing this from negative reinforcement where an unpleasant stimulus would be removed. Unlike classical conditioning where responses are reflexive, operant conditioning involves voluntary behaviors that are strengthened or weakened by their consequences. The treat serves as a reinforcer that makes the sitting response more likely to occur in future situations when the command is given. This differs from punishment, which would decrease the behavior, and from reinforcement schedules, which describe patterns of reinforcement delivery rather than the basic reinforcement process itself.

10

A worker checks email most after random message arrivals. Which reinforcement schedule best explains this pattern?

Fixed-ratio schedule: reinforcement after a set number of checks, producing bursts of checking separated by post-reinforcement pauses.

Variable-ratio schedule: reinforcement occurs after an unpredictable number of checks, so each check is rewarded after random response counts.

Variable-interval schedule: the first check after unpredictable time intervals is reinforced by message arrival, supporting steady checking.

Fixed-interval schedule: the first check after a set time is reinforced, causing a predictable surge only near the interval’s end.

Explanation

Email checking maintained by random message arrivals follows a variable-interval (VI) schedule, where the first response (checking) after varying, unpredictable time intervals is reinforced by finding new messages. This schedule produces steady, moderate response rates without the post-reinforcement pauses seen in fixed schedules. Workers check consistently because they never know when the next message might arrive, but checking multiple times in quick succession won't produce more messages. This differs from variable-ratio schedules where more responses could yield more reinforcement. The unpredictable timing maintains persistent checking behavior, demonstrating how operant principles shape modern workplace behaviors differently than classical conditioning's stimulus-response associations.

Page 1 of 7