Operant Conditioning and Reinforcement Schedules (7C)

Help Questions

MCAT Psychological and Social Foundations › Operant Conditioning and Reinforcement Schedules (7C)

Questions 1 - 10
1

In a study of smartphone checking, participants receive a brief, pleasant sound the first time they unlock their phone after a fixed 15-minute period has passed since the last sound (fixed interval 15 min). Which checking pattern is most likely to emerge over days of use?

Checking increases as 15 minutes approaches, with relatively little checking right after the sound

Checking decreases because the sound acts as punishment for unlocking

Checking occurs in unpredictable bursts because the schedule is variable ratio

Checking is constant and very high throughout the day because each unlock is reinforced

Explanation

This question tests understanding of fixed interval schedules and their characteristic response patterns. Fixed interval (FI) schedules reinforce the first response after a fixed time period, producing a distinctive scalloped pattern where responding is minimal immediately after reinforcement and gradually increases as the interval progresses. In this FI-15min smartphone study, participants learn that checking immediately after hearing the sound is pointless, leading to reduced checking early in each interval. As 15 minutes approaches, checking increases because the probability of reinforcement rises. Option B incorrectly suggests constant high checking, which would be inefficient given the temporal constraint. The FI schedule teaches temporal discrimination, creating the characteristic pattern of low responding post-reinforcement followed by acceleration.

2

A customer service center tracks how often employees check an internal knowledge base. Management introduces a rule: the first time an employee checks the knowledge base after a random amount of time has passed (ranging from 5 to 25 minutes) earns a small raffle entry (variable interval schedule). Which outcome is most likely?

Employees show very high rates of checking with almost no pauses because reinforcement is based on number of checks

Employees check at a relatively steady rate throughout the shift because reinforcement availability is time-based and unpredictable

Employees check less often over time because raffle entries remove an aversive stimulus and thus reduce checking

Employees check mostly near the end of each 25-minute block because the interval is fixed and predictable

Explanation

This question tests understanding of variable interval (VI) reinforcement schedules in an applied workplace setting. In operant conditioning, a VI schedule provides reinforcement for the first response after a variable time period has elapsed - here, the first knowledge base check after a random 5-25 minute interval earns a raffle entry. VI schedules characteristically produce steady, moderate response rates because the unpredictability of reinforcement availability makes consistent checking the most effective strategy. The correct answer is A because employees cannot predict when checking will be rewarded, leading to relatively constant checking behavior throughout their shift. Answer C is incorrect because the predictable checking pattern near interval endpoints only occurs with fixed interval schedules, not variable ones. The critical insight is that when reinforcement timing is unpredictable, organisms maintain steady response rates rather than clustering responses at specific times.

3

A study tests how quickly children clean up toys. In Group 1, a child gets a sticker after every 5 toys put away (fixed ratio 5). In Group 2, a child gets a sticker after the first toy put away after 2 minutes have passed (fixed interval 2 min). Assuming both groups value stickers similarly, which group is most likely to show a higher overall rate of toy pickup during the session?

Group 2, because reinforcement is delivered for every response once 2 minutes pass

Neither group, because reinforcement schedules do not affect response rate once a behavior is learned

Group 2, because fixed interval schedules produce the highest response rates

Group 1, because fixed ratio schedules directly reinforce higher response rates

Explanation

This question compares response rates between fixed ratio and fixed interval schedules in operant conditioning. Operant conditioning strengthens behaviors via reinforcements; fixed ratio (FR) rewards after a set number of responses, promoting high rates, while fixed interval (FI) rewards after time, often yielding lower overall rates with scalloping. Children in Group 1 get stickers on FR5 for toys, while Group 2 uses FI 2 minutes. The correct answer is B because FR directly ties reinforcement to response count, encouraging faster pickup to earn more stickers quickly. A distractor like A fails, as FI typically produces lower rates than FR, not higher. For a transferable check, compare if output is response-driven (FR) versus time-driven (FI). This applies to piecework pay versus hourly wages in productivity studies.

4

In a lab study on operant conditioning, rats press a lever to receive a food pellet. During Phase 1, a pellet is delivered after every 10 lever presses (a fixed ratio schedule). In Phase 2, the schedule is changed so that, on average, a pellet is delivered after 10 presses, but the exact number varies unpredictably from trial to trial (a variable ratio schedule with mean 10). Which change in lever-pressing is most likely when the rats move from Phase 1 to Phase 2?

A sharp decrease in responding because variable schedules function like punishment and suppress behavior

A pattern of “scalloping,” with responding increasing as time passes since the last pellet

A lower overall response rate with longer pauses after each pellet because reinforcement is less predictable

A higher, steadier response rate with fewer post-reinforcement pauses because reinforcement is tied to responding but is unpredictable

Explanation

This question tests understanding of how different reinforcement schedules affect response patterns in operant conditioning. In operant conditioning, reinforcement schedules determine when a behavior is rewarded - fixed ratio (FR) schedules provide reinforcement after a set number of responses, while variable ratio (VR) schedules provide reinforcement after an unpredictable number of responses averaging around a set value. When rats switch from FR-10 to VR-10, they experience unpredictability in when the next pellet will arrive, which typically produces higher, steadier response rates with minimal post-reinforcement pauses. The correct answer is B because VR schedules eliminate the predictability that causes post-reinforcement pauses in FR schedules, leading to more consistent responding. Answer A is incorrect because VR schedules actually produce higher, not lower, response rates than FR schedules. A key principle to remember is that variable schedules generally produce more persistent behavior than fixed schedules because the organism cannot predict exactly when reinforcement will occur.

5

A language-learning app gives a user a “streak bonus” only when the user completes a lesson on the 7th day of a week (no bonus is given on days 1–6, even if lessons are completed). This is a fixed interval schedule with a 7-day interval. If the app wants to increase lesson completion on earlier days without changing the size of the bonus, which modification would be expected to produce more consistent daily studying?

Switch to a variable interval schedule where the first lesson after an unpredictable number of days earns the bonus

Add a response cost so that missed days remove points, because punishment increases desired behavior

Remove the bonus entirely so studying becomes intrinsically motivated and therefore more frequent

Keep the fixed interval but deliver the bonus only after an unpredictable number of lessons averaging seven

Explanation

This question evaluates strategies to modify fixed interval schedules for more consistent behavior in operant conditioning. Operant conditioning uses reinforcements to increase desired behaviors, with fixed interval (FI) schedules rewarding the first response after a fixed time, often leading to low early responding and a surge near the interval's end. The app's FI 7-day schedule likely results in procrastination, with lessons clustered on day 7. The correct answer is A because switching to variable interval (VI) promotes steady daily engagement, as the unpredictable timing encourages consistent checking or responding. A distractor like D fails, as removing reinforcement typically leads to extinction, not intrinsic motivation. To verify transferably, note if unpredictability in timing boosts regular behavior without end-loaded surges. This principle applies to improving habits like exercise through apps with random rewards.

6

A supervisor praises an employee after the employee completes exactly 8 data-entry forms (fixed ratio 8). The supervisor then changes to praising the employee after the employee completes the first form submitted after exactly 10 minutes have passed (fixed interval 10 min). Which change in performance is most likely following the switch?

No change, because fixed ratio and fixed interval schedules produce identical response patterns

A shift from output controlled by time to output controlled by number of forms

A dramatic decrease because praise removes a negative stimulus and therefore punishes data entry

A shift from high, consistent output to output that speeds up near the end of each 10-minute period

Explanation

This question tests understanding of how behavior changes when switching from ratio to interval schedules. Fixed ratio (FR) schedules produce high, consistent output with brief post-reinforcement pauses, as every response brings the employee closer to praise. Fixed interval (FI) schedules produce scalloped patterns where responding is low after reinforcement and increases as the interval progresses. When switching from FR-8 to FI-10min, the employee learns that completing forms immediately after praise is ineffective, leading to reduced output early in each interval. Output then accelerates near the 10-minute mark when praise becomes available. Option B incorrectly suggests the schedule remains response-based when it's now time-based. The key insight is that FI schedules promote temporal discrimination rather than consistent responding.

7

A coffee shop loyalty program gives a free drink after exactly 12 purchases (fixed ratio 12). Management changes the program so that a free drink is given after an unpredictable number of purchases averaging 12 (variable ratio). Which customer behavior change is most likely?

Customers will stop buying because the program change removes a negative reinforcer

Customers will buy less often because variable schedules reduce motivation compared with fixed schedules

Customers will show increasing purchases as the 12th purchase approaches, then pause after the free drink

Customers will buy coffee at a steadier rate because rewards are less predictable

Explanation

This question explores behavioral shifts from fixed ratio to variable ratio in consumer loyalty programs via operant conditioning. Operant conditioning uses schedules to maintain behaviors; fixed ratio (FR) often includes post-reward pauses, while variable ratio (VR) generates steady, high-rate responding due to unpredictability. The coffee program changes from FR12 to VR averaging 12 for free drinks. The correct answer is A because VR eliminates predictable pauses, encouraging more consistent purchases as each could trigger the reward. A distractor like B fails, as it reflects FR patterns with acceleration and post-reward lulls, not VR. For a transferable check, see if behavior steadies without post-reward drops, typical of VR. This is evident in variable-reward apps boosting user engagement.

8

A researcher trains mice to nose-poke for a sucrose drop. The mice are first reinforced on a fixed ratio 1 schedule (every nose-poke earns sucrose). The researcher then shifts to a fixed ratio 10 schedule (every 10th nose-poke earns sucrose) without changing sucrose amount. Which change in responding is most likely immediately after the shift?

Responding becomes timed to 10-second intervals because the schedule is now interval-based

A temporary drop in responding because reinforcement is less frequent, followed by higher response rates once learned

An immediate increase in responding because fewer rewards per response always strengthens behavior

Responding stops because sucrose now functions as punishment when delivered after 10 responses

Explanation

This question tests understanding of ratio strain and schedule thinning effects. When organisms are shifted from dense reinforcement (FR-1) to lean reinforcement (FR-10), they typically show ratio strain - a temporary disruption in responding as they adjust to the new requirement. Initially, mice may pause or show erratic responding because the expected reinforcement doesn't occur after single responses. However, once they learn the new contingency, response rates typically increase to maintain reinforcement frequency, often exceeding the original rate. Option B incorrectly suggests immediate increases, ignoring the adjustment period. Option C wrongly implies the schedule is interval-based when it remains ratio-based. The key principle is that abrupt increases in ratio requirements produce temporary disruptions before higher sustained rates emerge.

9

A dog is trained to sit. The trainer delivers a treat for the first correct sit that occurs after a variable amount of time has passed (sometimes 10 s, sometimes 45 s, averaging 30 s). This is a variable interval schedule. Which response pattern is most consistent with this schedule after training is established?

Moderate, steady sitting attempts over time because reinforcement depends on time, unpredictably

A scalloped pattern: little sitting right after a treat, increasing sharply as 30 s approaches

Very high response rates with minimal pauses because each sit may be the one that earns the treat

No sitting because the unpredictability of timing functions as punishment

Explanation

This question examines response patterns under variable interval schedules in operant conditioning. Operant conditioning reinforces behaviors through schedules; variable interval (VI) delivers rewards for the first response after an unpredictable time interval, typically producing moderate, steady responding without pronounced pauses. The dog's sitting is reinforced on a VI averaging 30 seconds, varying from 10 to 45 seconds. The correct answer is C because VI encourages consistent attempts over time, as the dog can't predict when the next opportunity arises. A distractor like B fails, as it describes fixed interval scalloping, not VI's steady pattern. To check transferably, observe if responses remain even without time cues, indicative of VI. This mirrors behaviors like checking email sporadically for unpredictable replies.

10

A dog is trained to sit. During training, the owner gives a treat for the first sit that occurs after exactly 30 seconds have passed since the last treat (fixed interval 30 s). The owner then switches to giving a treat for the first sit after a variable amount of time averaging 30 seconds (variable interval). Which change is most likely after the switch?

Sitting becomes more evenly distributed over time, with fewer long pauses after treats

Sitting becomes dependent on completing a fixed number of sits rather than time

Sitting stops because the treat is now delivered as negative reinforcement

Sitting becomes clustered right after each treat, with longer pauses that grow over time

Explanation

This question tests understanding of how switching from fixed to variable interval schedules affects response distribution. Fixed interval (FI) schedules produce scalloped patterns where responding clusters near the end of each interval, with minimal responding immediately after reinforcement. Variable interval (VI) schedules eliminate this temporal predictability, producing more evenly distributed responding throughout time. When the owner switches from FI-30s to VI-30s, the dog can no longer predict when sitting will be reinforced, leading to more consistent sitting behavior with fewer long pauses. Option A incorrectly describes increased clustering, which is characteristic of FI, not VI schedules. The key insight is that variability in interval schedules promotes steady responding by removing temporal cues.

Page 1 of 6