# How Variable Interval Schedules Influence Behavior

In operant conditioning, a variable-interval schedule is a schedule of reinforcement where a response is rewarded after an unpredictable amount of time has passed, which is the opposite of a fixed-interval schedule. This schedule produces a slow, steady rate of response.﻿﻿

As you probably recall, operant conditioning can either strengthen or weaken behaviors through the use of reinforcement and punishment. This learning process involves forming an association with behavior and the consequences of that action.﻿﻿

Psychologist B.F. Skinner is credited with the introduction of the concept of operant conditioning. He observed that reinforcement could be used to increase a behavior, and punishment could be used to weaken behavior. He also noted that the rate at which a behavior was reinforcement had an effect on both the strength and frequency of the response.

## How Does a Variable-Interval Schedule Work?

To understand how a variable-interval schedule works, let's start by taking a closer look at the term itself. Schedule refers to the rate of reinforcement delivery, or how frequently the reinforcement is given. Variable indicates that this timing is not consistent and may vary from one trial to the next. Finally, interval means that delivery is controlled by time. So, a variable-interval schedule means that reinforcement is delivered at varying and unpredictable intervals of time.﻿﻿

Imagine that you are training a pigeon to peck at a key to receive a food pellet. You put the bird on a variable-interval 30 (VI-30) schedule. This means that the pigeon will receive reinforcement an average of every 30 seconds. It is important to note that this is an average, however. Sometimes the pigeon might be reinforced after 10 seconds; sometimes it might have to wait 45 seconds. The key is that the timing is unpredictable.

## Characteristics of the Variable-Interval Schedule

• Very resistant to extinction﻿﻿
• The rate of response is moderate but steady
• Very minimal pause after reinforcement is given

## Examples of Variable-Interval Schedules

• Checking Your Email: Typically, you check your email at random times throughout the day instead of checking every time a single message is delivered. The thing about email is that in most cases, you never know when you are going to receive a message. Because of this, emails roll in sporadically at completely unpredictable times. When you check and see that you have received a message, it acts as a reinforcer for checking your email.
• Your Employer Checking Your Work: Does your boss drop by your office a few times throughout the day to check your progress? This is an example of a variable-interval schedule. These check-ins occur at unpredictable times, so you never know when they might happen. The chances are good that you work at a fairly steady pace throughout the day since you are never quite sure when your boss is going to pop in, and you want to appear busy and productive when she does happen to stop by. Immediately after one of these check-ins, you might briefly pause and take a short break before resuming your steady work pace.
• Pop Quizzes: Your psychology instructor might issue periodic pop quizzes to test your knowledge and to make sure you are paying attention in class. While these exams occur with some frequency, you never really know exactly when he might give you a pop quiz. One week you might end up taking two quizzes, but then go a full two weeks without one. Because you never know when you might receive a pop quiz, you will probably pay attention and stay caught up in your studies to be prepared.