Main content

## MCAT

### Course: MCAT > Unit 11

Lesson 9: Learning- Learning questions
- Classical and operant conditioning article
- Classical conditioning: Neutral, conditioned, and unconditioned stimuli and responses
- Classical conditioning: Extinction, spontaneous recovery, generalization, discrimination
- Operant conditioning: Positive-and-negative reinforcement and punishment
- Operant conditioning: Shaping
- Operant conditioning: Schedules of reinforcement
- Operant conditioning: Innate vs learned behaviors
- Operant conditioning: Escape and avoidance learning
- Observational learning: Bobo doll experiment and social cognitive theory
- Long term potentiation and synaptic plasticity
- Non associative learning
- Biological constraints on learning

© 2023 Khan AcademyTerms of usePrivacy PolicyCookie Notice

# Operant conditioning: Schedules of reinforcement

Created by Jeffrey Walsh.

## Want to join the conversation?

- schedules of reinforcement

fixed ratio = productive vs fixed interval = complacent

commission sales <-----------------------> salary worker or min wage workers

what societal benefit is there for fixed interval schedules of reinforcement?(3 votes)- Fixed intervals provide stability. For example most employees budget based on what they are going to make. They need a guarantee they can pay their rent, phone bill, buy groceries ect. Additionally, depending on the occupation commission sales may create an unwanted tension between co-workers as they compete for customers or cause customers to distrust the salesperson. That is why salary and hourly wage is still benefits society.

Finally, these reinforcement techniques can be mixed, for example bonuses given to the employee with the most sales. Thus the best workers are rewarded but this doesn't interfere with their social stability.(26 votes)

- If the amount of responses increases at a fixed rate (for example: bonus after 3+2x number of responses, where x is the number of times a bonuses already given), is this still a variable ratio schedule?(2 votes)
- It would never be a variable ratio schedule. It would be a fixed ratio schedule because it happens at predicted intervals starting from 5 then going to 7,9,etc. you see a pattern of +2.(3 votes)

- Which schedule would cause a decrease in responses right before reinforcement?(1 vote)
- Both Fixed Ratio and Fixed interval can show a decrease in the the behaviour right after it is reinforced. There is a decrease in the behaviour because the person/animal knows what behaviour they need to perform to get the reward and they won't start the behaviour until they want the reward. For example the car salesman may have a decrease in sales until he wants another bonus at which point he will start making sales again.(3 votes)

- How would you describe Variable Interval better because I could not really understand the definition.(2 votes)
- when is extinction easiest and hardest to do?(2 votes)
- Do partial reinforcement schedules apply to punishment as well? I read that punishment is most effective when applied consistently, and was confused because partial reinforcement schedules cause behaviors to be more resistant to extinction.(2 votes)
- why is continuous reinforcement less resistance than partial reinforcement?(1 vote)
- Continuous reinforcement makes it easier to achieve desirable behavior but is also more susceptible to extinction if the reinforcement is no longer provided.(2 votes)

- Does the variable times interval(variable interval) average also be same when the time interval was fixed (Fixed interval)? imitating what occurs in variable ratio at (5;40)(1 vote)
- Can variable ratio schedules be used in training kids in classrooms? if i have say 80 questions for training them in a specific subject, how can i use VR schedule to keep them motivated to learn?(1 vote)
- You could so by giving them such-and-so reward for an average of every five or so questions either answered or just answered.(1 vote)

- Hi I have a psychology final on tuesday 11/20/18 and I need these questions answered: When is the best time to use

partial reinforcement? and when is the best time to use continuous reinforcement?(1 vote)- Continous to establish behavior and partial/ intermittent to maintain the behavior. Intermittent reinforcement also provides greater resistance to exctintion.(1 vote)

## Video transcript

Many of our behaviors are on a
partial reinforcement schedule. Partial reinforcement
refers to a situation in which a behavior is
reinforced only some of the time. These partial schedules
of reinforcement are important because they
are generally more resistant to extinction than
continuous reinforcement. As we discussed,
behaviors are shaped through a process of
continuous reinforcement of successive approximations
of the target behavior. However, continuous
reinforcement eventually becomes less reinforcing. So there's a need for
these partial schedules of reinforcement, which
vary in their ability to maintain learned behaviors. And these were actually
discovered by B.F. Skinner through observation of reward
schedules with animals. However, they apply
to humans too. So there are four schedules
of partial reinforcement and each one has
a different effect on controlling and
maintaining behaviors. As you're watching
this video, you'll probably think of
situations in your life where your behavior
was reinforced on each of these schedules. And by the end of
the video, you'll be able to label those
situations with the terminology used in operant conditioning. So here you can see
the four schedules of partial reinforcement. If these terms are new
to you, don't worry. They'll start to
make a lot more sense when you break them apart. So for our purposes, I want
you to associate the word "ratio" with amount
of responses. The word ratio looks
similar to ration. And a ration of
food, for example, is a certain amount of food. So ratio means amount. And for our purposes,
an amount of responses. Now when you see
the word "interval," I want you to associate the
word interval with time. Think of that phrase,
an interval of time, like maybe a long
interval of time passed, or we were only given
a short interval of time to answer the question. So interval means time. So here's ratio, which means
the amount of responses. And here's interval,
which means time. Now, each of these
categories can either be fixed or variable, meaning
that they can be fixed, as in consistent, or variable,
as in there is a variation. And if you combine
these words together, you come up with
the four schedules of partial reinforcement,
fixed ratio, fixed interval, variable ratio, and
variable interval. So let's talk about a
fixed ratio schedule. So pretend this
car salesmen gets a bonus for every
five cars he sells. That bonus is a reinforcer
placed on a fixed ratio schedule. I a fixed ratio schedule,
reinforcement only occurs after a fixed
number of responses. So this car salesman has
to sell five cars in order to get a bonus. And if he sells five cars in
a week, he'll get a bonus. If he sells five cars in a
month, he'll get a bonus. And if he sells five cars
in a day, he'll get a bonus. You get the idea. What I'm illustrating here
is that the reinforcement, in this case the bonus, is
contingent on the number of cars he sells, regardless of
how long it takes him to do it. So since the only barrier
between the car salesman and his bonus is the
number of cars he sells, you might imagine he'll
work at a furious pace to earn as many
bonuses possible. Jobs that demand you to work at
a fast paced manner like this, they often pay their employees
on a fixed ratio schedule. Think of like factory
workers and fruit pickers, for instance. That's the benefit of
a fixed ratio schedule. It tends to emit a
high rate of behavior because the frequency of getting
the reward with a reinforcer pretty much depends
on the person. So that's a fixed
ratio schedule. Now, let's talk about a
fixed interval schedule. So pretend this car salesman
receives a paycheck every two weeks as long as
he sells one car. The paycheck is on a
fixed interval schedule because the reinforcement
occurs after a consistent amount of time has passed, in
this case two weeks. So in this case,
his paycheck doesn't change if he sells
one car or 100 cars during that time interval. So his paycheck is dependent on
the amount of time that passes. So as you might
imagine, he probably doesn't have much
of an incentive to sell more than
one car if he'll make the same amount
of money anyway. And that's the classic rate of
responding for fixed interval schedules. It's much slower than, say,
the fixed ratio schedule like we discussed earlier. Now, here we have the
variable ratio schedule. A variable ratio schedule
means that the reinforcer is delivered after an average
number of correct responses has occurred. So a variable ratio schedule
is similar to a fixed ratio schedule except the
number of responses needed to receive the reinforcement
changes after each reinforcer is presented. So put simply, a
variable ratio schedule is literally a series
of fixed ratio schedules that just change. What matters in the end
is the average number of correct responses. So using an example we used
for the fixed ratio schedule, there's a car salesmen receiving
a bonus for every five cars he sells. If the number needed to receive
a bonus was always fix at five, then that would be a
fixed ratio schedule. But a variable ratio
schedule would vary. So maybe he must
sell like five cars to get the first bonus,
and then three cars to get the second bonus,
and then seven cars to get the third bonus, and then
six cars to get the fourth one, and then four cars to
get the fifth bonus. If you add up all the
cars sold and then divide it by the five
bonuses he received, you'd find out that the
average number of cars sold to receive a bonus is five,
which is what the fixed ratio schedule was above,
five cars per bonus. The difference here was that
the variable ratio schedule has a lot of uncertainty. The car salesman cannot predict
when he'll receive a bonus in this case. But with every
car sold, he comes closer to getting that bonus. So the classic example used when
it comes on a variable ratio schedule is a slot machine. If you ever played
a slot machine, you understand the power of
a variable ratio schedule. A slot machine is
programmed to pay out after an average
number of pulls. But since you never really know
when the payout will occur, you keep playing and hoping
that you'll win something. And that's one
reason it's so hard to walk away from
a slot machine. You always wonder
like what would happen if the next
pull is the jackpot. You don't want to
miss that, obviously. So a slot machine is a
variable ratio schedule because the reinforcement is
dependent on your behavior. That is, you have to bet money
and pull a lever in order to have a chance to
receive anything. And it doesn't matter
how long you wait in between pulling that lever. It simply matters
whether you performed the behavior of pulling a lever. So that is a variable
ratio schedule. A variable interval
schedule means that the responses
are reinforced after a variable amount
of time has passed. So to use the car
salesman again, imagine a supervisor
randomly showing up without notice to
give him a bonus. As long as the supervisor
sees him actively talking to a customer,
he'll give him that bonus. Now since a car salesman never
knows when a supervisor will drop by, he has to consistently
engage with a customer in order to increase the chance
that a supervisor will notice. The difference here
is that he could have sold one car or even 100
cars on average that month. But all that matters in
order to receive the bonus is whether or not he's
actively engaging in a sale when his supervisor
happens to come by. So a variable interval
schedule like this results in a more regular rate
of responding than does a fixed interval schedule. So those are the four types
of partial reinforcement schedules. And as you might
imagine, each one has a tendency to emit a
different pattern of response.