Learning Principles Unit 4

Pataasin ang iyong marka sa homework at exams ngayon gamit ang Quizwiz!

On a fixed interval (FI) schedule, what is the "rule" that describes whether a particular response will produce the reinforcer? (see book pp. 200-202)

First response after 15 seconds have passed means reinforcer

On a ____ (FI) schedule of reinforcement, each reinforcer becomes available after a constant, unchanging amount of time has passed.

Fixed Interval

Review the four simple schedules of reinforcement

Fixed ratio (FR) Variable Ration (VR) Fixed interval (FI) Variable Interval (VI)

On a ____ (FR) schedule of reinforcement, the animal must perform the response a constant, unchanging number of times for each reinforcer.

fixed ratio

Whenever Chuck mispronounces a word, his wife, Evelyn, calls him an "idiot." However, despite this, Chuck still mispronounces words just as frequently as he always has (i.e., there is no evidence that Evelyn's attempts to reduce the frequency of Chuck's mispronunciations have been effective). We can therefore conclude that: A) Chuck's behavior has not been successfully punished B) Chuck is deliberately mispronouncing words in order to annoy Evelyn C) Deep down, Evelyn unconsciously likes it when Chuck mispronounces words D) Evelyn's attempts to positively punish Chuck's behavior have actually resulted in negative punishment

A) Chuck's behavior has not been successfully punished

The schedule of reinforcement that you would use if you wanted to produce the most rapid learning of a new behavior would be: A) continuous reinforcement B) intermittent reinforcement C) a fixed interval schedule D) a variable ratio schedule

A) continuous reinforcement

Punishers are most likely to be effective when they are: A) delivered on a continuous schedule B) initially mild, then progressively more intense C) not contingent on the performance of any particular behavior D) (all of the above)

A) delivered on a continuous schedule

According to the _____ hypothesis, behaviors are more resistant to extinction following intermittent reinforcement than they are following continuous reinforcement because the animal finds it more difficult to distinguish between intermittent reinforcement and extinction than it does to distinguish between continuous reinforcement and extinction. A) discrimination B) frustration C) sequential D) response unit

A) discrimination

Jessica often whines for extra helpings when she is given dessert at the dinner table. In order to put an end to the whining, Jessica's father decides that, whenever Jessica whines, she will not receive extra dessert. (In other words, Jessica's father has decided to stop reinforcing Jessica's whining behavior.) The strategy that Jessica's father is attempting to implement would be best described as a(n) _____ procedure. A) extinction B) positive punishment C) response prevention D) negative reinforcement

A) extinction

John spent his summer picking cantaloupes for a farmer. The farmer paid John $1 for every 20 cantaloupes he picked. John worked on a _____ schedule of reinforcement. A) fixed ratio B) variable ratio C) fixed interval D) variable interval

A) fixed ratio

The reappearance, during extinction, of a previously reinforced (and extinguished) behavior is known as: A) resurgence B) ratio strain C) progression D) matching

A) resurgence

The word "positive" in "positive punishment" refers to the fact that: A) something is added as a consequence of the behavior B) something positive is removed as a consequence of the behavior C) the procedure has beneficial results D) the procedure is used with good intentions

A) something is added as a consequence of the behavior

In a "cumulative record," the horizontal (X) axis represents _____, and the vertical (Y) axis represents _____. A) the passage of time; the total number of times the behavior has occurred B) the total number of times the behavior has occurred; the passage of time C) the number of responses required for each reinforcer; the magnitude of the reinforcer D) the magnitude of the reinforcer; the number of responses required for each reinforcer

A) the passage of time; the total number of times the behavior has occurred

"Differential reinforcement of ____ behavior" (DRA) is a technique that involves extinguishing an unwanted behavior while reinforcing another, more appropriate behavior.

Alternative

How is the sequential hypothesis similar to the frustration hypothesis? How is it different?

An animal on a partial reinforcement schedule learns to associate its memories of non-reinforced responses with the eventual delivery of the reinforcer.

Which of the following schedules does NOT require the performance of a particular behavior? A) Continuous reinforcement B) A fixed time (FT) schedule C) A fixed interval (FI) schedule D) A fixed duration (FD) schedule

B) A fixed time (FT) schedule

Which of the following has NOT been shown to increase the effectiveness of a punisher? A) Providing a verbal explanation for why the punisher was administered. B) Using a mild punisher at first and then gradually increasing its intensity. C) Strengthening the contingency between the response and the punisher. D) Offering an alternative way to obtain the reinforcer that previously maintained the behavior.

B) Using a mild punisher at first and then gradually increasing its intensity.

"Given a choice between two alternative behaviors, each on its own schedule of reinforcement, animals tend to distribute their responses in a way that corresponds to the frequency with which reinforcers are available on each schedule." This statement best describes the: A) "break point" B) matching law C) response unit hypothesis D) partial reinforcement effect (a.k.a. partial reinforcement extinction effect)

B) matching law

Delaying the delivery of a punisher is likely to: A) increase its effectiveness B) reduce its effectiveness C) have no impact on its effectiveness D) have a variable and unpredictable impact on its effectiveness

B) reduce its effectiveness

The term "_____" refers to the rate at which a behavior occurs once it has resumed following reinforcement. A) latency B) run rate C) cumulative record D) post-reinforcement pause

B) run rate

Why can responding on a fixed interval schedule be described as having a "scalloped" pattern?

Because it's more gradual, these curves are scalloped shaped

Why is continuous reinforcement an example of a fixed ratio schedule?

Because you keep doing the response in order to get the rewards

When a previously reinforced behavior first stops getting reinforced, the immediate effect is often an abrupt increase in that behavior. This is called an "extinction ____."

Burst

Which of the following best describes the pattern of responding that variable ratio schedules tend to produce? A) A high run rate followed by long pauses after reinforcers B) A slow run rate followed by long pauses after reinforcers C) A high, steady rate of responding with little or no pausing after reinforcers D) A slow, steady rate of responding with little or no pausing after reinforcers

C) A high, steady rate of responding with little or no pausing after reinforcers

Often, the initial effect of extinction is an increase in the frequency of the behavior that is being extinguished. This phenomenon is called: A) ratio strain B) stretching the ratio C) an extinction burst D) spontaneous recovery

C) an extinction burst

David stayed out too late one evening and missed his curfew. After being punished by his parents, David walked into his younger brother's room, picked a fight, and then physically attacked his brother. This example best illustrates that _____ can occur as a "side effect" of punishment. A) conditioned fear B) extinction bursts C) displaced aggression D) learned helplessness

C) displaced aggression

When reinforcers are contingent on the continuous performance of a behavior for a certain period of time, a _____ schedule of reinforcement is in effect. A) fixed time (FT) B) fixed interval (FI) C) fixed duration (FD) D) progressive ratio (PR)

C) fixed duration (FD)

A cumulative record tends to show a "scalloped" pattern of responding when the subject is on a _____ schedule of reinforcement. A) fixed ratio B) variable ratio C) fixed interval D) variable interval

C) fixed interval

True or false? On a fixed interval schedule, reinforcers occur periodically regardless of what the animal does.

False

In a _____ schedule, a behavior is under the influence of different schedules at different times, and each schedule is associated with a particular stimulus. However, in _____ schedules, two or more schedules are available at once (e.g., a pigeon may have the option of pecking one key on a VI 20-sec. schedule or pecking a different key on a VI 40-sec. schedule). A) mixed; tandem B) tandem; mixed C) multiple; concurrent D) concurrent; multiple

C) multiple; concurrent

The more work required for each reinforcer, the longer the post-reinforcement pause. Thus, pauses are longer in an FR 100 schedule than in an FR 20 schedule. Research with "multiple" schedules of reinforcement has shown that these pauses do not occur due to fatigue; rather, the length of the pause depends on the size of the upcoming ratio (not the size of the just-completed ratio). Because of this, "post-reinforcement pauses" are now often referred to as "_____." A) bliss points B) rest periods C) pre-ratio pauses D) variable intervals

C) pre-ratio pauses

For several days, a rat in a Skinner box has received food after every 10 presses of a bar. However, when the researchers suddenly increase the number of required bar-presses from 10 to 200, they observe a severe disruption of the rat's bar-pressing behavior. This phenomenon is known as: A) regression B) resurgence C) ratio strain D) the partial reinforcement effect (a.k.a. partial reinforcement extinction effect)

C) ratio strain

George is training a pigeon to peck a key. He begins by reinforcing each key-peck that the pigeon performs. Once the response is learned, however, George stops reinforcing every key-peck and begins reinforcing every other key-peck. Later, George reinforces every third key-peck, then every fifth key-peck, then every tenth key-peck, and so on. Because he is gradually increasing the requirements for each reinforcer, George is using a procedure that researchers call "_____." A) matching B) resurgence C) stretching the ratio D) continuous reinforcement

C) stretching the ratio

Suppose that Shawna is a bowler who gets a "strike" on approximately one out of every three attempts. However, sometimes, Shawna bowls two or three strikes in a row, and other times, she goes four attempts or more without bowling a strike. This is an example of a _____ schedule of reinforcement. A) fixed ratio B) continuous C) variable ratio D) fixed interval

C) variable ratio

In ____ schedules, two or more schedules are available at once (e.g., a pigeon may have one key that it can peck on a FR 10 schedule and another key that it can peck on a VR 15 schedule).

Concurrent

Immediate punishers are more effective than delayed punishers. This illustrates the importance of the ____ between the response and the punisher.

Contiguity

There is a "____" between a response and a punisher if performing the response makes the punisher more likely to occur.

Contingency

In ____ reinforcement, a behavior is reinforced every time it occurs.

Continuous

continuous reinforcement vs. Partial Reinforcement: Which one results in faster acquisition of new behaviors?

Continuous

Robert was severely scolded by Mrs. Valentine for asking inappropriate questions during class. Which of the following problem(s) could arise as a result of Mrs. Valentine's use of punishment? A) Robert might become fearful of Mrs. Valentine B) Robert might begin to scold his siblings and peers C) Robert might refuse to say anything at all during class D) (all of the above)

D) (all of the above)

In one alternative to punishment, called "_____," the environment is altered in such a way as to block the occurrence of an unwanted behavior. A) extinction B) "Time Out" C) penalty training D) response prevention

D) response prevention

Behaviors tend to be more resistant to extinction following intermittent reinforcement than they are following continuous reinforcement. This is known as: A) spontaneous recovery B) the progressive ratio effect C) the extinction burst phenomenon D) the partial reinforcement effect (a.k.a. partial reinforcement extinction effect)

D) the partial reinforcement effect (a.k.a. partial reinforcement extinction effect)

A hungry pigeon is in a Skinner box and is pecking a key for access to food. The first food delivery occurs for the first peck after 1 minute has elapsed. The second food delivery occurs for the first peck after 3 additional minutes have elapsed. The third food delivery occurs for the first peck after 2 additional minutes have elapsed. This pigeon is being reinforced on a _____ schedule of reinforcement. A) continuous B) variable ratio C) fixed interval D) variable interval

D) variable interval

What is an "extinction burst," and why can it be a problem in behavioral therapy?

Definition: a sudden increase in the rate of behavior during the early stages of extinction. Problem: when attempting to extinguish an undesirable behavior, the behavioral problem may initially get worse

According to the ____ hypothesis, extinction takes longer after partial reinforcement because it is more difficult to distinguish between partial reinforcement and extinction than it is to distinguish between continuous reinforcement and extinction.

Discrimination

Review the "3-term contingency" that operant conditioning involves

Discriminative stimulus, Response, outcome

An animal experiences "____" when a previously reinforced behavior stops being reinforced.

Extinction

How do "fixed" schedules differ from "variable" schedules?

Fixed schedules happen at the same time. In Variable, the reward will be given after the animal gives the response.

The ____ hypothesis claims that, during partial reinforcement, animals learn to associate their emotional reaction to nonreinforced responses with the eventual delivery of the reinforcer.

Frustration

How does the "matching law" suggest that an animal will behave when it is given a choice between two behaviors, each of which is on its own schedule of reinforcement?

Given a choice situation involving 2 alternatives, the frequency with which the animal chooses one alternative will "match" the frequency with which the animal can obtain reinforcers via this alternative. Know the basic equation that summarizes the matching law (see below)

What is represented on the horizontal and vertical axes of the cumulative record?

Horizontal: the passage of time Vertical: total responses

is Providing a verbal rationale for why the punisher occurred helpful in punishment?

If a verbal reprimand was given, organisms are less likely to do it again.

How can "two-process theory" be applied to punishment?

If an animal performs a response and then experiences the punisher, It associates the stimuli that were present before it performed the response with the aversive aspects of the punisher. So, the rat stops pressing the lever, because it fears the lever

How does the intensity of the punisher tend to affect the reduction of the punished response?

If an organism is exposed to a mild shock first, their response to the larger shock is less so then those who weren't exposed to a previous shock. The more intense punisher seemed worked well. Decreasing the behavior and producing more of a suppression then mild punisher

How is resistance to extinction related to the predictability of reinforcers?

If it's more predictable, you are more likely to give up once it stops the first time. Whereas if it varies, you will keep trying because you don't know when it will produce the reinforcement.

How did Jenkins (1962) cast doubt on the discrimination hypothesis?

If partial reinforcement is followed by continuous reinforcement, animals still display greater resistance to extinction

How does continuous reinforcement differ from partial (or "intermittent") reinforcement?

In continuous rein. Every response gets reinforced, while in intermittent rein. Not all responses get reinforced.

continuous reinforcement vs. Partial Reinforcement: Which one results in behaviors that are more "resistant to extinction"?

Intermittent

How does the variability of the animal's behavior change during extinction?

It is increased! Animals seems to realize, during extinction, that it needs to "try something else."

What effect (if any) does the size of the ratio have on the run rate and the length of post-reinforcement pauses?

Longer ratio means a longer post-reinforcement pause, but no effect on run rate.

"Differential reinforcement of ____" (DRL) is a technique that involves reinforcing a behavior only if a certain amount of time has passed since the behavior was last performed.

Low rate

What effect does extinction have on emotional behavior (e.g., aggression)?

May increase the frequency of it (ex) we pay for vending machine food, and it gets stuck. What do we do? Kick it)

What is the pattern of responding that VI schedules tend to produce?

Moderate response rate with steady responding

Which hypothesis tends to be supported by research?

Multiple schedule reinforcement

On a variable ratio (VR) schedule, what is the "rule" that describes whether a particular response will produce the reinforcer?

On average, reinforcer occurs after 10 responses - but, each individual reinforcer can occur after a different number of responses.

In ____ (or "intermittent") reinforcement, a behavior is reinforced on some occasions and not reinforced on other occasions.

Partial

How do "ratio" schedules differ from "interval" schedules?

Ratio describes how many while interval focuses on when it's happening.

The term "____ to extinction" refers to how long a behavior persists after it stops being reinforced.

Resistance

An alternative to punishment that involves altering the environment in such a way as to block an unwanted behavior from occurring.

Response prevention

According to the ____ hypothesis, the partial reinforcement extinction effect is merely a by-product of the fact that reinforcers require more responses on partial schedules than they do on continuous schedules.

Response unit

What effects does extinction have on the animal's behavior?

Responses become weaker, extinction burst

When one behavior is extinguished, previously extinguished behaviors may reappear. This phenomenon is known as "____."

Resurgence

What is a schedule of reinforcement? What are "schedule effects"? (see book p. 194)

Schedule of reinforcement: A rule that specifies which occurrence of a particular response will be followed by a reinforcer.

The ____ hypothesis claims that, during partial reinforcement, animals learn to associate their memories of nonreinforced responses with the eventual delivery of the reinforcer.

Sequential

Last name of the American psychologist who invented the "operant chamber" and conducted extensive research on schedules of reinforcement.

Skinner

What does the slope (steepness) of the line (on a cumulative record) tell us about the animal's behavior?

Steep line means the animal is responding rapidly Flatter line means the animal is responding more slowly

Why might immediate punishers be more effective than delayed punishers?

The "delay" could cause the animal to forget that it performed the behavior that it is getting punished for. When the punishment is delayed, the animal is doing other things, so it may think it's getting punished for these things

How does the discrimination hypothesis account for PRE?

The difference between continuous reinforcement and extinction is greater than the difference between partial reinforcement and extinction. So, it's more difficult for the animal to distinguish (discriminate) between partial reinforcement and extinction.

How does the frustration hypothesis explain the PRE?

The frustrated emotional state becomes a SD (discriminative stimulus) for the animal, it is a signal to it that if they are frustrated, the reward will eventually come.

On a variable interval (VI) schedule, what is the "rule" that describes whether a particular response will produce the reinforcer?

The interval changes from one reinforcer to the next

What happens (or, rather, does not happen) during "extinction" in operant conditioning?

The reinforcer no longer follows the response (even when SD is present)

On a fixed ratio (FR) schedule, what is the "rule" that describes whether a particular response will produce the reinforcer?

The reinforcer occurs after a certain number of responses have been made

What is the "partial reinforcement effect" (PRE; also known as the "partial reinforcement extinction effect," or PREE)?

The schedule of reinforcement that was in effect prior to extinction.

Why can the pattern of behavior on FR schedules be described as "stop-and-go"?

There will be a period where the animal stops responding to the reinforcer, but once responding resumes, it will continue at a high rate until the animal learns the next reinforcer.

What did Thorndike and Skinner conclude regarding the effectiveness of punishment?

They viewed that punishers didn't work. Thye thought that extinction was just as effective as punishment

How does the overall response rate differ on ratio vs. interval schedules?

This has a high response rate with steady responding

True or false? A continuous schedule of punishment is typically more effective than a partial schedule in weakening a behavior.

True

On a ____ (VR) schedule of reinforcement, the animal must perform the response a different, changing number of times for each reinforcer.

VAriable Ratio

Which type of schedule produces behavior that occurs at a steadier, more consistent rate (with less "pausing")—fixed or variable?

Variable

On a ____ (VI) schedule of reinforcement, each reinforcer becomes available after a different, changing amount of time has passed.

Variable Interval

What is a "cumulative recorder"? How does it work?

What it is: a device that was used to create graphs called "cumulative records" How it works: Anytime the animal poked a thing or press a bar, the pen would move.

What are "concurrent" schedules of reinforcement, and why are concurrent schedules often used in research on "choice"?

What they are: when an animal s presented with 2+ schedules Why are they used: Because two different schedules are in affect at the same time, so the animal is free to choose one over the other

An alternative to a punishment called "_____" involves extinguishing an unwanted behavior while also reinforcing a behavior that cannot be performed at the same time as the unwanted behavior. A) penalty training B) response prevention C) differential reinforcement of low rate (DRL) D) differential reinforcement of incompatible behavior (DRI)

differential reinforcement of incompatible behavior (DRI)

"Differential reinforcement of ____ behavior" (DRI) is a technique that involves extinguishing an unwanted behavior while reinforcing another behavior that cannot be performed at the same time as the unwanted behavior.

incompatable

The ____ is a principle of choice which says that, given the opportunity to respond on two or more schedules, the rates of responding on each schedule will match the rates at which reinforcers can be obtained on each schedule.

matching law

In a ____ schedule, a behavior is under the influence of two or more simple schedules at different times, with each schedule being associated with a particular stimulus.

multiple

What is a "post-reinforcement pause"?

no responses for a while after delivery of a reinforcer

What is the "run rate"?

once responding resumes, it continues at a high rate until the animal earns the next reinforcer.

What is "resurgence"?

the reappearance, during extinction, of other, previously reinforced (and extinguished) behaviors.

Which type of schedule produces behavior that is more resistant to extinction— fixed or variable?

variable


Kaugnay na mga set ng pag-aaral

Chapter 32 serious Mental Illness

View Set

Gastrointestinal Disorders (prep U additional)

View Set

Pathogenesis, Disease, and Epidemiology

View Set