Chapter 7 – Instrumental Conditioning: Motivational Mechanisms

Slides:



Advertisements
Similar presentations
Schedules of reinforcement
Advertisements

Classical Conditioning in Dating
January 26 th, 2010 Psychology 485.  History & Introduction  Three major questions: What is learned? Why learn? How does learning happen?
Transposition: Spence’s Behavioral Solution Transposition seems to support the cognitive view that organisms learn by discovering how things are related,
Lectures 14: Instrumental Conditioning (Basic Issues) Learning, Psychology 5310 Spring, 2015 Professor Delamater.
Instrumental Conditioning Also called Operant Conditioning.
Learning Unit 5. Topics in Learning Unit Defining Learning Classical Conditioning Operant Conditioning Cognitive Learning.
PSY402 Theories of Learning Chapter 9, Theories and Applications of Aversive Conditioning.
Lecture 17: Instrumental Conditioning (Associative Structures) Learning, Psychology 5310 Spring, 2015 Professor Delamater.
Classical Conditioning Pavlov’s experiment - psychic secretions. Pavlov was a Russian physiologists who studied digestion. He won the Nobel prize in physiology.
Learning How do we learn through our environment? Classical Conditioning – Neutral stimulus acquires ability to produce a response Operant Conditioning.
Chapter 9 – Extinction of Conditioned Behavior Outline –Effects of Extinction Procedures Decreased responding Increased variability in responding –Extinction.
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
PSY 402 Theories of Learning Chapter 9 – Motivation.
PSY 402 Theories of Learning Chapter 9 – Motivation.
PSY402 Theories of Learning Wednesday, November 19, 2003 Chapter 6 -- Traditional Theories (Cont.)
PSY 402 Theories of Learning Chapter 10 – A Synthetic Perspective on Instrumental Learning.
Chapter 4 Classical Conditioning: Mechanisms
Theoretical Approaches and questions in Operant Conditioning Psychology 3306.
Classical Conditioning, Operant Conditioning, and Modeling,
Ninth Edition 5 Burrhus Frederic Skinner.
Operant Conditioning: Schedules and Theories of Reinforcement
EXPLORING PSYCHOLOGY EIGHTH EDITION IN MODULES David Myers PowerPoint Slides Aneeq Ahmad Henderson State University Worth Publishers, © 2011.
© 2013 by McGraw-Hill Education. This is proprietary material solely for authorized instructor use. Not authorized for sale or distribution in any manner.
Operant Conditioning Unit 4 - AoS 2 - Learning. Trial and Error Learning An organism’s attempts to learn or solve a problem by trying alternative possibilities.
Chapter 7 Learning. Classical Conditioning Learning: a relatively permanent change in behavior that is brought about by experience Ivan Pavlov: – Noticed.
Operant Conditioning  B.F. Skinner ( ) elaborated Thorndike’s Law of Effect developed behavioral technology.
PSY 402 Theories of Learning Chapter 10 – A Synthetic Perspective on Instrumental Learning.
Reinforcement & Punishment: What is an S R ? Lesson 11.
CHAPTER 4 Pavlovian Conditioning: Causal Factors.
Learning. A. Introduction to learning 1. Why do psychologists care about learning? 2. What is and isn’t learning? IS: A relatively permanent change in.
Psychology 2250 Last Class Characteristics of Habituation and Sensitization -time course -stimulus-specificity -effects of strong extraneous stimuli (dishabituation)
Behavioral Approaches to Personality What is behavior?
Neobehaviorists. Neobehaviorism Life after Watson Life after Watson Optimism Optimism But…. But….
PSY402 Theories of Learning Chapter 6 – Appetitive Conditioning.
Copyright McGraw-Hill, Inc Chapter 5 Learning.
LEARNING  a relatively permanent change in behavior as the result of an experience.  essential process enabling animals and humans to adapt to their.
Learning Experiments and Concepts.  What is learning?
Instrumental Conditioning: Motivational Mechanisms.
Learning and Conditioning. I. The Assumptions of Behaviorism A. Behaviorists are deterministic. B. Behaviorists believe that mental explanations are ineffective.
Module 10 Operant & Cognitive Approaches. OPERANT CONDITIONING Operant conditioning –Also called _________________________________ –Kind of learning in.
Learning Chapter 5.
Theoretical Approaches and Questions in Operant Conditioning Psychology 3306.
Theories of Reinforcement Why is a reinforcer effective? Why do reinforcers increase the probability of a response?
Steven I. Dworkin, Ph.D. 1 Basic Principles of Operant Conditioning Chapter 6.
Module 9 Classical Conditioning. THREE KINDS OF LEARNING Learning –A relatively enduring or permanent change in behavior that results from previous experience.
Instrumental Conditioning II. Delay of Reinforcement Start DelayChoice Correct Incorrect Grice (1948) Goal Reward or No Reward.
Chapter 11 Inferences and the Representations of Knowledge in Operant Conditioning.
PSY402 Theories of Learning Chapter 4 – Appetitive Conditioning.
Unit 3.  In Unit 2, we explored the many ways in which biological processes can influence behavior.  However, as future ABA practitioners, our foremost.
Chapter 6: Behaviorist and Learning Aspects This multimedia product and its contents are protected under copyright law. The following are prohibited by.
The Associative Structure of Instrumental Conditioning Simple, Binary Associations  S-R association.
Unit 1 Review 1. To say that learning has taken place, we must observe a change in a subject’s behavior. What two requirements must this behavioral change.
Extinction of Conditioned Behavior Chapter 9 Effects of Extinction Extinction and Original Learning What is learned during Extinction.
Unlearned Reinforcers and Aversive Conditions Chapters 9.
Chapter 7 The Associative Structure of Instrumental Conditioning.
CHAPTER 5: Behavioral and Social Theories of Learning © (2015, 2012, 2009) by Pearson Education, Inc. All Rights Reserved Educational Psychology: Theory.
Chapter 5 Learning. What is Learning?  A relatively permanent change in behavior that results from experience  Learning is adaptive  Three major types.
Basic Learning Processes Robert C. Kennedy, PhD University of Central Florida
Chapter 6 LEARNING. Learning Learning – A process through which experience produces lasting change in behavior or mental processes. Behavioral Learning.
Conditioning and Learning Unit 6 Conditioning and Learning Modules
Becoming Extinct. Extinction of Conditioned behavior Extinction is a typical part of the learning process – Pavlovian or Instrumental contingency can.
S-R; S-O; S-(R-O); OH! OH!. What motivates and directs instrumental behavior? Two different approaches: – Associative Structure used by Thorndike and.
Allocating your Behavior. The Response Allocation Approach There are many possible activities that you could engage in – Sleeping, eating, drinking, sex,
Context Cues and Conditional Relations
Choice Behavior One.
Theories of Reinforcement
PSY402 Theories of Learning
Learning.
Theories of Reinforcement
Presentation transcript:

Chapter 7 – Instrumental Conditioning: Motivational Mechanisms Outline The Associative Structure of Instrumental Conditioning S-R association and the Law of Effect S-O Association expectancy of reward R-O relations in Instrumental Conditioning Behavioral Regulation Early Behavioral Regulation theories Consummatory-Response Theory The Premack Principle The Behavioral Bliss Point

What Motivates Instrumental Responding? Two different perspectives. 1. The associative structure of instrumental conditioning A molecular perspective Similar to the tradition of Pavlov Relationships among specific stimuli 2. Behavioral Regulation A molar perspective Skinnerian tradition Concerned with how instrumental conditioning sets limits on the organisms free flow of activity

The associative structure of instrumental conditioning Thorndike Instrumental conditioning involves more than just a response and reinforcer It occurs in a specific context (S) Three events 1) Stimulus context (S) 2) The instrumental response (R) 3) The response outcome (O) can be associated in a variety of ways. Figure 7.1

Figure 7.1 – Diagram of instrumental conditioning

The S-R Association and the Law of Effect Behaviors that are followed by a satisfying state of affairs become more probable. Behaviors that are followed by an annoying state of affairs become less probable Thorndike thought that the key association was the S-R association. The role of the outcome (O) was to stamp in the association between the contextual cues (S) and the instrumental response (R) instrumental conditioning did not involve learning about the reinforcer (O), or the relationship between R-O.

This view was hit pretty hard by the cognitive revolution. Thorndike did not believe that animals “knew” why they were running the maze (or pressing the lever) They don’t “expect” reward. behaviors were robotic (stamped in) by O (the reinforcer). This view was hit pretty hard by the cognitive revolution. Some resurgence in subcategories of human behavior Habit formation Drugs Infidelity gambling Context (S) can induce drug seeking (R) The important point is that from an S-R perspective the response is automatic Out of their control

Expectancy of Reward and the S-O association Clark Hull (1931) Kenneth Spence (1956) Thought that animals may come to expect reward Expectancy perhaps established through Pavlovian Conditioning  

Perhaps organisms learn two things about the Stimulus (S) Two-Process theory 1) S comes to evoke the response directly by association with R S-R association O (RF) stamps in R in the context of S 2) Instrumental Activity also comes to be made in response to expectancy of reward S-O association. S  Food CS US

Figure 7.1 – Diagram of instrumental conditioning

Modern Two-Process Theory (Rescorla & Soloman, 1967) There are two distinct kinds of learning Pavlovian Instrumental They are related, however, in a special way   During Instrumental conditioning As S-R learning progresses a Pavlovian process kicks in S becomes associated with O S (context) --------- O(response outcome) = Emotion Chamber ------- Food = Hope maze ------------ Shock = Fear

This S-O association further motivates responding. Implication rate of instrumental responding will be modified by the presentation of a classically conditioned stimulus. Tone  Food = hope Making the tone a CS+ for food   Presentation of a food CS+ while an animal is responding for food RF should increase hope and thus increase response rate

Results Consistent with Modern Two-Process Theory Pavlovian-Instrumental Transfer Test Phase 1 Instrumental training Barpress  food Phase 2 Pavlovian training CS – US Tone - Food Phase 3 Transfer phase CS from phase 2 is periodically presented to observe its effect on barpressing. If two process theory is correct when should animals respond the fastest?

Table 7.1 – Experimental Design for Pavlovian Instrumental Transfer Test

Does this procedure look familiar? Conditioned emotional response Conditioned suppression Pavlovian fear conditioning to the tone disrupted Instrumental responding Thus two-process theory works in either case Positive emotions increase motivation to respond when good outcome Negative emotions decrease motivation to respond when bad outcome

R-O Relations Thorndike’s S-R explanation of instrumental responding and Two-Process theories ignore R-O Relations Common sense implies that animals may associate outcomes with particular responses Push button on remote  expect visual reward Open door on fridge  expect food reward

Evidence for R-O relations Example: Colwill and Rescorla (1986) Outcome devaluation studies Example: Colwill and Rescorla (1986) Phase 1 Train rat to push a vertical rod Left (VI 60s) = food pellets Right (VI 60s) = sugar solution Phase 2 Devalue food or sugar (depending on rat) Sugar  LiCl Test Which way does the rat push the bar? The response is altered by changing the value of the outcome. Implies that animals expect that outcome when they make the response. An R-O relation Don’t want sugar so make the response associated with food

Behavioral Regulation This view of instrumental behavior is quite different from the associative account we just discussed. Does not focus on molecular stimuli how does reinforcement of responding in the presence of a particular stimuli affect behavior? The focus is molar how do instrumental contingencies put limitations on an organisms activity and cause redistributions of those activities?

Early Behavioral Regulation Theories Consummatory Response Theory Sheffield Is it the food that is reinforcing or the behavior (eating) that is reinforcing?  Consummatory responses Chewing, licking, swallowing Consummatory responses are special Represent consumption (or completion) of an instinctive behavior sequence. Getting food and then consuming it. fundamentally different from other instrumental behaviors, such as running, jumping, or lever pressing. A big change in the view of RF RF no longer a stimulus RF is a behavior

David Premack According to Premack disagreed with Sheffield consummatory responses are not necessarily more reinforcing than other behaviors  According to Premack consummatory responses are special only because they occur more often than other behaviors (e.g., lever pressing) Free environment with a lever and food A rat that knows nothing about lever pressing (naïve) is likely to spend more time eating than pressing the lever

The Differential Probability Principle Premack Principle Of any two responses the more probable response will reinforce the less probable one. Two responses of different probabilities H – high likelihood L – low likelihood The opportunity to perform H after L will result in reinforcement of L LH reinforces L The opportunity to perform L after H will not result in reinforcement of H HL does not reinforce H

Simply get a baseline measurement of both activities. Behaviors that an animal does a lot, will reinforce behaviors that an animal does not perform as much. strictly empirical. does not posit that some behaviors are enjoyed more than others.   Simply get a baseline measurement of both activities. A kid may engage in video game playing behavior quite often, but engage in homework activity much less.

If you make access to the video game contingent on homework activity do you think that home work activity will increase? Do homework  get to play video games? If you make homework activity contingent on video game activity do you think that video game activity will increase? Play video games  get to do homework?

Premack deprived rats of water Empirical Evidence Premack deprived rats of water if given a choice between water and running in a wheel the rat would now spend more time drinking water What if you make water drinking activity contingent on running in a wheel? The rat runs in the wheel more than it normally would. What if you could make running in a wheel more valuable than water? How would you do this? Allow the rat all the water it wants Restrict the opportunity in a wheel. Now make access to the running wheel contingent on drinking water. what happens? the rats drink three times as much water as the baseline rate

Premack principle in kids first graders eat candy or play pinball get the baseline some prefer candy, some prefer pinball How would Premack increase pinball playing for children who preferred to eat candy? Make access to candy contingent on playing pinball Play pinball get candy How would Premack increase candy eating for children who preferred to play pinball? Make access to the pinball machine contingent on eating candy Eat candy get to play pinball

Figure 7.5 – Diagram of Premack’s (1965) study.

What is nice about Premack’s theory is that it is strictly empirical. it contains no hypothetical constructs. No references to unobservables like hunger No reference to pleasurable vs. nonpleasurable things.

The Behavioral Bliss Point If we have several activities that we can engage in we distribute our behavior among those activities in a way that is optimal The bliss point can be determined like Premack did Time spent engaging in each activity Student Time spent watching TV Time spent studying

Figure 7.8 – Allocation of behavior between watching TV and studying.

The line in Fig 7.8 represents an instrumental contingency. In Figure 7.8 the students Bliss point is to spend much more time watching TV (60m) than studying (15m) The line in Fig 7.8 represents an instrumental contingency. Now the student is only allowed to watch TV for the same amount of time that they study They can no longer achieve the Bliss Point They will now redistribute their behavior

How do they redistribute? Must make a compromise Minimum-deviation model (Staddon) The rate of one response is brought as close to its preferred level as possible without moving the other response too far away from its preferred level Filled circle on Fig. 7.8 37.5 minutes of each activity 22.5 more minutes of studying 15 + 22.5 = 37.5 studying 22.5 less minutes of TV = 37.5 TV 60 - 22.5 = 37.5 TV

Application of Bliss-Point to Behavior Therapy Figure 7.9 Left to his own devices the child likes a lot of social RF from parents, while eliciting very few positive behaviors Bliss point The parents have been trying to RF positive behaviors, so they provide social rewards only after the child has engaged in two positive behaviors (2:1 ratio) Dotted line If not going well a therapist might be tempted to tell the parents to RF every positive behavior (1:1 to ratio) Solid line

Figure 7.9 – Hypothetical data on parental social reinforcement and positive child behavior.

Note - the minimum-deviation model actually predicts fewer positive behaviors after RF is increased The two solid dots Certainly an important consideration Things are not always as simple as they seem.