Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
419 Cards in this Set
- Front
- Back
Environmental stimuli or events an organism escapes or avoids |
Aversive stimuli |
|
Behavioral contingency that decreases the rate of response |
Punishment |
|
An aversive stimulus acquired by a history of conditioning |
Conditioned aversive stimuli |
|
An event or stimulus that decreases the rate of operant behavior |
Punisher |
|
When the delivery of a stimulus or event decreases the rate of a response |
Positive punishment |
|
Positive punishment procedure that includes practicing an appropriate response multiple times |
Overcorrection |
|
A stimulus or event the removal of which decreases the rate of a response |
Negative punishment |
|
The contingent removal of access to positive reinforcers following a problem behavior |
Timeout procedure |
|
Negative punishment procedure where reinforcers are removed based on behavior |
Response cost |
|
Presentation of a lower frequency operant will punish a higher frequency behavior |
Relativity of punishment |
|
Maintenance of response suppression over time |
Permanence of punishment |
|
Time between shock presentations |
Shock-shock interval |
|
Time between a response and the presentation of a shock |
Response-shock interval |
|
Avoidance behavior emitted to a warning stimulus |
Discriminated avoidance |
|
Training avoidance with no presented warning stimulus |
Nondiscriminated (Sidman) avoidance |
|
Focus on small moment-to-moment relationships |
Molecular perspective |
|
Focus on large-scale factors that regulate responding |
Molar perspective |
|
Negative reinforcement of behavior that prevents/postpones avoidance contingencies |
Timeout from avoidance |
|
When an animal gives up avoiding or escaping an aversive situation |
Learned helplessness |
|
Aggressive responses elicited by an aversive stimuli |
Reflexive aggression |
|
Aggressive behavior reinforced by the removal of an aversive stimuli |
Operant aggression |
|
Person who delivers the punishment and the context becomes conditioned aversive stimuli |
Social disruption |
|
Use of punishment to get others to act as we like |
Coercion |
|
Aversive stimuli based on phylogeny |
Primary aversive stimuli |
|
Reflexive behavior that prevents the occurrence of operant behaviors |
Positive punishment elicits... |
|
The observed decrease in effectiveness of punishers not delivered immediately |
Delayed punishment |
|
Form of punishment to suppress self-injury where water is sprayed in the face of a participant |
Water misting |
|
Debate over if, when, and how punishment should be used |
Punishment debate |
|
Punishers that have to be used repeatedly |
Ineffective punishers |
|
What a social agent that frequently uses punishment becomes |
Conditioned punishing stimulus |
|
A neutral stimulus is paired with an aversive US and becomes a conditioned aversive stimulus |
Conditioned suppression |
|
A CS paired with a US interferes with a subsequent CS-US association |
Blocking |
|
When a CS that accompanies a drug is presented causing cravings |
Conditioned withdrawal |
|
The tendency for a system to remain stable |
Homeostasis |
|
When more of a US is needed to obtain the same effect |
Tolerance |
|
Pairing a second CS with an already functional CS rather than a CS and a US |
Second-order conditioning |
|
The US comes on and goes off before the CS comes on |
Backward conditioning |
|
The CS is presented and removed prior to presentation of the US |
Trace conditioning |
|
The CS and US are presented at the same time |
Simultaneous conditioning |
|
The CS is presented a few seconds before the US occurs |
Delayed conditioning |
|
An organism showing a CR to one stimulus but not another |
Respondent discrimination |
|
A graph that plots stimulus values against magnitude of response |
Generalization gradient |
|
When an organism shows a CR to values of the CS that were not trained |
Respondent generalization |
|
Observation of an increase in the CR after respondent extinction has occurred |
Spontaneous recovery |
|
Baseline; strength of the target response |
Respondent level |
|
Presenting the CS without the US repeatedly |
Respondent extinction |
|
The result of pairing a CS with a US over repeated trials |
Respondent acquisition |
|
A response elicited by the presentation of an arbitrary stimulus associated with a US |
Conditioned response |
|
An arbitrary stimulus that is associated with a US to elicit reflexive behavior |
Conditioned stimulus |
|
A life history that contributes to behavior |
Ontogenetic |
|
With repeated presentation of the US, the UR gradually declines in magnitude |
Habituation |
|
As the intensity of the US increases the latency to appearance of the UR decreases |
The law of latency |
|
As the intensity of the US increases so does the magnitude of the UR |
The law of intensity-magnitude |
|
A point below which no response occurs and above which a response always occurs |
The law of the threshold |
|
The phylogenetic behavior elicited by the unconditioned stimulus |
Unconditioned response |
|
The phylogenetically endowed eliciting event for a reflex |
Unconditioned stimulus |
|
Phylogentic patterns of behavior that are flexible and adaptable |
Modal action patterns |
|
Sequences of behavior that are phylogenetic in origin |
Fixed-action pattern |
|
Behavior relations that are based on the genetic endowment of organisms |
Phylogenetic |
|
A stimulus-response pairing theory of respondent conditioning |
Rescorla-Wagner model |
|
States how and when stimuli and behavioral consequences will be presented |
Schedule of reinforcement |
|
Behavior occurring at a steady operant level/rate |
Steady-state performance |
|
escribes what the experimenter does, not the behavior of the organism |
Mechner notation |
|
Every operant required by the contingency is reinforced |
Continuous reinforcement |
|
The increase in behavioral variability during extinction |
Resurgence |
|
Schedules of reinforcement based on the number of emitted responses |
Ratio schedules |
|
Schedules of reinforcement based on time since the last consequence occurred |
Interval schedules |
|
A schedule that delivers reinforcement after a fixed number of responses |
Fixed ratio |
|
Steep period of responding followed by reinforcement then a pause in responding |
Break-and-run |
|
The pause in responding following a consequence |
Postreinforcement pause |
|
A schedule in which the number of responses required for reinforcement changes after each reinforcer is presented |
Variable-ratio |
|
A schedule in which the operant is reinforced after a fixed amount of time has passed |
Fixed-interval |
|
A cumulative record pattern that shows an increasing response rate as reinforcement approaches (associated with fixed-interval schedules) |
Scalloping |
|
The idea that the effects of contingencies of reinforcement extend across species |
Assumption of generality |
|
Schedule in which the first operant to occur after a variable amount of time is reinforced |
Variable-interval |
|
Contingency where the reinforcer is only available for a set time after an interval schedule has timed out |
Limited hold |
|
A schedule where the number of responses required for reinforcement are increased systematically |
Progressive ratio schedule |
|
The highest ratio value completed on a progressive ratio schedule |
Breakpoint |
|
The period between initial steady state performance and the next steady state |
Transition state |
|
The time that passes between reinforcers |
Interreinforcement interval |
|
Systematic use of reinforcement to establish desired behavior |
Contingency management |
|
Analysis of the small moment-to-moment relations between behavior and consequences |
Molecular accounts of schedule performance |
|
The change in behavior allocation across time |
Behavioral dynamics |
|
Quick burst of response followed by a pause |
Pause-and-run pattern of behavior |
|
the number of responses emitted during extinction before a behavior disappears |
Resistance to extinction |
|
A schedule of reinforcement where a response-independent reinforcer is delivered after a set time |
Fixed time schedule |
|
Behavior that persists in the presence of a stimulus despite disruption |
Behavioral momentum |
|
A schedule of reinforcement where the previous response requirement is multiplied by a set number |
Geometric progressive ratio |
|
A schedule of reinforcement where a set number is added to the previous response requirement |
Arithmetic progression |
|
Published Schedules of Reinforcement with Skinner |
Charles Bohris Ferster |
|
The acquisition, maintenance, and change of an organism's behavior as a result of lifetime events |
Learning |
|
Behavior is due to an interaction between genetic influence and environmental experience |
Behavior theory |
|
Using experimentation to break down environment-behavior relations into principles of behavior |
The experimental analysis of behavior |
|
Experimental analysis that includes assumptions about how to study behavior, techniques, and practical implications |
Science of behavior |
|
A comprehensive natural science approach to studying the behavior of organisms |
Behavior analysis |
|
An organism will respond differently to two different situations |
The principle of discrimination |
|
When a behavior has been reinforced in one setting but not in the other |
Differential reinforcement |
|
The use of behavior principles to solve practical problems |
Applied behavior analysis |
|
When an organism learns new ways of behaving in reaction to environmental changes |
Conditioning |
|
A behavior that is elicited by a biologically relevant stimulus |
Reflex |
|
When a neutral or meaningless stimulus is paired with an unconditioned stimulus |
Respondent conditioning |
|
A behavior that is elicited by a conditioned stimulus |
Respondent |
|
Demonstrated respondent conditioning by pairing a dog's salivation with a bell |
Ivan Pavlov |
|
A change in operant response as a function of consequences |
Operant conditioning |
|
A behavior that operates on the environment to produce an effect |
Operant |
|
Applies at three levels: natural selection, selection by operant conditioning, and cultural selection |
Selection by consequences |
|
Immediately producing or resulting in the occurrence |
Immediate causation |
|
Explaining a phenomenon by pointing to remote events that make it likely |
Remote causation |
|
Changes in the interconnections of neurons due to experience |
Neuroplasticity |
|
The interrelationship of the endogenous opiate and dopamine systems |
Neural basis of reward |
|
The integration of the science of behavior with neuroscience |
Behavioral neuroscience |
|
The conditions, events, and stimuli arranged by other people that regulate human action |
Culture |
|
The intellectual force behind behavior analysis |
B. F. Skinner |
|
Focused on a stimulus-response approach to behavior |
John B. Watson |
|
In one of Watson's famous experiments he was conditioned to fear a white rat |
Little Albert |
|
Originated the Law of Effect, stating that behaviors become stronger and weaker based on consequences |
Thorndike |
|
Termed by Thorndike to describe behavior in his puzzle-box experiments |
Trial-and-error learning |
|
Behavior only accessible to the person doing it |
Private behavior |
|
Everything an organism does including private and covert actions like thinking and feeling |
Behavior |
|
Classifying behavior and analyzing the environment in terms of functions |
Functional analysis |
|
Refers to the effect produced by a behavioral or environmental event |
Function |
|
Behavior is categorized by age to infer stages of development |
Structural approach |
|
Physical form or characteristics of the response |
Topography |
|
Behavior that increases or decreases by the presentation of a CS that precedes the CR |
Respondent |
|
Respondent behavior that occurs by presentation of the CS or US |
Elicited |
|
Operant behavior that likely occurs by the presence of the stimulus |
Emitted |
|
Operates on the environment to produce a change, effect, or consequence |
Operant |
|
Refers to all the topographic forms of the performance with a similar function |
Response class |
|
Events and stimuli that change behavior |
Environment |
|
When the occurrence of an event changes the behavior of an organism |
Stimulus function |
|
An event or stimulus that has acquired a function based on respondent conditioning |
Conditioned-stimulus function |
|
An event that follows a response and increases its frequency |
Reinforcement function |
|
The ability of an antecedent event to set the occasion for a behavior |
Discriminative function |
|
An event that sets the occasion for behavior |
Discriminative stimuli |
|
Stimuli that vary physically but have a common effect on behavior |
Stimulus class |
|
Events that increase or maintain the rate of behavior when present |
Positive reinforcers |
|
Events that increase or maintain the rate of behavior when removed |
Negative reinforcers |
|
Environmental events that temporarily increase effectiveness of reinforcement or responses that produce reinforcement |
Establishing operation |
|
Environmental events that temporarily decrease the effectiveness of behavioral consequences and reduce behaviors that result in those consequences |
Abolishing operation |
|
An event that alters the reinforcement effectiveness of behavioral consequences |
Motivating operation |
|
Condition changed by the experimenter |
Independent variable |
|
The measured effect in an experiment |
Dependent variable |
|
An experimental method that demonstrates causation by turning the treatment on and off |
A-B-A-B reversal |
|
The A-phase of a study that measures behavior before manipulation |
Baseline |
|
Schedule-controlled behavior that is stable and does not change over time |
Steady-state performance |
|
A behavioral baseline that varies with small increases in the independent variable |
Baseline sensitivity |
|
A single individual is exposed to the independent variable |
Single-subject research |
|
Defines the relationship between the occasion, the operant class, and consequences |
Contingency of reinforcement |
|
Systematic rise or decline in the values of the scores |
Trend |
|
Behavior that operates on the environment to produce change |
Operant |
|
A set of responses that vary in topography but produce a common environmental consequence |
Operant class |
|
An event or stimulus that precedes an operant and sets the occasion |
Discriminative stimulus |
|
An event that precedes a response and produces a lower probability of emitting an operant |
S-delta |
|
A stimulus or event the presentation of which increases or maintains the rate of a response |
Positive reinforcement |
|
A stimulus or event the removal of which increases or maintains the rate of the response |
Negative reinforcement |
|
A stimulus or event the presentation of which decreases the rate of the response |
Positive punishment |
|
A stimulus or event the removal of which decreases the rate of the response |
Negative punishment |
|
A higher frequency behavior will function as reinforcement for a lower frequency behavior |
The Premack principle |
|
Differences in frequency of different responses in a free-choice setting |
Response hierarchy |
|
Behavior that produces the opportunity to engage in some activity |
Instrumental response |
|
Activity obtained by making the instrumental response |
Contingent response |
|
When access to the contingent behavior is restricted and falls below baseline |
Response deprivation |
|
Time from the onset of one event to the onset of another |
Latency |
|
Stamping in or out of a response coined by Thorndike |
Law of effect |
|
The number of responses in a specified interval |
Rate of response |
|
Reinforcement at the neuron level |
In-vitro reinforcement |
|
Organism is free to respond or not to respond over a period of time |
Free operant method |
|
Procedure for restricting access to a reinforcing event |
Deprivation operation |
|
Click of the feeder with the presentation of food |
Magazine training |
|
An event or stimulus that is effective because of the organism's life history |
Conditioned reinforcer |
|
Each response produces reinforcement |
Continuous reinforcement |
|
Differential reinforcement of successive approximations of a desired response |
Shaping |
|
The tendency for the effectiveness of a consumable reinforcer to decrease after repeated presentations |
Satiation |
|
Withholding reinforcement for a previously reinforced response |
Extinction |
|
A low rate of operant behavior as a function of S-delta |
Discriminated extinction |
|
The continuation of operant behavior when it is placed on extinction |
Resistance to extinction |
|
Only some responses are reinforced |
Intermittent schedule of reinforcement |
|
The more intermittent reinforcement, the greater the resistance to change |
Partial reinforcement effect |
|
An increase in the magnitude of a response after operant extinction has occurred |
Spontaneous recovery |
|
Species-characteristic behavior that becomes invasive during conditioning |
Instinctive drift |
|
Approaching a sign/stimulus that signals a biologically relevant event |
Sign tracking |
|
A respondent conditioning procedure that encourages operant behavior |
Autoshaping |
|
When a CS is paired with a US, the CS is said to substitute for the US |
Stimulus substitution |
|
Ontogenetic and phylogenetic histories of an organism |
Context for conditioning |
|
When a distinctive taste is paired with induced nausea or sickness |
Taste aversion learning |
|
When an association is more likely because of phylogenetic history |
Preparedness |
|
When a taste is preferred as a result of learning history |
Conditioned taste preference |
|
When a place is preferred as a result of learning history |
Conditioned place preference |
|
When a specific learning history causes a place to be avoided |
Conditioned place aversion |
|
Adjunctive behavior; behaviors that occur within the interreinforcement period |
Interim behavior |
|
Non-functional behaviors generated by properties of a reinforcement schedule |
Facultative behavior |
|
Stereotypical behaviors that occur as the occurrence of reinforcement approaches |
Terminal behavior |
|
Stereotypical behavior that occurs during the interreinforcement period of a schedule of reinforcement |
Schedule-induced behavior |
|
Excessive drinking |
Polydipsia |
|
An observed increase in physical activity following food restriction |
Activity anorexia |
|
A procedure that demonstrates the role of respondent conditioning in autoshaping |
Negative automaintenance |
|
An autoshaping procedure where key pecking cancels the delivery of food |
Omission procedure |
|
Instinctive processes activated by US |
Behavior system |
|
Changes in the interconnections of neurons due to experience |
Neuroplasticity |
|
Demonstrated the effects of reflexive behaviors on operant conditioning |
Breland & Breland |
|
Differential reinforcement of successive approximations of a desired response |
Shaping |
|
Demonstrated that respondent behavior is modifiable by reinforcers using curare to immobilize skeletal muscles |
Miller experiments |
|
Demonstrated taste aversion learning |
Garcia experiments |
|
An observed increase, peak, then drop in schedule-induced behavior |
Bitonic function |
|
Provided evidence of adjunctive behavior in people |
Hollis (1973) |
|
First researchers to demonstrate self-starvation in rats |
Routtenberg & Kuznesof (1967) |
|
Natural selection favored animals travelling when food was scarce |
Natural selection explanation of activity anorexia |
|
Demonstrated that respondent conditioning effects can outweigh reinforcement in negative auotmaintenance |
Williams & Williams (1969) |
|
Activities that occur in the natural environment that are unrelated to the current contingencies of reinforcement |
Displacement behavior |
|
A stimulus or event that alters the probability of an operant when presented |
Controlling stimulus |
|
A differential response to two or more stimuli |
Discrimination |
|
Reinforcement for any behavior other than a target operant |
Differential reinforcement |
|
Change in behavior that occurs when either an SD or S∆ is presented |
Stimulus control |
|
Two or more simple schedules are presented one after the other, each with a unique discriminative stimulus |
Multiple schedule |
|
A measure of the degree of stimulus control that compares the rate in the SD component to the sum of the rates in both SD and S∆ phases |
Discrimination index |
|
Behavior that is accidentally reinforced |
Superstitious behavior |
|
A schedule where reinforcers are provided for any response that is not the target behavior |
Differential reinforcement of other behavior |
|
The tendency for a change in one component of a schedule to affect the rate of response in the unchanged component |
Behavioral contrast |
|
When the response rate in the unchanged component increases as a result of a change in another component of a schedule |
Positive contrast |
|
When the response rate in the unchanged component decreases as a result of a change in another component of a schedule |
Negative contrast |
|
A change in the rate of response caused by a change in the upcoming reinforcement schedule |
Anticipatory contrast |
|
The spread of stimulus control from one stimulus to another |
Generalization |
|
When a reinforced operant is emitted in the presence of stimuli not present during training |
Stimulus generalization |
|
A graph that shows the relation between probability of response and stimulus value |
Generalization gradient |
|
Change in the peak of a generalization gradient away from an extinction (S-delta) stimulus |
Peak shift |
|
The probability of response is highest in the presence of the actual stimulus used in training |
Absolute stimulus control |
|
An organism responds to differences among the values of two or more stimuli |
Relative stimulus control |
|
A training procedure where only one controlling stimulus is present at a time |
Successive discrimination |
|
A training procedure where the SD and S∆ are presented at the same time |
Simultaneous discrimination |
|
A training procedure where the organism is not allowed to respond to the extinction stimulus |
Errorless discrimination |
|
Transferring stimulus control from one value of a stimulus to another by slowly decreasing the presence of the controlling stimulus |
Fading |
|
A schedule where a reinforcer is provided when behavior is emitted toward a stimulus that is the same as a target stimulus |
Matching to sample |
|
A schedule where a reinforcer is provided when behavior is emitted toward a stimulus that is the same as a previously presented target stimulus |
Delayed matching to sample |
|
The time between offset of the sample stimulus and onset of the comparison stimuli |
Retention interval |
|
The process of discriminating relevant events from alternative possibilities |
Remembering |
|
Differential response to stimuli that depends on the stimulus context |
Conditional discrimination |
|
A stimulus that signals the consequence is not available (extinction) |
S-delta |
|
A schedule of reinforcement with a sequence of responses required for reinforcement |
Response chain |
|
Procedures that interfere with rehearsal |
Retroactive interference |
|
The distribution of operant behavior among alternatives |
Choice |
|
The alternative of reinforcement chosen more frequently |
Preference |
|
Two or more simultaneous schedules of reinforcement each with a unique discriminative stimulus |
Concurrent schedules of reinforcement |
|
On a concurrent schedule, a response required to switch alternatives |
Changeover response |
|
On concurrent schedules, a control procedure to stop rapid switching between alternatives by imposing a minimal delay before a reinforcer is available |
Changeover delay |
|
The tendency for the relative rate of response between options to match the relative rate of reinforcement for the options |
Matching law |
|
Measure of the distribution of behavior between two or more alternative sources of reinforcement |
Relative rate of response |
|
Measure of the distribution of reinforcement between two or more alternatives |
Relative rates of reinforcement |
|
The relationship between behavior and reinforcement in choice situations |
Matching |
|
Response distributions that maximize reinforcement at the moment |
Melioration |
|
The use of concepts to predict, control, and analyze behavior |
Behavioral economics |
|
The tendency for reinforcers to lose effectiveness the longer the delay to their occurrence |
Delay discounting |
|
A mathematical curve describing the value of a consequence as the time to its occurrence is increased |
Discounting curve |
|
Selecting a small-immediate outcome over a larger-delayed outcome |
Impulsive behavior |
|
Selecting a larger-delayed outcome over a small-immediate outcome |
Self-control behavior |
|
The idea that reinforcement value decreases as the delay between the choice and reinforcer increases |
Ainslie-Rachlin principle |
|
A change in response from one option to another as the delay to the consequences of behavior change |
Preference reversal |
|
Behavior emitted prior to choice that reduces impulsive behavior |
Commitment response |
|
Rate of reinforcement is a function of reinforcement on the schedule relative to total reinforcement |
Quantitative law of effect |
|
Unknown contingencies that support the behavior of the organism |
Extraneous sources of reinforcement |
|
Ba/(Ba + Bb) = Ra/(Ra + Rb) |
Matching law equation |
|
Found time spent talking matched rate of agreement from the listener |
Conger & Killeen (1974) |
|
Found that behavior-change based on interval schedules are more effective than ratio |
Myerson & Hale (1984) |
|
Obtaining the highest overall rate of reinforcement from foraging behavior |
Optimal foraging |
|
As price increases consumption decreases because of the increase in demand |
Demand curve |
|
A commodity where the consumption changes with changes in price |
Elastic commodity |
|
A commodity where the consumption does not change much with changes in price |
Inelastic commodity |
|
A commodity that decreases in consumption as the consumption of another commodity increases |
Substitutability |
|
A commodity that rate of consumption is unaffected by changes in consumption for another commodity |
Independent commodities |
|
Mazur's equation that shows hyperbolic decay |
Hyperbolic discounting equation |
|
A stimulus or event that strengthens or maintains a behavior based on the organism's conditioning history |
Conditioned reinforcement |
|
A stimulus or event that strengthens or maintains a behavior due to the organism's conditioning history |
Conditioned reinforcer |
|
A stimulus or event that is able to strengthen a response without any previous learning history |
Unconditioned reinforcer |
|
Two or more simple schedules presented sequentially with separate discriminative stimuli and a single terminal reinforcer |
Chain schedule of reinforcement |
|
Two or more schedules presented sequentially without unique discriminative stimuli and a single terminal reinforcer |
Tandem schedule |
|
Chain schedules of reinforcement where the topography of response is similar in each component |
Homogeneous chains |
|
Chain schedules of reinforcement where the topography of response is different in each component |
Heterogeneous chains |
|
A technique for training complex behaviors by starting at the response closest to the primary reinforcer |
Backward chaining |
|
A technique for testing the effectiveness of a conditioned reinforcer by using it to train a new operant |
New-response method |
|
A technique for testing the effectiveness of a conditioned reinforcer by presenting the stimulus that accompanied the primary reinforcer without the primary reinforcer present |
Established-response method |
|
The idea that conditioned reinforcement is a product of classical conditioning |
S-S account of conditioned reinforcement |
|
Multiple schedules of reinforcement where requirements of one schedule are reinforced according to requirements of the other |
Second-order schedule |
|
The idea that it is necessary for a stimulus to be an SD in order to be a conditioned reinforcer |
Discriminative-stimulus account |
|
A schedule of reinforcement where one of two or more simple schedules are presented without separate discriminative stimuli |
Mixed schedule of reinforcement |
|
Topographically different operant that produces an SD or S∆ depending on the active schedule |
Observing response |
|
Stimuli that signal a decrease in the overall time to reinforcement will serve as conditioned reinforcers |
Delay-reduction hypothesis |
|
Event/stimulus associated with more than one unconditioned reinforcement |
Generalized conditioned reinforcer |
|
A special class of generalized conditioned reinforcer where the stimulus/event is social interaction or feedback |
Generalized social reinforcement |
|
A contingency where the conditioned reinforcers are tokens that can be stored and exchanged |
Token economy |
|
Critical for conditioned reinforcement in primates (not unconditioned) |
The amygdala |
|
Releases dopamine to both positive and aversive conditioned stimuli |
Nucleus accumbens septi |
|
Demonstrated that a stimulus can become a conditioned reinforcer if it provides information about unconditioned reinforcement |
Egger & Miller (1962) |
|
Provides no new information about the unconditioned reinforcement |
Redundant stimuli |
|
Evaluated the strength of a CR that predicted good or bad news |
Wyckoff (1952, 1969) |
|
The social reactions to aggression, such as submission, that maintain aggressive behavior |
Contingencies of aggression |
|
A victim's emotional attachment to their abuser; negative reinforcement of aggression |
Stockholm syndrome |
|
Contingencies, such as laws and social norms, that exert pressure against problem behavior |
Countercontrol |
|
A type of reinforcer that can be stored and later exchanged for other reinforcers |
Token reinforcement |
|
Demonstrated that the behavior of schizophrenics can be modified using a token economy |
Schaefer & Martin (1966) |
|
Demonstrated the efficacy of token reinforcement |
Kelleher (1958) |
|
Survival or reinforcement contingencies where the SD is the action of another and the response is topographically similar to the SD |
Correspondence relations |
|
A novel response that occurs by observing a model emit a similar response |
Imitation |
|
Imitation of a modeled response after a delay |
Delayed imitation |
|
Neurons that fire both when the organism performs the action and when they watch the action performed by another |
Mirror neurons |
|
Imitative behavior that is controlled by its consequences |
Operant imitation |
|
A higher-order operant emerging from repeated reinforcement of imitative behavior |
Generalized imitation |
|
Classic experiment that demonstrated imitation of observed aggressive acts |
Bobo doll experiment |
|
Observer's behavior is influenced by both the observed responses and consequences |
Observational learning |
|
Verbal stimuli that regulate the behavior of listeners by describing active contingencies |
Contingency-specifying stimuli |
|
Behavior that is under the control of contingency-specifying stimuli |
Rule-governed behavior |
|
Operant behavior that precedes some future response |
Precurrent behavior |
|
A major function of precurrent behavior where people construct stimuli to regulate future action |
Construction of SDs |
|
Operant behavior under the control of existing contingencies |
Contingency shaped |
|
Famous social psychology demonstration of the control of rule-governed behavior |
Milgram experiment |
|
Experimenter that found evidence that following instructions is rule-governed behavior |
Mark Galizio |
|
Rules and instructions that alter the function and strength of other stimuli |
Function-altering events |
|
Occurs when two verbal stimuli exert control over a common verbal topography |
Joint control |
|
Demonstrated observational learning through work with monkeys |
Kawai (1965) |
|
Actions that reduce the inconsistency between verbal and physical actions |
Dissonance reduction |
|
An instance of imitative behavior that occurs without any previous learning history |
Spontaneous imitation |
|
Contingencies of reinforcement from social situations and interactions |
Social conditioning |
|
Showed observational learning in birds in Europe |
Fisher & Hinde (1949) |
|
Proposed that observational learning is a special case of operant conditioning |
Miller & Dollard (1941) |
|
Showed imitation in rhesus monkeys using puzzles and raisins |
Warden, Fjeld, & Koch (1940) |
|
Provided evidence that spontaneous imitation is a form of phylogenetic behavior |
Epstein's pigeon and a ball (1984) |
|
Imitative behavior occurs because it has been important to species survival |
Phylogenetic basis of spontaneous imitation |
|
Demonstrated spontaneous imitative behavior in 12- to 21-day-old infants |
Meltzoff & Moore (1977) |
|
An infant's replication of an adult face when it cannot see its own face |
Opaque imitation |
|
Monitoring facial movements through proprioceptive feedback and comparing to what they see |
Active intermodal mapping |
|
Rules we give to ourselves |
Self-generated rules |
|
The performance of a speaker and environmental conditions that maintain this performance |
Verbal behavior |
|
The contingencies that regulate verbal behavior and originate from practices of the verbal community |
Verbal community |
|
Verbal operants whose form is regulated by motivational conditions |
Manding |
|
Verbal operants regulated by nonverbal discriminative stimuli and maintained by generalized reinforcement from the verbal community |
Tacting |
|
Verbal behvior where the response exactly corresponds to the stimulus |
Echoic |
|
Topographically similar behaviors that serve distinct behavioral functions |
Functional independence |
|
A class of establishing operations that depends on a history of reinforcement |
Conditioned establishing operation |
|
When a reinforcer for one response does not control the form of the next response |
Nonspecific reinforcement |
|
Verbal operants regulated by verbal discriminative stimuli |
Intraverbal behavior |
|
Requires that the verbal stimulus and response be in the same mode and physical resemblance |
Formal similarity |
|
Verbal operants regulated by verbal stimuli with correspondence between stimulus and response but not formal similarity |
Textual behavior |
|
Verbal behavior that emerges from the combination of the behavior of listening and speaking |
Naming relation |
|
Differential response to stimuli depending on stimulus context |
Conditional discrimination |
|
Behavior of one person causes reinforcement for the behavior of the other |
Interlocking contingencies |
|
Interlocking contingencies between speaker and listener |
Social episode |
|
Presentation of one class of stimuli that occasions response to other classes |
Stimulus equivalence |
|
When class A is interchangeable with class B |
Symmetry |
|
Responding to a one-to-one relationship between classes |
Reflexivity |
|
Responding as A=C after being trained A=B and B=C |
Transitivity |
|
When a reinforcer is delivered for choosing the stimulus that is a direct match to the sample stimulus |
Identity matching |
|
When a reinforcer is delivered for choosing a stimulus that is different from, but related to, the sample stimulus |
Symbolic matching |
|
Testing for symmetry by switching the roles of the sample and test stimuli |
Reversal test |
|
Audible sources of speech sound |
Phonation |
|
Oral and nasal airways that provide the phonetic quality of sound |
The supralaryngeal vocal tract |
|
Both writing and speaking that is modifiable by consequences |
Verbal operants |
|
The effects of words in the forms of instructions, advice, maxims, and laws on the listener's behavior |
Rule-governed behavior |
|
The rules that govern the construction and grammar of language |
Syntax |
|
The meanings of words |
Semantics |
|
Used blocked-response CEO to train manding by deaf subjects |
Hall & Sundberg (1987) |
|
Demonstrated the functional independence of manding and tacting in children 3-5 |
Lamarre & Holland (1985) |
|
Focuses on the application of principles, methods, and procedures of behavior to socially significant problems |
Applied behavior analysis |
|
Reinforcement delivered for any behavior other than a target operant |
Differential reinforcement of other behavior |
|
Reinforcement delivered for a behavior that cannot occur at the same time as the target behavior |
Differential reinforcement of alternative behavior |
|
A formal, written statement of expected behaviors and their consequences |
Behavioral contract |
|
When the strengthening of a target response also strengthens similar responses |
Response generalization |
|
The tendency for behavior to persist for a time after the contingencies are removed |
Behavior maintenance |
|
Once learned, a new behavior is maintained by naturally occurring contingencies |
Behavior trapping |
|
A research tool that demonstrates experimental control by gradually introducing the treatment across behaviors, participants, or contexts |
Multiple baseline design |
|
A measurement procedure that splits a given block of time into smaller observation periods |
Interval recording |
|
Samples behavior over a long period with measurements occurring at specified times |
Time sampling |
|
A measurement procedure where how long a behavior occurs is recorded |
Duration recording |
|
Behavior that occurs over a long period of time rather than in discrete chunks |
Continuous behavior |
|
The use of operant principles to arrange contingencies to promote positive behaviors |
Contingency management |
|
Organized course method where students can move at their own pace |
Personalized system of instruction |
|
A method of systematic instruction that involves targeting specific behaviors, counting, and graphing performance |
Precision teaching |
|
A measure of behavior that focuses on accuracy and frequency |
Fluency |
|
The selection of a larger-later reinforcer over a smaller-sooner outcome |
Self-control |
|
A procedure for reinforcement where the work requirement is systematically adjusted up or down |
Changing criterion design |
|
The delivery of a behavioral intervention for 40+ hours a week starting early in life |
Early intensive behavioral intervention |
|
An observed suppression in food consumption and increase in physical activity following food restriction |
Activity anorexia |
|
A distinctive taste is a CS followed by the physiological consequences of wheel running (US) |
Activity induced taste aversion |
|
The ability to learn that food taste predicts the amount of calories ingested |
Conditioned overeating |
|
A behavioral program for teaching women to effective breast self-examination for cancer |
MammaCare |
|
Verbal operant behavior used to label private events |
Feelings and thoughts |
|
A period during which treatment is withheld so that a later change can be evaluated |
Baseline |
|
Change in behavior that occurs when either an SD or S∆ is presented |
Stimulus control |
|
The relations between behavior and their consequent events |
Contingencies of reinforcement |
|
The set of target responses, consequences that follow actions, and long-term goals |
Behavioral plan of action |
|
Designing interventions that produce changes in behavior that occur in all relevant settings |
Generality of behavior change |
|
Response definition should be based on observable features of behavior |
Objectivity |
|
The idea that behavior is shaped by natural selection, behavioral selection, and cultural selection |
Selection by consequences |
|
The evolutionary history of a species |
Phylogeny |
|
Differing reproduction of members of a species based on their genetic make-up |
Natural selection |
|
Change in genetic make-up of a species over time |
Evolution |
|
Environmental relations that result in the differential success of members of a species |
Contingencies of survival |
|
Characteristics observed during the lifetime of an individual |
Phenotype |
|
Genetic make-up of the organism |
Genotype |
|
Spontaneous changes in the genetic make-up of an individual |
Mutation |
|
The evolved ability to adjust behavior based on experience |
Behavioral flexibility |
|
The interlocking social contingencies of many people |
Cultural practice |
|
Contingent relations between cultural practices and effect on the members of the group |
Metacontingencies |
|
A set of interlocking contingencies with different individual and group outcomes |
Social traps |
|
A group contingency that establishes individual avoidance of certain events |
Cultural taboo |
|
Variations in individual behavior that are adopted by the entire group |
Cultural evolution |
|
An evolutionary theory that genotype may be related to an animal's survival of food-related challenges |
Thrifty gene theory |
|
Described the metacontingencies of cultural practices |
Sigrid Glenn |
|
A type of behavior that occurs approximately the same way in all members of the species |
Behavioral rigidity |
|
Found evidence of genetic control of a complex behavioral sequence in Aplysia |
Scheller & Axel (1984) |
|
Found Phormia regina showed heritability of the process of classical conditioning |
Hirsch & McCauley (1977) |
|
Outlined an evolutionary theory of behavior dynamics |
McDowell (2010) |
|
Sexual recombination of existing genes and mutations |
Sources of heritable genetic variation |
|
Sets the foundation for biologically based behavior initiated by the environment |
Genetic blueprint |
|
A rule that states how mother and father behavior are combined in successive generations |
Reproduction rule |
|
Reoccurrence of a behavior sometime after its last reinforcement |
Memory |
|
The response forms that make contact with the environment and unit of selection at the behavioral level |
Operant |
|
Increased activation, integration, and consolidation of neurons |
Neurocellular effects of operant learning |
|
Withholding the controlling stimuli weakens operants and generates behavioral variation |
Extinction |
|
Research on attachment that suggests "contact comfort" functions as reinforcement |
Harlow & Zimmerman (1959) |
|
A population in the Canary Islands that communicates by whistling |
The Gomeros |
|
Allows people to profit from what others say |
Rule-governed behavior |