What is the term for the rules that determine whether responses lead to the presentation of a reinforcer?

The famous behavioral scientist B. F. Skinner believed that, in order to experimentally analyze human and animal behavior, each behavioral act can be broken down into three key parts. These three parts constitute his three-term contingency: discriminative stimulus, operant response, and reinforcer/punisher. The three-term contingency is fundamental to the study of operant conditioning.

To illustrate the operation of behavioral analysis, the behavior of leaving class when the school day is over can be broken down into the parts of the three-term contingency. The bell, which serves as the discriminative stimulus, is sounded at the end of the school day. When the bell rings, students exit the classroom. Exiting the classroom is the operant response. Reinforcement of leaving the classroom at the proper time results from the other behaviors in which students can engage now that the school day is over.

But, if the same behavior of exiting the classroom occurs prior to the bell's ring (that is, in the absence of the discriminative stimulus), a student now faces punishment. Punishment of leaving class early would result because this behavior violates school rules and leads to a variety of adverse consequences, like staying after school.

Glossary Index | Quotations

Premack's principle (or the differential probability hypothesis) refers to reinforcing a target behavior by awarding some privilege to engage in a more desired behavior afterward.

From: Encyclopedia of Psychotherapy, 2002

Contingency Management

Christopher A. Kearney, Jennifer Vecchio, in Encyclopedia of Psychotherapy, 2002

IV.F. Premack's Principle

Premack's principle (or the differential probability hypothesis) refers to reinforcing a target behavior by awarding some privilege to engage in a more desired behavior afterward. For example, a child may be told that he or she can have dessert after eating his or her vegetables, or be told that he or she can play baseball after completing a certain chore. The more highly desirable behavior is used to entice a person to engage in a usually productive, albeit less enjoyable, activity. Typically, Premacks principle is used by parents in addressing their children, and therapists often build this principle into their parent-training regimens. Premack's principle is used loosely in other clinical settings as well. For example, a person who successfully loses weight may then be allowed to shop for new clothes.

Premack's principle is often extended in parent-training programs to encourage appropriate child behavior and discourage inappropriate behavior. For example, parents may establish a set of house rules that carry rewards if followed and punishments if not. The establishment of this consequence system is often linked to other areas of parent training such as modifying parental commands, setting up daily routines, and increasing supervision of the child.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B012343010000060X

Learning Theory and Behaviour

E.J. Capaldi, A. Martins, in Learning and Memory: A Comprehensive Reference, 2008

1.03.3.5 The Response Deprivation Hypothesis

There is evidence that, contrary to the Premack principle, low-probability responses can reinforce higher-probability responses. It is a matter of depriving the animal of the less-probable response. For example, if ingesting X is less probable than ingesting Y, the tendency to ingest Y can be increased above its baseline by depriving the animal of X and making the availability of X contingent on ingesting Y (e.g., Timberlake and Allison, 1974; Allison, 1989). There are many interesting applications of the response deprivation hypothesis. For example, 3-year-old children who loved to run around and scream were induced in a fairly short time to sit still, a not highly probable activity for 3 year olds. Essentially, after the children sat quietly for a specified time period, they were allowed to run and scream for a few minutes (Homme et al., 1963).

The response deprivation hypothesis allows us, like the Premack principle, to determine in advance how reinforcing some activity might be. It also allows us to adjust reinforcers to particular animals or people. For example, a person who enjoys TV will perform other behaviors if those behaviors allow TV viewing, which is otherwise reduced. But if a person enjoys reading more than TV watching, let reading be contingent on TV watching.

A more subtle implication of the response deprivation hypothesis bears mention. Normally we think of instrumental behavior such as bar pressing as different from consummatory behavior such as eating. This distinction is violated by the response deprivation hypothesis. That is, any behavior can serve as a reinforcer if it is made to occur below its normal baseline. Animals deprived of food will run in order to eat, and an animal deprived of activity will eat in order to run.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780123705099000802

Token Economy: Guidelines for Operation

Teodoro Ayllon, Michael A. Milan, in Encyclopedia of Psychotherapy, 2002

III.C. Backup Reinforcers

The backup reinforcers are the natural reinforcers that give value to the tokens. The token economy should not deprive any clients of anything to which they are entitled by the Constitution or by statute. The backup reinforcers therefore are additions to these legally mandated minimums. They may involve access to additional amounts or enhancements of entitlements, such as additional exercise time or special foods, as well as anything to which the client is not legally entitled. In general, backup rein-forcers fall in several categories. These consist of edibles or consumables, such as special food or drink; activities, such as extra recreational or computer game time; material objects, such as special athletic shoes or posters; and independence, such as reduced supervision or movement through progressively less restrictive levels in a structured levels program. Each can be considered a to-be-earned privilege.

A number of strategies may be used to identify backup reinforcers. The most straightforward consists of asking the clients what they would like and would work for. Another is based on the Premack principle and involves observing clients during free or unstructured time to identify the things they are most likely to do. These may then be used as backup reinforcers for the typically lower probability target behaviors. Finally, the staff can be asked what they believe will serve as backup reinforcers for their clients. It is important to note, however, that the information gathered through these strategies results only in the identification of potential backup reinforcers. The reinforcing properties of the potential backup reinforcers must be tested to determine whether they are true backup reinforcers. This is done by assessing whether clients will engage in target behavior to earn tokens and then exchange the tokens for the potential backup reinforcers.

What is reinforcing to one client may or may not be available to another client. The token economy must therefore offer a variety of backup reinforcers to ensure that there are sufficient reinforcers for all participants. Although some backup reinforcers may motivate most or all clients, additional backup reinforcers that function as such for only a few or, perhaps, only one client should also be made available to ensure that the unique interests of all clients are addressed. The backup reinforcers should be available to clients on a regular basis. Some should be available several times a day, others several times a week, and perhaps still others several times a month or only on a monthly basis. The schedule of availability of backup reinforcers is dictated, in part, by the characteristics of the clients, and a token economy cannot err in the direction of making backup reinforcers too frequently available to clients. Finally, the backup reinforcers must be available only through the token economy. The availability of “bootleg” backup reinforcers, that is, backup reinforcers that may be obtained through other means, will dilute or negate the reinforcing properties of backup reinforcers and should be prevented.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B0123430100002257

Respondent and operant behavior1

Peter Sturmey, ... Erica Doran, in Functional Analysis in Clinical Treatment (Second Edition), 2020

Application of operant learning

Many interventions are based on operant models of psychopathology. Operant-based interventions are exemplified by identifying reinforcers; teaching new behaviors; increasing existing adaptive behaviors with reinforcement including concurrent schedules of reinforcement; decreasing undesirable behaviors with reinforcement of other behavior, punishment, and extinction; and manipulating motivating operations for reinforcers and punishers.

Application: Identifying reinforcers

The therapist must identify reinforcers that will be effective and can be readily manipulated. An accurate functional assessment of a client's problem may identify reinforcers maintaining the problematic behavior that may also be suitable for inclusion in an intervention. Verbally competent humans often have a wide range of reinforcers and may report reinforcers in a largely accurate manner. Simple interviews are often adequate to identify reinforcing consequences in some circumstances, although the therapist should be cautious about using such self-report data, as the accuracy or completeness of client verbal behavior is readily influenced by many variables.

A wide range of psychometric instruments has been developed to assess potential reinforcers. Cautela and Kastenbaum (1967) developed the Reinforcer Survey Schedule for use in adult mental health settings. There are now at least 23 variants of this scale, designed to be used with children, adolescents, seniors, psychiatric patients, people with autism and/or intellectual disabilities, people with visual impairments and in such varied contexts as marriage, social behavior, sex, work, parent-child relationships, and school (Cautela & Lynch, 1983, Table 1). These scales have served as models for the subsequent development and elaboration of a wide range of similar psychometric tools. These tools often, but not always, identify functional reinforcers that may be incorporated into treatment plans. They often have to be updated as society offers novel stimuli and changes in fashions, which results in reinforcers changing quickly over time (King & Houlihan, 2018).

In some cases, identification of reinforcers may be quite challenging, such as when working with individuals with poor verbal skills, young children, or individuals with disorders that are characterized by few effective reinforcers, such as depression and chronic schizophrenia, or that have one powerful but problematic reinforcer, such as in addictions and obsessions. In such cases, direct assessments tend to be more accurate than survey methods. Potential reinforcers can also be identified by observation of approach and avoidance behavior in the natural environment.

A variety of stimulus preference assessment technologies address this issue. For example, Ortega, Iwata, Nogales-Gonzalez, and Frades (2012) identified reinforcers for 14 adults with mild to severe dementia. They identified four leisure activities and four edible items to include in a paired stimulus preference assessment. The procedure consisted of presenting two items at a time and asking the individuals to select one. Across choice presentations, leisure and edible items were presented together an equal number of times, and selection of one item resulted in brief access to the selected item. The authors identified a preference ranking for the items within 30–60 min, such that the item selected most often received the highest ranking. They then validated the preference assessment with three of the 14 participants by conducting a reinforcer assessment whereby access to the highest ranked item was provided contingently upon completing a simple task. Low rates of task completion were observed during baseline and rate of task completion increased when the preferred item was delivered for task completion.

Therapists can also use the Premack Principle to identify reinforcers. High probability behaviors may function as reinforcers if they are made contingent upon low probability behaviors (see Mitchell & Stoffelmayr [1973] for an example of the Premack Principle as applied to very inactive people with schizophrenia). Progressive ratio schedules have also been used to identify reinforcers. In progressive ratio schedules, participants must emit more and more responses for access to stimuli until they no longer respond. The number of responses emitted is a parametric measure of the stimulus' strength as a reinforcer (Russell, Ingvarsson, & Haggar, 2018).

When one is assessing stimuli as potential reinforcers, it is important to recognize the dynamic nature of reinforcement. When a primary reinforcer is delivered some degree of satiation occurs which may weaken the effectiveness of that stimulus as a reinforcer in the future. A stimulus might function as a reinforcer for one response, but not another. For example, a stimulus might function as a reinforcer for merely reaching to pick up the reinforcer, but not for completing an effortful or aversive task associated with therapy. Whether and to what extent a stimulus functions as a reinforcer may also depend on the recent deprivation of that stimulus. For example, activity might function as a powerful reinforcer after a period of sitting, but not after a period of activity (Laraway, Snycerski, Michael, & Poling, 2003; Michael, 1982, 2000). Finally, the effectiveness of a stimulus as a reinforcer may depend on the availability of other reinforcers: In a reinforcer-deprived environment, a stimulus might function as a reinforcer, whereas in a reinforcer-rich environment, the same stimulus might not. Likewise, one might be able to enhance the effectiveness of a stimulus as a reinforcer by reducing the overall availability of reinforcers. Cautela (1984) referred to this latter idea when describing the “general level of reinforcement” in a client's life.

These observations have important implications for practitioners. To circumvent the problem of reinforcer satiation, therapists should: (a) prefer to use generalized reinforcers, such as tokens, points, or money, to reduce the likelihood of satiation; (b) establish secondary reinforcers; (c) vary the reinforcers used; and (d) employ client choice of stimuli on a regular basis.

Application: increasing existing adaptive behavior

A common clinical problem occurs when an important behavior, although in the client's repertoire, is emitted at an inappropriate frequency. For example, a shy person may interact with others, but only rarely. Alternatively, a person might complain about minor health problems constantly, but rarely talk about other topics. In such situations, it may be that there is an inappropriate schedule of reinforcement supporting the inappropriate frequency of the target behavior. Alternatively, the schedule of reinforcement for other behaviors may be inappropriate. For example, perhaps the schedule of reinforcement maintaining healthy behavior is too weak, resulting in higher rates of unhealthy behavior.

In such situations, one possible approach is to enhance the schedule of reinforcement for the desirable target behavior. An example of this approach, applied to a very challenging problem that has not been dealt with effectively by other methods, comes from Olseon and Baker (2014). These researchers attempted to increase speaking in two elderly people with dementia. The target behavior consisted of the participants requesting to engage in conversation or requesting to obtain an item needed to complete a task. In baseline, participants rarely emitted requests. The introduction of a simple reinforcement contingency in addition to textual prompts resulted in an increase in requesting for one participants, but the treatment had an inconsistent effect for the other. These results are promising given the severity of the problem and the absence of effective alternate methods of intervention.

Application: decreasing undesirable behavior with reinforcement

The acquisition of operant behavior has a broad impact on the entire behavioral repertoire. Not only does the reinforced operant increase; other responses decrease and the entire behavioral repertoire becomes more predictable and organized in ways that may be useful (Catania, 2007). These properties of operant behavior have obvious implications for treatment of psychopathology; often the problem referred is one in which the person's behavioral repertoire is “out of control,” “disorganized,” or “fragmented.” For some people, the problem may be the absence of appropriate behavior due to skills deficits, or the lack of environmental support for the existing appropriate behavior. In either case, the establishment and support of some operant behavioral repertoire may be helpful.

Dixon et al. (2004) used differential reinforcement of alternative behavior (DRA) to reduce the inappropriate verbal behavior, including profanity, aggressive remarks, and suicidal and sexual utterances, of four adults aged 20–62 years with acquired brain injuries from traumatic accidents. Prior to treatment, the conducted functional analysis sessions to determine the function of inappropriate verbal behavior. For two participants, escape from demands maintained the inappropriate vocalizations, and for the other two attention maintained the inappropriate vocalizations. For the participants whose utterances were maintained by escape from demand, treatment consisted of preventing escape for inappropriate utterances, ignoring inappropriate utterances, and allowing escape contingent upon appropriate verbal utterances. For the participants whose utterances were maintained by attention, treatment consisted of withholding attention for inappropriate utterances and providing contingent attention for appropriate utterances. In all cases, the DRA treatment decreased the frequency of inappropriate utterances. For two participants, the frequency of appropriate behavior also increased.

A second example comes from Holtyn et al. (2014) who examined the effects of contingent reinforcement for clean urine samples in 33 adults addicted to opiates. The authors used a therapeutic workplace intervention in which participants could work to earned money for completing computer typing tasks, providing enrollment in drug treatment and, depending on treatment phase, abstinence of drug use. Initially, participants could work and earn money if they were enrolled in methadone treatment. Following three weeks of the methadone contingency for accessing the work environment, the authors implemented an opiate-abstinence contingency in which participants could continue to work if they also provided opiate negative urine tests in addition to continued enrollment in methadone treatment. During the final treatment phase, access to reinforcement for work completion was contingent upon cocaine, opiate negative urine samples, and continued enrollment in methadone treatment. Across all treatment phases, failure to meet the contingency to work—dropping out of methadone treatment or producing a positive urine sample—resulted in a decrease in hourly pay. Once the participants met the work contingency again, the base pay increased by $1 per hour for each hour of work until they reached the maximum pay. Twenty-four participants remained enrolled in methadone treatment for the 26-week intervention and that the percentage of participants with opiate and cocaine negative urine samples increased only after the introduction of the respective abstinence contingency.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780128054697000024

Case Conceptualization and Treatment: Children and Adolescents

A. Anderson, A. Hudson, in Comprehensive Clinical Psychology (Second Edition), 2022

5.05.5.1.1 Systematic Use of Reinforcement

The most straightforward procedure for strengthening behavior is simply to wait for the behavior to occur, and then to reinforce it. However, there are some important issues related to this seemingly straightforward process. Perhaps the most important of these is to select an appropriate reinforcer. It must be remembered that a reinforcer is defined in terms of its effect on the behavior it is contingent upon. Clinicians can err in what they think will be reinforcing for particular children. The information collected in the assessment phase is critical for the right choice of an effective reinforcer, and this is not always an easy task. Some clinicians develop what are termed reinforcement menus, or lists of things that are potentially reinforcing, and then have the child or caregiver select from the menu. Another very helpful hint is to use what is known as the Premack principle. Premack (1959) noted that the activities children chose to engage in during their free time could be used to reinforce other, less-preferred activities. He technically defined this as using higher-probability behaviors to reinforce lower-probability behaviors. The importance of the principle is that if you are having trouble identifying what is reinforcing for children, observe what they do in their free time.

Reinforcers vary in nature and can be classified in a number of ways. One of the earliest distinctions made was between primary and secondary reinforcers. A primary reinforcer, sometimes referred to as an unconditioned reinforcer, is one that does not depend on learning to have taken place for it to develop its reinforcing properties. Food, water, and warmth are examples of such primary reinforcers. In contrast, learning has been involved in secondary or conditioned reinforcers developing their reinforcing properties. For example, some children find playing games on a computer to be very reinforcing. For this to have happened, some learning about the computer and the particular game must have taken place. It can be argued that there is limited clinical usefulness in the distinction between primary and secondary reinforcers as ethical considerations dictate that children have a fundamental right to unrestricted access to most primary reinforcers, and hence they should not be used in ABA programs. It is the case, however, that edible reinforcers such as candy have been used successfully to motivate children who do not respond to secondary reinforcers.

A second distinction that is made in regard to types of reinforcers is that between naturally occurring and artificial reinforcers. A naturally occurring reinforcer is one that is readily available and typically used in the setting in which the ABA program is being conducted. In most homes, reinforcers such as parent attention, pocket money, and access to television are available to children. Unfortunately, these reinforcers are often given to the children in an unsystematic or unplanned manner. An artificial reinforcer is one that is not usually available in the setting, but is brought in for the purposes of the program. Candy is not usually used as a reinforcer in most classrooms but could be introduced for use in a program for a special purpose. It is generally advisable for clinicians to use artificial reinforcers only when absolutely necessary, as the fading of such reinforcers can be more difficult than is the case for naturally occurring reinforcers.

A third distinction that is made is between extrinsic and intrinsic reinforcers. An extrinsic reinforcer is one that is quite separate from the behavior that it is reinforcing in that the behavior could occur and not be followed by the reinforcer. If a clinician is using praise to reinforce a child's behavior, the behavior can be allowed to occur without the reinforcement being administered. An intrinsic reinforcer can be viewed as one that is actually part of the behavior. The behavior is reinforcing by itself, a process often referred to as automatic reinforcement, and does not need additional reinforcement delivered by another person to maintain it. For most individuals no additional reinforcement is required to keep them watching their favorite show on television. They do it because they like it. The distinction between intrinsic and extrinsic reinforcement can be clinically important. Some behaviors, for example reading, may initially need to be extrinsically reinforced to get them going, but once established, can be maintained by the enjoyment they produce.

Yet another way of distinguishing among reinforcers is to classify them based on the broad type of reinforcer. The types listed can vary among authors in the field, but a common list includes social, food, activity, and token reinforcers. Social reinforcers include praise, hugs, or attention from another person. As the name suggests, food reinforcers are things the child can eat. Rather than being basic food requirements, which, as mentioned earlier, are problematical from an ethical viewpoint, they are special treats that the person does not have regular access to. Activity reinforcers are those things that the person likes to do, for example, playing games and watching television. Finally, token reinforcers do not have value in themselves but derive their value from what they can be exchanged for. Many classroom teachers have children earning points for completion of academic work, and then allow the children to exchange those points for back-up reinforcers such as extra time at recess or reduced homework. Money is certainly the most frequently used token reinforcer.

A final point to be made in regard to the systematic use of reinforcement to strengthen a target behavior is that the fundamental principles derived from learning theory need to be adhered to. Reinforcement usually needs to be immediate, particularly with children at a low developmental level. If another behavior occurs between the target behavior and the delivery of the reinforcement, the more recent behavior might be the one that is strengthened. The second behavior under these circumstances is usually referred to as superstitious behavior. Also, attention needs to be paid to what is known about schedules of reinforcement. A continuous schedule needs to be used to maximize the rate of development of the new behavior, and then there should be a move to an intermittent schedule, referred to as a thinning of the reinforcement, to make the behavior more resistant to extinction.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780128186978000881

Food refusal in children: A review of the literature

Keith E. Williams, ... Laura Seiverling, in Research in Developmental Disabilities, 2010

The behavioral procedures and participant demographics in the food refusal intervention studies.

StudyIntervention proceduresParticipants and ages% with medical problems% with developmental delays
1. Hatcher (1979) Premack Principle 2-year-old girl 100% 100%
2. Riordan, Iwata, Finney, Wohl, and Stanley (1984) Differential reinforcement, planned ignoring, simultaneous presentation, physical guidance 4 children, 16–40 months of age 100% 100%
3. Blackman and Nelson (1985) Forced oral feeding, differential reinforcement 9 children, aged 4–40 months 89% 89%
4. Handen et al. (1986) Demand fading, short meal duration, contingent access to reinforcers, time out, planned ignoring 7 children, aged 10–66 months 100% Unknown
5. Blackman and Nelson (1987) Forced oral feeding, escape extinction, differential reinforcement, demand fading, texture fading 11 children, aged 10–29 months 91% 73%
6. Luiselli and Gleason (1987) Texture fading, sensory stimulation, contingent reinforcement 4-year-old girl 100% 100%
7. Lamm and Greer (1988) Graduated prompts, eliciting stimulus, positive reinforcement 3 children, aged 10–13 months 100% 100%
8. Dunbar, Jarvis, and Breyer (1990) Positive reinforcement, planned ignoring, verbal reprimand, shaping, play facilitation 3 children, aged 18–48 months 100% 67%
9. Chamberlin et al. (1991) Group program providing nutrition and behavior management information 6 children, aged 12–27 months 100% Unknown
10. Hoch et al. (1994) Positive reinforcement, escape extinction, planned ignoring 2 children, aged 25 and 41 months 100% 50%
11. Cooper et al. (1995) Contingent attention, escape extinction, representation, noncontingent access to toys, choice procedure 4 children, aged 1 year 8 months to 6 years 2 months 100% 100%
12. Hoch, Babbitt, Coe, Duncan, and Trusty (1995) Positive reinforcement, escape extinction, representation, swallow induction 3.5-year-old girl 100% 100%
13. Kerwin et al. (1995) Differential reinforcement of incompatible behavior, escape extinction, planned ignoring, representation 3 children, aged 2.5–5 years 100% Unknown
14. Ahearn et al. (1996) Positive reinforcement, escape extinction, representation 3 children, aged 33–42 months 67% 67%
15. Coe et al. (1997) Differential reinforcement, escape extinction, representation 5-year-old boy and a 2-year-old girl 100% 50%
16. Foy et al. (1997) Physical restraint, praise, tube feed reduction, planned ignoring 19 children, aged 9–51 months 100% 53%
17. Freeman and Piazza (1998) Stimulus fading, positive reinforcement, escape extinction, planned ignoring 6-year-old girl 100% 100%
18. Didden, Seys, and Schouwink (1999) Positive reinforcement, escape extinction, shaping, planned ignoring 1.5-year-old girl 100% 100%
19. Benoit et al. (2000) Escape extinction, other behavioral procedures used but not specified 32 children, aged 4–36 monthsa 100% 100%
20. Gutentag and Hammer (2000) Positive reinforcement, planned ignoring, texture fading 3-year-old girl 100% 100%
21. Luiselli (2000) Demand fading, visual cues, positive reinforcement 3-year-old boy 100% 0%
22. Kahng, Tarbox, and Wilke (2001) Differential reinforcement, response cost, planned ignoring 5-year-old boy 0% 100%
23. O’Reilly & Lancioni (2001) Planned ignoring, positive reinforcement, escape extinction 5-year-old girl 100% 100%
24. Patel, Piazza, Martinez, Volkert, and Santana (2002) Differential reinforcement, escape extinction, planned ignoring 3 children, aged 2–3 years 100% 100%
25. Dawson et al. (2003) High-p instructional sequence, escape extinction, planned ignoring, representation 3-year-old girl 100% 100%
26. Piazza et al. (2003) Differential reinforcement, escape extinction, planned ignoring 4 children, aged 23 months to 4 years 100% 50%
27. Zangen et al. (2003) Drip feeding, medication for motility, visceral pain and non-specific arousal, unspecified cognitive-behavioral therapy 14 children, aged 1.5–6 years 100% Unknown
28. Mueller, Patel, Kelly, and Pruett (2004) Blending together preferred and novel foods, differential reinforcement or noncontingent reinforcement, escape extinction, planned ignoring 2 children, aged 3–4 years 100% 100%
29. Reed et al. (2004) Noncontingent reinforcement, escape extinction, representation, planned ignoring 4 children, aged 15 months to 4 years 75% 50%
30. Gulotta et al. (2005) Food redistribution, escape extinction, representation, positive reinforcement, planned ignoring 4 children, aged 2–5 years 100% 75%
31. Patel et al. (2005) Positive reinforcement, escape extinction, representation, planned ignoring, texture fading 3 children, aged 3–4 years 100% 100%
32 DeMoor et al. (2005) Texture fading, demand fading, positive reinforcement, planned ignoring 3 children, aged 1 year 6 months to 6 years 8 months 100% 100%
33 Clawson, Palinski, and Elliott (2006) Oral-motor exercises, positive reinforcement, escape extinction, response cost, representation 3 children, aged 15–42 months 100% 100%
34. DeMoor et al. (2007) Stimulus fading, positive reinforcement, verbal reprimand, time out, escape extinction, planned ignoring, appetite manipulation 5 children, aged 2.5–3.3 years 100% 100%
35. Gibbons et al. (2007) Positive reinforcement, planned ignoring, representation, escape extinction, tongue thrust reduction 6-year-old girl 100% 100%
36. Williams, Riegel, Gibbons, and Field (2007) Positive reinforcement, escape extinction, representation, swallow induction, texture fading, response cost, token economy, exit criterion 46 children, aged 16–133 months 100% 59%
37. Kindermann et al. (2008) Positive reinforcement, appetite manipulation 10 children, aged 9–21 months 100% Unknown
38. Casey et al. (2008) Enriched environment, behavioral momentum, differential reinforcement, functional communication training, escape extinction, contingent restraint, exit criterion 8-year-old boy Unknown 100%

aBenoit et al. used a group design with a behavioral treatment group and a nutrition counseling only group. The participant demographics for the behavioral treatment group were included.

Read full article

URL: https://www.sciencedirect.com/science/article/pii/S0891422210000028

Facilitating or undermining? The effect of reward on food acceptance. A narrative review

Lucy J. Cooke, ... Jane Wardle, in Appetite, 2011

Theoretical approaches to reward in food acceptance

The idea that contingent rewards increase performance of the rewarded behaviour is the central tenet of instrumental learning (Thorndike, 1911). Incentive effects on performance are equally fundamental to classical economic accounts of human behaviour (Benabou & Tirole, 2003). The Premack principle offers another way to understand improvements in performance, predicting that higher probability behaviours will reinforce lower probability behaviours. Classical learning may also affect performance through conditioned preferences and expectations (Baeyens, Eelen, Crombez, & Van den Bergh, 1992), ultimately resulting from the rewarded food taking on the positive motivational properties of the reward (Bindra, 1978; Toates, 1986). On the face of it, there is no reason to suppose that rewarding children for eating should not have these same positive effects.

However, social psychologists and economists have observed a paradoxical effect of rewards in some situations. Behaviour is believed to be determined by the combination of intrinsic motivation (e.g. enjoyment of the task) and extrinsic motivation (e.g. incentives) (Ryan & Deci, 2000). A meta-analysis of reward studies indicated that under certain conditions, external rewards may actually reduce intrinsic motivation (Deci, Koestner, & Ryan, 1999); indicated by a decrease in performance of or enjoyment of the behaviour when rewards are withdrawn (Cameron & Pierce, 1994). In the classic study, preschoolers who were promised a reward for drawing with magic markers showed less intrinsic motivation for drawing at the end of the reward period than their peers who had received either an identical but unexpected reward or no reward (Lepper, Greene, & Nisbett, 1973). Economists have also noted that offering a reward for pro-social behaviour can reduce compliance to below the spontaneous level; for example, local objection to the construction of a nuclear waste depository was running at 49% before incentives were offered and at 75% after the Swiss Government's offered financial compensation. Increasing the amount of offered compensation had no influence on rates of objection (Frey & Oberholzer-Gee, 1997). This has been termed ‘Motivation Crowding’ (Frey & Jegen, 2001).

Adverse effects of reward have been explained by social psychologists both in terms of ‘over-justification’ theory (Lepper et al., 1973) and self-determination theory (Deci et al., 1999). Both propose that rewarding people for doing something they already enjoy undermines intrinsic motivation and creates conditions for a net decline in motivation when rewards are subsequently withdrawn. Over-justification theory explains it in terms of the individual attributing their behaviour to the external reward and therefore effectively overlooking the intrinsic pleasantness of the task. Self-determination theory focuses on external rewards compromising feelings of competence and autonomy.

Applying over-justification theory in the food context, and construing liking as a form of intrinsic motivation, suggests that while incentives for eating less palatable foods may increase liking, incentives for eating palatable foods should show an adverse effect. Premack's principle makes similar predictions about rates of behaviour. If consumption of the rewarded food is already a highly probable behaviour then it is not likely that it will be reinforced by other behaviours. Conversely, if consumption of a disliked food or ‘lower probability’ behaviour is reinforced with a liked activity, an increase in response can be expected. Classical conditioning offers another possible explanation for drops in preferences for the rewarded food: if being rewarded for a task is experienced as negative (e.g. if the reward is enjoyed less than the rewarded task) then the task itself may take on these hedonic properties and become less liked. It would seem that in both operant and classical conditioning, the initial preference for the rewarded food and the reward itself are important factors in determining a positive outcome.

To some extent that is how the experimental literature plays out. The early studies rewarded consumption of novel but palatable food or drinks (Birch et al., 1982; Mikula, 1989; Newman & Taylor, 1992), and typically found that liking decreased. In contrast, the more recent school studies have rewarded consumption of less palatable foods (often vegetables) and found increased liking (Hendy et al., 2005; Horne et al., 2004; Lowe et al., 2004, 2006).

Simple sensory over-satiation may also explain these effects (Hendy et al., 2005; Rolls, Rolls, Rowe, & Sweeney, 1981). If the offer of rewards pushes consumption of a single food beyond the point of satiation, it can produce negative associations with the target food and hence negative shifts in preference. This idea was tested by Birch et al. (1984) who found no additional negative effects of rewards when children were required to consume their target drink beyond satiation. It is also noteworthy that researchers have avoided over-satiation by typically rewarding consumption of fairly small quantities of foods.

Read full article

URL: https://www.sciencedirect.com/science/article/pii/S0195666311005150

When a response is strengthened because it is followed by the presentation?

Chapt 6 vocab -Learning.

When a response is strengthened because it is followed by the removal?

Negative Reinforcement In an attempt to increase the likelihood of a behavior occurring in the future, an operant response is followed by the removal of an aversive stimulus. This is negative reinforcement.

When a response is strengthened because it is followed by the removal of an aversive stimulus it is called?

Negative reinforcement: The strengthening of a response through the removal of a stimulus after the response occurs.

When an animal's innate response tendencies interfere with the conditioning process it is referred to as?

Instinctive Drift. When an animal's innate response tendencies interfere with conditioning processes.