Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống
1
/ 13 trang
THÔNG TIN TÀI LIỆU
Thông tin cơ bản
Định dạng
Số trang
13
Dung lượng
263,32 KB
Nội dung
Animal Learning & Behavior 2001, 29 (1), 66-78 Rats don’t always respond faster for more food: The paradoxical incentive effect LEWIS A BIZO, LAUREN C KETTLE, and PETER R KILLEEN Arizona State University, Tempe, Arizona Rats’ leverpressing was reinforced on variable-ratio (VR) schedules As ratio values increased, response rates initially increased with them, then eventually decreased In Experiment 1, rates were uniformly higher with one-pellet reinforcers than with two-pellet reinforcers—the paradoxical incentive effect Killeen’s (1994) mathematical principles of reinforcement (MPR) described the data quantitatively but failed to predict the advantage for the one-pellet condition In Experiment 2, rats received one-, two-, and three-pellet reinforcers with counterbalanced preloads of pellets; the continued superiority of the smaller reinforcers ruled out a satiation explanation Experiment introduced a 20-sec intertrial interval (ITI), and Experiment filled the ITI with an alternate response to test a memorial/overshadowing explanation In Experiment 5, the rats received one or two standard grain pellets or one sucrose pellet as reinforcers over an extended range of ratios Once again, rates were higher for one than for two pellets at short to moderate VR values; thereafter, two pellets supported higher response rates The diminution of the effect in Experiment and its reversal in Experiment and in Experiment at large ratios provided evidence for overshadowing and reconciled the phenomenon with MPR C(·)aR B }} , d (1 + aR) Killeen’s (1994, 1998; Killeen & Bizo, 1998) mathematical principles of reinforcement (MPR) are based on three assumptions: (1) An incentive can sustain only a limited amount (a seconds) of responding; (2) response rates are constrained by the time required for a single response (d ); (3) reinforcement only occurs when both a response and an incentive are contained within the same memory window—that is, when the response is coupled to the incentive The effectiveness of reinforcement is a joint function of the motivational factor, characterized by a (which depends on qualities of the incentive and state of the organism, such as its deprivation level), and the tightness of the coupling between the reinforcement and the target response This last factor is characterized by the coupling coefficient, which captures what historically have been called the contingencies of reinforcement These principles, instantiated in a mathematical model, were able to account for the ability of reinforcement to govern behavior in a variety of situations They are recapitulated here (1) where B is the rate of responding and R the rate of reinforcement The coupling coefficient was originally written as z, but here is given as C(·), to indicate that it is specific to a particular schedule (·) Notice that dividing by a renders this equivalent to Herrnstein’s hyperbola (Herrnstein, 1979), with C(·) / d k and l /a R The specific activation parameter, a, increases with the magnitude of the incentive: Doubling the incentive value (which may require more than doubling the amount of reinforcement; Killeen, 1985b) will have the same effect on response rates as doubling the rate of reinforcement (Leon & Gallistel, 1998) Reinforcement strengthens more than just the immediately preceding response; it strengthens whatever is in the organism’s memory at the time of reinforcement This may include other target responses or other interresponses that occur between measured target responses—pausing, “superstitious” responding, and stylistic ways of making an extended target response (Herrnstein, 1966) Interresponses are more likely on interval schedules than on ratio schedules, since the former nondifferentially reinforce any response (including another target response) that occurs before the final target response Interresponses may include consummatory behavior when that precedes the target response (as on schedules of continual reinforcement, CRF) In general, the reach of a reinforcer decays exponentially with time, and its contact with prior target responses may be overshadowed by other substantial activities, such as consummation of a reinforcer (Killeen & Smith, 1984; Shimp & Moffitt, 1977) or other intervening behavior (Kramer, 1982) The coupling coefficient has been derived for the basic reinforcement schedules Model and Predictions Killeen, Hanson, and Osborne (1978) showed that the motivational level of an organism is asymptotically proportional to the product of the rate of reinforcement (R) and the parameter a Because of ceilings on response rate, this proportionality is reflected in a hyperbolic relation between reinforcement rate and response rate The coupling coefficient C(·) multiplies that hyperbolic function: This research was supported by NSF Grant IBN 9408022 and NIMH Grant K05 MH01293 Correspondence should be addressed to L A Bizo, Department of Psychology, University of Southampton, Highfield, Southampton, S017 1BJ, England (e-mail: lewisb@soton.ac.uk) Copyright 2001 Psychonomic Society, Inc d > 0, 66 THE PARADOXICAL INCENTIVE EFFECT 67 (Killeen, 1994) In the appendix it is rederived for variable ratio (VR) schedules and shown to be n C(VR n) } } , n + (1 b ) / b 0 100) Here, we increase the range of VR values substantially to see whether the predicted incentive effects will be observed at larger ratio values We also manipulate reinforcer quality: If reinforcer quantity is held constant while reinforcer quality is varied, incentive effects should predominate Method Figure Top panel: response rates averaged across animals as a function of variable-ratio (VR) value for Experiment 4, which scheduled an intervening response reinforced during the intertrial interval (ITI) The rats received either one pellet or three pellets per reinforcer The curves are directed by Equation Bottom panel: response rates for nose poking averaged across animals as a function of the VR in operation for leverpresses Responses during the ITI were reinforced according to a random ratio (RR) p 1/30 The average standard errors of the mean are shown on the legends twice as high when one pellet was delivered as a reinforcer for leverpressing (empty circles) than when three pellets were delivered for leverpressing (filled circles) There was also a slight increase in nose-poking rate as the VR requirement for leverpressing increased An alternative to the memorial explanation of these results is a contrast explanation: The higher rates of reinforcement for nose poking under the one-pellet condition might have generated a contrast effect, differentially reducing rates of leverpressing for one pellet One remedy is to reconduct this experiment with constant rates of reinforcement for nose poking An alternative tactic is employed in Experiment Animals Six experimentally naive male Sprague-Dawley rats served as subjects They were maintained at 85% ± 10 g of their adlib weight and, if necessary, were postfed supplemental rodent chow to sustain their weight within the specified range The rats were housed individually, and all the other conditions of treatment were the same as those for the other experiments Apparatus Two operant chambers were utilized Rats 33, 34, and 38 completed all sessions in a chamber measuring 31 cm wide, 17 cm high, and 22 cm deep, enclosed within a Lehigh Valley (Laurel, MD) sound-attenuating chamber A white houselight was centered on the work panel 14 cm above the metal-rod floor A Coulbourn Instruments (Allentown, PA) retractable lever was located on the work panel 2.5 cm above the floor and cm from the left wall A force of 0.22 N applied to the lever activated a microswitch and registered with the computer A pellet dispenser (Davis Scientific Instruments, Model PD-104) clicked when delivering 45-mg pellets into a food tray centered on the work panel 2.5 cm above the floor Rats 35, 36, and 37 completed all the sessions in a chamber measuring 30 cm wide, 26 cm high, and 24 cm deep, enclosed within a Lehigh Valley (Laurel, MD) sound-attenuating chamber A white houselight was centered on the work panel 19 cm above a wiremesh floor A BRS/LVE (Laurel, MD) retractable lever was located on the work panel 3.5 cm above the floor and cm from the left wall A force of 0.45 N applied to the lever activated a microswitch and registered with the computer A Med Associates (Georgia, VT) pellet dispenser delivered 45-mg pellets into a food tray that was centered on the work panel cm above the floor Procedure The rats were initially trained to leverpress for food under a concurrent FT 90-sec FR schedule One pellet was delivered every 90 sec or after a leverpress occurred Once leverpressing occurred reliably, the rats were exposed to a VR schedule of reinforcement for three sessions, after which the experiment began Before each session, the rats were given 12 to habituate to the chamber with the houselight illuminated Following this habituation time, the lever was extended, and leverpressing was reinforced on a constant-probability VR schedule Sessions terminated after 90 reinforcers or 90 min, whichever occurred first Each reinforcer was followed by a 1-sec blackout, during which the lever was retracted A single session exposure was given to each of the fol- THE PARADOXICAL INCENTIVE EFFECT 75 two pellets supported much higher response rates, as is predicted by MPR The data from the sucrose pellets reinforce these interpretations Rates were highest for the sucrose pellet until ratios of around 300, when the two-Noyes-pellet condition caught up A sucrose pellet has incentive motivation (a) approximately equivalent to two pellets (and thus their congruence at high ratios), but it overshadows prior responses less than two Noyes pellets The reduced value of e permits the greater motivational value to carry rates above the two-pellet condition at moderate ratio values The effects shown in Figure are significant, both theoretically and statistically To show this, the data were detrended by fitting the model to the data for ratios between and 100, pooled across conditions: first, pooled across sucrose and one pellet, then across sucrose and two pellets, and finally across one pellet and two pellets Residual deviations were calculated Since each subject experienced each condition, it is possible to a paired t test for deviations from the model as a function of condition All the comparisons were significant beyond the p 01 level GENERAL DISCUSSIO N Figure Response rates from Experiment averaged across animals and plotted as a function of variable-ratio (VR) value (top panel) or its logarithmic transform (bottom panel) The rats received one Noyes pellet (empty circles), two Noyes pellets (filled circles), or one sucrose pellet (squares) per reinforcer The curves are directed by Equation The average standard errors of the mean are shown on the legends lowing VR values: 3, 5, 8, 11, 17, 26, 38, 50, 75, 100, 125, 150, 200, 250, 300, 400, 500, 600, 700, 800, 900, and 1,000 This series was repeated with three different reinforcers Reinforcement consisted of one 45-mg pellet (Noyes Formula A/1 Rodent Pellets), two 45-mg pellets, or one 45-mg sucrose pellet (Noyes Formula F Sucrose Pellet) These are referred to as Noyes, Noyes, and Sucrose The rats were exposed to the conditions in a counterbalanced order The second pellet, if scheduled, was delivered 0.33 sec after the first Conditions (reinforcement type) were changed after the completion of the VR progression or when the rat obtained no reinforcers for two consecutive sessions Results and Discussion The PIE was replicated, although at a magnitude somewhat diminished from the previous experiments The top panel of Figure shows the data plotted over a linear axis, and the bottom panel shows them plotted over a logarithmic axis One Noyes pellet generated higher response rates on ratios up to 100, close to the upper limit of ratios studied in the previous experiments Thereafter, Ours is not the first set of experiments to show that larger reinforcers not necessarily entail higher response rates: Bonem and Crossman (1988) reviewed the effects of reinforcer magnitude in a wide variety of animal learning studies and found exceptions to every generalization But only a few studies have shown an inverse effect of magnitude Svartdal (1993) found that humans would respond less vigorously for large magnitude payoffs than for small, as Di Lollo, Ensminger, and Notterman (1965) found to be the case for rats Black and Elstad (1964) found that rats ran faster in a straight alley for a 10-sec than for a 30-sec access to food Leslie and Toal (1994) found little effect of one versus four pellets on interval schedules and an inverse effect on ratio schedules They noted Bonem and Crossman’s “pessimistic conclusions” and recommendation that “further experiments should be of a theory-testing nature rather than be simply directed at the collection of further data” (p 119) Larger amounts of food often increase the postreinforcement pause (see, e.g., Harzem, Lowe, & PriddleHigson, 1978; Lowe, Davey, & Harzem, 1974), although in contrasted conditions of reinforcement, smaller amounts of food may cause longer pauses, a negative contrast effect (Blakely & Schlinger, 1988) Hatton and Shull (1983) found larger feedings to presage longer pausing than did smaller feedings on fixed-interval schedules—but only when they were intermixed, not blocked These results are similar to those found by Keesey and Kling (1961) on variable-interval schedules, and by Perone and Courtney (1992) and Baron, Mikorski, and Schlund (1992) on ratio schedules In the present study, pausing was longer after larger feedings (e.g., in all the phases of Experiment 3, median PRPs were 2.6 sec for the three-pellet conditions, 1.25 sec for the 2-pellet conditions, and 1.0 sec for the 1-pellet conditions) However, running rates—response 76 BIZO, KETTLE, AND KILLEEN rates calculated from the end of the pause until the next reinforcer—show the same effects as those reported here for overall rates Therefore, although a contributing factor, longer pauses after larger reinforcers cannot solely account for the data in this experiment Ours is not the first model to suggest that reinforcement has negative as well as positive effects on response rates Catania (1971) studied the effect of reinforcement on prior responding and later (Catania, 1973; Catania, Sagvolden, & Keller, 1988) derived a model, analogous to Equation 1, based on the inhibitory effects of reinforcement These equivocal effects of magnitude should not be taken as an indication that it is a weak controlling variable Pigeons can discriminate small differences in sizes of grain and strongly prefer larger sizes (Killeen, Cate, & Tran, 1993) When animals are given a choice between different amounts, preference is strongly controlled by amount (Neuringer, 1967), and the effect increases with the delay to reward (e.g., Ito, 1985) Rather than a weak independent variable, the problem is that absolute response rates are an insensitive dependent variable, especially at high rates of reinforcement One of the reasons for this insensitivity are ceilings on rate: Osborne (1978) showed that general activity increased to a much greater extent with increases in amount of reinforcement than did keypecking or leverpressing Incentive effects will be most clearly differentiated when larger reinforcers not also differentially reduce the transreinforcer effect of reinforcement, thus undermining their effectiveness at high rates of reinforcement This was shown in Experiment 4, where increased erasure by interposed tasks permitted a large nonparadoxical incentive effect to be manifest In choice experiments, there is often additional erasure by the trials procedure, changeover delays, and so forth, which will tend to level the erasure and let incentive effects predominate This analysis predicts that incentive effects will be greatest when the reinforcer is delayed, which will tend to equate transreinforcer effects This has been found (e.g., Green & Snyderman, 1980) Specific activation (a), the carrier of incentive effects in this theory, is the inverse of Herrnstein’s R o (see, e.g., de Villiers & Herrnstein, 1976), which covaries substantially with motivational operations—both incentive operations, such as changes in magnitude of reinforcers, and drive operations, such as satiation (see Petry & Heyman, 1997, for a recent review) Variations in a will affect response rates primarily at high ratio values: The derivative of Equation with respect to a is n /a 2, showing that a’s effect on rates is proportional to the mean ratio value n The incentive motivational effects of larger reinforcers were visible only at the highest ratio values At high rates of reinforcement (small n), the effects of variations in amount will be small At high incentive values (small 1/a ), the effects of variations in amount will be small Effects of manipulations in amount will be most noticeable when both rate of reinforcement and levels of motivation are low The equation of motion for interval schedules has similar properties: Its derivative with re- spect to a is an inverse function of both a and the square of the reinforcement rate Consistent with these predictions, Heyman and Monaghan (1994) showed that effects of changes in sucrose concentration had little effect on response rates at high rates of reinforcement, and Dallery, McDowell, and Lancaster (2000) found the effects of sucrose magnitude to be greatest at very low concentrations Increases in reinforcer magnitude will change a number of mediating factors of different importance in different experimental contexts: Incentive motivation will generally increase, hunger drive will sometimes decrease, and response–reinforcer coupling may decrease When animals can eat to satiation, different magnitudes of reinforcement have large effects on meal spacing (Collier, Johnson, & Morgan, 1992) In this case there is an inverse—but not paradoxical—incentive effect, because very large quantities of reinforcer satiate the animal for substantial periods Reed (1991) found that a normal magnitude effect (four-pellet response rates greater than onepellet response rates) on a VR 30 schedule was greatly attenuated when the hopper was illuminated during reinforcement and that a paradoxical magnitude effect (onepellet response rates greater than four-pellet response rates) on a random-interval 30-sec schedule was abolished when the hopper was illuminated during reinforcement Some portion of these effects might have been due to increased blocking in the one-pellet case when the hopper was illuminated As Leslie and Toal (1994) suggested, disentangling such convoluted effects may benefit more from theoretically driven experiments than from purely empirical studies The present set of experiments constitutes a step in that direction REFERENCES Baron, A., Mikorski, J., & Schlund, M (1992) Reinforcement magnitude and pausing on progressive-ratio schedules Journal of the Experimental Analysis of Behavior, 58, 377-388 Baum, W M (1993) Performances on ratio and interval schedules of reinforcement: Data and theory Journal of the Experimental Analysis of Behavior, 59, 245-264 Bizo, L A., Bogdanov, S V., & Killeen, P R (1998) Satiation causes within-session decreases in instrumental responding Journal of Experimental Psychology: Animal Behavior Processes, 24, 439-452 Bizo, L A., & Killeen, P R (1997) Models of ratio schedule performance Journal of Experimental Psychology: Animal Behavior Processes, 23, 351-367 Black, R W., & Elstad, P (1964) Instrumental and consummatory behavior as a function of length of reward-period Psychonomic Science, 1, 301-302 Blakely, E., & Schlinger, H (1988) Determinants of pausing under variable-ratio schedules: Reinforcer magnitude, ratio size, and schedule configuration Journal of the Experimental Analysis of Behavior, 50, 65-73 Bonem, M., & Crossman, E K (1988) Elucidating the effects of reinforcement magnitude Psychological Bulletin, 104, 348-362 Campbell, L S., & Dougan, J D (1995) Within-session changes in the VI-response function: Separating food density from elapsed session time Journal of the Experimental Analysis of Behavior, 64, 95-110 Catania, A C (1971) Reinforcement schedules: The role of responses preceding the one that produces the reinforcer Journal of the Experimental Analysis of Behavior, 15, 271-287 Catania, A C (1973) Self-inhibiting effects of reinforcement Journal of the Experimental Analysis of Behavior, 19, 517-526 THE PARADOXICAL INCENTIVE EFFECT Catania, A C., & Reynolds, G S (1968) A quantitative analysis of the responding maintained by interval schedules of reinforcement Journal of the Experimental Analysis of Behavior, 11, 327-383 Catania, A C., Sagvolden, T., & Keller, K J (1988) Reinforcement schedules: Retroactive and proactive effects of reinforcers inserted into fixed-interval performances Journal of the Experimental Analysis of Behavior, 49, 49-73 Collier, G., Johnson, D F., & Morgan, C (1992) The magnitudeof-reinforcement function in closed and open economies Journal of the Experimental Analysis of Behavior, 57, 81-89 Dallery, J., McDowell, J J., & Lancaster, J S (2000) Falsification of matching theory’s account of single-alternative responding: Herrnstein’s k varies with sucrose concentration Journal of the Experimental Analysis of Behavior, 73, 23-43 Deutsch, J A (1983) Dietary control and the stomach Progress in Neurobiology, 20, 313-332 de Villiers, P A., & Herrnstein, R J (1976) Toward a law of response strength Psychological Bulletin, 83, 1131-1153 Di Lollo, V., Ensminger, W D., & Notterman, J M (1965) Response force as a function of amount of reinforcement Journal of Experimental Psychology, 70, 27-31 Ferguson, S A., & Paule, M G (1995) Lack of effect of prefeeding on food-reinforced temporal response differentiation and progressive ratio responding Behavioural Processes, 34, 153-160 Ferguson, S A., & Paule, M G (1997) Progressive ratio performance varies with body weight in rats Behavioural Processes, 40, 177-182 Green, L., Kagel, J H., & Battalio, R C (1982) Ratio schedules of reinforcement and their relationship to economic theories of labor supply In M L Commons, R J Herrnstein, & H Rachlin (Eds.), Quantitative analysis of behavior: Vol Matching and maximizing accounts (pp 395-429) Cambridge, MA: Ballinger Green, L., & Snyderman, M (1980) Choice between rewards differing in amount and delay: Toward a choice model of self-control Journal of the Experimental Analysis of Behavior, 34, 135-147 Harzem, P., Lowe, C F., & Priddle-Higson, P J (1978) Inhibiting function of reinforcement: Magnitude effects on variable-interval schedules Journal of the Experimental Analysis of Behavior, 30, 1-10 Hatton, J L., & Shull, R L (1983) Pausing on fixed-interval schedules: Effects of the prior feeder duration Behaviour Analysis Letters, 3, 101-111 Herrnstein, R J (1966) Superstition: A corollary of the principles of operant conditioning In W K Honig (Ed.), Operant behavior: Areas of research and application (pp 33-51) New York: AppletonCentury-Crofts Herrnstein, R J (1979) Derivatives of matching Psychological Review, 86, 486-495 Heyman, G M., & Monaghan, M M (1994) Reinforcer magnitude (sucrose concentration) and the matching law theory of response strength Journal of the Experimental Analysis of Behavior, 64, 505-516 Ito, M (1985) Choice and amount of reinforcement in rats Learning & Motivation, 16, 95-108 Keesey, R E., & Kling, J W (1961) Amount of reinforcement and free-operant responding Journal of the Experimental Analysis of Behavior, 4, 125-132 Killeen, P R (1985a) The bimean: A measure of central tendency that accommodates outliers Behavior Research Methods, Instruments, & Computers, 17, 526-528 Killeen, P R (1985b) Incentive theory: IV Magnitude of reward Journal of the Experimental Analysis of Behavior, 43, 407-417 Killeen, P R (1994) Mathematical principles of reinforcement Behavioral & Brain Sciences, 17, 105-172 Killeen, P R (1995) Economics, ecologics, and mechanics: The dynamics of responding under conditions of varying motivation Journal of the Experimental Analysis of Behavior, 64, 405-431 Killeen, P R (1998) The first principle of reinforcement In C D L Wynne & J E R Staddon (Eds.), Models of action: Mechanisms for adaptive behavior (pp 127-156) Mahwah, NJ: Erlbaum Killeen, P R., & Bizo, L A (1998) The mechanics of reinforcement Psychonomic Bulletin & Review, 5, 221-238 Killeen, P R., Cate, H., & Tran, T (1993) Scaling of pigeon’s choice of grains: Bigger is better Journal of the Experimental Analysis of Behavior, 60, 203-217 77 Killeen, P R., Hanson, S J., & Osborne, S R (1978) Arousal: Its genesis and manifestation as response rate Psychological Review, 85, 571-581 Killeen, P R., & Smith, J P (1984) Perception of contingency in conditioning: Scalar timing, response bias, and the erasure of memory by reinforcement Journal of Experimental Psychology: Animal Behavior Processes, 10, 333-345 Kramer, S P (1982) Memory for recent behavior in the pigeon Journal of the Experimental Analysis of Behavior, 38, 71-85 Lattal, K A., & Gleeson, S (1990) Response acquisition with delayed reinforcement Journal of Experimental Psychology: Animal Behavior Processes, 16, 27-39 Leon, M I., & Gallistel, C R (1998) Self-stimulating rats combine subjective reward magnitude and subjective reward rate multiplicatively Journal of Experimental Psychology: Animal Behavior Processes, 24, 265-277 Leslie, J C., & Toal, L (1994) Varying reinforcement magnitude on interval schedules Quarterly Journal of Experimental Psychology, 47B, 105-122 Lowe, C F., Davey, G C L., & Harzem, P (1974) Effects of reinforcement magnitude on interval and ratio schedules Journal of the Experimental Analysis of Behavior, 22, 553-560 Mazur, J E (1983) Steady-state performance on fixed-, mixed-, and random-ratio schedules Journal of the Experimental Analysis of Behavior, 39, 293-307 McDowell, J J., & Wixted, J T (1986) Variable-ratio schedules as variable-interval schedules with linear feedback loops Journal of the Experimental Analysis of Behavior, 46, 315-329 Mosteller, F., & Tukey, J W (1977) Data analysis and regression: A second course in statistics Reading, MA: Addison-Wesley Neuringer, A J (1967) Effects of reinforcer magnitude on choice and rate of responding Journal of the Experimental Analysis of Behavior, 10, 417-424 Nevin, J A (1994) Extension to multiple schedules: Some surprising (and accurate) predictions Behavioral & Brain Sciences, 17, 145-146 Osborne, S R (1978) A quantitative analysis of the effects of amount of reinforcement on two response classes Journal of Experimental Psychology: Animal Behavior Processes, 4, 297-317 Perone, M., & Courtney, K (1992) Fixed-ratio pausing: Joint effects of past reinforcer magnitude and stimuli correlated with upcoming magnitude Journal of the Experimental Analysis of Behavior, 57, 33-46 Petry, N M., & Heyman, G (1997) Rat toys, reinforcers, and response strength: An examination of the Re parameter in Herrnstein’s equation Behavioural Processes, 39, 39-52 Powell, R W (1968) The effect of small sequential changes in fixedratio size upon the post-reinforcement pause Journal of the Experimental Analysis of Behavior, 11, 589-593 Pubols, B H (1960) Incentive magnitude, learning and performance in animals Psychological Bulletin, 57, 89-115 Reed, P (1991) Multiple determinants of the effects of reinforcement magnitude on free-operant response rates Journal of the Experimental Analysis of Behavior, 55, 109-123 Reed, P., & Wright, J E (1988) Effects of magnitude of food reinforcement on free operant response rates Journal of the Experimental Analysis of Behavior, 49, 75-85 Roll, J M., McSweeney, F K., Johnson, K S., & Weatherly, J N (1995) Satiety contributes little to within-session decreases in responding Learning & Motivation, 12, 323-341 Shimp, C P., & Moffitt, M (1977) Short-term memory in the pigeon: Delayed-pair-comparison procedures and some results Journal of the Experimental Analysis of Behavior, 28, 13-25 Skinner, B F (1938) The behavior of organisms New York: AppletonCentury-Crofts Svartdal, F (1993) Working harder for less: Effect of incentive value on force of instrumental response in humans Quarterly Journal of Experimental Psychology, 46A, 11-34 Timberlake, W (2000) Motivational modes in behavior systems In R R Mowrer & S B Klein (Eds.), Handbook of contemporary learning theories (pp 155-209) Mahwah, NJ: Erlbaum Toates, F (1986) Motivational systems New York: Cambridge University Press 78 BIZO, KETTLE, AND KILLEEN APPENDIX The Coupling Coefficient for Variable-Ratio Schedules On idealized VR schedules requiring a mean number of responses n, there is a constant probability p / n that any response will terminate the count and provide reinforcement At least one response must be made, guaranteeing coupling of b The probability that the sequence of responses will survive until the second response is (1 p) and that it will survive until the j th response is (1 p) j21 The last response—the one that triggers reinforcement—must always occur and will always receive a strengthening of b, the first addend in the following equation The penultimate response will occur with probability p and will receive a strengthening of b (1 b ), the second addend: C(n) b b (1 b )(1 p) b (1 b ) (1 p) + Continue this process ad infinitum to arrive at the coupling associated with a VRn schedule, C(VRn), here abbreviated as C(n): ¥ C (n) = å b (1 - b ) j -1 (1 - p) j -1 j =1 ¥ C (n) = b å b [(1 - b )(1 - p )] j -1 j =1 C ( n) = b - (1 - b )(1 - p ) C ( n) = + p(1 - b ) / b Substitute n / p to get Equation in the text The size of the ratio schedule required to attain a given level of saturation of memory [C(n)] is é C ( n) ù é - b ù nc = ê úê ú ë - C ( n) û ë b û Thus, if b 1, the number of responses necessary to provide 75% saturation of memory is: [.75/.25][.9/.1] 27 Incomplete Erasure The above assumes that if the first response after a previous reinforcer is reinforced, it receives a maximum coupling of b This is because the consummation of a reinforcer fills memory with that consummatory behavior (and its sequalae, such as postprandial area-restricted search, etc.) and those, rather than the earlier target responses, fill the remainder of memory But brief reinforcers may not completely overshadow the memory of prior target responses Call the proportion by which memory of prior target responses is erased after a reinforcement e Then, the response that triggered the prior reinforcer receives, in addition to the usual strengthening of b, the residual from the current epoch—diminished by the erasure e and the additional weakening created by the intervening target responses in the current epoch Under the assumption of complete erasure, the amount of association allocated to responses is given above With incomplete erasure, another series must be interleaved with it The response that triggers reinforcement now receives an additional strengthening of C + p(1 b ) (1 e)C + With probability p, the next response will also be reinforced; it will be distant by one intervening response and one intervening reinforcer and therefore will receive the additional strengthening of (1 b )(1 e)C + The first term represents the decrement that is due to the next response, the second to erasure by reinforcement The third term gives the coupling to all additional reinforcers in the future, a quantity we are in the process of calculating The next potential additional source of strength for a response that precedes reinforcement is another reinforcer two responses later The expected strengthening then becomes C + = p (1 - b )(1 - e )C + + p (1 - p )(1 - b ) (1 - e )C + And in general, N C + = p (1 - b )(1 - e )C + å [(1 - b )(1 - p )] j -1 j =1 Combining this surplus with the direct effects of reinforcement gives N C + = [ b + p (1 - b )(1 - e )C + ]å [(1 - b )(1 - p )] j -1 j =1 We may solve this equation for C +: C+ = bå , - p (1 - b )(1 - e ) å where S is the sum over the first N responses: å= - [(1 - b )(1 - p)]N - [(1 - b )(1 - p )] This result shows that the expected level of coupling will increase over the beginning of a session as N increases This effect may look like “warm-up,” but it involves contingencies/coupling, not motivation Asymptotically, C+ = b , b + e p (1 - b ) which further reduces to C+ = n n + e (1 - b )/b Compare this with Equation in the text Thus, after all this bookkeeping, we see that incomplete erasure of memory does not change the form of the model but just modifies the apparent value of the rate of decay of memory Since both e and b are usually free parameters, for ordinary use the coupling coefficient for VR schedules is simply a hyperbolic function of the ratio requirement: n C(VR n) } n+k (Manuscript received November 2, 2000; revision accepted for publication December 1, 2000.) ... reinforcer Run rates were calculated for the time from the f irst response to the time of the last response of the VR requirement Run rates therefore excluded the PRP from their time base The. .. for the three-pellet reinforcer zero-prefeed condition are the same in each panel Notice that prefeeding did not eliminate the paradoxical incentive effect effect, it was far from the level they... expected that the pure incentive effect would generate higher response rates for the larger reinforcer amounts Method Animals The rats were the same as those used in Experiment Apparatus The apparatus