ing the fixed-interval schedule-were observed during the interval of delay. Similarly, Ferster

Similar documents
J. E. R. STADDON DUKE UNIVERSITY. The relative inability of the usual differential. to ask whether performance under DRL schedules

CAROL 0. ECKERMAN UNIVERSITY OF NORTH CAROLINA. in which stimulus control developed was studied; of subjects differing in the probability value

THE EFFECTS OF TERMINAL-LINK STIMULUS ARRANGEMENTS ON PREFERENCE IN CONCURRENT CHAINS. LAUREL COLTON and JAY MOORE University of Wisconsin-Milwaukee

SECOND-ORDER SCHEDULES: BRIEF SHOCK AT THE COMPLETION OF EACH COMPONENT'

on both components of conc Fl Fl schedules, c and a were again less than 1.0. FI schedule when these were arranged concurrently.

CONCURRENT CHAINS UNIVERSITY-IMPERIAL VALLEY CAMPUS. links differ with respect to the percentage of

REPEATED MEASUREMENTS OF REINFORCEMENT SCHEDULE EFFECTS ON GRADIENTS OF STIMULUS CONTROL' MICHAEL D. ZEILER

PREFERENCE FOR FIXED-INTERVAL SCHEDULES: AN ALTERNATIVE MODEL'

CONDITIONED REINFORCEMENT IN RATS'

FIXED-RATIO PUNISHMENT1 N. H. AZRIN,2 W. C. HOLZ,2 AND D. F. HAKE3

Examining the Constant Difference Effect in a Concurrent Chains Procedure

DISCRIMINATION IN RATS OSAKA CITY UNIVERSITY. to emit the response in question. Within this. in the way of presenting the enabling stimulus.

NIH Public Access Author Manuscript J Exp Psychol Anim Behav Process. Author manuscript; available in PMC 2005 November 14.

REINFORCEMENT AT CONSTANT RELATIVE IMMEDIACY OF REINFORCEMENT A THESIS. Presented to. The Faculty of the Division of Graduate. Studies and Research

ANTECEDENT REINFORCEMENT CONTINGENCIES IN THE STIMULUS CONTROL OF AN A UDITORY DISCRIMINA TION' ROSEMARY PIERREL AND SCOT BLUE

that simple discrimination training to compound given a set to react -to one aspect of a stimulus a "set", a discrimination is "pretrained" along

Within-event learning contributes to value transfer in simultaneous instrumental discriminations by pigeons

Sum of responding as a function of sum of reinforcement on two-key concurrent schedules

CRF or an Fl 5 min schedule. They found no. of S presentation. Although more responses. might occur under an Fl 5 min than under a

Sequences of Fixed-Ratio Schedules of Reinforcement: The Effect of Ratio Size in the Second and Third Fixed-Ratio on Pigeons' Choice

REINFORCEMENT OF PROBE RESPONSES AND ACQUISITION OF STIMULUS CONTROL IN FADING PROCEDURES

CS DURATION' UNIVERSITY OF CHICAGO. in response suppression (Meltzer and Brahlek, with bananas. MH to S. P. Grossman. The authors wish to

Stimulus control of behavior during the postreinforcement pause of FI schedules

Reports from the Research Laboratories. J

pain" counted as "two," and so forth. The sum of point, the hour point, and so forth. These data are

Discover Activity. Think It Over Inferring Do you think height in humans is controlled by a single gene, as it is in peas? Explain your answer.

Concurrent schedule responding as a function ofbody weight

Stimulus variation and dimensional contrast

RESPONSE PERSISTENCE UNDER RATIO AND INTERVAL REINFORCEMENT SCHEDULES KENNON A. LATTAL, MARK P. REILLY, AND JAMES P. KOHN

REINFORCEMENT OF DRINKING BY RUNNING:

UNIVERSITY OF FLORIDA 1972

postreinforcement pause for a minute or two at the beginning of the session. No reduction

KEY PECKING IN PIGEONS PRODUCED BY PAIRING KEYLIGHT WITH INACCESSIBLE GRAIN'

Operant response topographies of rats receiving food or water reinforcers on FR or FI reinforcement schedules

AUTOSHAPING OF KEY PECKING IN PIGEONS

Value transfer in a simultaneous discrimination by pigeons: The value of the S + is not specific to the simultaneous discrimination context

The generality of within-session patterns of responding: Rate of reinforcement and session length

Study of the FR-TO effect

Genetic Variations. F1 Generation. Mechanisms of Genetics W W. STAAR Biology: Assessment Activities

NEW YORK UNIVERSITY. as in matching-to-sample or oddity performances, the rate of other, concurrently reinforced responses,

INTRODUCING NEW STIMULI IN FADING

INTERACTIONS AMONG UNIT PRICE, FIXED-RATIO VALUE, AND DOSING REGIMEN IN DETERMINING EFFECTS OF REPEATED COCAINE ADMINISTRATION

PIGEONS CHOICES BETWEEN FIXED-RATIO AND LINEAR OR GEOMETRIC ESCALATING SCHEDULES PAUL NEUMAN, WILLIAM H. AHEARN, AND PHILIP N.

The Post-Reinforcement Pause and Terminal Rate In Fixed-Interval Schedules

STIMULUS FUNCTIONS IN TOKEN-REINFORCEMENT SCHEDULES CHRISTOPHER E. BULLOCK

Observing behavior: Redundant stimuli and time since information

LOG- LINEAR ANALYSIS OF FERTILITY USING CENSUS AND SURVEY DATA WITH AN EXAMPLE

between successive DMTS choice phases.

Behavioural Processes

A concurrent assessment of the positive and negative properties of a signaled shock schedule*

EFFECTS OF INTERRESPONSE-TIME SHAPING ON MULTIPLE SCHEDULE PERFORMANCE. RAFAEL BEJARANO University of Kansas

Context effects on conditioning, extinction, and reinstatement in an appetitive conditioning preparation

Stimulus control of foodcup approach following fixed ratio reinforcement*

IDENTIFICATION OF C-REACTIVE PROTEIN FROM GINGIVAL CREVICULAR FLUID IN SYSTEMIC DISEASE

ANGELO SANTI, LORRAINE WEISE, and DWAYNE KUIPER Wilfrid Laurier University, Waterloo, Ontario, Canada

Conditions under which amobarbital sodium influences contrast in consummatory behavior

THE BASIS OF SUPERSTITIOUS BEHAVIOR: CHANCE CONTINGENCY, STIMULUS SUBSTITUTION, OR APPETITIVE BEHA VIOR? WILLIAM TIMBERLAKE AND GARY A.

Variability as an Operant?

Processing of empty and filled time intervals in pigeons

Variability Is an Operant

Some Parameters of the Second-Order Conditioning of Fear in Rats

Measures of temporal discrimination in fixedinterval performance: A case study in archiving data

Schedule Induced Polydipsia: Effects of Inter-Food Interval on Access to Water as a Reinforcer

Animal memory: The contribution of generalization decrement to delayed conditional discrimination retention functions

Reports from the Research Laboratories

Reinforced variation and selection

Comfort, the Intelligent Home System. Comfort Scene Control Switch

PROBABILITY OF SHOCK IN THE PRESENCE AND ABSENCE OF CS IN FEAR CONDITIONING 1

Human Schedule Performance with Hypothetical Monetary Reinforcement

Stimulus control of topographically tagged responding

Attention shifts during matching-to-sample performance in pigeons

STEPHEN P. KRAMER. (Kojima, 1980; Lattal, 1975; Maki, Moe, &

The ability of different strains of rats to acquire a visual nonmatching-to-sample task

Exploratory Approach for Modeling Human Category Learning

UNIVERSITY OF WALES SWANSEA AND WEST VIRGINIA UNIVERSITY

A comparison of signaled vs. unsignaled free-operant avoidance in wild and domesticated rats

Confounding variables in the discriminated Irt procedure.

SUBSTITUTION EFFECTS IN A GENERALIZED TOKEN ECONOMY WITH PIGEONS LEONARDO F. ANDRADE 1 AND TIMOTHY D. HACKENBERG 2

Serial learning, interitem associations, phrasing cues, interference, overshadowing, chunking, memory, and extinction

UNIVERSITY OF NORTH CAROLINA CHAPEL HILL AND UNIVERSITY OF NORTH CAROLINA WILMINGTON

Pigeons transfer between conditional discriminations with differential outcomes in the absence of differential-sample-responding cues

RESPONSE-INDEPENDENT CONDITIONED REINFORCEMENT IN AN OBSERVING PROCEDURE

EFFECTS OF D-AMPHETAMINE ON SIGNALED AND UNSIGNALED DELAYS TO REINFORCEMENT. Lee Davis Thomas

The effects of two different states of food deprivation for 6 roosters was measured with a

Travel Distance and Stimulus Duration on Observing Responses by Rats

ON THE EFFECTS OF EXTENDED SAMPLE-OBSERVING RESPONSE REQUIREMENTS ON ADJUSTED DELAY IN A TITRATING DELAY MATCHING-TO-SAMPLE PROCEDURE WITH PIGEONS

Transitive inference in pigeons: Control for differential value transfer

Schedules of Reinforcement

Minimizing Stimulus-Reinforcer Contributions to Operant Latency

Signaled reinforcement effects on fixed-interval performance of rats with lever depressing or releasing as a target response 1

Acquisition of bar-pressing under interval and ratio-schedules in the presence and absence of stimuli correlated with water delivery

EFFECTS OF A LIMITED HOLD ON PIGEONS MATCH-TO-SAMPLE PERFORMANCE UNDER FIXED-RATIO SCHEDULING. Joseph Leland Cermak, B.A.

Behavioral Contrast: A New Solution to an Old Problem

edited by Derek P. Hendry'

Pausing time allocation and constraint on output

Pigeons memory for time: Assessment of the role of subjective shortening in the duration-comparison procedure

of behavioral contrast. To the extent that an account of contrast is a major concern, therefore, some alternative formulation seems desirable.

Remembering: The role of extraneous reinforcement

Size invariance in curve tracing

SERIAL CONDITIONING AS A FUNCTION OF STIMULUS, RESPONSE, AND TEMPORAL DEPENDENCIES

Transcription:

JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAIOR 1969, 12, 375-383 NUMBER 3 (MAY) DELA YED REINFORCEMENT ERSUS REINFORCEMENT AFTER A FIXED INTERAL' ALLEN J. NEURINGER FOUNDATION FOR RESEARCH ON THE NEROUS SYSTEM When interreinforcement intervals ere equated, pigeons demonstrated little or no preference beteen reinforcement after a delay interval and reinforcement presented on a fixed-interval schedule. The small preferences sometimes found for the fixed interval (a) ere considerably smaller than hen the delay and fixed intervals differed in duration, and (b) ere caused by the absence of light during the delay. These results suggest that the effects of delayed reinforcement on prior responding can be reproduced by imposing a temporally equal fixed-interval schedule in place of the delay; and, therefore, that the time beteen a response and reinforcement controls the probability of that response, hether other responses intervene or not. Reinforcement is delayed by interposing an interval beteen a response and the reinforcement for that response. The delay is found to eaken responding: rates of responding are loer than hen reinforcement immediately follos a response, pauses are longer, ne responses and discriminations take more time to learn, and a delayed reinforcement is less likely to be chosen than an immediate one (Skinner, 1938, p. 139 ff; Perin, 1943; Perkins, 1947; Grice, 1948; Chung, 1965; Smith, 1967). Such results cause most experimenters to treat the effects produced by schedules of delayed reinforcement as different from those produced by schedules of immediate reinforcement. Hoever, the ork of Ferster and his associates suggests that hen the temporal parameters of these schedules are equated, the effects produced ith delayed reinforcements are replicatedl ith immediate ones. For example, Ferster (1953) shoed that pigeons' rates of responding prior to a 1-min delay period ere comparable to their rates prior to reinforcement presented on a 1-min fixed-interval schedule; furthermore, superstitious responses-perhaps analogous to responses dur- 'Dedicated to B. F. Skinner in his sixty-fifth year. This research as supported by grant M. H. 1218 from the National Institutes of Health to the Foundation for Research on the Nervous System, Dr. Shin-Ho Chung, principal investigator. I gratefully acknoledge the cheerful assistance of ivian Goldman, Mary Miskella, Maria Oliva, and Andrea Schein. Reprints may be obtained from the author, Foundation for Research on the Nervous System, 36 The Fenay, Boston, Massachusetts 2215. ing the fixed-interval schedule-ere observed during the interval of delay. Similarly, Ferster and Hammer (1965) demonstrated that the behavior of monkeys prior to a 24-hr delay of reinforcement interval as approximately the same as that prior to a 24-hr fixed-interval schedule. Technically, these experiments employed chain schedules, ith responses in the initial link of the chain leading to a fixedinterval terminal link in one case and a delayof-reinforcement link in the other. The to terminal links differed in the folloing ays: responses ere emitted throughout the fixed interval and the final response as immediately folloed by reinforcement; during the dlelay interval, effective responses could not be emitted and a response-independent reinforcement (i.e., one presented ithout regard to the animal's behavior) occurred at the end of the delay. The to schedules ere identical ith respect to other parameters, most important of hich as reinforcement frequency. The present ork extended Ferster's experiments by permitting pigeons to choose beteen (lelayed reinforcement and reinforcement on a temporally equal fixed-interval schedule. If, hen other parameters are equated, delayed and immediate reinforcers have the same effects on responding, subjects should choose equally beteen these to alternatives. On the other hand, if delaying a reinforcer has special effects on behavior, a lesser preference for the delay ould be predicted. 375

376 ALLEN J. NEURINGER METHOD Subjects Ten male White Carneaux pigeons, all ith previous experience in a variety of experiments, ere fed enough after each experimental session to maintain them at approximately 85% of their free-feeding body eights. Only five of these (Subjects 28, 281, 282, 31, and 34) ere used in Exp. I. Apparatus A standard operant conditioning chamber consisted of to Plexiglas Gerbrands response keys and a Gerbrands food hopper belo the keys. The keys could be transilluminated by 7- hite or 7- red bulbs, and the hopper as illuminated by a 7- hite bulb henever grain as presented. Except in Exp. III, to be discussed belo, these lights provided the only illumination in the chamber (i.e., there as no houselight). Reinforcement consisted of 3-sec access to mixed grain from the hopper. Pecks of at least 15 g force on a lighted key ere recorded and produced feedback clicks from a dc relay mounted behind the front panel. The experiment as automatically controlled by relays, stepping sitches, counters, and timers. EXPERIMENT I EQUAL DELAY AND FIXED INTERALS Procedure In the preliminary condition, pecks on the to response keys ere reinforced according to to independent variable-interval (I) 9- sec schedules, one I scheduling reinforcement on the left key and the other on the right. The interreinforcement intervals, identical on the to schedules, ere 12, 14, 23, 148, 44, 7, 11, 162, 8, 96, and 222 sec. The Is ere presented concurrently so that subjects could freely respond on either key and thereby choose beteen the to schedules. The only restriction as that a sitch from one key to the other prevented reinforcement for 1.5-sec. (This restriction is called a changeover delay, or COD, and as scheduled in the same ay as in Herrnstein, 1961.) When the numbers of responses became stable and approximately equal on the to keys, the procedure as changed to a concurrent chain (Herrnstein, 1964). Figure 1 is a schematic diagram of this procedure. The same 9-sec Is and 1.5-sec COD as used above ere presented in the concurrent initial links of to chain schedules. Responses during the initial link on one key no resulted in the occasional presentation (as determined by the respective I) of a terminal link composed of a fixedinterval (FI) schedule of reinforcement. Responses during the initial link on the other key resulted in the occasional presentation of a delay-of-reinforcement terminal link. Whereas the initial links ere presented concurrently, the terminal links ere presented exclusive of one another, i.e., the subject as confronted ith either a fixed-interval schedule or a delay of reinforcement. During the initial links, both keys ere lighted by hite bulbs. During the fixed-interval terminal link, the light transilluminating the "fixed-interval key" as changed from hite to red, and the other key, the "delay key," became dark and inoperative. The first peck on the fixed-interval key after the fixed interval terminated produced immediate access to grain. During the delay-of-reinforcement terminal link, the chamber as totally dark (blackout) and both keys ere inactivated. The delay as terminated by the response-independent presentation of grain. After either a delayed reinforce- Fig. 1. Diagram of one cycle of the experimental procedure. Each box represents a possible condition in the chamber. A hite circle indicates that the key is operative and lighted by a hite bulb, a striped circle indicates that the key is operative and lighted by a red bulb, and a black circle indicates that the key is dark and inoperative. Responding in the choice link (left) leads either to a delay (upper-middle) hich is folloed by response-independent reinforcement (upper-right) or to a fixed interval (loer-middle) hich is folloed by response-produced reinforcement (loer-right). After reinforcement in either condition, the choice link is again presented.

DELAY ERSUS FIXED INTERAL 377 ment or reinforcement on the fixed-interval schedule, the initial links ere again presented. Thus, subjects could freely respond on to keys during the initial links and thereby choose to enter either a delay-of-reinforcement or a fixed-interval terminal link. The basic differences beteen the delay and fixed-interval periods ere: (1) the chamber as totally dark during the delay, hereas the fixed-interval key as red during the fixed interval; therefore, (2) as as expected, fe or no pecks ere emitted in the darkened chamber during the delay, hereas many responses ere emitted on the lighted fixed-interval key during the fixed-interval; and (3) reinforcement as presented ithout regard to the subject's behavior at the end of the delay period, hereas a response produced reinforcement at the end of the fixed interval. The duration of the delay alays equalled the duration of the fixed interval in this experiment; both ill be referred to as the "time to reinforcement". This time as varied in the folloing order: 8, 2, 3, 6, 18, and 45 sec. Subjects could first choose beteen an 8-sec delay-of-reinforcement and an 8-sec fixedinterval schedule, then beteen a 2-sec delay and 2-sec fixed interval, etc. A total of 35 to 5 sessions as given at each time-to-reinforcement value, ith the fu'nctions of the keys being interchanged after a minimum of 15 sessions; thus, a minimum of 15 sessions as given ith the delay on the left and a minimum of 15 ith the delay on the right. Each session as terminated after the fortieth reinforcement. Preference as determined from the number of responses emitted on each key during the concurrently presented initial links. The number of initial-link responses emitted on the fixed-interval key as divided by the total number of initial-link responses on both keys. This datum, Fl choices divided by total choices, ill be referred to as the "per cent choice of the fixed-interval schedule". According to this per cent choice measure, 5% indicates that an equal number of responses ere emitted on the to keys during the initial links, 67% indicates that the Fl key as chosen tice for each choice response on the delay key, 75% indicates three to one, etc. Also to be discussed are (a) the overall rate of responding during the initial links, defined as the total number of initial-link responses on both keys divided by the total time spent in these links; and (b) the rate of responding during the fixed-interval terminal link, defined as the number of responses emitted during the fixedinterval divided by the time spent in the interval. Arithmetic average performances over 14 sessions-the last seven sessions hen the delay as on the left plus the last seven hen the delay as on the right-ere used for all data to be discussed in all of the folloing experiments. RESULTS Figure 2 shos the per cent choices of the fixed-interval key as a function of time to reinforcement. Most points lie close to, but above, the 5% line, indicating a small preference for the Fl condition. This preference as approximately constant as time to reinforcement varied. The average of all points is 55%, shon by the dashed line, or an average preference for Fl over delay of about 1.2 to 1. Note that these preferences did not greatly influence the relative number of reinforcements obtained from the to conditions. At the 3-sec time-to-reinforcement value, here the difference as greatest, 52.6% of the reinforcements ere from the Fl condition and 47.4% ere from the delay. The approximate invariance of the choices might indicate that time to reinforcement had no effect in the present experiment. That this as not the case is seen in Fig. 3, here overall rates of responding in the initial links are shon to decrease as time to reinforcement in the terminal links increased. Similarly, Fig. 4 shos that the rates of responding ithin the terminal fixed-interval component decreased as the duration of the fixed interval increased. Thus, although preferences did not vary, the rates at hich choices ere emitted and the response rates ithin the fixed interval decreased as time to reinforcement increased. EXPERIMENT II UNEQUAL DELAY AND FIXED INTERALS This experiment compared the magnitude of the 55% choice value in Exp. I ith preferences for the shorter of to times to reinforcement. Subjects ere therefore permitted to choose beteen unequal delay and fixed intervals.

ALLEN J. NEURINGER -J a C] I- C) - (I) -J -J -i 1 9 8 7 6 5 4 3 2 1 X 28 31 281 @ 34 a 282 Q AE 1 2 3 4 5 6 TIME TO REINFORCEMENT (SEC) (TERMINAL LINKS) Fig. 2. Per cent choice of the fixed-interval key (initial-link responses on the fixed-interval key divided by the total number of initial link responses on both fixed-interval and delay keys) as a function of the times to reinforcement during the fixed-interval and delay-of-reinforcement terminal links. The fixed and delay intervals ere equal. a Procedure A concurrent-chain schedule as again used. As in Exp. I, the initial concurrent links ere identical I 9-sec schedules ith COD 1.5- sec, and the terminal links consisted of a fixedinterval schedule on one key and a delay-ofreinforcement schedule on the other. The only difference beteen this experiment and Exp. I is that no the durations of the delay and fixed intervals differed. The same subjects used in Exp. I, no referred to as Group I, first chose beteen a 1-sec fixed-interval and a 2-sec delay, and then beteen a 1-sec fixedinterval and 2-sec delay. Each of these comparisons as presented for a minimum of 35 sessions, ith the fixed interval on the left for at least 15 sessions and then on the right for another 15 sessions. A second group of pigeons, Subjects 1, 5, 11, 26, and 45, received the same experience as Group I, except that the 1-sec condition as a delay of reinforcement hile the fixed interval as 2 sec in one case and 2 sec in the other. Thus, hile the time-to-reinforcement values ere identical for the to groups, the schedule conditions ere reversed, i.e., the fixed interval as 1 sec for Group I hereas the delay as 1 sec for Group II. To groups ere used in order to factor out the effects of the type of sched-

DELAY ERSUS FIXED INTERAL 379 16 15 _ * I1 I- 1 i 9 C _% CL 8 -J 7 cn 4 _j O p 6 en i, C-5 o 4 I o 3 2 1 v Q X 28 v 31 281 34 A 282 AE 271 233' v x 1% 1 2 3, 4 5 TIME TO REINFORCEMENT (TERMINAL LINKS) W -i 21 195 18 ' 15 Z 135 Lu X 12 v x x x 6 (SEC) Fig. 3. Choice responses per minute (total initial-link responses on both keys divided by the total time in the initial link) as a function of the times to reinforcement during the terminal links. The durations of the fixedand delay-interval terminal links ere equal. ule (delay versus FI) from the effects of the times of reinforcement. The schedule effects can be seen by comparing Group I and Group II performances; the effects of time to reinforcement can be seen by comparing choices during the 1-sec versus 2-sec condition ith choices during the 1-sec versus 2-sec condition. RESULTS Figure 5 shos the percentage of initiallink choice responses on the "1-sec key", i.e., the key on hich the terminal link as either a 1-sec fixed interval (Group I) or 1-sec delay (Group II). The average Group I performances are shon by the open bars and 9 en 15 a9 (-cc n3-.o I&& 7I, U 6 45 3 A X 28 31 281 34 A 282 OAE X S a v o I I I I I 1 2 3 4 5 6 TIME TO REINFORCEMENT (FIXED INTERAL) (SEC) Fig. 4. Rate of responding during the fixed-interval terminal link as a function of the duration of the fixed interval. the average Group II performances by the striped bars. Consider first the effects of changing the comparison time to reinforcement from 2 sec to 2 sec. This change caused the average Group I choices of the 1-sec key to increase from 31% to 72%; Group II choices similarly increased from 25% to 67%. Consider next the effects of the type of schedule. Note again that any difference beteen the to groups' choices must be attributed to the difference in schedule conditions. The Group I choices of the 1-sec condition ere, on the average, six and five percentage points higher than the Group II choices. Thus, hereas an

38 ALLEN J. NEURINGER l) LL I. 1 F- 8 6 4 2 T GROUP I FI=lOo" xv x 1 GROUP Mr DELAY=O1" X 28 * I 281 M 5 A 282 A 11 31 v 26 34 45 2 SEC 2 SEC IME TO REINFORCEMENT ON COMPARISON KEY Fig. 5. Per cent choices of the 1-sec fixed-interval condition hen the delay condition as 2 sec and 2 sec, respectilvely (Group I, open bars), and of the 1-sec delay of reinforcement condition hen the fixed-interval as 2 sec and 2 sec (Group II, striped bars). The beteen-group differences sho the effects of the different schedule conditions (FI versus delay), hereas the ithin-group changes sho the effects of the increase in time to reinforcement. increase in time to reinforcement caused per cent choices to increase by approximately 42 points, the difference beteen Fl and delay schedules caused per cent choices to differ by only six points. These results are consistent ith the small preferences for Fl over delay found in Exp. I and, furthermore, suggest that those preferences ere indeed relatively small. 6 A m EXPERIMENT III CONTROLLING FOR THE BLACKOUT Attempts ere next made to determine hether the blackouts per se ere responsible for the small preferences for Fl over delay obtained in Exp. I and II. The blackout as removed from the delay condition in Exp. IIIA and added to the fixed interval in Exp. IIIB. Procedure Experiment IIIA: a 7- yello houselight as added to the experimental chamber, and as lighted throughout the session; thus, there ere no blackouts. Subjects 28, 281, 282, 31, and 34 chose beteen a 2-sec fixed-interval terminal link and a 2-sec delay of reinforcement. During the delay, hile both keys ere dark and inoperative, the chamber remained lighted by the houselight. Except for the absence of blackout, this procedure as identical to that in Exp. I. If the blackout caused the slightly loer preference for delay in Exp. I and II, the subjects ould no be expected to respond equally in the initial links of the chains; on the other hand, if the preferences ere due to some other characteristic of the delay operation (e.g., the response-independent presentation of grain), the same 55% choice value ould be predicted. Subjects received at least 3 sessions' experience under this condition, ith the position of the delay and fixed-interval conditions being interchanged after approximately 15 sessions. Experiment IIIB: Subjects 1, 5, 11, 26, and 45 chose beteen a 2-sec delay of reinforcement and a 2-sec "delay-plus-fr 1" (delay plus one response) terminal link. The delays in both conditions ere identical to those in Exp. I, i.e., the chamber as completely dark (blackout). The delay consisted of a 2-sec period of blackout at the end of hich response-independent reinforcement occurred. The delay-plus-fr 1 condition also contained a 2-sec period of blackout; hoever, at the end of this blackout the "delay-plus-fr 1 key" as transilluminated by a yello 7- bulb and a single peck on this lighted key produced immediate reinforcement. Except for the delay-plus-fr 1 contingency, the present procedure as identical to that used in Exp. I. If the blackout stimulus alone caused

DELAY ERSUS FIXED INTERAL 381 the obtained preferences, subjects ould no be expected to respond equally in the initial links. On the other hand, if another attribute of delayed reinforcement caused the previous results, the subjects should no prefer the immediate reinforcement in the delay-pius- FR 1 condition. Subjects received approximately 3 sessions under this condition, ith the positions of the keys being interchanged after approximately 15 sessions. RESULTS The basic data ere the percentages of initial-link responses emitted on the fixed-interval key in Exp. IIIA (or per cent choices of the fixed interval) and the percentages of initial-link responses to the delay-plus-fr 1 key in Exp. IIIB. The per cent choices of the Fl key in Exp. IIIA ere 56, 42, 48, 49, and 56 for the subjects in the order given in the procedure above. The average of these is 5%, indicating indifference beteen the to alternatives. Some subjects responded on the darkened and inoperative key during the delay (since the houselight as on) hereas other subjects responded little or not at all. There as no correlation beteen this responding and per cent choice. The per cent choices of the delay-plus-fr 1 key for subjects in Exp. IIIB, in the order given above, ere 57, 47, 5, 55, and 47, respectively, for an average of 51%. Once again, this average indicates an approximate indifference beteen the to alternatives. In neither Exp. IIIA, here blackouts ere absent in both delay-of-reinforcement and fixed-interval conditions, nor in Exp. IIIB, here blackouts ere present in both conditions, ere the Exp. I preferences (55%) obtained. These results therefore suggest that the loer preferences for delayed reinforcement in Exp. I and II ere caused by the blackout stimulus per se. DISCUSSION Hull (1952, p. 126 ff) distinguished beteen to basic types of delay-of-reinforcement experiments: in the first, response chains are required during the "delay" interval and the last response in the chain is immediately folloed by reinforcement, e.g., as in a multiple unit mae; in the second, no specific response is required during the delay and responseindependent reinforcement occurs at the end of the delay, e.g., as in an operant chamber here the operandum is removed during the delay interval. Most studies on delayed reinforcement use the second type of situation: the operandum is inactivated or removed, the chamber is darkened, or responding during the delay is punished (Perin, 1943; Ferster, 1953; Des, 196; Chung, 1965). Studies of Hull's first type of "delay" are no found under the rubric of "schedules of reinforcement" and, more specifically, under "chain schedules". The basic question raised in the present ork is hether these to situations, no to be referred to as delay of reinforcement and chain schedules of reinforcement, can be integrated ithin a single frameork. To put the question another ay, are the behavioral effects of a delayed reinforcement due to attributes unique to delay (e.g., an interval of "no responding" beteen response and responseindependent reinforcement) or to attributes common to all chain schedules (e.g., the interval beteen initiation of the chain's terminal link and reinforcement)? To anser this question, pigeons ere permitted to choose beteen a schedule of delayed reinforcement and a temporally equal fixed-interval schedule. It as found that an average of 55% of the choice responses ere emitted on the "fixed-interval key", indicating a 1.2 to 1 preference for fixed interval rather than delay. While this preference might suggest that attributes unique to delayed reinforcement affect responding, the preference as both approximately constant over a range of intervals (Exp. I) and relatively small (Exp. II), suggesting that some factor other than the delay might have been responsible. That factor as shon to be the blackouts present during the delay interval: removing (Exp. IIIA) or controlling for (Exp. IIIB) the blackouts caused the subjects to choose equally beteen the delay and fixed-interval alternatives. (Note that hereas some studies similarly suggest that blackouts are aversive, e.g., Ferster, 1958, other studies demonstrate the opposite effect, e.g., Neuringer and Chung, 1967). Thus, it is concluded that pigeons demonstrate relatively little or no preference in their choices beteen a delayed reinforcement and a temporally equal fixed-interval schedule of reinforcement. This conclusion does not imply that the interval beteen a response and reinforce-

382 ALLEN J. NEURINGER ment is an unimportant variable. To the contrary, Fig. 3 and 4 sho that as this interval increases, response rates decrease, and Fig. 5 shos that subjects prefer the shorter of to times to reinforcement. The present results, together ith those of Ferster (1953) and Ferster and Hammer (1965), do imply, hoever, that the main effects of a delay interval on behavior are due to the interval rather than to attributes unique to the delay operation; in other ords, hether or not further responding occurs during the interval, and hether or not a response is immediately folloed by reinforcement at the end of the interval, make little or no difference in subjects' preferences. Other studies similarly suggest that interreinforcement responses exert relatively little influence over behavior, hereas interreinforcement intervals exert a relatively great control. For example, Anger (1956) and Herrnstein (1964) shoed that responding by rats in a single operandum situation and pigeons in a choice situation, respectively, as better correlated ith reinforcements per unit time than ith reinforcements per response. Des (1962, 1965, 1966a, 1966b) argued that pigeons' rates of responding under fixed-interval schedules ere controlled by the time beteen reinforcements and did not depend upon the responses emitted during this time. Autor (196) and Killeen (1968) demonstrated that the choices of pigeons varied ith reinforcement rates hether or not responses ere required for reinforcement (see, hoever, Fantino, 1968 for different results). And Neuringer and Schneider (1968) found that the response latencies of pigeons increased linearly ith interreinforcement time but ere completely unrelated to the number of interreinforcement responses. These studies, together ith the present findings, suggest the folloing hypothesis: the probability (or rate, or latency) of a response is controlled by the interval beteen that response and reinforcement (a) independently of the number of other responses intervening in the interval, andl (b) independently of hether such intervening responses are required or prohibited. (Note that a single response, the one that produces reinforcement, is required under the fixed-interval schedule, hereas effective responses are prohibited under the delay schedtile.) More research is of course necessary to substantiate this hypothesis. For example, it must be determined hether the effects of delay of reinforcement intervals on the learning of ne responses and discriminations might also be explained by the intervals rather than by attributes unique to the delay. The series of experiments by Des (1962, 1965, 1966a, 1966b) has suggested that the scalloped patterns of responding found under fixed-interval schedules are caused not by hypothetical response chains but by the passage of time. The present ork suggests a similar alternative to the notion that "superstitious chains of responses" must be invoked to explain the control exerted by delayed reinforcers. As indicated in the Introduction, Ferster (1953), as ell as other experimenters (Blough, 1959; Hearst, 1962) have reported superstitious responses during the delay period. Hoever, the animal often is not observed during the delay (e.g., Chung and Herrnstein, 1967), and, hen observed, superstitious chains are sometimes not found (Mabry, 1965; Smith, 1967). Thus, the more general, as ell as more parsimonious, hypothesis is the present one, i.e., that the time beteen a response and reinforcement controls the probability of that response, hether other responses intervene or not. REFERENCES Anger, D. The dependence of interresponse times upon the relative reinforcement of different interresponse times. Jouirnal of Experimental Psychology, 1956, 52, 145-161. Autor, S. M. The strength of conditioned reinforcers as a function of the frequency and probability of reinforcement. Unpublished doctoral dissertation, Harvard University, 196. Blough, D. S. Delayed matching in the pigeon. Journal of the Experimental Analysis of Behavior, 1959, 2, 151-16. Chung, S-H. Effects of delayed reinforcement in a concurrent situation. Journal of the Experimental Analysis of Behavior, 1965, 8, 439-444. Chung, S-H. and Herrnstein, R. J. Choice and delay of reinforcement. Journal of the Experimental Analysis of Behavior, 1967, 1, 67-74. Des, P. B. Free-operant behavior under conditions of delayed reinforcement. I. CRF-type schedules. Journal of the Experimental Analysis of Behavior, 196, 3, 221-234. Des, P. B. The effect of multiple SA periods on responding on a fixed-interval schedule. Journal of the Experimental Analysis of Behavior, 1962, a, 369-374. Des, P. B. The effect of multiple SA periods on re-

DELAY ERSUS FIXED INTERAL 383 sponding on a fixed-interval schedule: III. Effect of changes in pattern of interruptions, parameters and stimuli. Journal of the Experimnental Analysis of Behavior, 1965, 8, 427-435. Des, P. B. The effect of multiple SI periods on responding on a fixed-interval schedule:i. Effect of continuous SI ith only short SD probes. Journal of the Experimental Analysis of Blehavior, 1966, 9, 147-151. (a) Des, P. B. The effect of multiple SI periods on responding on a fixed-interval schedule:. Effect of periods of complete darkncss anid of occasional omissions of food presentationis. Journal of the Experimental Analysis of Behavior, 9, 1966, 573-578. (b) Fantino, E. Effects of required rates of responding upon choice. Journal of the Experimental Analysis of Behavior, 1968, 11, 15-22. Ferster, C. B. Sustained behavior under delayed reinforcement. Journal of Experimental Psychology, 1953, 45, 218-224. Ferster, C. B. Control of behavior in chimpanees and pigeons by time out from positive reinforcement. Psychological Monographs, 1958, 72, No. 14 (Whole No. 461). Ferster, C. B. and Hammer, C. ariables determining the effects of delay in reinforcement. Journal of the Experimental Analysis of Behavior, 1965, 8, 243-254. Grice, G. R. The relation of secondary reinforcement to delayed reard in visual discrimination learning. Journal of Experimental Psychology, 1948, 38, 1-16. Hearst, E. Delayed alternation in the pigeon. Journal of the Experimental Analysis of Behavior, 1962, 5, 145-161. Herrnstein, R. J. Relative and absolute strength of response as a function of frequency of reinforcement. Journal of the Experimental Analysis of Behavior, 1961, 4, 267-272. Herrnstein, R. J. Secondary reinforcement and rate of primary reinforcement. Journal of the Experinental Analysis of Behavior, 1964, 7, 27-36. Hull, C. L. A behavior system. Ne Haven: Yale University Press, 1952. Killeen, P. Response rate as a factor in choice. Psychonomic Science, 1968, 12, 34. Mabry, J. H. Discriminative functions based on a delay in the reinforcement relation. Journal of the Experimental Analysis of Behavior, 1965, 8, 97-13. Neuringer, A. J. and Chung, S-H. Quasi-reinforcement: control of responding by a percentage-reinforcement schedule. Journal of the Experimnental Analysis of Behavior, 1967, 1, 45-54. Neuringer, A. J. and Schneider, B. A. Separating the effects of interreinforcement time and number of interreinforcement responses. Journal of the Experimental Analysis of Behavior, 1968, 11, 661-667. Perin, C. T. The effect of delayed reinforcement upon the differentiation of bar responses in hite rats. Journal of Experimental Psychology, 1943, 32, 95-19. Perkins, C. C. Jr. The relation of secondary reard to gradients of reinforcement. Journal of Experimental Psychlology, 1947, 37, 377-392. Skinner, B. F. The behavior of organismns. Ne York: Appleton-Century-Crofts, 1938. Smith, L. Delayed discrimination and delayed matching in pigeons. Journal of thte Experimental Analysis of Behavior, 1967, 1, 529-533. Received 16 October 1968.