Inference in Dynamic Environments Mark Steyvers & Scott Brown, UC Irvine

Size: px
Start display at page:

Download "Inference in Dynamic Environments Mark Steyvers & Scott Brown, UC Irvine"

Transcription

1 Inference in Dynamic Environments Mark Steyvers & Scott Brown, UC Irvine 1 Introduction 1.1 Background Understanding how people make decisions, from very simple perceptual to complex cognitive decisions, is an important area of research in psychology. A great deal of effort has gone into finding empirical regularities of decision making behavior, and into developing models of underlying decision processes. A remarkable similarity of almost all previous research into decision making is the assumption of stationarity. With few exceptions (e.g., Kac, 1962; Rabbit, 1981; Triesman & Williams, 1984; Vickers & Lee, 1988, 2000), models of decision making assume that successive decisions are independent. This assumption probably cannot be true, but has proven useful in keeping models of decision making simple and tractable. Assuming stationarity also seems quite reasonable given that most every decision making experiment has employed a stationary decision making environment. Of course, real decision making environments are dynamic rather than stationary. Consider the case of a military observer making decisions about the identity (friend vs. enemy) of noisy stimuli from reconnaissance pictures. The difficulty of these decisions will change throughout the task, as more or less clear pictures are used, or more or less uniform terrain is observed. An ideal observer must dynamically adjust their decision making process to reflect changes in the environment. For example, if it becomes easier to identify friendly stimuli in new terrain, observers should relax their criterion for identifying enemy stimuli. Some prior research has attempted to address this classic criterion setting problem. Triesman and Williams (1984) developed a dynamic variant of signal detection theory (SDT) in which the decision criterion changed from decision to decision, based on the previous stimuli and responses. Vickers and Lee (1988, 2000) developed a similar idea for a much more complex decision making model. Our paradigm departs from previous work in some important ways, illustrated in Table 1. In terms of dynamics, the simplest kind of study is one in which the experimental design is static. A static (or stationary) design is one in which the properties of the task do not change during the experiment, hence participants do not need to change their behavior during the experiment in order to remain optimal. The critical feature of a static experiment is that the experimental conditions are drawn from a stationary distribution throughout the experiment, at least from the point of view of the participant. The upper left cell in Table 1 represents static experiments, in which the analyses are also static (i.e, no consideration of sequential effects, or the effects of task history). Experiments with between-subjects designs are typical of this category the task for a given participant does not change during the experiment, and there is no compelling reason to consider sequential effects. Static experiments are limited in their design, so researchers typically use dynamic experiments, meaning that experimental conditions change with time, forcing participants to adjust their decision making processes in order to remain optimal. The lower left cell of Table 1 represents dynamic experiments with static analyses: i.e., analyses that neglect the effect of task history and sequential dependencies. Static Analyses (No sequential effects.) Dynamic Analyses Static Experiment Between-subjects designs; early SDT experiments. Chaos, nonlinear dynamics, prior research into criterion setting. Dynamic Experiment Within-subjects blocked designs, including much of cognitive psychology research. Our research proposal. Table 1: Types of dynamic research

2 Research of this kind is very common in psychology: blocked designs are used, making the task dynamic, but static analyses are applied because researchers (often implicitly) assume that sequential dependencies between blocks are either unimportant or unmeasurable. This kind of research includes some of our own work (e.g., Heathcote, Brown & Mewhort, 2000; Shiffrin & Steyvers, 1997). Some researchers acknowledge that the effects of task history may be important, but still decide to use static analyses, and control for dynamic effects by censoring some data (e.g., from the start of each block). Static experiments (those with unchanging tasks) can also be subjected to dynamic analyses (top right cell of Table 1). These cases have so far represented almost all dynamic research, including most research into the presence of chaos or nonlinear dynamics in behavioral data (e.g., Kelly, Heath & Longstaff, 2001; Van Orden, Moreno & Holden, 2003), and previous seminal examinations of the criterion setting problem (e.g., Kac, 1962; Rabbit, 1981; Triesman & Williams, 1984). For example, in Triesman and Williams' research, the decision criterion was allowed to vary based on previous responses, but these variations were essentially spontaneous and unnecessary: an ideal observer in Triesman and Williams experiments would not have varied their decision making criterion at all. This is in contrast to many real-world decision making situations, where optimal behavior necessitates adjusting to ( tracking ) the changing environment. 1.2 Proposed direction of research Applying dynamic analyses to dynamic experimental designs (bottom right cell of Table 1) is a new direction we investigate in this proposal. The extra complexity introduced by having both dynamic models and experiments means that we are forced to keep both the experimental design and the modeling as simple as possible, to maintain tractability. We propose here new research that avoids both the assumptions of stationary decision processes and stationary decision environments. We develop an experimental paradigm in which dynamic decision making environments force participants to change their decision making processes in order to remain (approximately) ideal. This paradigm allows us to observe decision makers tracking changes in the environment. We also propose the continued development of two models for the decision process in dynamic environments. One model is an ideal observer system in which statistical evidence for a changed environment is weighed in optimal fashion against evidence for a stable environment. The ideal observer analysis results in estimates for the (optimal) number of trials it takes to detect and adjust to new decision environments. Our other model is a dynamic SDT model that estimates how long it actually takes for individual decision makers to adapt to novel decision environments. By comparing predictions from the ideal observer model to the parameter estimates from the decision model (from individual decision makers), we can quantify the degree of mismatch between ideal and actual observer. Taking the above reconnaissance picture example, a good observer would accurately track changes in the difficulty of the pictures, so it seems logical that the best person to select for this task is one who can change their decision environment as quickly as the ideal observer. It is also of interest to measure the direction of mismatch between ideal and actual observer. For example, some observers might change their decision processes too quickly or too late (compared to the ideal observer) given a change in the decision environment. A change in either direction will lead to suboptimal behavior. In tasks where there is a high penalty for falsely detecting a change, selection of good observers might favor late switchers. Similarly, if the task requires quick change detection with a severe penalty of detecting a change too late, selection of good observers might favor early switchers. 1.3 Overview of proposal In Section 2, we develop a new paradigm for studying dynamic decision making; this paradigm forms the basis for much of the research outlined in this proposal. In Section 3, we sketch a theory of change detection, based on the principles of ideal observer models. Section 4 gives an overview of our experiments and presents data from four completed experiments (over 400 participants) using the new paradigm, investigating individuals' abilities to track dynamic decision making environments. Section 4.2 develops a simple dynamic version of signal detection theory that we use as a measurement model, and shows its application to existing data. Section 4.3 develops our ideal observer model of change detection in greater detail, and illustrates its application by - 2 -

3 Difficult Decision Environment Easy Difficult Easy Ideal Observer decision parameter Subject s estimated decision parameter Lag between ideal and actual switch point Lag between subject s and ideal switch point Time Figure 1: Basic paradigm. Decision environments change (top) leading to changes in the ideal observer decision processes (middle). Participants actual processes (bottom) lag behind the ideal observer model. comparison with data from our preliminary experiments. Section 5 outlines a series of new experiments we propose to extend our research. These experiments are intended to have both basic scientific value (as investigations of decision making processes) and applied value (as new measurement tools). 2. Decision-making paradigm The decision paradigm we use is based on a simple two-alternative forced-choice design. In theory, any kind of decision stimuli could be used; for the first illustration we will continue with the reconnaissance picture example, but we describe many other examples below. To illustrate the paradigm, suppose there are a series of reconnaissance pictures, with a single target in each. The observer's task is to classify each target as an enemy stimulus or a friendly stimulus, based on physical characteristics (shape, color, markings, etc.). Two different decision environments could be defined by changing the properties of the enemy stimuli. In one environment, the enemy stimuli may look quite dissimilar from the friendly stimuli, making decisions relatively easy. In the other, enemy and friendly stimuli may be much more alike, resulting in relatively hard decisions. We can then construct a dynamic decision environment by alternating sequences of easy and hard decisions, as shown in the top panel of Figure 1. An ideal observer would have to raise the bar for what is considered an enemy stimulus in the hard condition, as enemy stimuli look very similar to friendly stimuli there, and they should lower the bar for what is considered an enemy stimulus when in the easy decision environment. As illustrated in the middle row of Figure 1, even our ideal observer model cannot immediately tell that a change in the environment has occurred. The ideal observer model must instead observe several pictures before it gathers enough statistical evidence to decide that the environment has changed, and that it must therefore change its decision making processes. As illustrated in the bottom part of Figure 1, an actual observer might deviate from the ideal observer by taking more time before adapting to the new environment. The comparison between actual and ideal observer results in a theoretically interesting and practically relevant measure: the lag between the switch point as estimated by the ideal observer and by the actual observer. An important property of our paradigm is that one stimulus class (the friendly stimuli, in this case) does not change during the experiment. Thus, any changes observed in response to those stimuli must be caused by changes in the decision processes of the observer. Continuing our reconnaissance picture example, if an observer raises their criterion in response to finding more difficult enemy stimuli, this should result in an observable increase in the number of friendly stimuli incorrectly identified as enemies. 3. Ideal Observers in Change Detection Tasks The paradigm outlined above is realistic in that participants are required to make a decision about every stimulus. They are thus forced to respond to changes in the environment only as a consequence of this task to maintain optimality in decision making, they must adjust to changes in the decision environment. This contrasts with other - 3 -

4 Signal Strength Signal Strength Signal Strength Signal Strength Actual Switch Point Time Best Hindsight Switch Point Early Switcher Hindsight error = -60 trials Ideal Switcher Hindsight error = +10 trials Late Switcher Hindsight error = +60 trials With all data, best hindsight switch point = 205 trials Figure 2. Illustration of switch point detection task, and ideal observer models. statistical research investigating change detection. Such research usually begins with the aim of detecting a change point in a complete sequence of data (e.g., by application of the Viterbi algorithm and estimation of a hidden Markov model). In our paradigm, participants must detect the change in a sequence as the sequence unfolds. We envisage the ideal observer model for our paradigm as follows. For simplicity, imagine that an observer is presented with only a sequence of numbers, perhaps representing the strength of sonar echo signals. They are told that, at exactly one point during the sequence, the distribution of these numbers will change, perhaps signaling the presence of a sonar target. Observers could operate by slowly gathering evidence about the distribution of the signals, and responding when they think this distribution has changed. This decision task is illustrated in Figure 2. The sequence of black circles represents a sequence of samples from the distributions. These are shown to the observer, one at a time. Note that the switch point is half way through the series at trial 200. The data in the first 200 trials sampled from N(0,1). The data in the last 200 trials are sampled from N(1.25,1). The observer s task is to identify when the change has occurred, as the sequence unfolds. That is, after each data point is presented, the observer is asked Has the switch point passed yet?, and they are allowed to answer yes only one time. An ideal observer model for this task may proceed using a nested model comparison framework. For the moment, suppose that the ideal observer knows the theoretical distribution of the non-signal data (occurring in the early parts of the sequence) and of the signal data (arriving after the switch point), but doesn t know the location of the switch point. After each data point arrives, the ideal observer could evaluate the evidence for two nested models. The first model is that the data so far have been generated by just one distribution (the non-signal distribution). The likelihood of the data under this hypothesis is easy to compute. The second hypothesis is that the data so far were generated from the non-signal distribution before some hypothetical switch point, and then from the signal distribution after that point. The likelihood of this hypothesis could be calculated by evaluating the likelihood separately for each possible switch point (from the beginning of the sequence to the current time) and choosing the most likely switch point. The ideal observer would then be in a position to respond that the switch had occurred if the likelihood of the switch model exceeded the likelihood of the one-state model by some criterion level. Under standard assumptions, twice the difference in log-likelihood values will be distributed as a χ 2 variable, so the criterion amount of difference required to detect a change would be determined by the ideal observer s desired Type I error rate. If the ideal observer has a very strict Type I error rate (e.g., p=.0001) they will only decide that a change has occurred when the two-state model has very much greater likelihood than the one-state model. An example is the late switcher in Figure 2. Conversely, if the ideal observer has a very lax Type I error rate (e.g., p=.2) they will decide that change has occurred when the twostate model fits only a little better than the one - 4 -

5 state model (e.g., the early switcher in Figure 2). Adjustments in the Type I error rate implement a trade off between the confidence that a change actually has occurred when a response is made, and the lateness of that response. The tradeoff between confidence and lateness can be resolved by appealing to the concept of hindsight. That is, at the end of the stimulus sequence, when all the data are available, the ideal observer can make a best hindsight estimate of the location of the switch point. The location of this estimate is more accurate than that made during the sequence see the vertical line marked best hindsight switch point in Figure 2. The ideal observer can then evaluate their own performance with the benefit of hindsight. That is, the actual response of the observer (i.e., the switch point indicated during the unfolding of the sequence) can be compared with the best hindsight estimate: we will refer to the difference as hindsight error. Then, the ideal level of evidence (i.e., Type I error rate) for detecting switch points could be calculated, by identifying the level that minimizes discrepancy between the actual response and the best hindsight value, over repeated sequences. In the example of Figure 2, this ideal level of evidence turns out to be p= For the particular example in Figure 2, the location of the switch point detected using this level of evidence is illustrated as the ideal switcher. We consider this ideal observer model for switch detection in greater detail below. We quantify the model, and compare the performance of ideal observers with performance of our participants in preliminary experiments. The correspondence between our preliminary experiments and the ideal observer analysis is not simple, since our experiments require more than just change detection from the participants. We therefore propose some additional experiments that provide more straightforward correspondence with the ideal observer model. 4. Overview of Experiments 4.1 Decision making situations So far, we have used just one decision making situation in our examples: classifying parts of pictures as friendly or enemy. Of course, this is not the only decision making example that is suitable for our work. We propose experiments using a variety of decision types, with the aim of generalizing our results across different levels of decision making processes, from low level sensory decisions to much higher level process. The experiments below include the following decision making situations: Lexical decision, i.e., deciding whether a character string is a valid English word or not. We manipulate decision difficulty by changing the wordiness of non-words (e.g., XXQZ is a less word-like non-word than STIP ). Numerosity judgment, i.e., deciding whether one symbol or another is more common in a given display. Our version of this classic task is designed to represent some aspects of realworld tasks, such as reading warning gauges (e.g., does the needle of a particularly noisy gauge point more to the left or the right?). We manipulate difficulty by changing how close to equal are the proportions of the two symbols. Recognition memory for real-world pictures and/or faces. Participants decisions are about whether pictures they are shown match ones they were previously instructed to remember. We manipulate difficulty by using distractor pictures that are more or less similar to those in memory. We plan to use both pictures of faces and pictures of outdoor scenery. Threat detection. Given only partial information about a possible assailant, participants must decide whether to respond with deadly or more moderate force. We manipulate decision difficulty by making the partial information more or less ambiguous. 4.2 Preliminary Experiments We have collected data from over 400 subjects in a preliminary series of three experiments. We aimed to study the way in which participants changed their decision criteria in response to environmental changes, on the trial-by-trial level. Before we began our experiments, it wasn't clear whether we would be able to observe these changes experimentally, as they could conceivably occur within one or two trials after the stimulus properties change. Therefore, our preliminary research aimed to collect sufficient data to quantify the nature of behavioral changes in response to stimulus changes by employing a decision environment as illustrated in Figure 3. The left and center panels of Figure 3 show how the stimulus environment changed: the expected value of the signal stimuli (top line in top left plot) were kept constant while the properties of the noise stimuli (lower line, left plot) were altered at random points

6 Experimental Sequence Predicted Criterion Predicted Hit/False Alarm Rates Figure 3. Conceptual design for Experiments #1-#3. The properties of the signal stimuli (triangles) are kept constant. The properties of the noise stimuli (circles) are varied. The left hand panel in Figure 3 shows the design of the experiment, with alternating noise stimulus properties and fixed signal properties. The middle panel shows the optimal decision criterion, and that the observer's decision criterion lags behind somewhat. This lag is a direct consequence of the difficulty in detecting the changes in stimulus properties. The right panel shows how false alarm and hit rates change under this model. The false alarm rates change both when the properties of the noise stimuli change, and also on a slower scale, as the observer adjusts their criterion. The hit rates change as well, and these changes can only be due to changes in decision criteria, as the signal properties are fixed throughout the experiment. Below, when we graph data and model fits from our experiments, we graph hit and false alarm rates as in the right panel of Figure 3. We average across stimulus context changes, so we show just one hard-to-easy context change and one easy-tohard change. These changes occurred either once only within each block (Experiment #1) or else between blocks (Experiments #2 and #3). Ideally, the decision task employed for these experiments would have stimulus changes that are not self-evident: if one were to present all red stimuli, then switch to all green stimuli, a participant's tracking of that change would be trivial. The decision tasks used in our experiments all fit the conceptual design in Figure 3. In Experiments #1 and #2 we used lexical decision where the subject decides whether a letter string forms an English word (signal) or a nonsense word (noise). By varying the 'wordiness' of the nonwords, we vary the signal strength of the noise distribution Rythmic Signal to Respond Procedure The LDT task produces bivariate response data: response latency (RT) and accuracy for each decision. When participants are forced to make decisions under great time pressure, without sufficient time to accurately weigh their options, variables such as speed-accuracy tradeoff (SAT) settings would be affected (McElree & Dosher, 1989). In order to accurately track dynamic changes in the amount of time available for decision making, participants would have to constantly adjust their SAT settings. Such adjustments could be modeled using decision boundaries in sequential sampling choice RT models (e.g., Heathcote & Brown, 2001; Mozer, Colagrosso, & Huber, 2002; Ratcliff, 1978; Usher & McClelland, 2001). However, No generally accepted method for dealing with the SAT exists because the definition of a suitable decision model is an unresolved question, and corrections for SAT are model dependent. To avoid such problems at a methodological rather than analytic level, we employed a new variant of the signal to respond procedure. This procedure restricted variability in response latency as much as possible, so that observed changes in accuracy were difficult to attribute to a SAT. The canonical signal to respond procedure presents participants with a stimulus and then, a certain amount of time later, presents them with a signal. Subjects are required to make their response within a certain window of time following the response signal

7 This procedure is often difficult for participants, resulting in much wasted data and occasionally ineffective control over the SAT. Our Experiments employed a variant of the signal to respond procedure, similar to the tempo naming procedure used by Kello and Plaut (2000, 2003; see also Steyvers, Wagenmakers, Shiffrin, Zeelenberg & Raaijmakers, 2001). This procedure sought to make responding within the acceptable window easier for participants by employing a rhythmic auditory cue. Our procedure involved keeping the rhythmic cue present and consistent throughout entire blocks. Specifically, a 256Hz tone sounded for 50msec every 400msec throughout Experiments #1-#3 (except for during between-block breaks). This rhythm anchored the temporal structure of the experiment for the participant. Stimuli were always presented on every (say) 10 th tone, and a response was always required between the two tones following the stimulus. Analysis of the RT data from this procedure confirmed that it helped subjects to maintain an accurate idea of when the responses are required, and to keep their response latencies tightly controlled Experiment #1 Lexical Decision with Random Switch Points Using our rhythmic procedure, the first experiment examined the effect of changing stimulus properties without informing participants. We used a standard lexical decision task (LDT) and kept the properties of the words constant across the whole experiment, while varying the wordiness of the non-words. We created non-words that were difficult to distinguish from words, by changing just one letter in seven-letter words, for example: SUBVIRT, LIBFARY and PETWIFY; these nonwords defined the difficulty decision environment. The easy decision environment was defined by non-words three letters different from words from seven-letter words, for example: FOMLERS, EPPAASI, and LERTINE. In Experiment #1 we changed the properties of the nonwords midway through blocks, at random points. This is not typical of psychological experiments, but we expected that in this design subjects would change their criteria relatively slowly, allowing us to better study those changes. Participants were not told that the decision environment would change during the experiment they were given only the usual instructions to respond as accurately as possible while keeping their response times within the acceptable response signal window. Our data from Experiment #1 (n=135) suggested that changing stimuli led to relatively long transients in hits and false alarm rates. The top left panel of Figure 4 graphs the data from Experiment #1. The vertical dashed line represents the point at which the decision context changed from either difficult to easy (open symbols) or easy to difficult (filled symbols). The units on the abscissa are trials within each block, with zero representing the point in the block where the properties of the nonwords changed. Note that this point was randomly distributed across blocks and the data are aligned to this point for graphing only. It is clear that the change in properties of the nonwords had large and rapid changes on the data from non-word stimuli, as expected. More interestingly, the data from word stimuli show changes even though these stimuli were unchanged throughout the experiment. These changes are shown in detail in the lower left panel of Figure 4, which graphs the mean difference in hit rates between the easy and hard decision environments. Naively, one would expect a higher hit rate in the easy environment, and hence a positive difference in this plot. Notice however, that after the environment changes the difference remains negative for around 10 trials. This blocking effect is well known from lexical decision tasks (e.g. Kinoshita & Lupker, 2002; Ratcliff Van Zandt & McKoon, 1999; Stone & Van Orden, 1993; Wagenmakers et al., in press), but the trial-by-trial dynamics have not previously been investigated. Interestingly, the change in the data from the word stimuli is gradual; on average it 14 trials for the cross over to occur. A gradual change strongly suggests that it is the long-term accumulation of task information that is driving this change. Below, we develop a simple measurement model to describe these data, based on a dynamic version of signal detection theory (SDT)

8 Exp. 1 Exp. 2 Exp. 3 Exp P( "target" ) HR( easy ) - HR( difficult ) Trials before and after switch Trials within blocks Trials within blocks Trials within blocks Figure 4: Data from Experiments #1-#4. In the top row: circles and triangles represent hit and false alarm rates respectively; filled and open symbols correspond to hard and easy decision environments respectively. The bottom row of plots show the difference between hit rates in easy and hard conditions, and how this changes during a block. The model fits shown in the top panels of Figure 4 are from dynamic variant of SDT outlined below. Note that the model captures the broad qualitative patterns in the data, such as the crossovers in both hit and false alarm rates, as well as steady changes in false alarm rates after the stimulus switch point Experiment #2 Lexical Decision with Blocked Stimulus Conditions Experiment #1 was interesting because it demonstrated that stimulus history has a strong effect on responses. Using random switch points within blocks reduces the realism of our design, as most experiments contain environment changes only between blocks. Researchers who wish to pool their data across blocks may well claim that the block break between blocks operates to minimize the effect of stimulus history. That is, the block break may cause subjects to reset their ideas of the structure of the stimuli somewhat. Experiment #2 tested this theory by replicating Experiment #1, except with the structure of the non-word stimuli changing precisely between blocks; that is, blocks always contained homogeneous stimuli, as is the usual practice. Additionally, since we did not need to have sufficiently long blocks to be able to randomly place stimulus switch points, we shortened the blocks from 100 trials to 40, and increased their number from 10 to 20. Once again, participants were not told that the experiment consisted of two different kinds of decision environment. The data from Experiment #2 (n=106) suggest that slow transient properties similar to those observed in Experiment #1 appeared even when changes in stimulus properties are synchronized with block breaks. Note that the hit rates (top middle panel of Figure 4) for easy and hard conditions look very similar at the start of the block, suggesting a criterion resetting during the block break. However, the hit rate for the difficult condition is statistically significantly higher than that for the easy condition, which suggests that the criterion at the start of the block is at least partially copied from the end of the previous block. This conclusion is supported by the graph of hit rate differences for Experiment #2, in the lower middle panel of Figure 4. Those data show that it takes an average of 10 trials (one quarter of each block) for the hit rates in easy and hard decision contexts to cross over to their asymptotic values, and that there are clear mirror effects whereby the hit rates change smoothly from easy to hard levels even though those conditions are blocked

9 4.2.4 Experiment #3 Lexical Decision With Prior Warning The results of Experiments #1 and #2 suggested that people take many trials to adjust their behavior after changes in the decision environment. Experiment #3 investigates whether the speed of this change can be influenced by conscious control. Experiment #3 used exactly the same design as Experiment #2, but with different instructions. Participants were explicitly told that there would be difficulty and easier blocks of trials, and that these would alternate. They were also shown examples of easy and difficulty nonwords. To ensure that they were aware of the current decision environment, throughout difficult blocks the words HARD BLOCK were displayed inside a large red patch on screen. During easy blocks, the words EASY BLOCK were displayed in a large green patch. During each block break, participants were reminded what type of block was coming next. The data from Experiment #3 are remarkable for their similarity to those from Experiment #2. Even with extensive knowledge of the experimental design, and forewarning, participants did not adjust to new decision environments noticeably quicker than in Experiment #2. There was still a clear crossover in hit rates during the early trials of each block (see bottom panel of Experiment #3 plots). This indicates that participants behavior during the early part of each block was strongly influenced by the previous block. The mean estimate of the time taken to adjust to the new decision environment (the lag parameter from our simple dynamic SDT model, below) was 14 trials. This indicates that participants took, on average, more than one third of each block to overcome the effects of task history from the previous block Experiment #4 Numerosity Experiments #1 and #2 both employed a lexical decision paradigm, leaving open the possibility that our results would not generalize to other decision tasks. Experiment #4 extends these results to a lower-level task: numerosity judgment. Classic numerosity judgment tasks (e.g., Palmeri, 1997) require decision makers to view patterns of symbols, typically dots, and decide how many symbols are in each. We used a categorization version of numerosity judgment, where participants always saw displays with exactly 10 symbols. The symbols were always of two types, and the participant's task was to decide which of the two symbols was in the majority. Our particular version of this task was constructed to emulate the process of reading a noisy gauge, or vibrating indicator needle. The two types of <<<><>><<> Figure 5: Example numerosity stimulus. Correct answer is left, as more symbols point left than right in this display. symbols we used were left and right pointing arrows, presented in a row, as in Figure 5. We manipulated decision difficulty by changing the distribution of proportions used: a display with 4/6 left/right elements is much more difficult than one with 2/8. We always had right-favoring display with only 4/6 or 3/7 proportions, and varied the left-facing displays from easy (9/1 or 8/2) to hard (7/3 or 6/4). To avoid biases, we reversed the left/right assignment for half of the participants. As in Experiment #2, switches between easy and hard decision environments only occurred between blocks. The data from Experiment #4 are graphed in the right-hand panels of Figure 4. Even though the numerosity task employed in Experiment #4 was very different to the lexical decision task used for Experiment #2, the data are very similar. A hit rate difference between easy and hard decision environments builds up slowly throughout each block, and blocks begin with a negative difference (bottom right panel of Figure 4) indicating that participants carried over their decision making processes between blocks; a sub-optimal strategy. 4.3 Simple Dynamic Measurement Model We have developed a simple dynamic version of signal detection theory (SDT) as a way to approximate our ideal observer model, with as few assumptions about the nature of the data as possible. Given a particular set of parameters, the ideal observer model discussed above leads to a distribution of switch detection points over repeated presentations of the same sequence. Our simple dynamic SDT model approximates this distribution with a single switch point. That is, we estimate a single lag for each participant, estimating how long they take (on average) to notice that an environment change has occurred. The simplicity of this model is especially - 9 -

10 Figure 6: Predictions from our dynamic SDT model. Left panel shows static SDT submodels. Right panel shows predicted criterion changes (top) and predicted hit rate and false alarm rate changes (bottom). important for its purpose as a measurement tool, as simplicity helps to decrease uncertainty in parameter estimates. This is crucial as we wish to interpret parameter estimates as meaningful measures of cognitive and behavioral phenomena at the level of individual decision makers. In its simplest form, our dynamic SDT model is designed to apply to decision-making tasks where there are two different decision-making environments that alternate throughout the task. The model is based on two static SDT models, one for each decision-making environment. For ease of explication, we assume that one decisionmaking environment is more difficult than the other. In that case, our model assumes that there is an SDT model operating in the difficult environment, defined by a sensitivity parameter (d' H, H for hard ) and a decision criterion (C H ) and another SDT model operating in the easy decision environment, defined by d' E and C E ( E for easy ). The crucial addition that allows us to model dynamic behavior is that we assume a lag when changing from one decision environment to the other. For example, when the decision environment changes from easy to difficult, we assume that the sensitivity of decisions changes immediately, from d' E to d' H. Such immediacy makes sense given that only the stimuli themselves define decision difficulty. By contrast, the decision criterion is under the control of the decision-maker, and thus will not change until they notice the change in decision environment, or some correlated variable (e.g., changed error rates). In our example, when changing from an easy to a hard decision environment, we assume that the decision criterion only changes from C E to C H after some lag, L. This assumption of a stepwise change in criterion may seem unreasonable. We examined other assumptions, such as a smooth exponential approach from the old to the new criterion, or a piecewise-linear approach, and found that they provided no significant improvement in fit. We chose the stepwise criterion change for its simplicity, and the interpretability of its parameter L (which simply measures the number of decisions after an environment change before the participant changes their criterion) Model Predictions Figure 6 makes the predictions of this model clearer. The SDT model illustrated in the top-left corner, t 1, illustrates behavior during easy decisions: note that d E is relatively large (the signal and noise distributions are relatively far apart) and that the criterion C E approximates the optimal criterion. This model leads to the hit rate (HR) and false alarm rate (FAR) predictions at the left of the right-hand panel, with high HR and low FAR. At time t 2, as shown in the lower right-hand panel, the decision environment changes from easy to hard. We assume that this change was implemented by increasing the similarity of the distractor stimuli to the signal stimuli, while keeping signal stimulus properties constant (so noise distributions change, but signal distributions don t). The SDT model then operating is shown in the lower left corner, with label t 2. Note that sensitivity has decreased due to the harder stimuli (d E has changed to d H ), but the criterion has not yet changed. This leads to the HR and FAR predictions shown under t 2 in the lower right-hand panel in dashed lines: no immediate change in HR, but a large increase in FAR. After some lag, L, the

11 decision-maker updates their criterion to C H, the appropriate criterion for hard decision environments (shown by the dashed line in the upper left plot). The SDT model then operating is shown as t 3 on the left hand side of Figure 6, and its predictions are shown by the dashed lines under the label t 3 in the lower right-hand plot: a decrease in both HR and FAR. Finally, the decision environment again changes, back to the easy condition, changing sensitivity but not immediately changing the decision criterion. This corresponds to SDT model t 4 and a predicted decrease in FAR, with no change in HR. Again, after some lag, the decision criterion is changed to C E, bringing us back to the SDT model t 1. As described, the dashed lines in Figure 6 show predictions for an individual subject s HR and FAR: note that our assumption of a stepwise criterion change implies only stepwise changes in predictions. In the text, we show fits to large groups of participants, where each participant is fit individually, and the observed and expected HR and FAR are averaged over subjects for the purposes of graphing only. Those graphs show smooth changes in FAR and HR, as illustrated by solid lines in the right-hand panels of Figure 6. Such smooth transitions are merely the result of averaging over many individual stepwise transitions, where the position of the step is variable Estimating Model Parameters We estimate the parameters of our model using maximum likelihood. Unlike static SDT models, where closed form analytic solutions are available, maximum likelihood parameter estimates must be obtained by search for the dynamic model. Conditional on a given value of the lag parameter, L, the model specifies one of four SDT models ( t 1, t 2, t 3, or t 4 above) as operating on every decision. These models naturally specify the probability of making each kind of response (hits, misses, false alarms and correct rejections), and so a likelihood for the parameter values under consideration can be obtained by multiplying probabilities across all trials in an experimental sequence. A practical difficulty is introduced as the lag parameter can take on only integer values, while the other parameters are continuous. We solved this problem by searching for maximum likelihood fits separately for every possible value of the lag parameter, and choosing the global maximum out of these (the lag parameter is constrained to be greater than zero, and less than the length of the longest decision environment) Unequal Variance Dynamic SDT A bonus from our use of a dynamic model is that we can also estimate an unequal variance version of our SDT model. This model has seven rather than five parameters, with two new parameters describing the ratio of variances in the signal and noise distributions, for hard and easy decision environments. A static SDT model for hard and easy decision environments (i.e., L=0, immediate criterion change) is unable to estimate unequal variance models, as there are effectively no data from models t 2 and t 4. We do not discuss the unequal variance model in the text, for two reasons: to maintain our focus on model simplicity, and because we have found that the unequal variance model provides no significant improvement in fit over the equal variance model. 4.4 Applied Ideal Observer Model The ideal observer model developed in Section 1.4 can be applied to data from our preliminary experiments. Naturally, there are degrees of idealization for such a model. We have chosen to idealize only the change detection aspect of these data, leaving levels of imperfection in the decision making aspect equal to those displayed by our participants. The alternative idealizing both aspects leads to a less interesting class of models with instant change detection and perfect decision making. Our choice represents an attempt to answer the question Given the observed imperfection in decision making, how quickly could an ideal observer reliably detect switch points?. Other ways of implementing an ideal observer model for our paradigm are certainly possible, and we will continue developing and testing different ideal observer models as our research progresses. We implemented the ideal observer model individually for each participant, as follows. Firstly, for each participant, measures of decision sensitivity and bias in both easy and hard decision environments (d E, d H, C E and C H ) were obtained from our dynamic SDT measurement model. These d values were then used to simulate data sequences. With these simulated sequences, and a given a particular level of evidence required for change detection (Type I error rate), the change points detected by an ideal observer model can be calculated, along with the hindsight change points,

12 premature switch detection. An example of a late switching distribution is also shown (p=.001). The ideal switching distribution is the one that most closely approximates the hindsight switch points, in a least-squares sense for this example, the ideal level of evidence was p=.010. The lower panel of Figure 7 shows the RMS difference between the observed switch point distributions and the hindsight values, as a function of the required level of evidence. The ideal level of evidence leads to an RMS difference between observed and hindsight switch points of only 15.3 trials. Figure 7. Example of applied ideal observer. as discussed above. For each participant, we identified the ideal required level of evidence as the one that minimized mean squared difference between ideal observer switch points and the hindsight switch points. Note that the ideal level of evidence changes across individuals, as decision behavior (sensitivity and bias) change. An example of this process is shown in Figure 7. For any data sequence (simulated using an individual participant s d and C values) we calculated the best hindsight estimate of the switch point, and the ideal observer s online switch points corresponding to 150 different evidence levels, logarithmically spaced between p=.63 and p= We repeated this exercise with 10,000 different random data sequences. This resulted in a distribution of observed switch points for each of the 150 confidence levels, and for the best hindsight switch points. Example distributions are shown in the top panel of Figure 7. The hindsight switch values are symmetrically distributed close to the true switch point (50). An example of an early switching distribution is shown, corresponding to a low level of evidence (p=.1) note that this level of evidence leads to some We compared our participants behavior to that of the ideal observer model by comparing the estimated switch detection lag for each observer (from the dynamic SDT model) with the mean of the switch detection lags from the corresponding ideal observer. Histograms across participants of these differences for Experiments #1 to #4 are shown in Figure 8. Larger values on the x-axes indicate greater sub-optimality from our participants (i.e., greater difference between observed switch detection lags, and mean ideal observer detection lags). Our participants were slower to detect the change in decision environments than the ideal observer models by an average of 17.5 trials in Experiment #1. In that experiment, our participants took an average of 21.8 trials to detect the change, whereas the corresponding ideal observer models took only 4.4 trials on average. Experiments #2 and #3 both employed lexical decision with stimulus switch points only occurring between blocks. In both these experiments, the mean difference between participants estimated switch lags and those of the ideal observer models was close to zero (less than one trial difference in each experiment). The correlation between ideal observer lags and the estimated lags for our participants was also high: r=.555 in Experiment #2 and r=.502 in Experiment #3. It seems that, when stimulus changes occur during block breaks, our participants were able to detect those changes as quickly as an ideal observer. A possible explanation that we are currently investigating is that participants were sub-optimal during the early parts of the experiments, but then caught on to the design of the stimulus sequences and were able to adjust more quickly in the late parts of the experiments. It is also interesting that the histograms for Experiments #2 and #3 are very similar, indicating

13 Figure 8. Optimality of change detection. that informing participants about the nature of the task (in Experiment #3) had little or no effect. Experiment #4 used a numerosity categorization task. In this task, participants estimated switch detection lags were actually smaller on average than the corresponding ideal observer lags (7.7 trials vs 12.5 trials, respectively). This result has several possible interpretations, which require further modeling and analysis to tease apart. One explanation is that participants in Experiment #4 were simply using a more liberal level of evidence for change detection than was ideal. This would lead to the very fast change detections shown in Figure 8, at the expense of some premature change detections and corresponding performance penalties. Another explanation is that participants in this task were using more information in their change detection processing than we allowed the ideal observer model. Our ideal observer model operated on constraints provided by the sensitivity and bias of the participants in the decision task. This task used a signal to respond procedure, and it is possible that participants continued to process stimulus information after giving their responses. This would provide them with greater information on which to base their change detection processing than just the information used to make decision task responses. 5 Proposed New Directions We propose several new research directions, based on the dynamic decision making paradigm and the dynamic SDT and ideal observer models developed above. These new directions include many new experiments, several of which are detailed below. Some of these experiments already have well developed methodological and stimulus-related details. Other experiments are less mature, and precise details of stimulus properties etc. are not finalized. These experiments are in areas where we feel that input from AFOSR researchers may be valuable in helping us to tailor the research towards Air Force applications. For example, one of the threat assessment experiments below is described as using quite unrealistic cartoon-like stimuli. We have chosen those stimuli because they allow tight experimental control, and parametric manipulations of factors such as similarity. However, in collaboration with AFOSR researchers, we can use precisely the same experimental design but with more realistic stimuli (e.g., real pictures, or satellite photos, rather than cartoons). We also plan to collaborate with AFOSR researchers on experimental aspects other than stimulus classes. For example, below we describe three experiments that assess the effects of fatigue, distraction and instruction on decision making. Input from AFOSR researchers may result in changing the emphasis of these experiments to other related variables (e.g., the effects of conflicting or inappropriate orders). We propose the development of research aspects other than experiments. We will continue to develop and refine the dynamic SDT model proposed above. One direction of development is to produce a hybrid model that incorporates elements of the dynamic SDT model and the ideal observer model. This model would then provide an actual generative mechanism for the detection of change in decision environments, rather than a solely descriptive account as at present. We propose the construction of a deliverable package from such a model. We envisage a software package that administers an experimental test to an individual (around 45 minutes) and then analyzes their data using our model. These analyses would return easily interpreted quantities, such as the lag parameter estimate, that describes how quickly that individual adjusts to dynamic environments, and the level of evidence required by a participant in order to detect change. Some of the experiments we propose would improve the managerial value of this system by providing normative data, and possibly correlations with existing psychometric measures

14 The feasibility of such a system is currently limited by the numerical difficulties inherent in estimating the parameters of the dynamic SDT model. It is possible for analyses of an individual data set (around 1000 observations) to take several hours on a reasonably fast workstation. We plan to fix this problem by developing estimation algorithms suitable for use on large-scale multiprocessor computers, such as exist at AFOSR. We have already begun this task, using the OpenMP (v2) standard for multi-threading compiler directives. Our goal is to develop software that analyzes data in real time: quickly enough that managers and team leaders can make use of the information. 5.2 Varying Instructions & payoffs Changing decision difficulty is not the only way that dynamic decision environments can be instantiated. For example, one experiment we propose below involves dynamically changing the payoff matrix for decisions. To make this concrete, imagine that incorrectly identifying a friend as an enemy (a false alarm) carries much greater penalties than incorrectly identifying an enemy as a friend (a miss). If these payoffs are dynamically altered, ideal decision makers will adjust their decision criteria appropriately, in order to minimize overall penalties. Changes in decision instructions can be naturally handled within our ideal observer model. Recall that the ideal level of evidence required to detect a change in that model (the Type I error rate) was calculated by minimizing the discrepancy between observed change points and those calculated with the benefit of hindsight. We defined discrepancy using a least squares error function. This function is symmetric the same penalty is given to early, premature, detection as to late detection. If the payoff matrix in a particular decision making environment were modified to be asymmetric (e.g., by penalizing premature change detection more heavily than late detection), the discrepancy function used to calculate the ideal Type I error rate could be changed accordingly. 5.3 Experiment #5: Simple Change Detection This experiment is designed for simple compatibility with our ideal observer model described above, allowing us to assess exactly how sub-optimal participants behavior was. To simplify the task as far as possible, we remove altogether the need for observers to make decisions on each trial. Instead, we provide observers with a sequence of data, and ask them to identify the switch point at which the properties of the data sequence change. To improve compatibility with our other experiments, and with our ideal observer model, participants are told that there are exactly two different states for a sequence, and that each sequence will begin in one state, and then switch to the other at exactly one point during the sequence. They will be given elements of the sequence one at a time, and asked to identify the change point as accurately as they can. Many different types of sequence can be employed. Participants could be shown the daily temperatures recorded in a (fictional) city, and told that, on one day during the month, a cold change lasting the rest of the month will arrive. With sufficient variability in daily temperature, this will be difficult to detect. Alternatively, participants could be shown a sequence of outcomes from rolling a pair of suspect dice, and asked to identify when the dice become less than fair. Another possibility is to show participants very noisy simulated radar pictures, with a very weak stimulus appearing at one point during the sequence. Participants are to identify that point as accurately as possible. An important aspect of these experiments is that participants are required to identify the point of stimulus change on line, while the stimuli are still being presented. They are also allowed to make only one response once they have indicated that the change has occurred they are unable to change their response. This situation provides compatibility with the ideal observer model, and with the dynamic SDT model. This constraint also provides a rationale for our definition of our ideal observer model as one that minimizes hindsight error. If an observer makes a too-hasty decision, they will discover this fact and have time to regret it while observing the passing of the remainder of the sequence. 5.4 Experiment #6a: Memory for Landscape Pictures Experiments #5a and #5b aim to generalize the results of our Experiments to more realistic stimuli (pictures of faces and outdoor scenery) and to higher level decisions (recognition memory). Recognition memory research has identified many mirror effects, similar in structure to that evident in the lexical decision data from our experiments. We will use a standard study-test recognition

15 Figure 9. Illustration of recognition memory stimuli (Experiment #6a). Easier stimuli (C and D) do not overlap, harder stimuli (A and B) do, and so are more confusable. memory design, using either pictures of human faces or fragments of landscape or pictures as stimuli. As in Experiments #1-#3, we will alter decision environments by alternating easy and difficult decision contexts. In Experiment #6a we will use landscape pictures and will construct easy and difficult distractor pictures by taking fragments of larger pictures and manipulating the amount of overlap between them. An example is given in Figure 9. The top half of that figure shows a landscape photograph with two fragments (A and B) cut out. These fragments share a great deal of overlap and so are very similar. The bottom row of Figure 9 shows another landscape picture with two fragments (C and D) cut out, but with less overlap and so lower similarity. If fragments A and C were in a decision maker's study list, and their task is to decide whether newly presented fragments are new or old (i.e., in the to-beremembered study list) we could create a very difficult decision making environment by presenting them with fragment B, which is very similar to the memorized fragment A. Conversely, we could ask them to decide whether fragment D is new or old, which would be much easier as it s less similar to items in memory. large database of black-and-white face pictures, from the FERET database. We have begun to pair these faces by subjective similarity rating. That is, we defined several major categories of face (by gender, age and ethnicity). Then, within each category, we identified pairs of very similar faces. For example, the top half of Figure 10 shows a pair of very similar faces from within one category (younger Asian women). These two faces are very similar to each other and so are highly confusable, like the A-B picture fragments of Experiment #6a. The bottom row of Figure 10 shows a random pairing from the same category: these two pictures are less similar, and so less confusable, like the C- D landscape fragments of Experiment #6a. 5.5 Experiment #7: Threat Assessment Experiment #7 continues the progression established in Experiments #5a and #5b towards more realistic decision making environments, including the assessment of threat level depicted in semi-realistic graphic displays. Those images could be of real people, or satellite reconnaissance photos, for instance. The decision making task used in this and all further experiments is to decide whether a stimulus presented in a noisy image represents a threat or not, and then respond accordingly. This task has several advantages over the tasks used above. Firstly, immediacy: the (imaginary) threat will be directed at the decision maker, which provides a 5.4 Experiment #6b: Memory for Human Faces We could perform the same recognition memory experiment with human face pictures as stimuli, but the similarity manipulation would be implemented differently. We have access to a Figure 10: Example faces for Experiment #6b. Top two faces are very similar, bottom two are less similar

The Dynamics of Experimentally Induced Criterion Shifts

The Dynamics of Experimentally Induced Criterion Shifts Journal of Experimental Psychology: Learning, Memory, and Cognition 2005, Vol. 31, No. 4, 587 599 Copyright 2005 by the American Psychological Association 0278-7393/05/$12.00 DOI: 10.1037/0278-7393.31.4.587

More information

Modeling Experimentally Induced Strategy Shifts Scott Brown, 1 Mark Steyvers, 2 and Pernille Hemmer 2

Modeling Experimentally Induced Strategy Shifts Scott Brown, 1 Mark Steyvers, 2 and Pernille Hemmer 2 PSYCHOLOGICAL SCIENCE Research Report Modeling Experimentally Induced Strategy Shifts Scott Brown, 1 Mark Steyvers, 2 and Pernille Hemmer 2 1 School of Psychology, University of Newcastle, Callaghan, New

More information

SUPPLEMENTAL MATERIAL

SUPPLEMENTAL MATERIAL 1 SUPPLEMENTAL MATERIAL Response time and signal detection time distributions SM Fig. 1. Correct response time (thick solid green curve) and error response time densities (dashed red curve), averaged across

More information

Lecturer: Rob van der Willigen 11/9/08

Lecturer: Rob van der Willigen 11/9/08 Auditory Perception - Detection versus Discrimination - Localization versus Discrimination - Electrophysiological Measurements - Psychophysical Measurements 1 Three Approaches to Researching Audition physiology

More information

A Race Model of Perceptual Forced Choice Reaction Time

A Race Model of Perceptual Forced Choice Reaction Time A Race Model of Perceptual Forced Choice Reaction Time David E. Huber (dhuber@psyc.umd.edu) Department of Psychology, 1147 Biology/Psychology Building College Park, MD 2742 USA Denis Cousineau (Denis.Cousineau@UMontreal.CA)

More information

A Race Model of Perceptual Forced Choice Reaction Time

A Race Model of Perceptual Forced Choice Reaction Time A Race Model of Perceptual Forced Choice Reaction Time David E. Huber (dhuber@psych.colorado.edu) Department of Psychology, 1147 Biology/Psychology Building College Park, MD 2742 USA Denis Cousineau (Denis.Cousineau@UMontreal.CA)

More information

A Memory Model for Decision Processes in Pigeons

A Memory Model for Decision Processes in Pigeons From M. L. Commons, R.J. Herrnstein, & A.R. Wagner (Eds.). 1983. Quantitative Analyses of Behavior: Discrimination Processes. Cambridge, MA: Ballinger (Vol. IV, Chapter 1, pages 3-19). A Memory Model for

More information

(Visual) Attention. October 3, PSY Visual Attention 1

(Visual) Attention. October 3, PSY Visual Attention 1 (Visual) Attention Perception and awareness of a visual object seems to involve attending to the object. Do we have to attend to an object to perceive it? Some tasks seem to proceed with little or no attention

More information

Lecturer: Rob van der Willigen 11/9/08

Lecturer: Rob van der Willigen 11/9/08 Auditory Perception - Detection versus Discrimination - Localization versus Discrimination - - Electrophysiological Measurements Psychophysical Measurements Three Approaches to Researching Audition physiology

More information

A model of parallel time estimation

A model of parallel time estimation A model of parallel time estimation Hedderik van Rijn 1 and Niels Taatgen 1,2 1 Department of Artificial Intelligence, University of Groningen Grote Kruisstraat 2/1, 9712 TS Groningen 2 Department of Psychology,

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

Section 6: Analysing Relationships Between Variables

Section 6: Analysing Relationships Between Variables 6. 1 Analysing Relationships Between Variables Section 6: Analysing Relationships Between Variables Choosing a Technique The Crosstabs Procedure The Chi Square Test The Means Procedure The Correlations

More information

Chapter 11. Experimental Design: One-Way Independent Samples Design

Chapter 11. Experimental Design: One-Way Independent Samples Design 11-1 Chapter 11. Experimental Design: One-Way Independent Samples Design Advantages and Limitations Comparing Two Groups Comparing t Test to ANOVA Independent Samples t Test Independent Samples ANOVA Comparing

More information

MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and. Lord Equating Methods 1,2

MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and. Lord Equating Methods 1,2 MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and Lord Equating Methods 1,2 Lisa A. Keller, Ronald K. Hambleton, Pauline Parker, Jenna Copella University of Massachusetts

More information

Detection Theory: Sensitivity and Response Bias

Detection Theory: Sensitivity and Response Bias Detection Theory: Sensitivity and Response Bias Lewis O. Harvey, Jr. Department of Psychology University of Colorado Boulder, Colorado The Brain (Observable) Stimulus System (Observable) Response System

More information

Supplemental Information. Varying Target Prevalence Reveals. Two Dissociable Decision Criteria. in Visual Search

Supplemental Information. Varying Target Prevalence Reveals. Two Dissociable Decision Criteria. in Visual Search Current Biology, Volume 20 Supplemental Information Varying Target Prevalence Reveals Two Dissociable Decision Criteria in Visual Search Jeremy M. Wolfe and Michael J. Van Wert Figure S1. Figure S1. Modeling

More information

Types of questions. You need to know. Short question. Short question. Measurement Scale: Ordinal Scale

Types of questions. You need to know. Short question. Short question. Measurement Scale: Ordinal Scale You need to know Materials in the slides Materials in the 5 coglab presented in class Textbooks chapters Information/explanation given in class you can have all these documents with you + your notes during

More information

Imperfect, Unlimited-Capacity, Parallel Search Yields Large Set-Size Effects. John Palmer and Jennifer McLean. University of Washington.

Imperfect, Unlimited-Capacity, Parallel Search Yields Large Set-Size Effects. John Palmer and Jennifer McLean. University of Washington. Imperfect, Unlimited-Capacity, Parallel Search Yields Large Set-Size Effects John Palmer and Jennifer McLean University of Washington Abstract Many analyses of visual search assume error-free component

More information

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison Empowered by Psychometrics The Fundamentals of Psychometrics Jim Wollack University of Wisconsin Madison Psycho-what? Psychometrics is the field of study concerned with the measurement of mental and psychological

More information

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review Results & Statistics: Description and Correlation The description and presentation of results involves a number of topics. These include scales of measurement, descriptive statistics used to summarize

More information

UC Merced Proceedings of the Annual Meeting of the Cognitive Science Society

UC Merced Proceedings of the Annual Meeting of the Cognitive Science Society UC Merced Proceedings of the Annual Meeting of the Cognitive Science Society Title Criterion Setting and the Dynamics of Recognition Memory Permalink https://escholarship.org/uc/item/66q087vn Journal Proceedings

More information

The Clock Ticking Changes Our Performance

The Clock Ticking Changes Our Performance The Clock Ticking Changes Our Performance Shoko Yamane, Naohiro Matsumura Faculty of Economics, Kinki University; Graduate School of Economics, Osaka University syamane@kindai.ac.jp Abstract We examined

More information

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis? How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis? Richards J. Heuer, Jr. Version 1.2, October 16, 2005 This document is from a collection of works by Richards J. Heuer, Jr.

More information

Running head: INDIVIDUAL DIFFERENCES 1. Why to treat subjects as fixed effects. James S. Adelman. University of Warwick.

Running head: INDIVIDUAL DIFFERENCES 1. Why to treat subjects as fixed effects. James S. Adelman. University of Warwick. Running head: INDIVIDUAL DIFFERENCES 1 Why to treat subjects as fixed effects James S. Adelman University of Warwick Zachary Estes Bocconi University Corresponding Author: James S. Adelman Department of

More information

Classical Psychophysical Methods (cont.)

Classical Psychophysical Methods (cont.) Classical Psychophysical Methods (cont.) 1 Outline Method of Adjustment Method of Limits Method of Constant Stimuli Probit Analysis 2 Method of Constant Stimuli A set of equally spaced levels of the stimulus

More information

Study on perceptually-based fitting line-segments

Study on perceptually-based fitting line-segments Regeo. Geometric Reconstruction Group www.regeo.uji.es Technical Reports. Ref. 08/2014 Study on perceptually-based fitting line-segments Raquel Plumed, Pedro Company, Peter A.C. Varley Department of Mechanical

More information

Name Psychophysical Methods Laboratory

Name Psychophysical Methods Laboratory Name Psychophysical Methods Laboratory 1. Classical Methods of Psychophysics These exercises make use of a HyperCard stack developed by Hiroshi Ono. Open the PS 325 folder and then the Precision and Accuracy

More information

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Lec 02: Estimation & Hypothesis Testing in Animal Ecology Lec 02: Estimation & Hypothesis Testing in Animal Ecology Parameter Estimation from Samples Samples We typically observe systems incompletely, i.e., we sample according to a designed protocol. We then

More information

Virtual Reality Testing of Multi-Modal Integration in Schizophrenic Patients

Virtual Reality Testing of Multi-Modal Integration in Schizophrenic Patients Virtual Reality Testing of Multi-Modal Integration in Schizophrenic Patients Anna SORKIN¹, Avi PELED 2, Daphna WEINSHALL¹ 1 Interdisciplinary Center for Neural Computation, Hebrew University of Jerusalem,

More information

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE ...... EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE TABLE OF CONTENTS 73TKey Vocabulary37T... 1 73TIntroduction37T... 73TUsing the Optimal Design Software37T... 73TEstimating Sample

More information

Appendix B Statistical Methods

Appendix B Statistical Methods Appendix B Statistical Methods Figure B. Graphing data. (a) The raw data are tallied into a frequency distribution. (b) The same data are portrayed in a bar graph called a histogram. (c) A frequency polygon

More information

Evolutionary Programming

Evolutionary Programming Evolutionary Programming Searching Problem Spaces William Power April 24, 2016 1 Evolutionary Programming Can we solve problems by mi:micing the evolutionary process? Evolutionary programming is a methodology

More information

Supplementary Materials

Supplementary Materials Supplementary Materials 1. Material and Methods: Our stimuli set comprised 24 exemplars for each of the five visual categories presented in the study: faces, houses, tools, strings and false-fonts. Examples

More information

The Regression-Discontinuity Design

The Regression-Discontinuity Design Page 1 of 10 Home» Design» Quasi-Experimental Design» The Regression-Discontinuity Design The regression-discontinuity design. What a terrible name! In everyday language both parts of the term have connotations

More information

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1 INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1 5.1 Clinical Interviews: Background Information The clinical interview is a technique pioneered by Jean Piaget, in 1975,

More information

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis EFSA/EBTC Colloquium, 25 October 2017 Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis Julian Higgins University of Bristol 1 Introduction to concepts Standard

More information

Sawtooth Software. The Number of Levels Effect in Conjoint: Where Does It Come From and Can It Be Eliminated? RESEARCH PAPER SERIES

Sawtooth Software. The Number of Levels Effect in Conjoint: Where Does It Come From and Can It Be Eliminated? RESEARCH PAPER SERIES Sawtooth Software RESEARCH PAPER SERIES The Number of Levels Effect in Conjoint: Where Does It Come From and Can It Be Eliminated? Dick Wittink, Yale University Joel Huber, Duke University Peter Zandan,

More information

Categorical Perception

Categorical Perception Categorical Perception Discrimination for some speech contrasts is poor within phonetic categories and good between categories. Unusual, not found for most perceptual contrasts. Influenced by task, expectations,

More information

The Clock Ticking Changes Our Performance

The Clock Ticking Changes Our Performance Shikakeology: Designing Triggers for Behavior Change: Papers from the 2013 AAAI Spring Symposium The Clock Ticking Changes Our Performance Shoko Yamane, Naohiro Matsumura Faculty of Economics, Kinki University;

More information

20. Experiments. November 7,

20. Experiments. November 7, 20. Experiments November 7, 2015 1 Experiments are motivated by our desire to know causation combined with the fact that we typically only have correlations. The cause of a correlation may be the two variables

More information

Supplementary Materials: Materials and Methods Figures S1-S2 Tables S1-S17 References

Supplementary Materials: Materials and Methods Figures S1-S2 Tables S1-S17 References Supplementary Materials: Materials and Methods Figures S1-S2 Tables S1-S17 References Materials and Methods Simon Task Participants were randomly assigned to one of four versions of the task. Upon return

More information

Cultural Differences in Cognitive Processing Style: Evidence from Eye Movements During Scene Processing

Cultural Differences in Cognitive Processing Style: Evidence from Eye Movements During Scene Processing Cultural Differences in Cognitive Processing Style: Evidence from Eye Movements During Scene Processing Zihui Lu (zihui.lu@utoronto.ca) Meredyth Daneman (daneman@psych.utoronto.ca) Eyal M. Reingold (reingold@psych.utoronto.ca)

More information

Detection Theory: Sensory and Decision Processes

Detection Theory: Sensory and Decision Processes Detection Theory: Sensory and Decision Processes Lewis O. Harvey, Jr. Department of Psychology and Neuroscience University of Colorado Boulder The Brain (observed) Stimuli (observed) Responses (observed)

More information

EXPERIMENTAL RESEARCH DESIGNS

EXPERIMENTAL RESEARCH DESIGNS ARTHUR PSYC 204 (EXPERIMENTAL PSYCHOLOGY) 14A LECTURE NOTES [02/28/14] EXPERIMENTAL RESEARCH DESIGNS PAGE 1 Topic #5 EXPERIMENTAL RESEARCH DESIGNS As a strict technical definition, an experiment is a study

More information

Absolute Identification is Surprisingly Faster with More Closely Spaced Stimuli

Absolute Identification is Surprisingly Faster with More Closely Spaced Stimuli Absolute Identification is Surprisingly Faster with More Closely Spaced Stimuli James S. Adelman (J.S.Adelman@warwick.ac.uk) Neil Stewart (Neil.Stewart@warwick.ac.uk) Department of Psychology, University

More information

Learning Deterministic Causal Networks from Observational Data

Learning Deterministic Causal Networks from Observational Data Carnegie Mellon University Research Showcase @ CMU Department of Psychology Dietrich College of Humanities and Social Sciences 8-22 Learning Deterministic Causal Networks from Observational Data Ben Deverett

More information

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4 UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4 Algebra II Unit 4 Overview: Inferences and Conclusions from Data In this unit, students see how the visual displays and summary statistics they

More information

Project exam in Cognitive Psychology PSY1002. Autumn Course responsible: Kjellrun Englund

Project exam in Cognitive Psychology PSY1002. Autumn Course responsible: Kjellrun Englund Project exam in Cognitive Psychology PSY1002 Autumn 2007 674107 Course responsible: Kjellrun Englund Stroop Effect Dual processing causing selective attention. 674107 November 26, 2007 Abstract This document

More information

Chapter 5: Field experimental designs in agriculture

Chapter 5: Field experimental designs in agriculture Chapter 5: Field experimental designs in agriculture Jose Crossa Biometrics and Statistics Unit Crop Research Informatics Lab (CRIL) CIMMYT. Int. Apdo. Postal 6-641, 06600 Mexico, DF, Mexico Introduction

More information

The Hare and the Tortoise: Emphasizing speed can change the evidence used to make decisions.

The Hare and the Tortoise: Emphasizing speed can change the evidence used to make decisions. Emphasising Decision Speed 1 The Hare and the Tortoise: Emphasizing speed can change the evidence used to make decisions. Babette Rae 1, Andrew Heathcote 1, Chris Donkin 2, Lee Averell 1 and Scott Brown

More information

Supporting Information

Supporting Information 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 Supporting Information Variances and biases of absolute distributions were larger in the 2-line

More information

Interpreting Instructional Cues in Task Switching Procedures: The Role of Mediator Retrieval

Interpreting Instructional Cues in Task Switching Procedures: The Role of Mediator Retrieval Journal of Experimental Psychology: Learning, Memory, and Cognition 2006, Vol. 32, No. 3, 347 363 Copyright 2006 by the American Psychological Association 0278-7393/06/$12.00 DOI: 10.1037/0278-7393.32.3.347

More information

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES 24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter

More information

Technical Specifications

Technical Specifications Technical Specifications In order to provide summary information across a set of exercises, all tests must employ some form of scoring models. The most familiar of these scoring models is the one typically

More information

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc. Chapter 23 Inference About Means Copyright 2010 Pearson Education, Inc. Getting Started Now that we know how to create confidence intervals and test hypotheses about proportions, it d be nice to be able

More information

Supplementary Materials

Supplementary Materials Supplementary Materials Supplementary Figure S1: Data of all 106 subjects in Experiment 1, with each rectangle corresponding to one subject. Data from each of the two identical sub-sessions are shown separately.

More information

Automatic detection, consistent mapping, and training * Originally appeared in

Automatic detection, consistent mapping, and training * Originally appeared in Automatic detection - 1 Automatic detection, consistent mapping, and training * Originally appeared in Bulletin of the Psychonomic Society, 1986, 24 (6), 431-434 SIU L. CHOW The University of Wollongong,

More information

RECALL OF PAIRED-ASSOCIATES AS A FUNCTION OF OVERT AND COVERT REHEARSAL PROCEDURES TECHNICAL REPORT NO. 114 PSYCHOLOGY SERIES

RECALL OF PAIRED-ASSOCIATES AS A FUNCTION OF OVERT AND COVERT REHEARSAL PROCEDURES TECHNICAL REPORT NO. 114 PSYCHOLOGY SERIES RECALL OF PAIRED-ASSOCIATES AS A FUNCTION OF OVERT AND COVERT REHEARSAL PROCEDURES by John W. Brelsford, Jr. and Richard C. Atkinson TECHNICAL REPORT NO. 114 July 21, 1967 PSYCHOLOGY SERIES!, Reproduction

More information

Writing Reaction Papers Using the QuALMRI Framework

Writing Reaction Papers Using the QuALMRI Framework Writing Reaction Papers Using the QuALMRI Framework Modified from Organizing Scientific Thinking Using the QuALMRI Framework Written by Kevin Ochsner and modified by others. Based on a scheme devised by

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination Timothy N. Rubin (trubin@uci.edu) Michael D. Lee (mdlee@uci.edu) Charles F. Chubb (cchubb@uci.edu) Department of Cognitive

More information

Model evaluation using grouped or individual data

Model evaluation using grouped or individual data Psychonomic Bulletin & Review 8, (4), 692-712 doi:.378/pbr..692 Model evaluation using grouped or individual data ANDREW L. COHEN University of Massachusetts, Amherst, Massachusetts AND ADAM N. SANBORN

More information

Changing expectations about speed alters perceived motion direction

Changing expectations about speed alters perceived motion direction Current Biology, in press Supplemental Information: Changing expectations about speed alters perceived motion direction Grigorios Sotiropoulos, Aaron R. Seitz, and Peggy Seriès Supplemental Data Detailed

More information

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science 2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science Stanislas Dehaene Chair in Experimental Cognitive Psychology Lecture No. 4 Constraints combination and selection of a

More information

Why do Psychologists Perform Research?

Why do Psychologists Perform Research? PSY 102 1 PSY 102 Understanding and Thinking Critically About Psychological Research Thinking critically about research means knowing the right questions to ask to assess the validity or accuracy of a

More information

Differentiation and Response Bias in Episodic Memory: Evidence From Reaction Time Distributions

Differentiation and Response Bias in Episodic Memory: Evidence From Reaction Time Distributions Journal of Experimental Psychology: Learning, Memory, and Cognition 2010, Vol. 36, No. 2, 484 499 2010 American Psychological Association 0278-7393/10/$12.00 DOI: 10.1037/a0018435 Differentiation and Response

More information

On optimal decision-making in brains and social insect colonies Marshall, Bogacz, Dornhaus, Planque, Kovacs,Franks

On optimal decision-making in brains and social insect colonies Marshall, Bogacz, Dornhaus, Planque, Kovacs,Franks On optimal decision-making in brains and social insect colonies Marshall, Bogacz, Dornhaus, Planque, Kovacs,Franks Presented by Reffat Sharmeen, Kolli Sai Namratha Contents Introduction Optimal decision-making

More information

Evaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY

Evaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY 2. Evaluation Model 2 Evaluation Models To understand the strengths and weaknesses of evaluation, one must keep in mind its fundamental purpose: to inform those who make decisions. The inferences drawn

More information

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo Please note the page numbers listed for the Lind book may vary by a page or two depending on which version of the textbook you have. Readings: Lind 1 11 (with emphasis on chapters 10, 11) Please note chapter

More information

Pooling Subjective Confidence Intervals

Pooling Subjective Confidence Intervals Spring, 1999 1 Administrative Things Pooling Subjective Confidence Intervals Assignment 7 due Friday You should consider only two indices, the S&P and the Nikkei. Sorry for causing the confusion. Reading

More information

Lessons in biostatistics

Lessons in biostatistics Lessons in biostatistics The test of independence Mary L. McHugh Department of Nursing, School of Health and Human Services, National University, Aero Court, San Diego, California, USA Corresponding author:

More information

Congruency Effects with Dynamic Auditory Stimuli: Design Implications

Congruency Effects with Dynamic Auditory Stimuli: Design Implications Congruency Effects with Dynamic Auditory Stimuli: Design Implications Bruce N. Walker and Addie Ehrenstein Psychology Department Rice University 6100 Main Street Houston, TX 77005-1892 USA +1 (713) 527-8101

More information

OCW Epidemiology and Biostatistics, 2010 David Tybor, MS, MPH and Kenneth Chui, PhD Tufts University School of Medicine October 27, 2010

OCW Epidemiology and Biostatistics, 2010 David Tybor, MS, MPH and Kenneth Chui, PhD Tufts University School of Medicine October 27, 2010 OCW Epidemiology and Biostatistics, 2010 David Tybor, MS, MPH and Kenneth Chui, PhD Tufts University School of Medicine October 27, 2010 SAMPLING AND CONFIDENCE INTERVALS Learning objectives for this session:

More information

Attention and Concentration Problems Following Traumatic Brain Injury. Patient Information Booklet. Talis Consulting Limited

Attention and Concentration Problems Following Traumatic Brain Injury. Patient Information Booklet. Talis Consulting Limited Attention and Concentration Problems Following Traumatic Brain Injury Patient Information Booklet Talis Consulting Limited What are Attention and Concentration? Attention and concentration are two skills

More information

Sensory Cue Integration

Sensory Cue Integration Sensory Cue Integration Summary by Byoung-Hee Kim Computer Science and Engineering (CSE) http://bi.snu.ac.kr/ Presentation Guideline Quiz on the gist of the chapter (5 min) Presenters: prepare one main

More information

Assessing Modes of Interaction

Assessing Modes of Interaction Project 2 Assessing Modes of Interaction Analysis of exercise equipment Overview For this assignment, we conducted a preliminary analysis of two similar types of elliptical trainers. We are able to find

More information

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

3 CONCEPTUAL FOUNDATIONS OF STATISTICS 3 CONCEPTUAL FOUNDATIONS OF STATISTICS In this chapter, we examine the conceptual foundations of statistics. The goal is to give you an appreciation and conceptual understanding of some basic statistical

More information

In this chapter we discuss validity issues for quantitative research and for qualitative research.

In this chapter we discuss validity issues for quantitative research and for qualitative research. Chapter 8 Validity of Research Results (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.) In this chapter we discuss validity issues for

More information

Awareness in contextual cuing with extended and concurrent explicit tests

Awareness in contextual cuing with extended and concurrent explicit tests Memory & Cognition 28, 36 (2), 43-415 doi: 1.3758/MC.36.2.43 Awareness in contextual cuing with extended and concurrent explicit tests ANDREA C. SMYTH AND DAVID R. SHANKS University College London, London,

More information

Psychology Research Process

Psychology Research Process Psychology Research Process Logical Processes Induction Observation/Association/Using Correlation Trying to assess, through observation of a large group/sample, what is associated with what? Examples:

More information

The Pretest! Pretest! Pretest! Assignment (Example 2)

The Pretest! Pretest! Pretest! Assignment (Example 2) The Pretest! Pretest! Pretest! Assignment (Example 2) May 19, 2003 1 Statement of Purpose and Description of Pretest Procedure When one designs a Math 10 exam one hopes to measure whether a student s ability

More information

Decision Rules for Recognition Memory Confidence Judgments

Decision Rules for Recognition Memory Confidence Judgments Journal of Experimental Psychology: Learning, Memory, and Cognition 1998, Vol. 24, No. 6, 1397-1410 Copyright 1998 by the American Psychological Association, Inc. 0278-7393/98/S3.00 Decision Rules for

More information

Editorial Note: this manuscript has been previously reviewed at another journal that is not operating a transparent peer review scheme.

Editorial Note: this manuscript has been previously reviewed at another journal that is not operating a transparent peer review scheme. Editorial Note: this manuscript has been previously reviewed at another journal that is not operating a transparent peer review scheme. This document only contains reviewer comments and rebuttal letters

More information

Statistics and Probability

Statistics and Probability Statistics and a single count or measurement variable. S.ID.1: Represent data with plots on the real number line (dot plots, histograms, and box plots). S.ID.2: Use statistics appropriate to the shape

More information

SLAUGHTER PIG MARKETING MANAGEMENT: UTILIZATION OF HIGHLY BIASED HERD SPECIFIC DATA. Henrik Kure

SLAUGHTER PIG MARKETING MANAGEMENT: UTILIZATION OF HIGHLY BIASED HERD SPECIFIC DATA. Henrik Kure SLAUGHTER PIG MARKETING MANAGEMENT: UTILIZATION OF HIGHLY BIASED HERD SPECIFIC DATA Henrik Kure Dina, The Royal Veterinary and Agricuural University Bülowsvej 48 DK 1870 Frederiksberg C. kure@dina.kvl.dk

More information

Are Retrievals from Long-Term Memory Interruptible?

Are Retrievals from Long-Term Memory Interruptible? Are Retrievals from Long-Term Memory Interruptible? Michael D. Byrne byrne@acm.org Department of Psychology Rice University Houston, TX 77251 Abstract Many simple performance parameters about human memory

More information

7 Grip aperture and target shape

7 Grip aperture and target shape 7 Grip aperture and target shape Based on: Verheij R, Brenner E, Smeets JBJ. The influence of target object shape on maximum grip aperture in human grasping movements. Exp Brain Res, In revision 103 Introduction

More information

Choose an approach for your research problem

Choose an approach for your research problem Choose an approach for your research problem This course is about doing empirical research with experiments, so your general approach to research has already been chosen by your professor. It s important

More information

Empirical Formula for Creating Error Bars for the Method of Paired Comparison

Empirical Formula for Creating Error Bars for the Method of Paired Comparison Empirical Formula for Creating Error Bars for the Method of Paired Comparison Ethan D. Montag Rochester Institute of Technology Munsell Color Science Laboratory Chester F. Carlson Center for Imaging Science

More information

Chapter Three: Sampling Methods

Chapter Three: Sampling Methods Chapter Three: Sampling Methods The idea of this chapter is to make sure that you address sampling issues - even though you may be conducting an action research project and your sample is "defined" by

More information

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14 CS/NEUR125 Brains, Minds, and Machines Assignment 5: Neural mechanisms of object-based attention Due: Friday, April 14 This Assignment is a guided reading of the 2014 paper, Neural Mechanisms of Object-Based

More information

Dynamic Integration of Reward and Stimulus Information in Perceptual Decision-Making

Dynamic Integration of Reward and Stimulus Information in Perceptual Decision-Making Integration of reward and stimulus information Dynamic Integration of Reward and Stimulus Information in Perceptual Decision-Making Juan Gao,2, Rebecca Tortell James L. McClelland,2,, Department of Psychology,

More information

A BAYESIAN SOLUTION FOR THE LAW OF CATEGORICAL JUDGMENT WITH CATEGORY BOUNDARY VARIABILITY AND EXAMINATION OF ROBUSTNESS TO MODEL VIOLATIONS

A BAYESIAN SOLUTION FOR THE LAW OF CATEGORICAL JUDGMENT WITH CATEGORY BOUNDARY VARIABILITY AND EXAMINATION OF ROBUSTNESS TO MODEL VIOLATIONS A BAYESIAN SOLUTION FOR THE LAW OF CATEGORICAL JUDGMENT WITH CATEGORY BOUNDARY VARIABILITY AND EXAMINATION OF ROBUSTNESS TO MODEL VIOLATIONS A Thesis Presented to The Academic Faculty by David R. King

More information

Model Evaluation using Grouped or Individual Data. Andrew L. Cohen. University of Massachusetts, Amherst. Adam N. Sanborn and Richard M.

Model Evaluation using Grouped or Individual Data. Andrew L. Cohen. University of Massachusetts, Amherst. Adam N. Sanborn and Richard M. Model Evaluation: R306 1 Running head: Model Evaluation Model Evaluation using Grouped or Individual Data Andrew L. Cohen University of Massachusetts, Amherst Adam N. Sanborn and Richard M. Shiffrin Indiana

More information

Convergence Principles: Information in the Answer

Convergence Principles: Information in the Answer Convergence Principles: Information in the Answer Sets of Some Multiple-Choice Intelligence Tests A. P. White and J. E. Zammarelli University of Durham It is hypothesized that some common multiplechoice

More information

Encoding of Elements and Relations of Object Arrangements by Young Children

Encoding of Elements and Relations of Object Arrangements by Young Children Encoding of Elements and Relations of Object Arrangements by Young Children Leslee J. Martin (martin.1103@osu.edu) Department of Psychology & Center for Cognitive Science Ohio State University 216 Lazenby

More information

Appendix: Instructions for Treatment Index B (Human Opponents, With Recommendations)

Appendix: Instructions for Treatment Index B (Human Opponents, With Recommendations) Appendix: Instructions for Treatment Index B (Human Opponents, With Recommendations) This is an experiment in the economics of strategic decision making. Various agencies have provided funds for this research.

More information

Making Inferences from Experiments

Making Inferences from Experiments 11.6 Making Inferences from Experiments Essential Question How can you test a hypothesis about an experiment? Resampling Data Yield (kilograms) Control Group Treatment Group 1. 1.1 1.2 1. 1.5 1.4.9 1.2

More information

2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences

2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences 2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences Stanislas Dehaene Chair of Experimental Cognitive Psychology Lecture n 5 Bayesian Decision-Making Lecture material translated

More information

An Ideal Observer Model of Visual Short-Term Memory Predicts Human Capacity Precision Tradeoffs

An Ideal Observer Model of Visual Short-Term Memory Predicts Human Capacity Precision Tradeoffs An Ideal Observer Model of Visual Short-Term Memory Predicts Human Capacity Precision Tradeoffs Chris R. Sims Robert A. Jacobs David C. Knill (csims@cvs.rochester.edu) (robbie@bcs.rochester.edu) (knill@cvs.rochester.edu)

More information