Improved Randomization Tests for a Class of Single-Case Intervention Designs

Size: px
Start display at page:

Download "Improved Randomization Tests for a Class of Single-Case Intervention Designs"

Transcription

1 Journal of Modern Applied Statistical Methods Volume 13 Issue 2 Article Improved Randomization Tests for a Class of Single-Case Intervention Designs Joel R. Levin University of Arizona, jrlevin@u.arizona.edu John M. Ferron University of South Florida, ferron@usf.edu Boris S. Gafurov George Mason University, bgafurov@gmu.edu Follow this and additional works at: Part of the Applied Statistics Commons, Social and Behavioral Sciences Commons, and the Statistical Theory Commons Recommended Citation Levin, Joel R.; Ferron, John M.; and Gafurov, Boris S. (2014) "Improved Randomization Tests for a Class of Single-Case Intervention Designs," Journal of Modern Applied Statistical Methods: Vol. 13 : Iss. 2, Article 2. DOI: /jmasm/ Available at: This Invited Article is brought to you for free and open access by the Open Access Journals at DigitalCommons@WayneState. It has been accepted for inclusion in Journal of Modern Applied Statistical Methods by an authorized editor of DigitalCommons@WayneState.

2 Improved Randomization Tests for a Class of Single-Case Intervention Designs Erratum In the original published version of this article, Panels B and C of Figure 15, p. 38, were reversed, and references to Figure 11 on p. 32 should have referred to Figure 12. This has been corrected. This invited article is available in Journal of Modern Applied Statistical Methods: iss2/2

3 Journal of Modern Applied Statistical Methods November 2014, Vol. 13, No. 2, Copyright 2014 JMASM, Inc. ISSN Invited Article: Improved Randomization Tests for a Class of Single-Case Intervention Designs Joel R. Levin University of Arizona Tucson, AZ John M. Ferron University of South Florida Tampa, FL Boris S. Gafurov George Mason University Fairfax, VA Forty years ago, Eugene Edgington developed a single-case AB intervention design-andanalysis procedure based on a random determination of the point at which the B phase would start. In the present simulation studies encompassing a variety of AB-type contexts, it is demonstrated that by also randomizing the order in which the A and B phases are administered, a researcher can markedly increase the procedure s statistical power. Keywords: Single-case intervention research, design and statistical analysis, randomization tests, statistical power, internal validity, scientific credibility Introduction Single-case designs that focus on behavioral and academic interventions are prevalent in a variety of clinical and educational fields (see, for example, Kratochwill & Levin, 2014). In contrast to conventional group intervention designs, single-case designs typically include only one or a few units (e.g., individuals, small groups, classrooms) to whom the intervention is administered. In addition, single-case intervention designs are intensive and implemented over longer periods of time, with more numerous assessments of the outcome measures (Horner & Odom, 2014; Kratochwill et al., 2010). Single-case intervention designs that currently incorporate formal criteria to enhance their scientific Joel R. Levin is a Professor Emeritus in the Department of Educational Psychology. at jrlevin@u.arizona.edu. John M. Ferron is a Professor in the Department of Educational and Psychological Studies in the College of Education. at ferron@usf.edu. Boris S. Gafurov is an Assistant Professor in the College of Education and Human Development. at bgafurov@gmu.edu. Erratum: In the original published version of this article, Panels B and C of Figure 15 were reversed. This error is now corrected. 2

4 LEVIN ET AL. credibility (Levin, 1994) include ABAB designs, alternating treatment designs, and multiple-baseline designs (Kratochwill et al., 2013). As the methodological rigor of single-case intervention designs has evolved over the years (Kratochwill & Levin, 2010), so too have the formal statisticalanalysis procedures that accompany them (see, for example, Kratochwill & Levin, 2014; and Manolov, Evans, Gast, & Perdices, 2014). Although various visual/graphical approaches remain an analytic staple of single-case data (e.g., Auerbach & Zeitlin, 2014; Kratochwill, Levin, Horner, & Swoboda, 2014; Parker, Vannest, & Davis, 2014), improved statistical methods have increasingly been considered as viable supplements to visual analysis. These improved statistical methods include econometric time-series analyses (e.g., McCleary & Welsh, 1992), adapted regression- and hierarchical linear modeling procedures (e.g., Maggin et al., 2011; Manolov & Solanas, 2013; Moeyaert, Ferron, Beretvas, Van den Noortgate, & Beretvas, 2014; Shadish, Kyse, & Rindskopf, 2013), and nonparametric permutation and randomization tests (e.g., Edgington & Onghena, 2007; Ferron & Levin, 2014; Heyvaert & Onghena, 2014). The last of these statistical approaches is the focus of the present study. Overview of the Present Study The motivation for single-case researchers to adopt a randomization test as one component of their analytic armament is that randomization tests provide strict control of the Type I error rate (i.e., the probability of concluding that phase-tophase differences in level, trend, variability, etc. are present when those differences are simply chance fluctuations) as long as: (1) the design includes randomization; (2) the accompanying statistical test is conducted in a manner that is consistent with the design frame; and (3) the test statistic is chosen without knowledge of the results (Edgington, 1980; Ferron & Levin, 2014). In contrast, demonstration of Type I error control has been elusive in studies of visual analysis (e.g., Ferron & Jones, 2006; Fisch, 2001; Stocks & Williams, 1995). Moreover, with regression and hierarchical models, Type I error control hinges on a relatively strong set of assumptions (Ferron, Moeyaert, Van den Noortgate, & Beretvas, 2014). The modeling assumptions include: (1) the error distribution is correctly specified (e.g., normally distributed, homogeneous variances across phases, and a first-order autoregressive function); (2) the baseline trajectory is correctly specified; (3) the baseline trajectory can be extrapolated (i.e., had the intervention not been implemented, the baseline trajectory would have continued, implying that there were no confounding effects of external events on the time 3

5 IMPROVED RANDOMIZATION TESTS series); and (4) the treatment phase trajectory is correctly specified. Accordingly, a single-case researcher may plan a multicomponent analysis in which visual analysis serves as the primary analysis tool, a randomization test is employed to ensure that the Type I error rate is controlled, and a regression-based or hierarchical linear model is examined to summarize and estimate the size of the effect(s). A concern with the addition of randomization tests to the analytic plan is that such tests require the researcher to introduce randomization into the design, and if the randomization is not carefully planned it can lead to a design that falls short of single-case design standards (e.g., Ferron & Levin, 2014; Kazdin, 1980; Kratochwill et al., 2010). As a consequence, researchers are encouraged to reflect carefully on the practical constraints of the context in which the study is conducted, on the desired design features (e.g., minimum phase lengths), and then tailor the randomization strategy to meet these constraints. Restricted randomization schemes have been developed to ensure that: (1) the desired number of phases and minimum phase lengths are included in reversal designs (Onghena, 1992); (2) the treatment alternates quickly enough in an alternating treatment design (Onghena & Edgington, 1994); (3) the baseline series stabilizes prior to commencement of the intervention phase (Ferron & Ware, 1994); (4) the intervention start points are staggered by a minimum amount of time in multiplebaseline designs (Koehler & Levin, 1998), and (5) researchers are able to obtain visually acceptable patterns by extending phases in multiple-baseline designs (Ferron & Jones, 2006) and reversal designs (Ferron & Levin, 2014). The present Monte Carlo simulation study employs nonparametric randomization tests in the company of a recently proposed methodological addition that greatly enhances the internal validity of AB and ABAB single-case intervention designs (Ferron & Levin, 2014; Levin, Evmenova, & Gafurov, 2014). In these designs, A typically represents a baseline, control, or standard treatment phase containing repeated outcome measurements and B represents an intervention, experimental, or new treatment phase also containing repeated outcome measurements. Here we examine the methodological addition s effect on the statistical conclusion validity (manifested by both Type I error control and increased statistical power) of randomization tests in single-case AB and ABAB designs, in both their single-case (N = 1) and multiple-case (N > 1) forms. In the following section, we first describe the methodological addition that enhances the internal validity (scientific credibility) of single-case intervention research and then outline how the addition is incorporated into a randomization test to improve the test s statistical conclusion validity. Our decision to start our investigations 4

6 LEVIN ET AL. with a single-participant (N = 1) AB design was not because we are advocating for the use of such a design, but because it provides the simplest point to begin study of the impact of the methodological addition. Once we have established the effects on statistical conclusion validity in the simplest situation, we will progressively add complexities to strengthen the design, building to the multipleparticipant (N > 1) ABAB design. Edgington s (1975) Random Intervention Start-Point Model Of four different types of randomization that can be incorporated into randomization in single-case AB experimental studies (specifically, within-case phase randomization, between-case intervention randomization, case randomization, and intervention start-point randomization (see Ferron & Levin, 2014), the last, highly creative, type was originally developed by Edgington (1975) and requires that the researcher: (1) randomly select an intervention start point from two or more that had been previously deemed acceptable; and then (2) assign to the case the start point that was actually selected. Although not applied in the conventional treatment randomization manner, this unique form of randomization increases a single-case study s internal validity and, when accompanied by the statistical test described in the following paragraph, it can increase the study s statistical conclusion validity as well. Moreover, this randomized intervention start-point approach can function to provide a true (i.e., scientifically credible) experimental comparison of two or more intervention (or intervention and control) conditions based on either one case or multiple cases per condition (for examples and discussion, see Ferron & Levin, 2014; Koehler & Levin, 1998; Levin, Lall, & Kratochwill, 2011; Levin & Wampold, 1999; and Marascuilo & Busk, 1988). With the randomized intervention start-point model, a randomization statistical test is conducted on the difference between the means of all B and all A series outcomes for each of the intervention start-point divisions (or transitions) that could have resulted from the random-selection process (see also Edgington & Onghena, 2007). [Moreover, any other summary measure of relevance to the researcher s hypothesis about the nature of change from Phase A to Phase B (e.g., change in the series medians, slopes, variances) can also be the focus of a randomization-test analysis.] With the resulting set of mean differences yielding a randomization distribution, the mean difference associated with the actual intervention start point is examined to see where it falls within the set. The probability of obtaining a 5

7 IMPROVED RANDOMIZATION TESTS mean difference as extreme as or more extreme than the actual mean difference represents the unlikelihood of the outcome. Either signed or unsigned mean differences are considered for one- and two-tailed hypothesis tests, respectively. For example, for an AB design with one case, 25 outcome-assessment periods, and 20 potential intervention start points, if the actual start point were found to produce the largest mean difference (in the predicted direction) between the B and A series outcomes, then the one-tailed significance probability associated with that event would be given by p = 1/20 =.05. For a two-tailed test, as or more extreme opposite-sign mean differences would also need to be taken into account. For instance, if there were a mean difference equal in magnitude but opposite in sign to the one just indicated for the actual intervention start point, then the twotailed significance probability would be 2/20 =.10. In Edgington s (1975) random intervention start-point model for a one-case AB design, it is assumed that the A phase consists of a baseline series, the B phase consists of an intervention series, and that the former logically precedes the latter. With those assumptions, the number of possible outcomes (B A mean differences) in the randomization distribution is k, the number of potential intervention start points. Accordingly, with one case, 30 total observations, and k = 10 potential intervention start points, if the actual B A mean difference produced were the largest of the 10 and in the predicted direction, then the onetailed significance probability of that outcome would be p = 1/10 =.10. In order to achieve statistical significance at a traditional α =.05 level (one-tailed), one would need to include at least k = 20 potential intervention start points in the randomization distribution (i.e., so that if the most extreme mean difference in the predicted direction were obtained, then p would equal 1/20 =.05). To achieve statistical significance with α =.05 via a two-tailed test, a longer series with a minimum of k = 40 potential intervention start points would be required (i.e., so that p = 2/40 =.05 is possible). Randomized Order (Dual Randomization) Addition to the Edgington Model Edgington (1975) proposed his random intervention start-point design-andanalysis procedure 40 years ago. It has been incorporated into a variety of singlecase intervention designs (e.g., Koehler & Levin, 1998; Levin & Wampold, 1999; Marascuilo & Busk, 1988; Onghena, 1992) and is being implemented in its original form to this day. However, it will be shown here that an addition to the procedure (referred to here as a modified procedure), which enhances its internal 6

8 LEVIN ET AL. validity by eliminating bias due AB phase-order effects, is possible and one that is applicable in a number of single-case intervention investigations. To illustrate, suppose that instead of A representing a baseline or control phase, it represents one type of experimental intervention say, a behavioral intervention for combatting a particular phobia. In contrast, B might represent a cognitive intervention targeting the same phobia. Within that context, the case receives both interventions. To have a legitimate (unconfounded) comparison of Intervention A and Intervention B, it is imperative that the order in which the two interventions are administered to the case is randomly (rather than arbitrarily) determined. The preceding statement applies whether the investigation includes only one case or multiple cases (although in multiple-case situations, systematic counterbalancing of intervention orders across cases might be implemented to achieve the same goal). In addition, it is worth noting that A and B need not refer only to two competing interventions. Rather, suppose that A represents a baseline, standard, or control condition and B an intervention condition. As has been suggested previously (e.g., Kratochwill & Levin, 2010), further suppose that prior to the commencement of the actual experiment, a mandatory baseline (or adaptation/warm-up) phase (A') is required of all cases. With A' included, it would then be possible, appropriate, and presumably acceptable to researchers to begin the experiment proper by randomizing each case s subsequent A and B phases (i.e., an A randomly selected to be first means that the case remains in the baseline condition, followed by the B intervention condition; and a B randomly selected to be first means that the case begins with the intervention condition, followed by the A baseline condition). Accordingly, the modified orderrandomization procedure is applicable in either one- or two-intervention AB designs, with the prospect of improving both design (internal validity) and analysis (statistical-conclusion validity) of two-phase single-case intervention studies. With intervention-order randomization built into the just-discussed one-case example based on 30 total observations and 10 potential intervention start points, in addition to the intervention start points associated with the conventional AB order of intervention administration, one would also need to consider the possibility that Intervention B had been randomly selected to be administered first. If that had happened, there would be a corresponding 10 potential intervention start points for the BA order of intervention administration, resulting in a total of k = 20 potential start-point outcomes that would be included in the complete randomization distribution. 7

9 IMPROVED RANDOMIZATION TESTS Multiple-Case Extension of the Modified Edgington Model As we will show, the order-randomization procedure applies to multiple-case (replicated) AB situations as well, increasing the total number of possible randomization-distribution outcomes by a factor of 2 N, where N represents the number of cases. Specifically, with N cases and one of k i potential intervention start points randomly selected for each case, with Marascuilo and Busk s (1988) multiple-case extension of Edgington s (1975) single fixed-order intervention N start-point model, a total of k randomization-distribution outcomes are i 1 i possible, and in the special case for which all k i are equal to k, this quantity reduces to k N. With the addition of an order-randomization process to create the present dual randomization model, the total number of possible randomizationdistribution outcomes increases to k 2 and k N 2 N = (2k) N for the general N N i 1 i and special-case situations, respectively. Hypothetical example We illustrate the present random-order randomization-test procedure for a replicated single-case AB design by means of a hypothetical example. Suppose that a language researcher wishes to improve the baseline vocalization output (A phase) of two low word-producing children through some type of positive-reinforcement intervention (B phase). For the random-order version of the present example we assume that a mandatory A' baseline (warm-up) phase was initially administered, followed by a random determination of whether the first phase of the actual study would be a baseline (A) or an intervention (B) phase, thereby producing either an A'AB or A'BA design. Although in comparison to a traditional fixed-order AB design, this type of randomized AB design is more scientifically credible (especially when replicated across cases), the latter design was not considered in the current What Works Clearinghouse (WWC) single-case intervention design Standards (Kratochwill et al., 2010). Our hypothetical study is presented simply to illustrate both the original (Edgington, 1975) fixed-order and the present random-order randomization-test procedures, without taking into account the study s internalvalidity characteristics. Consideration of internal-validity issues is included later in the Discussion section. In this hypothetical study, the number of single-word vocalizations by each child during a 5-minute play period is recorded, with Child 1 observed in each of 25 daily sessions and Child 2 observed in each of 15 daily sessions, and where both children must be observed in at least 3 A sessions and 3 B sessions (thereby resulting in 20 and 10 potential intervention transition points for Child 1 and 8

10 LEVIN ET AL. Child 2, respectively). In addition, because the researcher wishes to randomize the intervention order (AB or BA) for each child, three preliminary five-minute A' warm-up sessions are provided prior to the start of the children s actual experimental sessions. An initial coin toss determines that Child 1 will be administered an AB intervention order, with the 20 potential intervention transition points specified from between the 4 th and 23 rd sessions inclusive and the randomly selected actual intervention transition point occurring just prior to Session 10. For Child 2, a BA intervention order results from a second coin flip, with the 10 potential intervention transitions specified from between the 4th and 13th sessions inclusive and an actual randomly selected intervention transition point just prior to the 7 th observation. The A- and B-phase observations are presented in Table 1. Given the present random-order AB intervention start-point randomization model, the data were analyzed with Gafurov and Levin s (2014) single-case ExPRT (Excel Package of Randomization Tests) package see Levin et al. (2014) for complete information about ExPRT. In Table 2 are presented the B A mean differences associated with each of the potential intervention transition points for the two children. The first Table 2 entry of 2.41 for Child 1, which corresponds to an A-to-B intervention transition point just prior to Observation 4, was calculated by taking the average of Child 1 s Observations 4 through 25 (mean B phase = 6.41) minus the average of that child s Observations 1 through 3 (mean A phase = 4.00). The same process was followed for each of the subsequent 19 potential intervention points for Child 1, which ends with the average of that child s Observations 23 through 25 (mean B phase = 8.00) minus the average of that child s Observations 1 through 22 (mean A phase = 5.86), resulting in Child 1 s final mean difference of 2.14 in Table 2. Next, and as indicated in Table 2 s Footnote a, 20 additional mean differences were calculated for Child 1 under the assumption that instead of an A B intervention order, the reverse B A order had been selected. Under that assumption, the first mean difference for Child 1 would be = 2.41, which is exactly the same numerically but opposite in sign to the previously calculated child s first value in Table 2. The same is true for all of Child 1 s calculated reverse-order values, including the 20th one, which is now The same process applied to Child 2 s data yields the 10 actual B A mean differences presented in Table 2 (i.e., = 1.08 for the first one), as well as 10 reverse-order and opposite-sign A B mean differences. 9

11 IMPROVED RANDOMIZATION TESTS Table 1. Hypothetical data for Child 1 s 25-observation series, with a randomly selected AB intervention order, 20 potential intervention transition points (between Observations 4 and 23 Inclusive), and a randomly selected actual intervention transition point just prior to Observation 10; and for Child 2 s 15-observation series, with a randomly selected BA intervention order, 10 potential intervention transition points (between Observations 4 and 13 Inclusive), and a randomly selected actual intervention transition point just prior to Observation 7 Child 1 Child 2 Observation Phase Vocalizations Observation Phase Vocalizations 1 A 4 1 B 6 2 A 3 2 B 5 3 A 5 3 B 7 4 A 5 4 B 5 5 A 2 5 B 6 6 A 5 6 B 5 7 A 3 7* A 4 8 A 4 8 A 5 9 A 4 9 A 3 10* B 5 10 A 5 11 B 6 11 A 4 12 B 7 12 A 5 13 B 6 13 A 6 14 B 7 14 A 5 15 B 8 15 A 6 16 B 7 17 B 9 18 B 8 19 B 6 20 B 8 21 B 9 22 B 8 23 B 7 24 B 9 25 B 8 *Actual intervention transition point. 10

12 LEVIN ET AL. Table 2. The B A mean difference associated with: (1) each of Child 1 s 20 potential intervention transition points (O4-O23) for a randomly selected AB intervention order; and (2) each of Child 2 s 10 potential intervention transition points (O4-O13) for a randomly selected BA intervention order Child 1 Child 2 Potential Intervention Point B-A Mean Difference a B-A Mean Difference b O O O O * O O O * O O O O O O O O O O O O O *Mean difference associated with the actual intervention transition point. a The 20 A B mean differences are also calculated and added to these to form a 40-outcome randomization distribution; all of the A B mean differences are the same as the corresponding B-A mean differences given here but opposite in sign. b The 10 A B mean differences are also calculated and added to these to form a 20-outcome randomization distribution; all of the mean A B differences are the same as the corresponding mean B A differences given here but opposite in sign. The resulting joint randomization distribution therefore contains 40 mean differences for Child 1 combined with 20 mean differences for Child 2, for a total of = 800 averaged mean differences (i.e., Child 1 s 1 st mean difference averaged with Child 2 s 1 st mean difference, Child 1 s 1 st mean difference averaged with Child 2 s 2 nd mean difference, all the way up to and including Child 1 s 40 th mean difference averaged with Child 2 s 20 th mean difference). When that is done by the ExPRT program, it is found that the actual joint mean 11

13 IMPROVED RANDOMIZATION TESTS difference that was obtained in the study is 2.19, which is Child 1 s mean difference associated with that child s actual intervention transition point of O 10 (3.49) averaged with Child 2 s actual intervention transition-point mean difference of O 7 (.89). Of the 800 outcomes in the joint randomization distribution, a value of 2.19 is the 10 th highest, which results in a one-tailed significance probability of p = 10/800 = For this example, had a one-tailed Type I error probability (α) of.05 been selected, it could be concluded that the positivereinforcement intervention (B) distribution values differed statistically from those in the baseline distribution (A), with the additional inference that the former distribution s values were higher. We note that both here and in the various simulations conducted in the present series of investigations, one-tailed tests are conducted because it is assumed that [especially in single-case A (baseline) B (intervention) research] the researcher has a clear and defensible rationale for the direction of change that is associated with the intervention. Insofar as randomization tests are not tailored to test for the equality of two populations specific parameters, all that can be tested for is the equality of the two population distributions per se. For the present randomization test, the test statistic involves sample-mean differences and because that is the test that produced a statistically significant result here (favoring the intervention phase over the baseline phase), a reasonable inference is that there was an A- to B-phase upward shift in the children s level of responding. Advantages of the Order Randomization Modification The present order-randomization approach enhances the internal validity of a single-case AB design by virtue of its removing bias stemming from interventionorder effects. As an important byproduct, the approach also elevates the status of the basic AB single-case intervention design from a WWC Standards acceptable design standpoint (Kratochwill et al., 2010), particularly when replicated across independent participants at different points in time. According to the WWC Standards, two-phase A (Baseline) B (Intervention) designs are not scientifically credible (and therefore unacceptable) because they suffer from too many potential sources of internal invalidity. For extended discussion of acceptable designs, see Kratochwill, et al. (2010, 2013). Including outcomes from both intervention-administration orders in the randomization distribution also provides fundamental pragmatic advantages for single-case intervention researchers. First, with the original Edgington (1975) model, a researcher would need to designate 20 potential intervention start points 12

14 LEVIN ET AL. (based on at least 21 total observations) to produce a randomization test that is capable of detecting an intervention effect with a one-tailed Type I error probability less than or equal to.05. With the present procedure, a researcher would need to designate only half as many potential intervention start points (here, 10, based on a total of 11 total observations, resulting in 20 possible outcomes) to detect an intervention effect. A related reason why the present procedure has practical importance for single-case intervention researchers is that (and as will be demonstrated here) relative to the original Edgington (1975) model, the modified approach may produce statistical-power advantages as well. Thus, for no more expense than a coin to flip, a researcher might reap both methodological and statistical benefits by adopting the present dual-randomization procedure rather than either the original single-randomization Edgington model or Marascuilo and Busk s (1988) multiple-case extension of it. Relationship to Traditional Experimental Designs and Statistical Analyses Although unrecognized at the time that the present order-randomization approach was initially conceptualized, its logic maps directly onto a statistical procedure in the traditional group randomized treatment-design literature. In particular, consider a randomized two-treatment correlated-samples (or within-subjects) design based on N participants, to which a nonparametric randomization test is applied as an appropriate alternative in (especially small-sample) situations where the normality assumption of a correlated-samples t test (or a one-sample repeatedmeasures analysis) is questionable. To illustrate that situation, we revisit an example that was recently presented by Ferron and Levin (2014, p. 174). Suppose that in a sample of N = 8 adults, each participant is administered two different fear-reducing treatments, A (a behavioral treatment) and B (a cognitive intervention), with the former posited to be more effective than the latter. It is determined in advance that the equaleffectiveness hypothesis will be tested with a randomization test based on a onetailed α of.05. To produce a scientifically credible experiment, the order in which the two treatments are administered is again randomly determined on a case-bycase basis by means of coin flips: say, heads represents an AB order and tails a BA order. On the basis of that process, let us suppose that 5 participants ended up in the AB condition and 3 in the BA condition. Following the administration of each treatment, participants fear responses are assessed on a 7-point Likert scale, with higher numbers indicating greater fear. With the measure of interest defined 13

15 IMPROVED RANDOMIZATION TESTS as the difference between each participant s B and A ratings (i.e., B A), the following outcomes were obtained for the 8 participants: The observed test statistic is given by the average of these differences, which is equal to +17/8 = A randomization distribution is created from the 2 N = 2 8 = 256 possible ways in each + and signs could be attached to these 8 numerical values. For example, the first outcome in the randomization distribution (with all + signs) would be: yielding a mean difference of +24/8 = 3.000, and the last (with all minus signs) would be: yielding a mean difference of 24/8 = The remaining 254 possible outcomes would fall somewhere between these two extremes. The actually obtained mean difference of appears to be on the higher side of this distribution. In fact, it turns out to be among the 9 highest possible outcomes (specifically, an outcome that is exceeded by only 5 outcomes and that is tied with 3 others). Accordingly, a one-tailed test of the hypothesis that the A and B treatments have equal distributions would be associated with a p-value (consistent with the alternative hypothesis that Treatment B is producing higher fear ratings than Treatment A) that is equal to 9/256 =.035. Because this value is less than the predetermined α of.05, it would be concluded that the actually obtained mean difference of is statistically significant. Note that for this conventional-group design and associated randomization test, the all-possible assignment of + and signs to the 8 absolute B A differences corresponds exactly to the logic and operationalization of the singlecase AB order-randomization procedure to be investigated here. In particular, the procedure incorporates two separate forms of randomization for each of the N participating cases, Edgington s intervention start-point randomization and AB order randomization. In the simplest situation where there is only one potential intervention start point for each case (as in the just-presented N = 8 example), the total number of possible start-point randomizations is equal to k N = 1 8 = 1. The 14

16 LEVIN ET AL. present order-randomization procedure involves each of the 8 participants contributing two differences (i.e., B A and A B) to the randomization distribution, resulting in 2 N = 2 8 = 256 joint randomization outcomes, and which, according to the previously given special-case dual-randomization formula, k N 2 N, yields a total of = 256 possible randomization outcomes. This total is identical to the number of possible randomization-distribution outcomes associated with the just-presented example. It is instructive to note that the total number of possible randomization outcomes associated with order randomization N N can be alternatively expressed as x 0 x, where N = the number of cases and x = the number of positive B A differences that could be associated with the N 8 8 actual outcomes. For the present example, this expression is equal to x 0 x, or Thus, when there is only one potential intervention point for each case and the AB design includes multiple observations, the present randomized-order test based on the difference between the A- and B-phase means maintains the same correspondence with a conventional-group correlated-samples randomization test as was shown here. Implicit in the conventional correlated-samples test is that with random assignment to treatment conditions, outcomes representing both orders of treatment administration need to be considered in the randomization test distribution. As such, the present order-randomization procedure is not really a special case at all, but rather the single-case analog of a correlated-samples randomization t test. Focus of the Present Investigations The focus of our series of simulation investigations was to examine the Type I error and statistical power characteristics of the dual-randomization modification (intervention start-point plus intervention order) relative to those of Edgington s (1975) and Marascuilo and Busk s (1988) original single-randomization (intervention start-point) test procedures. In this study we present randomized intervention-order findings not just for a basic two-phase AB design, but also for a randomized pairs variation of that design (Levin & Wampold, 1999), a single- 15

17 IMPROVED RANDOMIZATION TESTS case adaptation of the conventional-group crossover design, and Onghena s (1992) four-phase ABAB design. Investigations 1-3: Randomized Intervention Order for the Basic AB Design Investigation 1 Method In Investigation 1, the focus was on 30-observation designs for a single participant (i.e., N = 1), where the intervention start point was randomly selected from the middle 20 observations. The series length of 30 was chosen for initial examination because: (1) 20 start points is the minimum number needed to obtain a statistically significant result with a one-tailed α of.05 for an AB randomized start-point design with one case; and (2) the WWC Standards require a minimum of five observations in each phase (Kratochwill et al., 2010, 2013). Data were generated using SAS IML (SAS, 2013), where the time-series data were obtained by adding an error vector to an effect vector. The error vector was created such that it was distributed normally and had an autocorrelation of 0 or.3 by using SAS s autoregressive moving-average simulation function (ARMASIM). The autocorrelation values of 0 and.3 were motivated by a survey of actual single-case studies where it was reported that the average autocorrelation was.2, after adjusting for bias in the estimates (Shadish & Sullivan, 2011). To obtain simulated errors based on an autocorrelation of.3, the autoregressive parameter matrix was set to {1.3}, the moving average parameter matrix was set to {1 0}, and a standard deviation of the independent portion of the error was set to 1.0 (for details on the simulation algorithm see Woodfield, 1988). The effect vector was coded to have values of 0 for all baseline observations, and values of d for all intervention phase observations, and thus d corresponds to the mean shift between intervention and baseline observations in standard deviation units, (μ B μ A )/σ (see Busk & Serlin, 2005), where the standard deviation is based on the independent portion of the within-case error term (see, for example, Levin, Ferron, & Kratochwill, 2012) (for an alternative operationalization of d that corresponds mathematically to a conventional groups effect-size measure, see Shadish et al. (2014)). The value of d was varied to examine the one-tailed Type I error probability for d = 0 and the powers for ds ranging from.5 to 5 in increments of.5. For reference, if the d used for the present data generation is estimated for each of the 200 Phase A-to-Phase B contrasts examined in the survey of singlecase interventions reported by Parker and Vannest (2009), the empirically 16

18 LEVIN ET AL. observed values of d (assuming no autocorrelation for simplicity) for the 10 th, 50 th, and 90 th percentile ranks are estimated to be 0.46, 1.70, and 3.88, respectively. By crossing each design (single, dual), with each level of autocorrelation (r = 0,.3), and each effect size (d = 0 to 5, in increments of.5), = 44 conditions were obtained, and for each of these conditions the data for 10,000 studies were simulated. The data for each simulated data set were analyzed using a randomization test in which the obtained test statistic (M B M A ) was compared to the complete randomization distribution. The proportion of simulated studies in which the randomization test led to a one-tailed p-value of.05 or less was determined to estimate the rejection rate (Type I error or power) of the randomization test for each of the 44 experimental conditions. Figure 1. Investigation 1: Comparison (α =.05, one-tailed) of randomization tests for a one-case (N = 1) AB randomized intervention start-point design (Single) and the randomized intervention start-point plus randomized intervention- order design (Dual), where the start point was randomly selected between the 6 th through the 25 th observations inclusive in a 30-observations study. The rejection rate of the null hypothesis is shown as a function of the effect size and level of autocorrelation. 17

19 IMPROVED RANDOMIZATION TESTS Results Results are shown in Figure 1 for Edgington s (1975) original procedure (single) and for the present randomized-order modification (dual). As may be seen in that figure, when the effect size is 0, all situations are associated with empirical powers (which, for d = 0 are equivalent to Type I error probabilities) that correspond to their nominal.05 values. Not surprisingly, based on previous findings (e.g., Ferron & Sentovich, 2002; Ferron & Ware, 1995; Levin et al., 2011), it may also be seen that for ds > 0 power is uniformly higher for r = 0 than for r =.3. As the effect size increases, so does power, although more rapidly for the dual-randomization procedure than for its singlerandomization counterpart. The largest power differences, favoring the former, reach.21 in the r = 0 situation for ds of 1.5 and 2.0; and in the r =.3 situation the largest power difference is.18 for a d of 2.5. Investigation 2 Method In Investigation 2, series length (i.e., the number of observations) was systematically varied for a single-participant (N = 1) design, while holding the effect size constant at d = 2. A d of 2 was chosen because it is a large enough effect to typically be of interest to a single-case researcher. Yet, a d of 2 is small enough that it is not readily detectable (power <.80) in a single-participant 30- observations design when there is a moderate autocorrelation of.30 and applying either the single- or dual-randomization approach (as may be seen in Figure 1, where powers are.50 and.67, respectively). The simulation methods paralleled those of the initial investigation (including a one-tailed α of.05), but d was held constant at 2.0 for all conditions and series length was varied from 20 to 150 in increments of 10. The number of potential intervention start points was always the series length minus 10 to ensure at least five observations in the baseline and intervention phases. Results Results for this set of simulations are provided in Figure 2, where with an autocorrelation of.30, power of at least.80 is attained for the dualrandomization approach with 60 observations (power =.81), in contrast to the single-randomization design where.80 power is not quite attained even with 150 observations (power =.79). For 30 to 100 observations, the power difference between the two randomization schemes (favoring dual) ranges from.13 to.31 when the autocorrelation is 0 and from.17 to.30 when the autocorrelation is

20 LEVIN ET AL. Figure 2. Investigation 2: Comparison (α =.05, one-tailed) of randomization tests for a one-case (N = 1) AB randomized intervention start-point design (Single) and the randomized intervention start-point plus randomized intervention-order design (Dual). The rejection rate of the null hypothesis is shown as a function of series length and level of autocorrelation. The effect size is 2.0 and the number of potential intervention start points (x) is equal to the series length minus 10 and encompasses the middle x observations. It should be noted that the power is 0 for the single-randomization scheme with 20 observations because there are only 10 possible intervention start points and thus statistical significance cannot be obtained at the one-tailed.05 level. In addition, the undulation in the power curves for the single-randomization approach makes sense when one recognizes that: (1) for a series length of 30, statistical significance with α =.05 can be attained only for the most extreme of the 20 permutations; and (2) with a series length of 40, statistical significance can again be attained only for the most extreme permutation, but now there are 30 permutations and so the most extreme is somewhat more difficult to achieve. Although power drops for the 40-observation series, with a series length of 50, statistical significance can be attained for either of the two most extreme permutations and thus power jumps back up again. 19

21 IMPROVED RANDOMIZATION TESTS Investigation 3a Method In Investigation 3a, the effect of multiple-case replications (i.e., N > 1) on the power of the single- and dual-randomization procedures was examined. More specifically, a design with 15 observations and 5 potential intervention start points, randomly selected from observations 6 through 10, was examined with 2, 3, 4, 5, and 6 participants based on a one-tailed α of.05. For the single-randomization approach, 7 and 8 participants were also included. These numbers of participants seemed reasonable given the survey by Shadish and Sullivan (2011), in which it was found that the number of cases in single-case studies averaged 3.64, with a range of 1 to 13. In the present study, effect sizes varied from 0 to 3 in increments of.5 and the autocorrelation was set either to 0 or.3. Figure 3. Investigation 3a: Comparison (α =.05, one-tailed) of randomization tests for the Single and Dual basic AB randomized designs replicated across N cases. The rejection rate of the null hypothesis is shown as a function of effect size and N, for a 15- observations design with 5 potential intervention start points designated from between the 6 th and 10 th observations inclusive and an autocorrelation of 0. 20

22 LEVIN ET AL. Results Results from simulations where the autocorrelation is 0 are shown in Figure 3, whereas those for an autocorrelation of.3 are shown in Figure 4. In both figures, it may be seen that for all sample sizes the empirical Type I error probabilities are well controlled at.05 for both the single- and dual-randomization approaches. The important thing to note is that in both figures, for all effect sizes the dual approach based on as few as N = 3 participants has associated power that is greater than or equivalent to the single approach based on N = 8 participants. For example, in Figure 4 it may be seen that with an autocorrelation of.3, N = 3 dual- and N = 8 single-randomization powers are.66 and.61, respectively, for an effect size of 1.0; and they are.90 and.89, respectively, for an effect size of 1.5. Figure 4. Investigation 3a: Comparison (α =.05, one-tailed) of randomization tests for the Single and Dual basic AB randomized designs replicated across N cases. The rejection rate of the null hypothesis is shown as a function of effect size and N, for a 15 observations design with 5 potential intervention start points designated from between the 6 th and 10 th observations inclusive and an autocorrelation of.3. 21

23 IMPROVED RANDOMIZATION TESTS Investigation 3b Method In this investigation, the simulations of Investigation 3a were replicated with the sole difference being that a two-tailed test with α =.05 was conducted, as opposed to a one-tailed test. Results The results are summarized in Figure 5 for an autocorrelation of 0 and in Figure 6 for an autocorrelation of.3. Again, it may be seen that all of the empirical Type I errors are at the expected.05 level for both autocorrelation values. Although the Investigation 3a results (i.e., the equivalence of dualrandomization N = 3 and single-randomization N = 8) were not identical here, the general pattern was. In this case, however, the appropriate power equivalence involves dual N = 4 and single N = 8. Specifically, in Figure 6 it may be seen that with an autocorrelation of.3, the former and latter powers are.65 and.61, respectively, for an effect size of 1.0; and they are.93 and.89, respectively, for an effect size of 1.5. Figure 5. Investigation 3b: Comparison (α =.05, two-tailed) of randomization tests for the Single and Dual basic AB randomized designs replicated across N cases. The rejection rate of the null hypothesis is shown as a function of effect size and N, for a 15 observations design with 5 potential intervention start points designated from between the 6 th and 10 th observations inclusive and an autocorrelation of 0. 22

24 LEVIN ET AL. Figure 6. Investigation 3b: Comparison (α =.05, two-tailed) of randomization tests for the Single and Dual basic AB randomized designs replicated across N cases. The rejection rate of the null hypothesis is shown as a function of effect size and N, for a 15 observations design with 5 potential intervention start points designated from between the 6 th and 10 th observations inclusive and an autocorrelation of.3. Thus, in the present investigation we observe that for two-tailed tests the dual-randomization power benefits (relative to single randomization) are comparable to those reported for Investigation 3a s one-tailed tests. It is important to point out, however, that the situations examined here were all based on multiple-case (N > 1) designs. It turns out that for the special-case N = 1 situation, although the dual- over single-randomization power advantage is evident when one-tailed tests are conducted (as was true in Investigations 1 and 2), the dualand single-randomization schemes yield equivalent power results with two-tailed tests. Because the two-tailed test is based on randomization-distribution absolutevalue outcomes, the dual-randomization distribution contains every outcome of the single-randomization distribution as well as its opposite-order complementary outcome, thereby yielding exactly the same p-value for each test. (To illustrate these notions, see Child 1 s hypothetical data, including Footnote a in Table 2. The 40 unsigned mean differences (i.e., 20 B A plus 20 A B ) would constitute the dual-randomization distribution for a two-tailed test). Because there are across-case combinations when N > 1, there is no longer a one-to-one 23

25 IMPROVED RANDOMIZATION TESTS correspondence between the single- and dual-randomization distributions and so their powers will generally differ, with the latter being greater (as was observed in Figures 5 and 6). Investigation 4: Randomized Intervention Order and/or Randomized Intervention Assignment in Levin and Wampold s (1999) AB Pairs Design Another type of dual-randomization strategy is possible when a case consists of a pair of participants, as in Levin and Wampold s (1999) simultaneous intervention start-point model. With the Levin-Wampold model, N participant (or other unit) pairs are created and the members of each pair are randomly assigned to two different intervention conditions (or to an intervention and control condition), X and Y. With this model, Levin and Wampold presented two hypotheses that would be of interest to researchers: (1) a general intervention effectiveness hypothesis, namely that averaged across the two intervention conditions, there is no difference between Phase A and Phase B performance (analogous to the time main effect in a conventional two-treatment pretest-posttest design); and (2) a comparative intervention effectiveness hypothesis, namely that the change in participants performance from Phase A to Phase B is the same in the two intervention conditions (analogous to the treatment-by-time interaction in a conventional two-treatment pretest-posttest design). Unrecognized by Levin and Wampold at the time, the randomization test of each of these hypotheses could potentially benefit from an additional randomization component. For the general intervention effectiveness hypothesis, that component is AB order randomization of the kind that we have considered in Investigations 1-3, either with or without a mandatory A' baseline phase; and for the comparative intervention hypothesis, that component consists of within-pair intervention randomization, wherein pair members are randomly assigned to the two intervention conditions. Implementing either of these randomization types increases the total number N of possible outcomes from k for Levin and Wampold s (1999) original i 1 i single randomization-test procedure (i.e., the number of potential intervention start points for each pair) to 2 N N k for the present dual approach (i.e., either i 1 i the number of possible random assignments of AB orders or the number of possible random assignments of interventions to pair members, times the number of potential intervention start points for each pair). In Investigation 4, we examine the statistical power consequences associated with the dual approach s additional 24

Statistical Randomization Techniques for Single-Case Intervention Data. Joel R. Levin University of Arizona

Statistical Randomization Techniques for Single-Case Intervention Data. Joel R. Levin University of Arizona Statistical Randomization Techniques for Single-Case Intervention Data Joel R. Levin University of Arizona Purpose of This Presentation (Dejà Vu?) To broaden your minds by introducing you to new and exciting,

More information

Confidence Intervals On Subsets May Be Misleading

Confidence Intervals On Subsets May Be Misleading Journal of Modern Applied Statistical Methods Volume 3 Issue 2 Article 2 11-1-2004 Confidence Intervals On Subsets May Be Misleading Juliet Popper Shaffer University of California, Berkeley, shaffer@stat.berkeley.edu

More information

Testing the Intervention Effect in Single-Case Experiments: A Monte Carlo Simulation Study

Testing the Intervention Effect in Single-Case Experiments: A Monte Carlo Simulation Study The Journal of Experimental Education ISSN: 0022-0973 (Print) 1940-0683 (Online) Journal homepage: http://www.tandfonline.com/loi/vjxe20 Testing the Intervention Effect in Single-Case Experiments: A Monte

More information

Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper.

Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper. Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper December 13, 2013 Jean Stockard Professor Emerita, University of Oregon Director

More information

Taichi OKUMURA * ABSTRACT

Taichi OKUMURA * ABSTRACT 上越教育大学研究紀要第 35 巻平成 28 年 月 Bull. Joetsu Univ. Educ., Vol. 35, Mar. 2016 Taichi OKUMURA * (Received August 24, 2015; Accepted October 22, 2015) ABSTRACT This study examined the publication bias in the meta-analysis

More information

Estimating individual treatment effects from multiple-baseline data: A Monte Carlo study of multilevel-modeling approaches

Estimating individual treatment effects from multiple-baseline data: A Monte Carlo study of multilevel-modeling approaches Behavior Research Methods 2010, 42 (4), 930-943 doi:10.3758/brm.42.4.930 Estimating individual treatment effects from multiple-baseline data: A Monte Carlo study of multilevel-modeling approaches JOHN

More information

The Impact of Continuity Violation on ANOVA and Alternative Methods

The Impact of Continuity Violation on ANOVA and Alternative Methods Journal of Modern Applied Statistical Methods Volume 12 Issue 2 Article 6 11-1-2013 The Impact of Continuity Violation on ANOVA and Alternative Methods Björn Lantz Chalmers University of Technology, Gothenburg,

More information

Evaluating Count Outcomes in Synthesized Single- Case Designs with Multilevel Modeling: A Simulation Study

Evaluating Count Outcomes in Synthesized Single- Case Designs with Multilevel Modeling: A Simulation Study University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Public Access Theses and Dissertations from the College of Education and Human Sciences Education and Human Sciences, College

More information

Meta-analysis using HLM 1. Running head: META-ANALYSIS FOR SINGLE-CASE INTERVENTION DESIGNS

Meta-analysis using HLM 1. Running head: META-ANALYSIS FOR SINGLE-CASE INTERVENTION DESIGNS Meta-analysis using HLM 1 Running head: META-ANALYSIS FOR SINGLE-CASE INTERVENTION DESIGNS Comparing Two Meta-Analysis Approaches for Single Subject Design: Hierarchical Linear Model Perspective Rafa Kasim

More information

The Regression-Discontinuity Design

The Regression-Discontinuity Design Page 1 of 10 Home» Design» Quasi-Experimental Design» The Regression-Discontinuity Design The regression-discontinuity design. What a terrible name! In everyday language both parts of the term have connotations

More information

WWC Standards for Regression Discontinuity and Single Case Study Designs

WWC Standards for Regression Discontinuity and Single Case Study Designs WWC Standards for Regression Discontinuity and Single Case Study Designs March 2011 Presentation to the Society for Research on Educational Effectiveness John Deke Shannon Monahan Jill Constantine 1 What

More information

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin Key words : Bayesian approach, classical approach, confidence interval, estimation, randomization,

More information

Fundamental Clinical Trial Design

Fundamental Clinical Trial Design Design, Monitoring, and Analysis of Clinical Trials Session 1 Overview and Introduction Overview Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics, University of Washington February 17-19, 2003

More information

Technical Specifications

Technical Specifications Technical Specifications In order to provide summary information across a set of exercises, all tests must employ some form of scoring models. The most familiar of these scoring models is the one typically

More information

Student Performance Q&A:

Student Performance Q&A: Student Performance Q&A: 2009 AP Statistics Free-Response Questions The following comments on the 2009 free-response questions for AP Statistics were written by the Chief Reader, Christine Franklin of

More information

Research Methods 1 Handouts, Graham Hole,COGS - version 1.0, September 2000: Page 1:

Research Methods 1 Handouts, Graham Hole,COGS - version 1.0, September 2000: Page 1: Research Methods 1 Handouts, Graham Hole,COGS - version 10, September 000: Page 1: T-TESTS: When to use a t-test: The simplest experimental design is to have two conditions: an "experimental" condition

More information

Investigating the robustness of the nonparametric Levene test with more than two groups

Investigating the robustness of the nonparametric Levene test with more than two groups Psicológica (2014), 35, 361-383. Investigating the robustness of the nonparametric Levene test with more than two groups David W. Nordstokke * and S. Mitchell Colp University of Calgary, Canada Testing

More information

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

3 CONCEPTUAL FOUNDATIONS OF STATISTICS 3 CONCEPTUAL FOUNDATIONS OF STATISTICS In this chapter, we examine the conceptual foundations of statistics. The goal is to give you an appreciation and conceptual understanding of some basic statistical

More information

1 The conceptual underpinnings of statistical power

1 The conceptual underpinnings of statistical power 1 The conceptual underpinnings of statistical power The importance of statistical power As currently practiced in the social and health sciences, inferential statistics rest solidly upon two pillars: statistical

More information

Using SAS to Calculate Tests of Cliff s Delta. Kristine Y. Hogarty and Jeffrey D. Kromrey

Using SAS to Calculate Tests of Cliff s Delta. Kristine Y. Hogarty and Jeffrey D. Kromrey Using SAS to Calculate Tests of Cliff s Delta Kristine Y. Hogarty and Jeffrey D. Kromrey Department of Educational Measurement and Research, University of South Florida ABSTRACT This paper discusses a

More information

Policy and Position Statement on Single Case Research and Experimental Designs from the Division for Research Council for Exceptional Children

Policy and Position Statement on Single Case Research and Experimental Designs from the Division for Research Council for Exceptional Children Policy and Position Statement on Single Case Research and Experimental Designs from the Division for Research Council for Exceptional Children Wendy Rodgers, University of Nevada, Las Vegas; Timothy Lewis,

More information

MEA DISCUSSION PAPERS

MEA DISCUSSION PAPERS Inference Problems under a Special Form of Heteroskedasticity Helmut Farbmacher, Heinrich Kögel 03-2015 MEA DISCUSSION PAPERS mea Amalienstr. 33_D-80799 Munich_Phone+49 89 38602-355_Fax +49 89 38602-390_www.mea.mpisoc.mpg.de

More information

CHAPTER THIRTEEN. Data Analysis and Interpretation: Part II.Tests of Statistical Significance and the Analysis Story CHAPTER OUTLINE

CHAPTER THIRTEEN. Data Analysis and Interpretation: Part II.Tests of Statistical Significance and the Analysis Story CHAPTER OUTLINE CHAPTER THIRTEEN Data Analysis and Interpretation: Part II.Tests of Statistical Significance and the Analysis Story CHAPTER OUTLINE OVERVIEW NULL HYPOTHESIS SIGNIFICANCE TESTING (NHST) EXPERIMENTAL SENSITIVITY

More information

Chapter 12. The One- Sample

Chapter 12. The One- Sample Chapter 12 The One- Sample z-test Objective We are going to learn to make decisions about a population parameter based on sample information. Lesson 12.1. Testing a Two- Tailed Hypothesis Example 1: Let's

More information

Issues and advances in the systematic review of single-case research: A commentary on the exemplars

Issues and advances in the systematic review of single-case research: A commentary on the exemplars 1 Issues and advances in the systematic review of single-case research: A commentary on the exemplars Rumen Manolov, Georgina Guilera, and Antonio Solanas Department of Social Psychology and Quantitative

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

Tests for the Visual Analysis of Response- Guided Multiple-Baseline Data

Tests for the Visual Analysis of Response- Guided Multiple-Baseline Data The Journal of Experimental Education, 2006, 75(1), 66 81 Copyright 2006 Heldref Publications Tests for the Visual Analysis of Response- Guided Multiple-Baseline Data JOHN FERRON PEGGY K. JONES University

More information

SPRING GROVE AREA SCHOOL DISTRICT. Course Description. Instructional Strategies, Learning Practices, Activities, and Experiences.

SPRING GROVE AREA SCHOOL DISTRICT. Course Description. Instructional Strategies, Learning Practices, Activities, and Experiences. SPRING GROVE AREA SCHOOL DISTRICT PLANNED COURSE OVERVIEW Course Title: Basic Introductory Statistics Grade Level(s): 11-12 Units of Credit: 1 Classification: Elective Length of Course: 30 cycles Periods

More information

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n. University of Groningen Latent instrumental variables Ebbes, P. IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

A Brief Introduction to Bayesian Statistics

A Brief Introduction to Bayesian Statistics A Brief Introduction to Statistics David Kaplan Department of Educational Psychology Methods for Social Policy Research and, Washington, DC 2017 1 / 37 The Reverend Thomas Bayes, 1701 1761 2 / 37 Pierre-Simon

More information

Comparison of the Null Distributions of

Comparison of the Null Distributions of Comparison of the Null Distributions of Weighted Kappa and the C Ordinal Statistic Domenic V. Cicchetti West Haven VA Hospital and Yale University Joseph L. Fleiss Columbia University It frequently occurs

More information

Six Sigma Glossary Lean 6 Society

Six Sigma Glossary Lean 6 Society Six Sigma Glossary Lean 6 Society ABSCISSA ACCEPTANCE REGION ALPHA RISK ALTERNATIVE HYPOTHESIS ASSIGNABLE CAUSE ASSIGNABLE VARIATIONS The horizontal axis of a graph The region of values for which the null

More information

Psychology Research Process

Psychology Research Process Psychology Research Process Logical Processes Induction Observation/Association/Using Correlation Trying to assess, through observation of a large group/sample, what is associated with what? Examples:

More information

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables?

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables? Research Questions, Variables, and Hypotheses: Part 2 RCS 6740 6/7/04 1 Review What are research questions? What are variables? Definition Function Measurement Scale 2 Hypotheses OK, now that we know how

More information

Sheila Barron Statistics Outreach Center 2/8/2011

Sheila Barron Statistics Outreach Center 2/8/2011 Sheila Barron Statistics Outreach Center 2/8/2011 What is Power? When conducting a research study using a statistical hypothesis test, power is the probability of getting statistical significance when

More information

Conditional Distributions and the Bivariate Normal Distribution. James H. Steiger

Conditional Distributions and the Bivariate Normal Distribution. James H. Steiger Conditional Distributions and the Bivariate Normal Distribution James H. Steiger Overview In this module, we have several goals: Introduce several technical terms Bivariate frequency distribution Marginal

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information

CHAPTER 8 EXPERIMENTAL DESIGN

CHAPTER 8 EXPERIMENTAL DESIGN CHAPTER 8 1 EXPERIMENTAL DESIGN LEARNING OBJECTIVES 2 Define confounding variable, and describe how confounding variables are related to internal validity Describe the posttest-only design and the pretestposttest

More information

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES 24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 10: Introduction to inference (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 17 What is inference? 2 / 17 Where did our data come from? Recall our sample is: Y, the vector

More information

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC Hypothesis Testing Richard S. Balkin, Ph.D., LPC-S, NCC Overview When we have questions about the effect of a treatment or intervention or wish to compare groups, we use hypothesis testing Parametric statistics

More information

Statistical inference provides methods for drawing conclusions about a population from sample data.

Statistical inference provides methods for drawing conclusions about a population from sample data. Chapter 14 Tests of Significance Statistical inference provides methods for drawing conclusions about a population from sample data. Two of the most common types of statistical inference: 1) Confidence

More information

Monte Carlo Analysis of Univariate Statistical Outlier Techniques Mark W. Lukens

Monte Carlo Analysis of Univariate Statistical Outlier Techniques Mark W. Lukens Monte Carlo Analysis of Univariate Statistical Outlier Techniques Mark W. Lukens This paper examines three techniques for univariate outlier identification: Extreme Studentized Deviate ESD), the Hampel

More information

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior 1 Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior Gregory Francis Department of Psychological Sciences Purdue University gfrancis@purdue.edu

More information

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data TECHNICAL REPORT Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data CONTENTS Executive Summary...1 Introduction...2 Overview of Data Analysis Concepts...2

More information

Chapter 1: Exploring Data

Chapter 1: Exploring Data Chapter 1: Exploring Data Key Vocabulary:! individual! variable! frequency table! relative frequency table! distribution! pie chart! bar graph! two-way table! marginal distributions! conditional distributions!

More information

Still important ideas

Still important ideas Readings: OpenStax - Chapters 1 13 & Appendix D & E (online) Plous Chapters 17 & 18 - Chapter 17: Social Influences - Chapter 18: Group Judgments and Decisions Still important ideas Contrast the measurement

More information

Chapter Three: Sampling Methods

Chapter Three: Sampling Methods Chapter Three: Sampling Methods The idea of this chapter is to make sure that you address sampling issues - even though you may be conducting an action research project and your sample is "defined" by

More information

Communication Research Practice Questions

Communication Research Practice Questions Communication Research Practice Questions For each of the following questions, select the best answer from the given alternative choices. Additional instructions are given as necessary. Read each question

More information

The SAGE Encyclopedia of Educational Research, Measurement, and Evaluation Multivariate Analysis of Variance

The SAGE Encyclopedia of Educational Research, Measurement, and Evaluation Multivariate Analysis of Variance The SAGE Encyclopedia of Educational Research, Measurement, Multivariate Analysis of Variance Contributors: David W. Stockburger Edited by: Bruce B. Frey Book Title: Chapter Title: "Multivariate Analysis

More information

Inferences: What inferences about the hypotheses and questions can be made based on the results?

Inferences: What inferences about the hypotheses and questions can be made based on the results? QALMRI INSTRUCTIONS QALMRI is an acronym that stands for: Question: (a) What was the broad question being asked by this research project? (b) What was the specific question being asked by this research

More information

Understandable Statistics

Understandable Statistics Understandable Statistics correlated to the Advanced Placement Program Course Description for Statistics Prepared for Alabama CC2 6/2003 2003 Understandable Statistics 2003 correlated to the Advanced Placement

More information

BIOL 458 BIOMETRY Lab 7 Multi-Factor ANOVA

BIOL 458 BIOMETRY Lab 7 Multi-Factor ANOVA BIOL 458 BIOMETRY Lab 7 Multi-Factor ANOVA PART 1: Introduction to Factorial ANOVA ingle factor or One - Way Analysis of Variance can be used to test the null hypothesis that k or more treatment or group

More information

Psychology Research Process

Psychology Research Process Psychology Research Process Logical Processes Induction Observation/Association/Using Correlation Trying to assess, through observation of a large group/sample, what is associated with what? Examples:

More information

A Case Study: Two-sample categorical data

A Case Study: Two-sample categorical data A Case Study: Two-sample categorical data Patrick Breheny January 31 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/43 Introduction Model specification Continuous vs. mixture priors Choice

More information

Please cite this article as: Heyvaert, M., Wendt, O., Van den Noortgate, W., & Onghena, P. (2015). Randomization and data-analysis items in quality

Please cite this article as: Heyvaert, M., Wendt, O., Van den Noortgate, W., & Onghena, P. (2015). Randomization and data-analysis items in quality Please cite this article as: Heyvaert, M., Wendt, O., Van den Noortgate, W., & Onghena, P. (2015). Randomization and data-analysis items in quality standards for single-case experimental studies. Journal

More information

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS Chapter Objectives: Understand Null Hypothesis Significance Testing (NHST) Understand statistical significance and

More information

Lecture 4: Research Approaches

Lecture 4: Research Approaches Lecture 4: Research Approaches Lecture Objectives Theories in research Research design approaches ú Experimental vs. non-experimental ú Cross-sectional and longitudinal ú Descriptive approaches How to

More information

Testing Means. Related-Samples t Test With Confidence Intervals. 6. Compute a related-samples t test and interpret the results.

Testing Means. Related-Samples t Test With Confidence Intervals. 6. Compute a related-samples t test and interpret the results. 10 Learning Objectives Testing Means After reading this chapter, you should be able to: Related-Samples t Test With Confidence Intervals 1. Describe two types of research designs used when we select related

More information

MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and. Lord Equating Methods 1,2

MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and. Lord Equating Methods 1,2 MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and Lord Equating Methods 1,2 Lisa A. Keller, Ronald K. Hambleton, Pauline Parker, Jenna Copella University of Massachusetts

More information

Lecture Notes Module 2

Lecture Notes Module 2 Lecture Notes Module 2 Two-group Experimental Designs The goal of most research is to assess a possible causal relation between the response variable and another variable called the independent variable.

More information

Examining differences between two sets of scores

Examining differences between two sets of scores 6 Examining differences between two sets of scores In this chapter you will learn about tests which tell us if there is a statistically significant difference between two sets of scores. In so doing you

More information

Empirical Formula for Creating Error Bars for the Method of Paired Comparison

Empirical Formula for Creating Error Bars for the Method of Paired Comparison Empirical Formula for Creating Error Bars for the Method of Paired Comparison Ethan D. Montag Rochester Institute of Technology Munsell Color Science Laboratory Chester F. Carlson Center for Imaging Science

More information

A Comparison of Shape and Scale Estimators of the Two-Parameter Weibull Distribution

A Comparison of Shape and Scale Estimators of the Two-Parameter Weibull Distribution Journal of Modern Applied Statistical Methods Volume 13 Issue 1 Article 3 5-1-2014 A Comparison of Shape and Scale Estimators of the Two-Parameter Weibull Distribution Florence George Florida International

More information

Introduction & Basics

Introduction & Basics CHAPTER 1 Introduction & Basics 1.1 Statistics the Field... 1 1.2 Probability Distributions... 4 1.3 Study Design Features... 9 1.4 Descriptive Statistics... 13 1.5 Inferential Statistics... 16 1.6 Summary...

More information

Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC

Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC Conducting Research in the Social Sciences Rick Balkin, Ph.D., LPC-S, NCC 1 Why we do research Improvement Description Explanation Prediction R. S. Balkin, 2008 2 Theory Explanation of an observed phenomena

More information

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity Measurement & Variables - Initial step is to conceptualize and clarify the concepts embedded in a hypothesis or research question with

More information

ELEMENTS OF PSYCHOPHYSICS Sections VII and XVI. Gustav Theodor Fechner (1860/1912)

ELEMENTS OF PSYCHOPHYSICS Sections VII and XVI. Gustav Theodor Fechner (1860/1912) ELEMENTS OF PSYCHOPHYSICS Sections VII and XVI Gustav Theodor Fechner (1860/1912) Translated by Herbert Sidney Langfeld (1912) [Classics Editor's note: This translation of these passages from Fechner's

More information

DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas

DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas Director, BU-CHED Zonal Research Center Bicol University Research and Development Center Legazpi City Research Proposal Preparation Seminar-Writeshop

More information

INTRODUCTION TO ECONOMETRICS (EC212)

INTRODUCTION TO ECONOMETRICS (EC212) INTRODUCTION TO ECONOMETRICS (EC212) Course duration: 54 hours lecture and class time (Over three weeks) LSE Teaching Department: Department of Economics Lead Faculty (session two): Dr Taisuke Otsu and

More information

Business Statistics Probability

Business Statistics Probability Business Statistics The following was provided by Dr. Suzanne Delaney, and is a comprehensive review of Business Statistics. The workshop instructor will provide relevant examples during the Skills Assessment

More information

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks)

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks) MBA SEMESTER III MB0050 Research Methodology- 4 Credits (Book ID: B1206 ) Assignment Set- 1 (60 Marks) Note: Each question carries 10 Marks. Answer all the questions Q1. a. Differentiate between nominal,

More information

Dr. Kelly Bradley Final Exam Summer {2 points} Name

Dr. Kelly Bradley Final Exam Summer {2 points} Name {2 points} Name You MUST work alone no tutors; no help from classmates. Email me or see me with questions. You will receive a score of 0 if this rule is violated. This exam is being scored out of 00 points.

More information

Still important ideas

Still important ideas Readings: OpenStax - Chapters 1 11 + 13 & Appendix D & E (online) Plous - Chapters 2, 3, and 4 Chapter 2: Cognitive Dissonance, Chapter 3: Memory and Hindsight Bias, Chapter 4: Context Dependence Still

More information

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison Empowered by Psychometrics The Fundamentals of Psychometrics Jim Wollack University of Wisconsin Madison Psycho-what? Psychometrics is the field of study concerned with the measurement of mental and psychological

More information

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA Data Analysis: Describing Data CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA In the analysis process, the researcher tries to evaluate the data collected both from written documents and from other sources such

More information

BIOSTATISTICS. Dr. Hamza Aduraidi

BIOSTATISTICS. Dr. Hamza Aduraidi BIOSTATISTICS Dr. Hamza Aduraidi Unit One INTRODUCTION Biostatistics It can be defined as the application of the mathematical tools used in statistics to the fields of biological sciences and medicine.

More information

Applied Statistical Analysis EDUC 6050 Week 4

Applied Statistical Analysis EDUC 6050 Week 4 Applied Statistical Analysis EDUC 6050 Week 4 Finding clarity using data Today 1. Hypothesis Testing with Z Scores (continued) 2. Chapters 6 and 7 in Book 2 Review! = $ & '! = $ & ' * ) 1. Which formula

More information

Experimental Psychology

Experimental Psychology Title Experimental Psychology Type Individual Document Map Authors Aristea Theodoropoulos, Patricia Sikorski Subject Social Studies Course None Selected Grade(s) 11, 12 Location Roxbury High School Curriculum

More information

Statistical Techniques. Masoud Mansoury and Anas Abulfaraj

Statistical Techniques. Masoud Mansoury and Anas Abulfaraj Statistical Techniques Masoud Mansoury and Anas Abulfaraj What is Statistics? https://www.youtube.com/watch?v=lmmzj7599pw The definition of Statistics The practice or science of collecting and analyzing

More information

Chapter 5: Producing Data

Chapter 5: Producing Data Chapter 5: Producing Data Key Vocabulary: observational study vs. experiment confounded variables population vs. sample sampling vs. census sample design voluntary response sampling convenience sampling

More information

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings Authors and Affiliations: Peter M. Steiner, University of Wisconsin-Madison

More information

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F Plous Chapters 17 & 18 Chapter 17: Social Influences Chapter 18: Group Judgments and Decisions

More information

Lesson 9: Two Factor ANOVAS

Lesson 9: Two Factor ANOVAS Published on Agron 513 (https://courses.agron.iastate.edu/agron513) Home > Lesson 9 Lesson 9: Two Factor ANOVAS Developed by: Ron Mowers, Marin Harbur, and Ken Moore Completion Time: 1 week Introduction

More information

CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH

CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH Why Randomize? This case study is based on Training Disadvantaged Youth in Latin America: Evidence from a Randomized Trial by Orazio Attanasio,

More information

WDHS Curriculum Map Probability and Statistics. What is Statistics and how does it relate to you?

WDHS Curriculum Map Probability and Statistics. What is Statistics and how does it relate to you? WDHS Curriculum Map Probability and Statistics Time Interval/ Unit 1: Introduction to Statistics 1.1-1.3 2 weeks S-IC-1: Understand statistics as a process for making inferences about population parameters

More information

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo Business Statistics The following was provided by Dr. Suzanne Delaney, and is a comprehensive review of Business Statistics. The workshop instructor will provide relevant examples during the Skills Assessment

More information

investigate. educate. inform.

investigate. educate. inform. investigate. educate. inform. Research Design What drives your research design? The battle between Qualitative and Quantitative is over Think before you leap What SHOULD drive your research design. Advanced

More information

Review of Veterinary Epidemiologic Research by Dohoo, Martin, and Stryhn

Review of Veterinary Epidemiologic Research by Dohoo, Martin, and Stryhn The Stata Journal (2004) 4, Number 1, pp. 89 92 Review of Veterinary Epidemiologic Research by Dohoo, Martin, and Stryhn Laurent Audigé AO Foundation laurent.audige@aofoundation.org Abstract. The new book

More information

NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES

NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES Amit Teller 1, David M. Steinberg 2, Lina Teper 1, Rotem Rozenblum 2, Liran Mendel 2, and Mordechai Jaeger 2 1 RAFAEL, POB 2250, Haifa, 3102102, Israel

More information

INVESTIGATING FIT WITH THE RASCH MODEL. Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form

INVESTIGATING FIT WITH THE RASCH MODEL. Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form INVESTIGATING FIT WITH THE RASCH MODEL Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form of multidimensionality. The settings in which measurement

More information

The essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV.

The essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV. EXPERIMENTAL DESIGNS I: Between-Groups Designs There are many experimental designs. We begin this week with the most basic, where there is a single IV and where participants are divided into two or more

More information

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Still important ideas Contrast the measurement of observable actions (and/or characteristics)

More information

Group Assignment #1: Concept Explication. For each concept, ask and answer the questions before your literature search.

Group Assignment #1: Concept Explication. For each concept, ask and answer the questions before your literature search. Group Assignment #1: Concept Explication 1. Preliminary identification of the concept. Identify and name each concept your group is interested in examining. Questions to asked and answered: Is each concept

More information

Appendix B Statistical Methods

Appendix B Statistical Methods Appendix B Statistical Methods Figure B. Graphing data. (a) The raw data are tallied into a frequency distribution. (b) The same data are portrayed in a bar graph called a histogram. (c) A frequency polygon

More information

CHAPTER ONE CORRELATION

CHAPTER ONE CORRELATION CHAPTER ONE CORRELATION 1.0 Introduction The first chapter focuses on the nature of statistical data of correlation. The aim of the series of exercises is to ensure the students are able to use SPSS to

More information

Chapter 1: Explaining Behavior

Chapter 1: Explaining Behavior Chapter 1: Explaining Behavior GOAL OF SCIENCE is to generate explanations for various puzzling natural phenomenon. - Generate general laws of behavior (psychology) RESEARCH: principle method for acquiring

More information

PSYCHOLOGY 300B (A01) One-sample t test. n = d = ρ 1 ρ 0 δ = d (n 1) d

PSYCHOLOGY 300B (A01) One-sample t test. n = d = ρ 1 ρ 0 δ = d (n 1) d PSYCHOLOGY 300B (A01) Assignment 3 January 4, 019 σ M = σ N z = M µ σ M d = M 1 M s p d = µ 1 µ 0 σ M = µ +σ M (z) Independent-samples t test One-sample t test n = δ δ = d n d d = µ 1 µ σ δ = d n n = δ

More information

Effects of Sequential Context on Judgments and Decisions in the Prisoner s Dilemma Game

Effects of Sequential Context on Judgments and Decisions in the Prisoner s Dilemma Game Effects of Sequential Context on Judgments and Decisions in the Prisoner s Dilemma Game Ivaylo Vlaev (ivaylo.vlaev@psy.ox.ac.uk) Department of Experimental Psychology, University of Oxford, Oxford, OX1

More information

Biostatistics 3. Developed by Pfizer. March 2018

Biostatistics 3. Developed by Pfizer. March 2018 BROUGHT TO YOU BY Biostatistics 3 Developed by Pfizer March 2018 This learning module is intended for UK healthcare professionals only. Job bag: PP-GEP-GBR-0986 Date of preparation March 2018. Agenda I.

More information