Running head: SELECTIVE REPORTING 1

Size: px
Start display at page:

Download "Running head: SELECTIVE REPORTING 1"

Transcription

1 Running head: SELECTIVE REPORTING 1 Please use the following citation when referencing this work: McGill, R. J., Ward, T. J., & Canivez, G. L. (2018). On the evidential value of school psychology intervention research. The School Psychologist, 72 (3), Retrieved from American Psychological Association, This paper is not the copy of record and may not exactly replicate the authoritative document published in the APA journal. Please do not copy or cite without author's permission. The final article is available, upon publication, at: On the Evidential Value of School Psychology Intervention Research Ryan J. McGill Thomas J. Ward William & Mary Gary L. Canivez Eastern Illinois University Author notes Correspondence concerning this article should be addressed to Ryan J. McGill, William & Mary School of Education, P.O. Box 8795, Williamsburg, VA rmcgill@wm.edu

2 SELECTIVE REPORTING 2 Abstract The present investigation examined the evidential value of intervention research published in school psychology journals from Although the nine journals that were reviewed favored studies that have a low probability of wrongly rejecting the so-called null hypothesis (H0), p-curve analyses did not yield evidence of selective reporting in the literature. Whereas the potential threat of publication bias was not supported, the prevalence rate of replication studies was relatively low. Nevertheless, the majority of published intervention studies in school psychology contained sufficient power to promote future replication of observed effects. It is believed that the present investigation will be instructive for broadening the replication conversation in the school psychology to avoid problems experienced in other psychological research areas. Keywords: replication crisis, p-curve, evidence-based practice

3 SELECTIVE REPORTING 3 On the Evidential Value of School Psychology Intervention Research There has been a longstanding debate in the field of school psychology regarding the role of research and its utility and applicability for school-based practitioners. Nevertheless, American Psychological Association and National Association of School Psychologists training and practice standards specify that practitioners should demonstrate skills to apply research findings as a foundation for service delivery regardless of their level of training, and the responsibility to acquire and apply accurate knowledge about effective practices are considered to be an epistemic responsibility of the clinician (e.g., O Donahue & Henderson, 1999). As stated by Lilienfeld et al. (2012), all school psychologists, regardless of the setting in which they operate, need to develop and maintain a skill set that allows them to distinguish evidencebased from non-evidence based practices (p. 8) and this notion is a foundational principle of the broader evidence-based practice movement. It stands to reason that implementing this skill set requires practitioners to place a tremendous amount of faith in the accuracy of published research and the integrity of the very publication process itself. Unfortunately, over the past decade, serious questions have been raised about commonly accepted methodologies (e.g., allegiance to null hypothesis significance testing [NHST]) in scientific research and the reproducibility of many published findings in psychological science. For example, Ioannidis (2005) suggested that half of all published research findings are likely false due to the prevalence of underpowered studies and the use of questionable research practices (QRPs) 1 and the results of a highly influential study published in 1 These are two separate but equally important confounds. The power of study is the probability that it will distinguish between a true effect and chance and is mediated by sample size and the strength of the observed effect. On the other hand, QRPs are a class of techniques in which

4 SELECTIVE REPORTING 4 Science seemed to confirm this contention. In that study, researchers associated with the Open Science Collaboration (2015) attempted to replicate 100 experiments reported in articles published in three high-ranking psychology journals in 2008 and were able to obtain a replication rate of only 39%. These results have prompted many to conclude that psychology is in the midst of a replication crisis 2. Though it should be noted that some catalyst scholars reject this notion (see Baumeister & Vohs, 2016) and subsequent debates on these issues have been acrimonious. Regardless of one s position on whether psychological findings are replicable, the principal takeaway from these debates is that we need to fundamentally change the way that we think about interpreting data and results. Low Statistical Power and the Prevalence of QRPs As noted by Nosek, Spies, and Motyl (2012), incentive structures in science prioritize novelty and a publication bias against research that reports null effects is well known. Sterling, Rosenbaum, and Weinkam (1995), examined the publication decisions for 11 major journals and found that that 94% of studies reporting statistical tests in psychology rejected the null hypothesis (H0) casting doubt on the representativeness of those findings. That is, if consumers accept these results at face value, they must conclude that virtually all studies that are published in the professional literature are performed with high power and under conditions in which investigators have formulated true hypotheses. Accordingly, the concept of statistical power (the probability of rejecting H0 when it is false) is critical for understanding the role that publication bias may play in the so-called replication crisis. In particular, troubles arise when one tries to researchers artificially increase the likelihood of rejecting the null hypothesis (i.e., data snooping, dropping cases, obtaining additional measurements until significance is obtained, etc.). 2 Concerns about reproducibility are not limited to psychology and are widespread across scientific disciplines.

5 SELECTIVE REPORTING 5 interpret a significant result from a study with low power. As an example, suppose a researcher reports a statistically significant effect in an intervention study with power at.50. If the same study were repeated with different samples under the same conditions, that effect would be observed in only 50% of the investigations. If all of these studies were submitted and accepted for publication, a pattern of contradictory findings would emerge and school psychologists conversant with this literature would be less likely to regard the intervention as an empirically supported practice. As a result, surveys (e.g., Szucs & Ioaannidis, 2017) indicating that the median estimated power in psychological research is approximately.30 are sobering and suggest that a non-trivial proportion of published studies are likely overestimating true effects or the product of a Type I statistical error. A file drawer problem can occur when the probability of publication becomes dependent on statistical significance. In this type of culture, negative results are selectively reported or in some cases discarded entirely resulting in a remarkable string of successes for psychological theories in published articles (Heene & Ferguson, 2017, p. 43). In response, researchers may resort to using a number of QRPs to increase their chances of attaining significant results that are more likely to be published. These include data snooping (data mining to uncover patterns in data that can be presented as statistically significant), hypothesizing after results are known (HARKing), and p-hacking (exploiting researcher degrees of freedom until a significant p-value is obtained). How prevalent are such practices? In a survey about their involvement in QRPs, the self-admission rate among 2,000 psychologists ranged from 27% to 40% across disciplines (John, Loewenstein, & Prelec, 2012). According to Simmons, Nelson, and Simonsohn (2011), the ubiquity of these practices make it unacceptably easy to publish statistically significant evidence consistent with any

6 SELECTIVE REPORTING 6 hypothesis (p. 1359) resulting in an epistemological confound they termed false positive psychology. At the root of this dilemma is the fact that QRPs increase the maximum falsepositive rate beyond conventional nominal levels (i.e., 5%). This is not a trivial statistical matter. Whereas practices that are universally regarded as unethical such as fabricating data increase Type I error by 100%, some estimates indicate that QRPs can increase the false-positive rate by up to 60% (Schimmack, 2012). Put simply, false positives are costly errors. Once published in the literature, they may be used by practitioners and researchers as evidence to support potentially ineffective practices. Correcting for Selective Reporting The prevalence of QRPs suggest that rather than discarding entire studies, researchers may merely eliminate (file) the subsets of analyses that produce negative findings. This selective reporting is particularly insidious because it upends assumptions about the number of failed attempts needed to produce a false-positive result and invalidates the traditional fail-safe calculations that are used to assess the file-drawer problem in meta-analyses. As a potential safeguard, Simonsohn, Nelson, and Simmons (2014) introduced the p-curve method for detecting effects associated with selective reporting. The purpose of the p-curve is to detect evidential value by distinguishing between sets of significant findings that are likely due to selective reporting. A p-curve is the distribution of statistically significant p values for a set of studies and the shape of the distribution helps to uncover selective reporting versus true effects. Interpreting results is fairly straight forward: right-skewed curves may indicate evidential value (i.e., findings that are likely replicable), flat curves indicate no evidential value, and left-skewed curves may indicate the presence of selective reporting in the literature. Although p-curve analyses are being

7 SELECTIVE REPORTING 7 increasingly used by researchers to defend and raise concern about the quality of research evidence in allied fields, they have yet to be reported in the school psychology literature. Purpose of the Current Investigation Unfortunately, substantive discussions of the replication crisis in the school psychology literature have been limited save a recent commentary by Shaw and D Intino (2017). Thus, the impact and prevalence of selective reporting in school psychology remains largely unknown. To remediate this gap in the literature, the goal of the present study was to examine the evidential value of intervention research published in nine school psychology journals over a two-year period ( ) with a specific emphasis on the potential threat of publication bias using the p-curve method and estimating the replication rate of published research in the field. Examination of these separate, but equally important issues, is important because it can lead to over estimates of effects in the empirical literature. Although a recent article by Villarreal and colleagues (2017) examined the characteristics of intervention research in school psychology journals, the evidential value of the studies was not assessed. It is believed that the results from the present investigation will be instructive for generating a much needed discussion about the quality of research practices in school psychology. Method Data collection and analyses for the present study occurred in several steps. First, the archives of nine school psychology journals (Contemporary School Psychology, International Journal of School and Educational Psychology, Journal of Applied School Psychology, Journal of School Psychology, Psychology in the Schools, School Psychology Forum, School Psychology International, School Psychology Quarterly, and School Psychology Review) were searched for all articles published from As a preliminary screening, the abstracts for the articles

8 SELECTIVE REPORTING 8 (N = 689) were reviewed to identify appropriate intervention articles. We focused specifically on locating articles that systematically evaluated intervention outcomes. That is, survey research examining the preferences and prevalence of practices among practitioners, and studies focused on the process of implementation 3 were excluded from further consideration. Intervention articles were extracted and evaluated in more detail to determine if they met a priori inclusionary criteria for the current study. In order to be included in the analysis, a statistical test result had to be associated with a determinable research hypothesis. In accordance with best practice, studies were not included if they were (a) commentary or editorial articles, (b) literature reviews or research summaries, (c) meta-analyses (to prevent reporting duplicate effects), (d) non-empirical case studies, or (c) reported results not compatible or able to be transformed to be compatible for p-curve analyses (i.e., exact p values). Next, we subjected the statistical effects from individual studies to p-curve analysis using the p-curve app version 4.06 ( to determine the evidential value of the studies as a whole. In a p-curve analysis, the p values from a set of studies are plotted along a curve and then statistically evaluated for potential bias using a binomial sign test. A right-side bias in a curve is considered to be evidence for the presence of a real effect (i.e., replicable) whereas a flat curve or left side bias suggests a questionable effect that may be an artifact of selective reporting and/or QRPs (Simonsohn et al., 2014). Supplementary tables (see Tables X.1-X.3) containing summary information for the statistical effects that were included in the present analyses and the studies that did not meet inclusionary criteria are available in an online supplement ( 3 Unless the purpose of the study was to evaluate the effect of an intervention designed to promote intervention implementation or integrity.

9 SELECTIVE REPORTING 9 Results Descriptive statistics for the initial article search are reported in Table 1. Of the 689 articles that were published across school psychology journals from , 27% (n = 189) were intervention articles where the evaluation of outcomes was a primary objective. Among these studies, 43% (n = 81) disclosed the result of a statistical test(s), among which, 94% reported one or more statistically significant outcomes. The articles were also inspected to estimate the replication rate of intervention research in school psychology. Studies were coded as a replication if replication was noted as an explicit goal of the research within the manuscript. The resulting replication rate among the school psychology journals that were reviewed (~6%) over this time period is relatively consistent with published estimates in other fields (e.g. Makel, Plucker, & Hegarty, 2012). Of the journals examined in the present study, School Psychology Review was the only journal that posted a replication policy on its website. That policy statement indicated that replication studies would be considered for publication as a part of a special section of the journal. Table 2 reports the results of p-curve analyses across the nine journals. Not surprisingly, the power estimates and percentage of statistically significant effects indicating evidential value (i.e., p <.025) varied significantly across the journals. Nevertheless, the Z-test for each p-curve was statistically significant indicating evidential value. The results of the omnibus p-curve analysis across journals is presented graphically in Figure 1 4. Among the 242 total effects that were extracted from 71 different intervention studies (M = 3.40 effects per study), 160 were statistically significant (i.e., p <.05) and 122 (76%) were indicative of evidential value (i.e., p <.025). Visual inspection of the graph in Figure 1 reveals the desired right side bias in the curve 4 Independent p-curve graphs for each journal are provided in the online supplement.

10 SELECTIVE REPORTING 10 resulting in a statistically significant binomial sign test (p <.05, one-tailed) indicating that effects associated with the present set of studies are not likely the result of selective reporting in the literature. To wit, the estimated power associated with the statistically significant effects included in the p-curve is.81 (90% CI [.74,.86]). Discussion Due to a host of high-profile failures to replicate studies in social and experimental psychology, methodologists are in the early stages of examining the credibility of traditional scientific practices in the discipline. Although we contend that school psychology has much to learn from these conversations, the field remains insulated from on-going efforts to improve the state of psychological science (Tackett et al., 2017). In an effort to broaden the replicability conversation, the present study utilized the p-curve method to examine the evidential quality of intervention research published in several school psychology journals in order to determine the degree to which statistically significant findings reflected selective reporting rather than true effects. To our knowledge, this is the first application of p-curve analyses reported in the school psychology literature. The present results are virtually identical to estimates furnished previously by Sterling, Rosenbaum, and Weinkam (1995). We found that the publication decisions in nine peer reviewed school psychology journals appear to disproportionality favor studies that observe effects that have a low value of incorrectly rejecting the null hypothesis. Of the studies that disclosed the results of a statistical test(s), 94% concurrently reported the results of at least one statistically significant finding. Although these results would seem to implicate the presence of publication bias, this hypothesis was not supported by results of the p-curve analyses.

11 SELECTIVE REPORTING 11 With regard to the issue of replication, our examination of 189 intervention articles across the school psychology journals indicated a relatively low percentage of replication studies. While this rate is not considerably different from other disciplines, it suggests that published intervention findings in school psychology are rarely subjected to systematic replication. Although we stipulate that the operational definition employed in the current study is likely a conservative estimate of the actual replication rate given the fact that many intervention studies could be classified as conceptual replications of previous work, the fact that so few authors reported replication as being an explicit goal in the studies suggests that the rate of direct replications, which has been regarded by some as the cornerstone of science (e.g., Coyne, Cook, & Therrien, 2016), in our field is likely quite low. Nevertheless, p-curve results indicated that the overwhelming majority of results reported in the intervention studies that were analyzed from were of evidential value and the estimated power in the overall sample (.81) far surpasses median estimates that have previously been reported in the literature (e.g., Suzcs & Ionnidis, 2017). As a result, it is unlikely that these results are artifacts of selective reporting. Despite these positive findings, it is important to note that the p-curve method focuses only on the effects of selective reporting in the literature and is not useful for identifying other important QRPs such as HARKing, which may be of greater concern to the field given the fact that school psychologists frequently have access to large datasets and samples of participants when conducting studies. In contrast to p-hacking, HARKing and data snooping are almost impossible to identify absent study pre-registration as readers only see the final results in a published article and have no way of knowing how those results were actually produced (Schimmack, 2012). Pre-registration is usually accomplished by posting research plans in an independent registry prior to data collection so that consumers are

12 SELECTIVE REPORTING 12 better able to distinguish exploratory from confirmatory research. Unfortunately, research preregistration in school psychology is virtually nonexistent. Study Limitations In spite of these results, the p-curve method has several limitations. Most notably, it is not possible to include studies that do not report results produced from alternative to exact tests. Accordingly, many single-case designs and studies primarily reporting effect sizes are not able to be included in the online app at the present time. In the current study, 57 studies reporting intervention outcomes were unable to be included in the p-curve analyses because the statistical information necessary for extracting exact p values was not available. Of the aforementioned studies, 82% employed single-case design (SCD). Given the prevalence of SCD research in the school psychology literature, this limitation is particularly notable. Additionally, the p-curve method is most often applied to investigate the quality of focal research programs and, in some cases, the results furnished by specific researchers and teams. Future investigations along these lines would be instructive. In doing so, it is important to keep in mind that selective reporting and other related QRPs are likely not the by-product of malicious intent and that they are a class of practices that are distinct from other behaviors such as data fabrication, which are clearly unethical (Nelson, Simmons, & Simonsohn, 2018). Conclusion The present study has substantive implications for school psychology research and practice. Given the recent high-profile replication failures in psychological science, efforts should be undertaken to encourage and promote a more robust culture of replication in the school psychology literature. Additionally, journal editors and reviewers can help to protect against the insidious effects of QRPs and selective reporting by giving equal consideration to high quality

13 SELECTIVE REPORTING 13 studies that report non-significant results and encouraging authors to pre-register their study protocols in open source forums such as the Open Science Framework (Kratochwill, Levin, & Horner, 2018). On the other hand, the issues raised in the present article suggest that practitioners should guard against overinterpreting the results from isolated intervention studies without considering the broader literature associated with the application of that intervention (i.e., literature that may report negative or in some cases contraindicated effects) and the degree to which those effects have been replicated in the school psychology literature. Additionally, all school psychologists are encouraged to become conversant with the broader replication crisis literature in psychology as well as other allied fields (i.e., evidence-based medicine). We believe these efforts are crucial for advancing our science and furthering efforts to make school psychology incorruptible.

14 SELECTIVE REPORTING 14 References Baumeister, R. F., & Vohs, K. D. (2016). Misguided effort with elusive implications. Perspectives on Psychological Science, 11, doi: / Coyne, M. D., Cook, B. G., & Therrien, W. J. (2016). Recommendations for replication research in special education: A framework for conceptual replications. Remedial and Special Education, 37, doi: / Heene, M., & Ferguson, C. J. (2017). Psychological science s aversion to the null, and why many of the things you think are true, aren t. In S. O. Lilienfeld & I. D. Waldman (Eds.), Psychological science under scrutiny: Recent challenges and proposed solutions (pp ). West Sussex, UK: Wiley. Ioannidis, J. P. A. (2005). Why most published research findings are false. PLos Medicine, 2, e124. doi: /journal.pmed John, L. K., Loewenstein, G., & Prelec, D. (2012). Measuring the prevalence of questionable research practices with incentives for truth telling. Psychological Science, 23, doi: / Kratochwill, T. R., Levin, J. R., & Horner, R. H. (2018). Negative results: Conceptual and methodological dimensions in single-case intervention research. Remedial and Special Education, 39, doi: / Lilienfeld, S. O., Ammirati, R., & David, M. (2012). Distinguishing between science pseudoscience in school psychology: Science and scientific thinking as safeguards against human error. Journal of School Psychology, 50, doi: /j.jsp Makel, M. C., Plucker, J. A., & Hegarty, B. (2012). Replications in psychology research: How

15 SELECTIVE REPORTING 15 often do they really occur? Perspectives on Psychological Science, 7, doi: / Nelson, L. D., Simmons, J., & Simonsohn, U. (2018). Psychology s renaissance. Annual Review of Psychology, 69, doi: /annurev- psych Nosek, B. A., Spies, J. R., & Motyl, M. (2012). Scientific Utopia II: Restructuring incentives and practices to promote truth over publishability. Perspectives on Psychological Science, 7, doi: / Schimmack, U. (2012). The ironic effect of significant results on the credibility of multiple-study articles. Psychological Methods, 17, Doi: /a Shaw, S. R., & D Intino, J. (2017). Evidence-based practice and the reproducibility crisis in psychology. Communique, 45 (5), Retrieved from Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychological Science, 22, doi: / Simonsohn, U., Nelson, L. D., & Simmons, J. P. (2014). P-curve: A key to the file-drawer. Journal of Experimental Psychology, 143, doi: /a Sterling, T. D., Rosenbaum, W. L., & Weinkam, J. J. (1995). Publication decisions revisited: The effect of the outcome of statistical tests on the decision to publish and vice versa. The American Statistician, 49, doi: / Szucs, D., & Ioannidis, J. P. A. (2017). Empirical assessment of published effect sizes and power in the recent cognitive neuroscience and psychology literature. PLos Biology, 3, e doi: /journal.pbio Tackett, J. L., Lilienfeld, S. O., Patrick, C. J., Johnson, S. L., & Shrout, P. E. (2017). It s time

16 SELECTIVE REPORTING 16 to broaden the replicability conversation: Thought for and from clinical psychological science. Perspectives on Psychological Science, 12, doi: / O Donohue, W., & Henderson, D. (1999). Epistemic and ethical duties in clinical decisionmaking. Behaviour Change, 16, doi: /bech Open Science Collaboration. (2015). Estimating the reproducibility of psychological science. Science, 349, 943. doi: /science.aac4716 Villarreal, V., Castro, M. J., Umana, I., & Sullivan, J. R. (2017). Characteristics of intervention research in school psychology journals: Psychology in the Schools, 54, doi: /pits Table 1

17 SELECTIVE REPORTING 17 Outcome of Tests of Significance and Prevalence of Replication Studies for Intervention Research Published in Nine School Psychology Journals from Journal k Intervention Articles a Articles Reporting Statistical Tests Statistically Significant b Replication Studies c CSP (32%) 9 (38%) 8 (89%) 0 (0%) IJSEP (25%) 1 (6%) 1 (100%) 1 (6%) JASP (39%) 5 (33%) 4 (80%) 1 (7%) JSP (27%) 18 (78%) 16 (88%) 1 (4%) PITS (30%) 16 (37%) 14 (88%) 2 (5%) SPF (37%) 6 (32%) 5 (83%) 2 (11%) SPI (18%) 8 (57%) 8 (100%) 0 (0%) SPQ (19%) 7 (44%) 7 (100%) 3 (19%) SPR (37%) 11 (59%) 11 (100%) 2 (11%) Total (27%) 81 (43%) 76 (94%) 12 (6%) Note. k = number of manuscripts published ( ). CSP = Contemporary School Psychology; IJESP: International Journal of School and Educational Psychology; JASP = Journal of Applied School Psychology; JSP = Journal of School Psychology; PITS = Psychology in the Schools; School Psychology Forum; SPI = School Psychology International; SPQ = School Psychology Quarterly, SPR = School Psychology Review. a Number of articles reviewed in the present study (i.e., intervention outcomes was a primary objective). b Articles reporting a statistically significant (i.e., p <.05) outcome(s). c Studies were coded as replication attempt if it was noted as an explicit goal of the research project. Table 2 Results of P-Curve Analyses of Intervention Research Published in School Psychology Journals from

18 SELECTIVE REPORTING 18 Evidential Value Based on P-Curve Evidential Value Based on Power Journal % Evidential Value a Z-Test of Half P- Curve Power Estimate [90% CI] 1. CSP 56% -2.71*.46** [.07,.82] Yes Yes 2. IJSEP Insufficient Data to Calculate Independent P-Curve 3. JASP 64% -4.79*.64** [.31,.86] Yes Yes 4. JSP 69% -5.72*.55** [.34,.73] Yes Yes 5. PITS 83% -3.49*.55** [.34,.73] Yes Yes 6. SPF 87% -5.25*.91** [.73,.97] Yes Yes 7. SPI 83% -9.44*.92** [.82,.97] Yes Yes 8. SPQ 84% *.99** [.99,.99] Yes Yes 9. SPR 77% -4.96*.56** [.31,.76] Yes Yes Total 76% *.81** [.74,.86] Yes Yes Note. CSP = Contemporary School Psychology; IJESP: International Journal of School and Educational Psychology; JASP = Journal of Applied School Psychology; JSP = Journal of School Psychology; PITS = Psychology in the Schools; SPF = School Psychology Forum; SPI = School Psychology International; SPQ = School Psychology Quarterly, SPR = School Psychology Review. Continuous test of the half p-curve based on the Stouffer method. As per Simonsohn, Simmons, & Nelson (2015), half curve values that p <.05 indicate the absence of right-side bias and thus evidential value. Conversely, evidential value is absent if the power test is p <.05 for the half-test. a Outcomes p <.025. * p <.05. ** p >.05.

19 SELECTIVE REPORTING 19 Figure 1. Results from a p-curve analysis examining the evidential value of intervention research published in nine school psychology journals from

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior 1 Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior Gregory Francis Department of Psychological Sciences Purdue University gfrancis@purdue.edu

More information

Running head: Estimating Replicability 1. Z-Curve: A Method for the Estimating Replicability Based on Test Statistics in Original Studies

Running head: Estimating Replicability 1. Z-Curve: A Method for the Estimating Replicability Based on Test Statistics in Original Studies Running head: Estimating Replicability 1 Z-Curve: A Method for the Estimating Replicability Based on Test Statistics in Original Studies Ulrich Schimmack and Jerry Brunner University of Toronto Mississauga

More information

Guidelines for reviewers

Guidelines for reviewers Guidelines for reviewers Registered Reports are a form of empirical article in which the methods and proposed analyses are pre-registered and reviewed prior to research being conducted. This format of

More information

Power Posing: P-Curving the Evidence

Power Posing: P-Curving the Evidence 658563PSSXXX10.1177/0956797616658563Simmons, SimonsohnP-Curving the Power-Posing Literature research-article2017 Commentary Power Posing: P-Curving the Evidence Joseph P. Simmons and Uri Simonsohn University

More information

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 Quentin Frederik Gronau 1, Monique Duizer 1, Marjan Bakker 2, & Eric-Jan Wagenmakers

More information

Registered Reports - Guidelines for reviewers

Registered Reports - Guidelines for reviewers Registered Reports - Guidelines for reviewers Registered Reports are a form of publication of empirical research in which the methods and proposed analyses are pre-registered and reviewed prior to the

More information

Christopher Chambers Cardiff University. Rob McIntosh University of Edinburgh. Zoltan Dienes University of Sussex

Christopher Chambers Cardiff University. Rob McIntosh University of Edinburgh. Zoltan Dienes University of Sussex 1 Christopher Chambers Cardiff University Zoltan Dienes University of Sussex Rob McIntosh University of Edinburgh Klaus Willmes-von Hinckeldey University of Aachen 2 Problems with everyday research practices

More information

Registered Reports: Peer reviewed study pre-registration. Implications for reporting of human clinical trials

Registered Reports: Peer reviewed study pre-registration. Implications for reporting of human clinical trials Registered Reports: Peer reviewed study pre-registration Implications for reporting of human clinical trials Chris Chambers Professor of Cognitive Neuroscience Cardiff University Brain Research Imaging

More information

Underreporting in Psychology Experiments: Evidence From a Study Registry

Underreporting in Psychology Experiments: Evidence From a Study Registry Article Underreporting in Psychology Experiments: Evidence From a Study Registry Social Psychological and Personality Science 2016, Vol. 7(1) 8-12 ª The Author(s) 2015 Reprints and permission: sagepub.com/journalspermissions.nav

More information

CJSP: On Reaffirming a Canadian School Psychology

CJSP: On Reaffirming a Canadian School Psychology 748212CJSXXX10.1177/0829573517748212Canadian Journal of School PsychologyEditorial editorial2017 Editorial CJSP: On Reaffirming a Canadian School Psychology Canadian Journal of School Psychology 1 5 The

More information

Considerations on determining sample size to give desired power: (pp. 18, 25, 33)

Considerations on determining sample size to give desired power: (pp. 18, 25, 33) 1 APPENDIX TO DAY 3 Considerations on determining sample size to give desired power: (pp. 18, 25, 33) The difference used in calculating sample size (i.e., the specific alternative used in calculating

More information

Less Story and more Reliability in cognitive neuroscience. David E. Huber, Kevin W. Potter, and Lucas D. Huszar. University of Massachusetts, Amherst

Less Story and more Reliability in cognitive neuroscience. David E. Huber, Kevin W. Potter, and Lucas D. Huszar. University of Massachusetts, Amherst Less Story and more Reliability in cognitive neuroscience David E. Huber, Kevin W. Potter, and Lucas D. Huszar University of Massachusetts, Amherst Contact Office: 425 Tobin Hall Office Phone: (413) 545-1559

More information

The Replication Crisis Explained N. WAGNER & R. TERNES MI AIR NOV. 2018

The Replication Crisis Explained N. WAGNER & R. TERNES MI AIR NOV. 2018 The Replication Crisis Explained N. WAGNER & R. TERNES MI AIR NOV. 2018 Replication as the Cornerstone of Science We all know that the scientific method relies on careful observation, prediction & hypothesis

More information

INADEQUACIES OF SIGNIFICANCE TESTS IN

INADEQUACIES OF SIGNIFICANCE TESTS IN INADEQUACIES OF SIGNIFICANCE TESTS IN EDUCATIONAL RESEARCH M. S. Lalithamma Masoomeh Khosravi Tests of statistical significance are a common tool of quantitative research. The goal of these tests is to

More information

Experimental Psychology

Experimental Psychology Title Experimental Psychology Type Individual Document Map Authors Aristea Theodoropoulos, Patricia Sikorski Subject Social Studies Course None Selected Grade(s) 11, 12 Location Roxbury High School Curriculum

More information

RISK AS AN EXPLANATORY FACTOR FOR RESEARCHERS INFERENTIAL INTERPRETATIONS

RISK AS AN EXPLANATORY FACTOR FOR RESEARCHERS INFERENTIAL INTERPRETATIONS 13th International Congress on Mathematical Education Hamburg, 24-31 July 2016 RISK AS AN EXPLANATORY FACTOR FOR RESEARCHERS INFERENTIAL INTERPRETATIONS Rink Hoekstra University of Groningen Logical reasoning

More information

ETHICS OF PUBLICATION

ETHICS OF PUBLICATION ETHICS OF PUBLICATION Ethics in Psychological Science Beyond how participants are treated, its of huge importance to be honest in reporting data Data most often comes from unverifiable sources (e.g., anonymous

More information

ARE RESEARCH FINDINGS RELIABLE? SAINT KIZITO OMALA MAKERERE UNIVERSITY

ARE RESEARCH FINDINGS RELIABLE? SAINT KIZITO OMALA MAKERERE UNIVERSITY ARE RESEARCH FINDINGS RELIABLE? SAINT KIZITO OMALA MAKERERE UNIVERSITY 2 0 1 6 E A S T A F R I C A I M P A C T E V A L U AT I O N W O R K S H O P A N D E V I D E N C E S U M M I T RELIABLE RESEARCH When

More information

Principles of publishing

Principles of publishing Principles of publishing Issues of authorship, duplicate publication and plagiarism in scientific journal papers can cause considerable conflict among members of research teams and embarrassment for both

More information

Measuring and Assessing Study Quality

Measuring and Assessing Study Quality Measuring and Assessing Study Quality Jeff Valentine, PhD Co-Chair, Campbell Collaboration Training Group & Associate Professor, College of Education and Human Development, University of Louisville Why

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

USE AND MISUSE OF MIXED MODEL ANALYSIS VARIANCE IN ECOLOGICAL STUDIES1

USE AND MISUSE OF MIXED MODEL ANALYSIS VARIANCE IN ECOLOGICAL STUDIES1 Ecology, 75(3), 1994, pp. 717-722 c) 1994 by the Ecological Society of America USE AND MISUSE OF MIXED MODEL ANALYSIS VARIANCE IN ECOLOGICAL STUDIES1 OF CYNTHIA C. BENNINGTON Department of Biology, West

More information

Replications of Forecasting Research

Replications of Forecasting Research University of Pennsylvania ScholarlyCommons Marketing Papers Wharton Faculty Research 2010 Replications of Forecasting Research J. Scott Armstrong University of Pennsylvania Follow this and additional

More information

What is an Impact Factor? How variable is the impact factor? Time after publication (Years) Figure 1. Generalized Citation Curve

What is an Impact Factor? How variable is the impact factor? Time after publication (Years) Figure 1. Generalized Citation Curve Reprinted with pemission from Perspectives in Puplishing, No 1, Oct 2000, http://www.elsevier.com/framework_editors/pdfs/perspectives 1.pdf What is an Impact Factor? The impact factor is only one of three

More information

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS Chapter Objectives: Understand Null Hypothesis Significance Testing (NHST) Understand statistical significance and

More information

Research Practices that can Prevent an Inflation of False-positive Rates. Kou Murayama University of California, Los Angeles

Research Practices that can Prevent an Inflation of False-positive Rates. Kou Murayama University of California, Los Angeles Running head: PREVENTION OF FALSE-POSITIVES Research Practices that can Prevent an Inflation of False-positive Rates Kou Murayama University of California, Los Angeles Reinhard Pekrun University of Munich

More information

Registered reports as a solu/on to publica/on bias and p-hacking

Registered reports as a solu/on to publica/on bias and p-hacking Registered reports as a solu/on to publica/on bias and p-hacking Chris Chambers Cardiff University Brain Research Imaging Centre (CUBRIC) School of Psychology, Cardiff University Email: chambersc1@cardiff.ac.uk

More information

A Preliminary P-Curve Meta-Analysis of Learned Categorical Perception Research

A Preliminary P-Curve Meta-Analysis of Learned Categorical Perception Research A Preliminary P-Curve Meta-Analysis of Learned Categorical Perception Research Janet Andrews (andrewsj@vassar.edu) Joshua de Leeuw (jdeleeuw@vassar.edu) Calais Larson (calarson@vassar.edu) Xiaoqing Xu

More information

STI 2018 Conference Proceedings

STI 2018 Conference Proceedings STI 2018 Conference Proceedings Proceedings of the 23rd International Conference on Science and Technology Indicators All papers published in this conference proceedings have been peer reviewed through

More information

Discovering Evidence-Based Practice: What Does Research Tell Us? John D. Molteni, Ph.D. BCBA-D Saint Joseph College

Discovering Evidence-Based Practice: What Does Research Tell Us? John D. Molteni, Ph.D. BCBA-D Saint Joseph College Discovering Evidence-Based Practice: What Does Research Tell Us? John D. Molteni, Ph.D. BCBA-D Saint Joseph College Outline for Today Discuss how one should read and understand research Discuss how one

More information

Biostatistics Primer

Biostatistics Primer BIOSTATISTICS FOR CLINICIANS Biostatistics Primer What a Clinician Ought to Know: Subgroup Analyses Helen Barraclough, MSc,* and Ramaswamy Govindan, MD Abstract: Large randomized phase III prospective

More information

How to use research to lead your clinical decision

How to use research to lead your clinical decision ] 6 How to use research to lead your clinical decision ASK ACQUIRE ANALYZE ANSWER APPLY ASSESS making.] Phil Page PhD, PT, ATC, CSCS, FACSM 1) ASK What is your question? Be very specific in what you want

More information

Fixed Effect Combining

Fixed Effect Combining Meta-Analysis Workshop (part 2) Michael LaValley December 12 th 2014 Villanova University Fixed Effect Combining Each study i provides an effect size estimate d i of the population value For the inverse

More information

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

CHAMP: CHecklist for the Appraisal of Moderators and Predictors CHAMP - Page 1 of 13 CHAMP: CHecklist for the Appraisal of Moderators and Predictors About the checklist In this document, a CHecklist for the Appraisal of Moderators and Predictors (CHAMP) is presented.

More information

THE UNIVERSITY OF WESTERN ONTARIO DEPARTMENT OF PHILOSOPHY Graduate Course Syllabus 2018-2019 Philosophy 9213B Philosophy of Scientific Experimentation Winter Term 2019 Thurs. 2:30 p.m.-5:30 p.m. STVH-1145

More information

Lesson 11.1: The Alpha Value

Lesson 11.1: The Alpha Value Hypothesis Testing Lesson 11.1: The Alpha Value The alpha value is the degree of risk we are willing to take when making a decision. The alpha value, often abbreviated using the Greek letter α, is sometimes

More information

Overview. Survey Methods & Design in Psychology. Readings. Significance Testing. Significance Testing. The Logic of Significance Testing

Overview. Survey Methods & Design in Psychology. Readings. Significance Testing. Significance Testing. The Logic of Significance Testing Survey Methods & Design in Psychology Lecture 11 (2007) Significance Testing, Power, Effect Sizes, Confidence Intervals, Publication Bias, & Scientific Integrity Overview Significance testing Inferential

More information

HOW CAN RESEARCH FINDINGS BECOME MORE RELIABLE? ANTHONY MVEYANGE WORLD BANK /EAST FELLOW

HOW CAN RESEARCH FINDINGS BECOME MORE RELIABLE? ANTHONY MVEYANGE WORLD BANK /EAST FELLOW HOW CAN RESEARCH FINDINGS BECOME MORE RELIABLE? ANTHONY MVEYANGE WORLD BANK /EAST FELLOW 2 0 1 6 E A S T A F R I C A I M PA C T E VA L U AT I O N W O R K S H O P A N D E V I D E N C E S U M M I T META-RESEARCH

More information

Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper.

Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper. Throwing the Baby Out With the Bathwater? The What Works Clearinghouse Criteria for Group Equivalence* A NIFDI White Paper December 13, 2013 Jean Stockard Professor Emerita, University of Oregon Director

More information

Promoting Transparent Reporting of Conflicts of Interests. and Statistical Analyses at the Journal of Sex Research

Promoting Transparent Reporting of Conflicts of Interests. and Statistical Analyses at the Journal of Sex Research 1 Promoting Transparent Reporting of Conflicts of Interests and Statistical Analyses at the Journal of Sex Research John K. Sakaluk, Statistical Consultant, Journal of Sex Research Department of Psychology

More information

It s Time to Broaden the Replicability Conversation: Thoughts for and from Clinical Psychological Science

It s Time to Broaden the Replicability Conversation: Thoughts for and from Clinical Psychological Science It s Time to Broaden the Replicability Conversation: Thoughts for and from Clinical Psychological Science Clinical Science Replicability 1 Jennifer L. Tackett 1, Scott O. Lilienfeld 2, Christopher J. Patrick

More information

COMMITTEE FOR PROPRIETARY MEDICINAL PRODUCTS (CPMP) POINTS TO CONSIDER ON MISSING DATA

COMMITTEE FOR PROPRIETARY MEDICINAL PRODUCTS (CPMP) POINTS TO CONSIDER ON MISSING DATA The European Agency for the Evaluation of Medicinal Products Evaluation of Medicines for Human Use London, 15 November 2001 CPMP/EWP/1776/99 COMMITTEE FOR PROPRIETARY MEDICINAL PRODUCTS (CPMP) POINTS TO

More information

Registered Reports guidelines for reviewers and authors. Guidelines for reviewers

Registered Reports guidelines for reviewers and authors. Guidelines for reviewers Registered Reports guidelines for reviewers and authors Guidelines for reviewers Registered Reports are a form of empirical article offered at NFS Journal in which the methods and proposed analyses are

More information

Regression Discontinuity Analysis

Regression Discontinuity Analysis Regression Discontinuity Analysis A researcher wants to determine whether tutoring underachieving middle school students improves their math grades. Another wonders whether providing financial aid to low-income

More information

Errors and biases in Structural Equa2on Modeling. Jelte M. Wicherts

Errors and biases in Structural Equa2on Modeling. Jelte M. Wicherts Errors and biases in Structural Equa2on Modeling Jelte M. Wicherts 1 De empirical cycle Observe Evaluate Theorize Test Predict 2 Grey zone map Remove outliers (Z > 2 ) P>.05? P>.05 P

More information

Is it possible to give a philosophical definition of sexual desire?

Is it possible to give a philosophical definition of sexual desire? Issue 1 Spring 2016 Undergraduate Journal of Philosophy Is it possible to give a philosophical definition of sexual desire? William Morgan - The University of Sheffield pp. 47-58 For details of submission

More information

Module 14: Missing Data Concepts

Module 14: Missing Data Concepts Module 14: Missing Data Concepts Jonathan Bartlett & James Carpenter London School of Hygiene & Tropical Medicine Supported by ESRC grant RES 189-25-0103 and MRC grant G0900724 Pre-requisites Module 3

More information

No Lie! Caveats and Cautions in the use of Polygraphy. John JB Allen, Ph.D. University of Arizona

No Lie! Caveats and Cautions in the use of Polygraphy. John JB Allen, Ph.D. University of Arizona No Lie! Caveats and Cautions in the use of Polygraphy John JB Allen, Ph.D. University of Arizona No Lie! Caveats and Cautions in the use of Polygraphy John JB Allen, Ph.D. University of Arizona jallen.faculty.arizona.edu/polygraph

More information

Is Leisure Theory Needed For Leisure Studies?

Is Leisure Theory Needed For Leisure Studies? Journal of Leisure Research Copyright 2000 2000, Vol. 32, No. 1, pp. 138-142 National Recreation and Park Association Is Leisure Theory Needed For Leisure Studies? KEYWORDS: Mark S. Searle College of Human

More information

9 research designs likely for PSYC 2100

9 research designs likely for PSYC 2100 9 research designs likely for PSYC 2100 1) 1 factor, 2 levels, 1 group (one group gets both treatment levels) related samples t-test (compare means of 2 levels only) 2) 1 factor, 2 levels, 2 groups (one

More information

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all.

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all. 1. RANDOMISED CONTROLLED TRIALS (Treatment studies) (Relevant JAMA User s Guides, Numbers IIA & B: references (3,4) Introduction: The most valid study design for assessing the effectiveness (both the benefits

More information

TRANSPARENCY AND OPENNESS IN SCIENCE

TRANSPARENCY AND OPENNESS IN SCIENCE TRANSPARENCY AND OPENNESS IN SCIENCE Marjan Bakker 24 May 2016 A CRISIS A CRISIS IN PSYCHOLOGY A CRISIS FRAUD 58 PAPERS Diederik Stapel See www.retractionwatch.com for the current count A CRISIS FEELING

More information

Supplementary Material for

Supplementary Material for www.sciencemag.org/content/349/6251/aac4716/suppl/dc1 This PDF file includes: Supplementary Material for Estimating the reproducibility of psychological science Materials and Methods Figs. S1 to S7 Tables

More information

Skipping the Replication Crisis in Visualization: Threats to Study Validity and How to Address Them

Skipping the Replication Crisis in Visualization: Threats to Study Validity and How to Address Them Skipping the Replication Crisis in Visualization: Threats to Study Validity and How to Address Them Robert Kosara * Tableau Research Steve Haroz Sorbonne Université ABSTRACT Replications are rare in visualization,

More information

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA Data Analysis: Describing Data CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA In the analysis process, the researcher tries to evaluate the data collected both from written documents and from other sources such

More information

The Impact of Relative Standards on the Propensity to Disclose. Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX

The Impact of Relative Standards on the Propensity to Disclose. Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX The Impact of Relative Standards on the Propensity to Disclose Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX 2 Web Appendix A: Panel data estimation approach As noted in the main

More information

Fixing the replicability crisis in science. Jelte M. Wicherts

Fixing the replicability crisis in science. Jelte M. Wicherts Fixing the replicability crisis in science Jelte M. Wicherts 1 Supporting responsible research practices Diminishing biases in research Lowering false positive ratios Fixing the replicability crisis in

More information

No Lie! Caveats and Cautions in the use of Polygraphy. No Lie! Caveats and Cautions in the use of Polygraphy. Roadmap. The Polygraph Test

No Lie! Caveats and Cautions in the use of Polygraphy. No Lie! Caveats and Cautions in the use of Polygraphy. Roadmap. The Polygraph Test Lie! Caveats and Cautions in the use of y John JB Allen, Ph.D. University of Arizona Lie! Caveats and Cautions in the use of y John JB Allen, Ph.D. University of Arizona The Test does not detect lies The

More information

Statistical Literacy in the Introductory Psychology Course

Statistical Literacy in the Introductory Psychology Course Statistical Literacy Taskforce 2012, Undergraduate Learning Goals 1 Statistical Literacy in the Introductory Psychology Course Society for the Teaching of Psychology Statistical Literacy Taskforce 2012

More information

Special issue article: Methods and statistics in social psychology: Refinements and new developments

Special issue article: Methods and statistics in social psychology: Refinements and new developments European Journal of Social Psychology, Eur. J. Soc. Psychol. 44, 701 710 (2014) Published online in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/ejsp.2023 Special issue article: Methods and

More information

Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection

Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection Author's response to reviews Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection Authors: Jestinah M Mahachie John

More information

Timing Your Research Career & Publishing Addiction Medicine

Timing Your Research Career & Publishing Addiction Medicine Timing Your Research Career & Publishing Addiction Medicine Jeffrey H. Samet, MD, MA, MPH Chief, General Internal Medicine John Noble MD Professor in General Internal Medicine & Professor of Public Health

More information

1 The conceptual underpinnings of statistical power

1 The conceptual underpinnings of statistical power 1 The conceptual underpinnings of statistical power The importance of statistical power As currently practiced in the social and health sciences, inferential statistics rest solidly upon two pillars: statistical

More information

EQUATOR Network: promises and results of reporting guidelines

EQUATOR Network: promises and results of reporting guidelines EQUATOR Network: promises and results of reporting guidelines Doug Altman The EQUATOR Network Centre for Statistics in Medicine, Oxford, UK Key principles of research publications A published research

More information

The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews

The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews David Moher 1, Alessandro Liberati 2, Douglas G Altman 3, Jennifer Tetzlaff 1 for the QUOROM Group

More information

Introduction to Clinical Trials - Day 1 Session 2 - Screening Studies

Introduction to Clinical Trials - Day 1 Session 2 - Screening Studies Introduction to Clinical Trials - Day 1 Session 2 - Screening Studies Presented July 23, 2018 Susanne J. May Department of Biostatistics University of Washington Daniel L. Gillen Department of Statistics

More information

Revised Cochrane risk of bias tool for randomized trials (RoB 2.0) Additional considerations for cross-over trials

Revised Cochrane risk of bias tool for randomized trials (RoB 2.0) Additional considerations for cross-over trials Revised Cochrane risk of bias tool for randomized trials (RoB 2.0) Additional considerations for cross-over trials Edited by Julian PT Higgins on behalf of the RoB 2.0 working group on cross-over trials

More information

False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant

False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant General Article False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant Psychological Science XX(X) 1 8 The Author(s) 2011 Reprints

More information

15 May 2017 Exposure Draft. Response Due Date 23 June Exposure Draft

15 May 2017 Exposure Draft. Response Due Date 23 June Exposure Draft 15 May 2017 Exposure Draft Response Due Date 23 June 2017 Exposure Draft Proposed Application Material Relating to: (a) Professional Skepticism Linkage with the Fundamental Principles; and (b) Professional

More information

POST GRADUATE DIPLOMA IN BIOETHICS (PGDBE) Term-End Examination June, 2016 MHS-014 : RESEARCH METHODOLOGY

POST GRADUATE DIPLOMA IN BIOETHICS (PGDBE) Term-End Examination June, 2016 MHS-014 : RESEARCH METHODOLOGY No. of Printed Pages : 12 MHS-014 POST GRADUATE DIPLOMA IN BIOETHICS (PGDBE) Term-End Examination June, 2016 MHS-014 : RESEARCH METHODOLOGY Time : 2 hours Maximum Marks : 70 PART A Attempt all questions.

More information

Fame: I m Skeptical. Fernanda Ferreira. Department of Psychology. University of California, Davis.

Fame: I m Skeptical. Fernanda Ferreira. Department of Psychology. University of California, Davis. Fame: I m Skeptical Fernanda Ferreira Department of Psychology University of California, Davis fferreira@ucdavis.edu February 20, 2017 Word count: 1492 1 Abstract Fame is often deserved, emerging from

More information

Research and science: Qualitative methods

Research and science: Qualitative methods Research and science: Qualitative methods Urban Bilstrup (E327) Urban.Bilstrup@hh.se 140922 2 INTRODUCTION TO RESEARCH AND SCIENTIFIC METHODS Outline Definitions Problem formulation? Aim and goal operational

More information

EPSE 594: Meta-Analysis: Quantitative Research Synthesis

EPSE 594: Meta-Analysis: Quantitative Research Synthesis EPSE 594: Meta-Analysis: Quantitative Research Synthesis Ed Kroc University of British Columbia ed.kroc@ubc.ca March 14, 2019 Ed Kroc (UBC) EPSE 594 March 14, 2019 1 / 39 Last Time Synthesizing multiple

More information

Denny Borsboom Jaap van Heerden Gideon J. Mellenbergh

Denny Borsboom Jaap van Heerden Gideon J. Mellenbergh Validity and Truth Denny Borsboom Jaap van Heerden Gideon J. Mellenbergh Department of Psychology, University of Amsterdam ml borsboom.d@macmail.psy.uva.nl Summary. This paper analyzes the semantics of

More information

Controlled Trials. Spyros Kitsiou, PhD

Controlled Trials. Spyros Kitsiou, PhD Assessing Risk of Bias in Randomized Controlled Trials Spyros Kitsiou, PhD Assistant Professor Department of Biomedical and Health Information Sciences College of Applied Health Sciences University of

More information

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Cochrane Pregnancy and Childbirth Group Methodological Guidelines Cochrane Pregnancy and Childbirth Group Methodological Guidelines [Prepared by Simon Gates: July 2009, updated July 2012] These guidelines are intended to aid quality and consistency across the reviews

More information

Kidane Tesfu Habtemariam, MASTAT, Principle of Stat Data Analysis Project work

Kidane Tesfu Habtemariam, MASTAT, Principle of Stat Data Analysis Project work 1 1. INTRODUCTION Food label tells the extent of calories contained in the food package. The number tells you the amount of energy in the food. People pay attention to calories because if you eat more

More information

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology ISC- GRADE XI HUMANITIES (2018-19) PSYCHOLOGY Chapter 2- Methods of Psychology OUTLINE OF THE CHAPTER (i) Scientific Methods in Psychology -observation, case study, surveys, psychological tests, experimentation

More information

Power, Effect Sizes, Confidence Intervals, & Academic Integrity. Significance testing

Power, Effect Sizes, Confidence Intervals, & Academic Integrity. Significance testing Power, Effect Sizes, Confidence Intervals, & Academic Integrity Lecture 9 Survey Research & Design in Psychology James Neill, 2018 Creative Commons Attribution 4.0 Overview 1Significance testing 2Inferential

More information

Generalization and Theory-Building in Software Engineering Research

Generalization and Theory-Building in Software Engineering Research Generalization and Theory-Building in Software Engineering Research Magne Jørgensen, Dag Sjøberg Simula Research Laboratory {magne.jorgensen, dagsj}@simula.no Abstract The main purpose of this paper is

More information

Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania

Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania Meta-Analysis Workshop David B. Wilson, PhD September 16, 2016 George Mason University Department of

More information

An introduction to power and sample size estimation

An introduction to power and sample size estimation 453 STATISTICS An introduction to power and sample size estimation S R Jones, S Carley, M Harrison... Emerg Med J 2003;20:453 458 The importance of power and sample size estimation for study design and

More information

In many healthcare situations, it is common to find

In many healthcare situations, it is common to find Interpreting and Using Meta-Analysis in Clinical Practice Cheryl Holly Jason T. Slyer Systematic reviews, which can include a meta-analysis, are considered the gold standard for determination of best practice.

More information

The idea of an essentially contested concept is incoherent.

The idea of an essentially contested concept is incoherent. Daniel Alexander Harris 1 The idea of an essentially contested concept is incoherent. Daniel Alexander Harris (2015) Daniel Alexander Harris 2 This essay will demonstrate the idea of an essentially contested

More information

Chapter 12. The One- Sample

Chapter 12. The One- Sample Chapter 12 The One- Sample z-test Objective We are going to learn to make decisions about a population parameter based on sample information. Lesson 12.1. Testing a Two- Tailed Hypothesis Example 1: Let's

More information

SOCQ121/BIOQ121. Session 8. Methods and Methodology. Department of Social Science. endeavour.edu.au

SOCQ121/BIOQ121. Session 8. Methods and Methodology. Department of Social Science. endeavour.edu.au SOCQ121/BIOQ121 Session 8 Methods and Methodology Department of Social Science endeavour.edu.au Endeavour College of Natural Health endeavour.edu.au 1 Review Progress report to the person next to you on

More information

Statistical reports Regression, 2010

Statistical reports Regression, 2010 Statistical reports Regression, 2010 Niels Richard Hansen June 10, 2010 This document gives some guidelines on how to write a report on a statistical analysis. The document is organized into sections that

More information

Kepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce

Kepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce Stats 95 Kepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce pure gold knowing it was an element, though

More information

Outline. Bioethics in Research and Publication. What is ethics? Where do we learn ethics? 6/19/2015

Outline. Bioethics in Research and Publication. What is ethics? Where do we learn ethics? 6/19/2015 Outline Bioethics in Research and Publication Gertrude Night, PhD Rwanda Agriculture Board What is ethics? History. The purpose of training in research ethics Some ethical principles addressed by research

More information

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings Authors and Affiliations: Peter M. Steiner, University of Wisconsin-Madison

More information

PsychDisclosure.org: Grassroots Support for Reforming Reporting Standards in Psychology

PsychDisclosure.org: Grassroots Support for Reforming Reporting Standards in Psychology 491437PPSXXX10.1177/1745691613491437LeBel et al.reforming Reporting Standards in Psychology research-article2013 PsychDisclosure.org: Grassroots Support for Reforming Reporting Standards in Psychology

More information

AOTA S EVIDENCE EXCHANGE CRITICALLY APPRAISED PAPER (CAP) GUIDELINES Annual AOTA Conference Poster Submissions Critically Appraised Papers (CAPs) are

AOTA S EVIDENCE EXCHANGE CRITICALLY APPRAISED PAPER (CAP) GUIDELINES Annual AOTA Conference Poster Submissions Critically Appraised Papers (CAPs) are AOTA S EVIDENCE EXCHANGE CRITICALLY APPRAISED PAPER (CAP) GUIDELINES Annual AOTA Conference Poster Submissions Critically Appraised Papers (CAPs) are at-a-glance summaries of the methods, findings and

More information

Comparing Direct and Indirect Measures of Just Rewards: What Have We Learned?

Comparing Direct and Indirect Measures of Just Rewards: What Have We Learned? Comparing Direct and Indirect Measures of Just Rewards: What Have We Learned? BARRY MARKOVSKY University of South Carolina KIMMO ERIKSSON Mälardalen University We appreciate the opportunity to comment

More information

This article, the last in a 4-part series on philosophical problems

This article, the last in a 4-part series on philosophical problems GUEST ARTICLE Philosophical Issues in Medicine and Psychiatry, Part IV James Lake, MD This article, the last in a 4-part series on philosophical problems in conventional and integrative medicine, focuses

More information

Psychology 205, Revelle, Fall 2014 Research Methods in Psychology Mid-Term. Name:

Psychology 205, Revelle, Fall 2014 Research Methods in Psychology Mid-Term. Name: Name: 1. (2 points) What is the primary advantage of using the median instead of the mean as a measure of central tendency? It is less affected by outliers. 2. (2 points) Why is counterbalancing important

More information

C2 Training: August 2010

C2 Training: August 2010 C2 Training: August 2010 Introduction to meta-analysis The Campbell Collaboration www.campbellcollaboration.org Pooled effect sizes Average across studies Calculated using inverse variance weights Studies

More information

An evidence rating scale for New Zealand

An evidence rating scale for New Zealand Social Policy Evaluation and Research Unit An evidence rating scale for New Zealand Understanding the effectiveness of interventions in the social sector Using Evidence for Impact MARCH 2017 About Superu

More information

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity Measurement & Variables - Initial step is to conceptualize and clarify the concepts embedded in a hypothesis or research question with

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information