was also my mentor, teacher, colleague, and friend. It is tempting to review John Horn s main contributions to the field of intelligence by

Similar documents
Extensions of the paper

Paul Irwing, Manchester Business School

Reliability, validity, and all that jazz

Identifying or Verifying the Number of Factors to Extract using Very Simple Structure.

Robert Pearson Daniel Mundfrom Adam Piccone

NORMATIVE FACTOR STRUCTURE OF THE AAMR ADAPTIVE BEHAVIOR SCALE-SCHOOL, SECOND EDITION

A Broad-Range Tailored Test of Verbal Ability

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz

TheIntegrationofEFAandCFAOneMethodofEvaluatingtheConstructValidity

LEDYARD R TUCKER AND CHARLES LEWIS

Manifestation Of Differences In Item-Level Characteristics In Scale-Level Measurement Invariance Tests Of Multi-Group Confirmatory Factor Analyses

Evaluating a proposed modification of the Guttman rule for determining the number of factors in an exploratory factor analysis

Reliability, validity, and all that jazz

The Development of Scales to Measure QISA s Three Guiding Principles of Student Aspirations Using the My Voice TM Survey

Factor Structure of the Pediatric Symptom Checklist with a Pediatric Gastroenterology Sample

LISREL analyses of the RIASEC model: Confirmatory and congeneric factor analyses of Holland's self-directed search

What can the Real World do for simulation studies? A comparison of exploratory methods

The Asian Conference on Education & International Development 2015 Official Conference Proceedings. iafor

Alternative Methods for Assessing the Fit of Structural Equation Models in Developmental Research

Measuring mathematics anxiety: Paper 2 - Constructing and validating the measure. Rob Cavanagh Len Sparrow Curtin University

Does factor indeterminacy matter in multi-dimensional item response theory?

PERSONAL SALES PROCESS VIA FACTOR ANALYSIS

Technical Specifications

Assessing Measurement Invariance in the Attitude to Marriage Scale across East Asian Societies. Xiaowen Zhu. Xi an Jiaotong University.

THE INTERPRETATION OF EFFECT SIZE IN PUBLISHED ARTICLES. Rink Hoekstra University of Groningen, The Netherlands

Confirmation of the structural dimensionality of the Stanford-Binet Intelligence Scale (Fourth Edition)

Introduction to Test Theory & Historical Perspectives

Crossing boundaries between disciplines: A perspective on Basil Bernstein s legacy

Reliability and Validity of the Divided

Sample Sizes for Predictive Regression Models and Their Relationship to Correlation Coefficients

The Logic of Causal Inference

Chapter 19: Categorical outcomes: chi-square and loglinear analysis

Principal Components Factor Analysis in the Literature. Stage 1: Define the Research Problem

College Student Self-Assessment Survey (CSSAS)

Exploratory Factor Analysis Student Anxiety Questionnaire on Statistics

Validating Measures of Self Control via Rasch Measurement. Jonathan Hasford Department of Marketing, University of Kentucky

VERY SIMPLE STRUCTURE : AN ALTERNATIV?E PROCEDURE FOR ESTIMATING THE OPTIMALl NUMBER OF INTERPRETABLE FACTORS

Proof. Revised. Chapter 12 General and Specific Factors in Selection Modeling Introduction. Bengt Muthén

Factor Analysis. MERMAID Series 12/11. Galen E. Switzer, PhD Rachel Hess, MD, MS

Estimating Individual Rater Reliabilities John E. Overall and Kevin N. Magee University of Texas Medical School

Teachers Sense of Efficacy Scale: The Study of Validity and Reliability

How!Good!Are!Our!Measures?! Investigating!the!Appropriate! Use!of!Factor!Analysis!for!Survey! Instruments!

CEMO RESEARCH PROGRAM

Historical increase in the number of factors measured by commercial tests of cognitive ability: Are we overfactoring?

Can We Assess Formative Measurement using Item Weights? A Monte Carlo Simulation Analysis

Audio: In this lecture we are going to address psychology as a science. Slide #2

Development and Psychometric Properties of the Relational Mobility Scale for the Indonesian Population

American Psychdqgical Association

Item Response Theory. Steven P. Reise University of California, U.S.A. Unidimensional IRT Models for Dichotomous Item Responses

Whose psychological concepts?

Structural Validation of the 3 X 2 Achievement Goal Model

Chapter 18: Categorical data

Context of Best Subset Regression

THE GLOBAL elearning JOURNAL VOLUME 5, ISSUE 3, 2016

Internal structure evidence of validity

International Women s Day 2014 Action Guide

EFFECTS OF ITEM ORDER ON CONSISTENCY AND PRECISION UNDER DIFFERENT ORDERING SCHEMES IN ATTITUDINAL SCALES: A CASE OF PHYSICAL SELF-CONCEPT SCALES

Study Evaluating the Alterations Caused in an Exploratory Factor Analysis when Multivariate Normal Data is Dichotomized

References. Embretson, S. E. & Reise, S. P. (2000). Item response theory for psychologists. Mahwah,

Modeling the Influential Factors of 8 th Grades Student s Mathematics Achievement in Malaysia by Using Structural Equation Modeling (SEM)

Using Effect Size in NSSE Survey Reporting

Investigating the robustness of the nonparametric Levene test with more than two groups

The Thesis Writing Process and Literature Review

The Effects of Societal Versus Professor Stereotype Threats on Female Math Performance

The Health Score: A summary

Achievement in Science as Predictors of Students Scientific Attitude

The complete Insight Technical Manual includes a comprehensive section on validity. INSIGHT Inventory 99.72% % % Mean

An Experimental Investigation of Self-Serving Biases in an Auditing Trust Game: The Effect of Group Affiliation: Discussion

Relationship, Correlation, & Causation DR. MIKE MARRAPODI

Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories

Regression CHAPTER SIXTEEN NOTE TO INSTRUCTORS OUTLINE OF RESOURCES

CIRCLE The Center for Information & Research on Civic Learning & Engagement. Service-Learning in K-12 Public Education

Impact and adjustment of selection bias. in the assessment of measurement equivalence

Rasch Versus Birnbaum: New Arguments in an Old Debate

Basic concepts and principles of classical test theory

How Widely is Blind Assessment Used in Scientific Research?

CALIFORNIA STATE UNIVERSITY STANISLAUS DEPARTMENT OF SOCIOLOGY ASSESSMENT MODEL

To link to this article:

2 Types of psychological tests and their validity, precision and standards

USAID Alumni Association 2019 Membership Survey Results Prepared by Sarah Clark

The Psychometric Development Process of Recovery Measures and Markers: Classical Test Theory and Item Response Theory

HOW MANY DIMENSIONS ARE NEEDED TO DESCRIBE TEMPERAMENT IN ANIMALS: A FACTOR REANALYSIS OF TWO DATA SETS. Sergey V. Budaev

Techniques for Explaining Item Response Theory to Stakeholder

Reliability Theory for Total Test Scores. Measurement Methods Lecture 7 2/27/2007

FACTOR ANALYSIS Factor Analysis 2006

Teacher Influences on Child Interest in STEM Careers

Behavioral genetics: The study of differences

Neurobiology and Information Processing Theory: the science behind education

Title: Do general practitioners and psychiatrists agree about defining cure from depression? The DESCRIBE survey

Professor MOHAMMAD RASHED ALFAGIH. Cardiac Research

Tintin and the Exploded Confetti Plant: Big One or Big Bang?

Wason's Cards: What is Wrong?

A Comparison of Several Goodness-of-Fit Statistics

Testing the Multiple Intelligences Theory in Oman

CHAPTER VI RESEARCH METHODOLOGY

Testing Means. Related-Samples t Test With Confidence Intervals. 6. Compute a related-samples t test and interpret the results.

Causal Attribution Beliefs of Success and Failure: A Perspective from Pakistan

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

Bayesian Tailored Testing and the Influence

Effect of Sample Size on Correlation and Regression Coefficients

Transcription:

Horn, J. L. (1965). A rationale and test for the number of factors in factor analysis. Psychometrika, 30, 179 185. (3362 citations in Google Scholar as of 4/1/2016) Who would have thought that a paper published over 50 years ago would still have an impact today. But this one still does! Such longevity must be due to the method being used the simulation of the random sampling base in factor analysis. Over the years, the journal Psychometrika has been a haven for what some might consider mathematical minutiae, but this paper is definitely not one of them. It has enjoyed almost 3,400 citations as of April, 2016; the question of why comes to mind. The solution to the burning question of the correct number of common factors in a matrix of correlations is easy. Henry Kaiser used to say that it was so simple that he solved it everyday before breakfast (from J.L. Horn, Personal Communication, June 1977). Horn recognized that a solution was needed; but he thought he was proposing just a temporary fix, put in place for others to develop more formal statistical procedures (which they soon did! For a full review, see Lawley & Maxwell (1963); Horn & Engstrom (1979); Jöreskog & Sörbom (1979)). In Horn (1965), it was suggested that Guttman s latent-root-one lower bound estimate for the rank of a correlation matrix be accepted as a psychometric upper bound, following the proofs and arguments of Kaiser and Dickman. Kaiser shows that for a principal component to have positive KR-20 internal consistency, it is necessary and sufficient that the associated eigenvalue be greater than one (Kaiser, 1960, p. 6); this was the source of the root-one criterion. This rule, however, was for a population, so Horn tried to add some sampling considerations. In the statistical calculation of the correlations and the roots, the rank for a sample matrix should be estimated by subtracting out the component in the latent roots that can be attributed to sampling error and the least-squares capitalization on this error. Horn promoted a procedure, termed Parallel Analysis (PA) by later authors, based on the generation of random variables for estimating the component that needs to be subtracted. Horn became an expert in the classical techniques of common factor analysis. He generally asked why any elegant mathematical-statistical theory 1

should be based on specific assumptions when we know these key assumptions are wrong and untestable. Specifically, Horn suggested that the number of common factors should not be determined simply by using the well-known eigenvalues greater than one criterion defined by one of his favorite advisors, Henry Kaiser (1960; see Horn, 1965; and McArdle, 2007). Instead, Horn proposed the use of computer simulation techniques, mainly based on model assumptions that were random, and advocated for their use whenever possible (see Horn & McArdle, 1980). Horn (1965) determined the number of common factors by selecting the number of the eigenvalues of a correlation matrix that were greater than or equal to those provided by data computer-simulated with known characteristics. In this very simple idea, all that was needed was to generate random data of similar size ; one could then calculate the latent roots and vectors of these random data to provide a criterion tailored to the particular data set being analyzed. As an example, Horn (1965) implemented this procedure for 297 people measured on 65 ability variables; he found evidence for 16 common factors by the Kaiser criterion but only 9 by the PA criterion. Of course, we now recognize the need to consider statistical fluctuations in these latent roots, but it is noteworthy that these kinds of calculations were done some 50 years ago! Horn s random variable approach has more recently been found to be the most accurate for determining the number of unrotated common factors (e.g., Ledesma & Valero-Mora, 2007; Montanelli & Humphreys, 1976; Velicer, Eaton, & Fava, 2000; Zwick & Velicer, 1986; see also Dinno (2009), Hayton (2009), and Courtney (2013) for recent evaluation, and hearty affirmation). This success also marks the beginning of Horn s fascination with the use of computer simulated data to solve the most complex problems in mathematical statistics (see Horn & McArdle, 1980). Notes on the Author John Horn (1928 2006) was a pioneer in multivariate thinking and the application of multivariate methods to research on intelligence and personality. His key works on individual differences in the methodological areas of factor analysis and the substantive areas of cognition are briefly reviewed here. John 2

was also my mentor, teacher, colleague, and friend. It is tempting to review John Horn s main contributions to the field of intelligence by highlighting his methods of factor analysis and his substantive debates about intelligence, but this is done elsewhere (in McArdle & Hofer, 2014). As a leader in multivariate methodology, Horn tried to reach the incredible heights of his well-known mentor, Raymond Cattell. As illustrated here, John believed strongly in a multivariate scientific approach, and questioned the typical use of unweighted sum scores as if they represented the best scores of the psychological constructs of interest. On a substantive basis, John believed that there were important individual differences among adults within the domains of cognition and personality (see Horn & Knapp, 1974; Horn & Donaldson, 1977). Some of Horn s early comments on the methods of factor analysis are worth repeating, especially the central concept of a functional unity (Horn, 1972, p. 161-162). He applied the ideas about a multivariate meta-theory to data on cognitive abilities, to create various testable hypotheses. This type of reasoning provides the basis for arriving at several substantive results on cognitive abilities, perhaps the most important being that Horn expanded on this initial work of his primary advisor, Raymond Cattell (see Horn, 1965; Horn & Cattell, 1966; 1967) to identify additional functional unities of primary mental abilities (see Woodcock, 1989). For these reasons I think that John Horn s major contributions to psychology, only some of which have been discussed here, continue to be ahead of their time, and to have a profound influence on our thinking and critical approach to answering complex questions. His contributions to factor analysis and the structure of intelligence, the important methodological debates of the 1970s and 1980s regarding age and cohort effects and related issues of sample selectivity, the innovative ideas underlying his approach to evaluating state, trait and trait-change (Horn, 1972), and his willingness and encouragement to engage in critical evaluation of fundamental ideas and accepted scientific approaches (i.e., the g-theory; see McArdle, 2012) are and will remain important contributions. Through his research and teaching he forced people to question popular assumptions, evaluate all the data available, and consistently challenged us to think longer, harder, and better. His work will 3

continue to inspire important research in the fields of multivariate analysis and human cognitive abilities for decades to come. The interested reader can see McArdle and Hofer (2014). References Courtney, M. G. R. (2013). Determining the number of factors to retain in EFA: Using the SPSS R-Menu v2.0 to make more judicious estimations. Practical Assessment, Research & Evaluation, 18 (8), 1 14. Dinno, A. (2009). Exploring the sensitivity of Horn s parallel analysis to the distributional form of random data. Multivariate Behavioral Research, 44, 362 388. Hayton, J. C. (2009). Commentary on Exploring the sensitivity of Horn s Parallel Analysis to the distributional form of random data. Multivariate Behavioral Research, 44, 389 395. Horn, J. (1972). State, trait and change dimensions of intelligence. British Journal of Educational Psychology, 42, 159 185. Horn, J. L., & Cattell, R. B. (1966). Refinement and test of the theory of fluid and crystallized general intelligences. Journal of Educational Psychology, 57, 253 270. Horn, J. L., & Cattell, R. B. (1967). Age differences in fluid and crystallized intelligence. Acta Psychologica, 26, 107 129. Horn, J. L., & Donaldson, G. (1977). Faith is not enough: A response to the Baltes-Schaie claim that intelligence does not wane. American Psychologist, 32, 369 373. Horn, J. L., & Engstrom, R. (1979). Cattell s scree test in relation to Bartlett s chi-square test and other observations on the number of factors problem. Multivariate Behavioral Research, 14, 283 300. Horn, J. L., & Knapp, J. R. (1974). Thirty wrongs do not make a right: Reply to Guilford. Psychological Bulletin, 81, 502 504. Horn, J.L., & McArdle, J.J. (1980). Perspectives on mathematical and statistical model building (MASMOB) in research on aging. In L. Poon (Ed.), Aging in the 1980 s: Psychological Issues (pp. 503 541). Washington, D. C.: American Psychological Association. 4

Jöreskog, K. G., & Sörbom, D. (1979). Advances in factor analysis and structural equation models. Cambridge, MA: Abt. Kaiser, H.F. (1960). The application of electronic computers to factor analysis. Educational and Psychological Measurement, 20, 141 151. Lawley, D.N. & Maxwell, A.E. (1963). Factor analysis as a statistical method. London: Butterworth. Ledesma, R., & Valero-Mora, P. (2007) Determining the number of factors to retain in EFA: an easy-to-use computer program for carrying out Parallel Analysis. Practical Assessment, Research & Evaluation, 12 (1), 1 11. McArdle, J. J. (2007). John L. Horn (1928 2006). American Psychologist, 62, 596 597. McArdle, J.J. (2012). Testing the idea of general intelligence. F & M Scientist, 1, 27 66. McArdle, J.J., & Hofer, S.M. (2014). Fighting for intelligence: A review of the academic research of John L. Horn. Multivariate Behavioral Research, 49, 1 16. Winner of the Tanaka Award as the best MBR Paper, October 2015. Montanelli, R. G., Jr., & Humphreys, L. G. (1976). Latent roots of random data correlation matrices with squared multiple correlations on the diagonal: A Monte Carlo study. Psychometrika, 41, 341 348. Velicer, W. F., Eaton, C. A., & Fava, J. L. (2000). Construct explication through factor or component analysis: A review and evaluation of alternative procedures for determining the number of factors or components. In Goffen, R. D. and Helms, E., (Eds.), Problems and Solutions in Human Assessment: Honoring Douglas N. Jackson at Seventy (pp. 41-71). New York: Springer. Woodcock, R.W. (1998). Extending Gf-Gc theory into practice. In McArdle, J.J. & Woodcock, R. W. (Eds.). Human abilities in theory and practice (pp. 137 156). Mahwah, NJ: Erlbaum. Zwick, W.R., & Velicer, W.F. (1986). Comparison of five rules for determining the number of components to retain. Psychological Bulletin, 99, 432 442. John J. McArdle, Department of Psychology, University of Southern California 5

Acknowledgements: We thank the National Institute on Aging (AG0713720 to the author) for support of this research and to Carol Prescott (University of Southern California) for helpful comments on the draft. The content is solely the responsibility of the author and does not necessarily represent the official views of the National Institute on Aging or the National Institutes of Health. 6