Internal Consistency: Do We Really Know What It Is and How to. Assess it?

Similar documents
Running head: CFA OF STICSA 1. Model-Based Factor Reliability and Replicability of the STICSA

Running head: CFA OF TDI AND STICSA 1. p Factor or Negative Emotionality? Joint CFA of Internalizing Symptomology

THE USE OF CRONBACH ALPHA RELIABILITY ESTIMATE IN RESEARCH AMONG STUDENTS IN PUBLIC UNIVERSITIES IN GHANA.

Bruno D. Zumbo, Ph.D. University of Northern British Columbia

GENERALIZABILITY AND RELIABILITY: APPROACHES FOR THROUGH-COURSE ASSESSMENTS

ESTABLISHING VALIDITY AND RELIABILITY OF ACHIEVEMENT TEST IN BIOLOGY FOR STD. IX STUDENTS

The effects of ordinal data on coefficient alpha

LANGUAGE TEST RELIABILITY On defining reliability Sources of unreliability Methods of estimating reliability Standard error of measurement Factors

Investigating the robustness of the nonparametric Levene test with more than two groups

Multivariable Systems. Lawrence Hubert. July 31, 2011

Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories

Basic concepts and principles of classical test theory

In many fields, such as education or second language acquisition,

INVESTIGATING FIT WITH THE RASCH MODEL. Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form

Internal structure evidence of validity


Examining the Psychometric Properties of The McQuaig Occupational Test

Running head: NESTED FACTOR ANALYTIC MODEL COMPARISON 1. John M. Clark III. Pearson. Author Note

Anumber of studies have shown that ignorance regarding fundamental measurement

Regression Discontinuity Analysis

Reliability, validity, and all that jazz

Construct Reliability and Validity Update Report

1. Evaluate the methodological quality of a study with the COSMIN checklist

How important is the General Factor of Personality?

SEM-Based Composite Reliability Estimates of the Crisis Acuity Rating Scale with Children and Adolescents

Statistical Methods and Reasoning for the Clinical Sciences

Skew and Internal Consistency

How important is the General Factor of Personality? A General Critique

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

1.4 - Linear Regression and MS Excel

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison

Introduction On Assessing Agreement With Continuous Measurement

2 Types of psychological tests and their validity, precision and standards

Significance Tests Based on Differences Donald W. Zimmerman, Carleton University

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

The c-bifactor model as a tool for the construction of semi-homogeneous upperlevel

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

CHAPTER ONE CORRELATION

Factor Indeterminacy in Generalizability Theory

The Impact of Continuity Violation on ANOVA and Alternative Methods

Empirical Formula for Creating Error Bars for the Method of Paired Comparison

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report

Introduction to the HBDI and the Whole Brain Model. Technical Overview & Validity Evidence

A Comparison of Methods of Estimating Subscale Scores for Mixed-Format Tests

Estimating Individual Rater Reliabilities John E. Overall and Kevin N. Magee University of Texas Medical School

Technical Specifications

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review

Standard Scores. Richard S. Balkin, Ph.D., LPC-S, NCC

Centre for Education Research and Policy

Issues That Should Not Be Overlooked in the Dominance Versus Ideal Point Controversy

Factorial Validity and Reliability of 12 items General Health Questionnaire in a Bhutanese Population. Tshoki Zangmo *

Standard Errors of Correlations Adjusted for Incidental Selection

Having your cake and eating it too: multiple dimensions and a composite

A PRELIMINARY EXAMINATION OF THE ICAR PROGRESSIVE MATRICES TEST OF INTELLIGENCE

Paul Irwing, Manchester Business School

Measurement and Descriptive Statistics. Katie Rommel-Esham Education 604

The complete Insight Technical Manual includes a comprehensive section on validity. INSIGHT Inventory 99.72% % % Mean

Development and Psychometric Properties of the Relational Mobility Scale for the Indonesian Population

Research Brief Reliability of the Static Risk Offender Need Guide for Recidivism (STRONG-R)

Jason L. Meyers. Ahmet Turhan. Steven J. Fitzpatrick. Pearson. Paper presented at the annual meeting of the

IRT Parameter Estimates

Comparing Pre-Post Change Across Groups: Guidelines for Choosing between Difference Scores, ANCOVA, and Residual Change Scores

Factors Affecting the Item Parameter Estimation and Classification Accuracy of the DINA Model

commentary Time is a jailer: what do alpha and its alternatives tell us about reliability?

Treatment Effects: Experimental Artifacts and Their Impact

Impact of Differential Item Functioning on Subsequent Statistical Conclusions Based on Observed Test Score Data. Zhen Li & Bruno D.

RELIABILITY RANKING AND RATING SCALES OF MYER AND BRIGGS TYPE INDICATOR (MBTI) Farida Agus Setiawati

The Effect of Guessing on Item Reliability

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

Using Analytical and Psychometric Tools in Medium- and High-Stakes Environments

Recommended Sample Size for Conducting Exploratory Factor Analysis on Dichotomous Data

1 The conceptual underpinnings of statistical power

EXPERIMENTAL RESEARCH DESIGNS

Differential Item Functioning

Estimating Reliability in Primary Research. Michael T. Brannick. University of South Florida

VARIABLES AND MEASUREMENT

Personal Style Inventory Item Revision: Confirmatory Factor Analysis

computation and interpretation of indices of reliability. In

The Influence of Test Characteristics on the Detection of Aberrant Response Patterns

EPSE 594: Meta-Analysis: Quantitative Research Synthesis

Answers to end of chapter questions

By Hui Bian Office for Faculty Excellence

Bipolar items for the measurement of personal optimism instead of unipolar items

Chapter -6 Reliability and Validity of the test Test - Retest Method Rational Equivalence Method Split-Half Method

Demonstrating validity

A Review on Sample Size Determination for Cronbach s Alpha Test: A Simple Guide for Researchers

Chapter 3 Psychometrics: Reliability and Validity

Psychometric properties of the PsychoSomatic Problems scale an examination using the Rasch model

Biserial Weights: A New Approach

Statistics for Psychosocial Research Session 1: September 1 Bill

Random Intercept Item Factor Analysis

Reliability Analysis: Its Application in Clinical Practice

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

An Alternative to the Trend Scoring Method for Adjusting Scoring Shifts. in Mixed-Format Tests. Xuan Tan. Sooyeon Kim. Insu Paek.

ASSESSING THE UNIDIMENSIONALITY, RELIABILITY, VALIDITY AND FITNESS OF INFLUENTIAL FACTORS OF 8 TH GRADES STUDENT S MATHEMATICS ACHIEVEMENT IN MALAYSIA

Mantel-Haenszel Procedures for Detecting Differential Item Functioning

A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale

EFFECTS OF ITEM ORDER ON CONSISTENCY AND PRECISION UNDER DIFFERENT ORDERING SCHEMES IN ATTITUDINAL SCALES: A CASE OF PHYSICAL SELF-CONCEPT SCALES

Transcription:

Internal Consistency: Do We Really Know What It Is and How to Assess it? Wei Tang, Ying Cui, CRAME, Department of Educational Psychology University of Alberta, CANADA The term internal consistency has been widely used but controversially defined. Cronbach (1951) used internal consistency and homogeneity interchangeably and claimed: an internally consistent or homogeneous test should be independent of test length. However, many other researchers (Green, Lissitz & Mulaik, 1977; McDonald, 1981; Miller, 1995; Schmitt, 1996) used the term internal consistency to refer to the interrelatedness of items, and distinguished internal consistency from homogeneity by claiming that homogeneity refers to the unidimensionality of a set of items. Another popular use of internal consistency is related to the reliability estimate of a test based on a single administration, which is traditionally called internal consistency reliability (Haertel, 2006). In this sense, internal consistency has been often used to denote a group of methods that are intended to estimate the reliability of a single-administrated test (Hattie, 1995). As illustrated above, the term internal consistency is associated with different meanings. Just as Sijtsma (2009) concluded, Internal consistency has not been defined that explicitly, far from it. Because of the vague and implicit definition of this term, the chaos in interpreting and using internal consistency has existed for a long time. The confusion in the concept of internal consistency may cause problems in understanding, measuring and applying other related psychometric properties of a test. Besides, it hampers the development of new and better indices for measuring test internal consistency. Therefore, an explicit and complete definition of the term 1

internal consistency is badly needed. To better understand and define this term, different definitions or interpretations of the term internal consistency and the corresponding measures for assessing internal consistency are reviewed and discussed in the following sections. Finally, an explicit definition of test internal consistency is provided and appropriate measures are recommended. What Is Internal Consistency? Internal Consistency = Internal Consistency Reliability? Internal consistency has long been used as the synonym of internal consistency reliability and thus the indices for measuring internal consistency reliability, such as Kuder and Richardson (1937) s KR 20 and coefficient alpha (Cronbach, 1951), have become widely used measures of internal consistency. Interpretation of alpha as a measure of internal consistency has gained more foothold in practical test construction and test use (Sijtsma, 2009). However, some researchers (Cortina, 1993; Green et al, 1977; Nunnally, 1978) have pointed out the inappropriateness of using internal consistency reliability coefficients to measure internal consistency. For example, Nunnally wrote: Estimates of reliability based on the average correlation among items within a test are said to concern the internal consistency. This is partly a misnomer, because the size of the reliability coefficient is based on both the average correlation among items (the internal consistency) and the number of items. Coefficient alpha is the basic formula for determining the reliability based on internal consistency. To better distinguish internal consistency and internal consistency reliability, we may first review the definition of reliability. The reliability of measurement results was initially defined by Spearman (1904) as the ratio of true score variance to 2

observed score variance in classical test theory. Reliability is the quantification of the consistency of results from a measurement procedure across replications. A test or measure is considered perfectly reliable if we get the same results repeatedly. The reliability of measurement results can be estimated by calculating the correlation between the total scores on two independent administrations or two parallel forms of the test. Although the definition of reliability requires data from repeated testing, such data are rarely available in practice. Thus, reliability coefficient is normally estimated based on scores from a single test administration, which is referred to as internal consistency reliability by Cronbach (1951). Intuitively, internal consistency reliability concerns the consistency of behavior within a very limited time interval, i.e., the time interval during which the items in the test are being responded to (Horst, 1953). And this short term reliability is comparable to the correlation between two tests taken within a short time interval. Internal consistency in the term internal consistency reliability does not work as the synonym of reliability but the qualifier of the reliability, which implies the reliability is estimated under the condition of internal consistency, that is, when the test is internally consistent, the reliability of the test can be estimated by these reliability coefficients. An internal consistent test can be interpreted as a test with all items measure the same underlying construct in the same way. In other words, test items are all parallel in an internally consistent test. Internal consistency reliability such as KR 20 and coefficient alpha are not designed to measure internal consistency but the reliability under the condition of the internal consistency. Using these reliability estimates for measuring internal consistency is inappropriate. Therefore, although both reliability and internal consistency are determined by the degree of interitem correlation, these two terms are distinctive in nature. 3

Reliability refers to the consistency of a whole test across replications whereas internal consistency refers to the consistency among the responses to the items within a test. A test that is not internally consistent could be reliable, and vice versa. The coefficients for internal consistency reliability should not be used directly as indices for internal consistency also because reliability estimates are all functions of item number. Specifically, reliability increases as item number increases. Different from reliability, internal consistency of a test should be independent of its length (Cronbach, 1951). For example, a test with 10 items may have the same internal consistency as it does when increased to 20 items, even though the reliability of the test is higher then. If test developers use many items to measure a single construct without increasing test internal consistency, this can serve only to increase reliability at the expense of precision and parsimony. Internal Consistency = Average Interitem Correlations? Conceptually, internal consistency measures whether items on the same test (or the same subscale on a larger test) that intend to measure the same construct produce consistent scores. If two items are designed to measure the same construct, a testee should answer both questions in the same way, which would indicate that the test has internal consistency. Thus, some researchers (Cortina, 1993) defined internal consistency as a measure based on the degree of correlations between different items on the same test (or the same subscale on a larger test). Since the correlations between items are, most often than not, different, using the average interitem correlation seems a simple and direct approach to represent the degree of correlations between different items. Cronbach (1951), to the authors knowledge, is the first one who proposed to use mean interitem correlation to measure internal consistency. He derived an index 4

(Cronbach s r ij) specifically for assessing internal consistency, though no explicit definition of internal consistency is provided. Cronbach s r ij (1951) is derived by applying the Spearman-Brown formula to alpha, thereby estimating the mean correlation between items. Since then, internal consistency has been interpreted and measured by the mean interitem correlation by some researchers and practitioners (e.g., Briggs & Cheek, 1986; Nunnally, 1978). However, we should be aware that the consistent scores does not necessarily mean the scores are same or similar across items. For example, in a psychological test, if all respondents expressed agreement with the statements "I like to study with a partner" and disagreement with the statement "I hate to study with a partner", this is still an indicative of perfect internal consistency of the test. Although the two items are negatively correlated, the degree of the correlation is as high as 1. Suppose there are more items like that in the test, then the mean interitem correlations could be close to zero because the positive interitem correlations cancel out the negative ones. To overcome that problem, the practitioners can either recode the raw scores in the opposite order or use the mean of the absolute value of the correlations between items as the indicator for internal consistency. One disadvantage of mean interitem correlation is that this measure is influenced by the extreme values or outliers. In addition, it does not reflect the variability among the interitem correlations. Therefore, the use of mean interitem correlation will be problematic when the interitem correlations follow a skewed distribution, especially with some extreme outliers. Internal Consistency = General Factor Saturation? An alternative way of interpreting internal consistency is that it is the extent to 5

which all of the items of a test measure the same construct, that is, the general factor saturation (Revelle, 1979) or the closeness to the unidimensionality (Sijtsma, 2009). The advantage of this perspective over the notion of the average degree of correlation among the items of a test is that the general factor saturation is not affected by the skewness of the distribution of item correlations or extreme values of item correlations. Besides, if the items of a test measure several unrelated latent traits, the average item correlation in such cases will be greater than zero whereas some subtests or subscales which measure the unrelated latent traits have the correlation of zero. Thus, using the average degree of item correlation to measure internal consistency is not an appropriate choice when a test is not unidimensional. Whereas the ideal of measurement is for all items of a test to measure the same latent variable, it is hard to achieve for a single test in real practice. For example, IELTS (International English Language Testing System) has four subscales which measure four distinctive features of English: listening, reading, writing and speaking. Although the whole test is designed to measure the proficiency level of English as a second language (the general trait/construct), each module taps a lower order trait (the group trait/construct). Tests or measures with items reflecting a single factor in the higher order while items reflecting a variety of aspects mark a number of factors in the lower order (one factor for each aspect) is not uncommon in psychological or educational measurement, and hence the assumption of unidimensionality of a test is often unsatisfied. Nevertheless, if the subscales are highly correlated, it is factorially homogeneous (Revelle, 1979). Since high correlations among the subscales indicate the high order factor (the general factor) can be extracted and be dominant, test internal consistency is interpreted as the general factor saturation or the closeness to the unidimensionality. That also explains why some researchers (Cronbach, 1951; 6

Revelle, 1979) use homogeneity and internal consistency as synonyms. As to the measures of test internal consistency defined in this way, different recommendations have been raised so far. Revelle (1979) recommended beta, the worst split-half reliability estimate, and pointed out: in the case of a lumpy test (one with several large group factors), alpha overestimates the general factor saturation of the test and underestimates the total common factor saturation. Beta, on the other hand, gives a more appropriate estimate of the general factor saturation but severely underestimates the common factor saturation. However, to find this worst split half requires trying all the possible splits, which is impractical when the test is just normal size, say, of 20 items. Since analytic method does not work, a heuristic is needed to estimate beta. ICLUST (Revelle, 1977; 1979), the program for hierarchical cluster analysis, was put forward to estimate beta. Sijtsma (2009) suggested using MRFA (minimum rank factor analysis, Ten Berge & Kiers, 1991) for assessing closeness of the covariance/ correlation matrix to unidimensionality. By using MRFA, closeness of the 1-factor solution to unidimensionality is assessed by the ratio of the first eigenvalue of the estimated true score covariance matrix to the sum of all J (J is the number of variables) eigenvalues (Ten Berge & Sočan, 2004). Sijtsma used the percentage of the above ratio, expressed as the ECV (explained common variance), to represent the closeness to unidimensionality. Revelle and Zinbarg (2009) suggested the hierarchical coefficient omega (ω h ) (McDonald, 1999; Zinbarg, Revelle, &Yovel, 2005) might be a more appropriate index of estimating internal consistency since ω h could indicate the extent to which all of the items in a test measure the general factor. ω h equals the ratio of the general factor variance to the total variance of the test, and thus a natural choice of measuring general factor saturation. 7

An Explicit Definition of Internal Consistency Based on the above review and discussion of the term internal consistency and the corresponding measures for assessing internal consistency, internal consistency can be defined as a psychometric property of a test that describes whether the items in the test that are proposed to measure the same construct produce consistent scores, which is associated with the degree of general factor saturation and the degree of interitem correlations, but irrelevant to test length. If defined more stringently, internal consistency means all the items are parallel in a test or subtest. It can be shown that internal consistency has a stronger connection with validity than reliability if we look at the formulas that depict the relationship between validity and reliability (Judd, Smith & Kidder, 1991; Streiner, 2003) below: and Reliability= σ CI 2 2 +σ SE σ2, (1) T Validity= σ 2 CI σ2. (2) T The CI represents the construct of interest, the SE the systematic error. Therefore, if internal consistency is simply defined as the general factor saturation, measuring internal consistency is no difference from measuring construct validity. However, as stated above, internal consistency is a more complex concept than general factor saturation because it reveals more information about the internal structure of a test. Comparison of the Indices for Measuring Internal Consistency Method To find out which of the recommended indices are appropriate measures of internal consistency, these indices are compared by using hypothetical data sets. The 8

main advantage of using artificial data is that the internal structures of hypothetical tests are known. Besides, tests with various internal structures can be analyzed and compared simultaneously, and thus the pattern or trend of the indices and the factors influencing the indices can be detected. The conditions are manipulated to resemble real psychological tests to overcome the disadvantage of hypothetical data: being artificial. Although psychological tests are diverse in terms of the content, such as achievement tests, aptitude tests, intelligence tests, personality tests, and interest inventories, they have several common points which can be utilized for data simulation. First, psychological tests often measure more latent traits or attributes instead of a single trait that is assumed in the classical test theory. For example, the questions on PDS (Personal Data Sheet) cover the topics such as excessive anxiety, depression, abnormal fears, impulse problems, etc (Woodworth, 1920). Second, the traits to be measured in a psychological test tend to be interrelated because they are the components of a more general construct. Considering these common characteristics in psychological tests, the model chosen for simulating real psychological tests is the bifactor model (Chen, West, & Sousa, 2006; Rindskopf & Rose, 1988; Schmid & Leiman, 1957; Yung, Thissen, & Mcleod, 1999), which includes not only a general factor F, but also K group factor(s),g k, expressed as: k X i = λ i F + α ik G k + E i k=1 Here, λ i is the factor loading for the general factor F and α ik is the factor loading for the kth group factor G k on the ith item; E i is the residual score of the ith item. To serve the purpose of comparison, the unifactor model is also included (i.e, α ik is set as zero). Totally, test data sets with five different internal structures are generated. 1) unifactor data with equal λ i ; 2) unifactor data with unequal λ i ; 3) bifactor data with larger λ i than α ik ; 4) bifactor data with equal λ i and α ik. 5) 9

bifactor data with smaller λ i than α ik. To generate a simple bifactor structure, only two group factors are considered with equal group factor loadings, and the general factor loadings are also the same in bifactor data sets. Each structure is manipulated with three levels of items number (10, 20 and 40), and two levels of average interitem correlation (0.3 and 0.6) for the unifactor data and two subscales of bifactor data. Results Unifactor Data Cronbach s r ij is exactly the same as the mean of interitem correlations. In fact, this index is equal to the average interitem correlation no matter whether the item covariance matrix is identical to the item correlation matrix as in the cases of this study. As shown in Table 1, Cronbach s r ij (shortened as C s r ij) is only affected by the degree of factor loadings of the item, that is, the degree of interitem correlations. If the average interitem correlation does not change, the coefficient r ij keeps constant as the item number increases from 10 to 40, regardless of the change in the discrepancy between interitem correlations. SD is the standard deviation of the correlations between items. When item factor loadings are the same, SD is zero. Under such condition, test length has no effect on SD. Under the conditions with heterogeneous loadings, test length has negligible effect on SD. As item number increases from 10 to 20, SD increases by 0.01. Whereas alpha, beta and ω h have the same values under the conditions with equal factor loadings, they are quite different under the conditions with unequal factor loadings. Among the three coefficients, beta has the lowest values while ω h has the highest values; the discrepancy among them decreases as the degree of the average interitem correlation increases. All of the three coefficients are positively affected by 10

item number and the degree of the average interitem correlation. However, as discrepancy among the interitem correlations occurs, only beta decreases. The decrease in beta becomes moderate as the average interitem correlation increases from 0.3 to 0.6. ECV is the only coefficient which does not change its value under all the unidimensional conditions, the values of which are expressed in proportions instead of percentages to serve the purpose of comparison. Under unidimensional conditions, the values of ECV are all ones, which indicates the perfect unidimensionality of a test. Although ECV works well for measuring the closeness to unidimensionality, it fails to reveal the change in the average interitem correlation and test internal structure. TABLE 1 Indices for measuring internal consistency for unidimensional data sets Test length 10 20 40 10 20 40 Equal Loadings λ i = 0.3 λ i = 0.6 C s r ij SD alpha beta ω h ECV 0.30 0.00 0.81 0.81 0.81 0.30 0.00 0.90 0.90 0.90 0.30 0.00 0.95 0.95 0.95 0.60 0.00 0.94 0.94 0.94 0.60 0.00 0.97 0.97 0.97 0.60 0.00 0.98 0.98 0.98 Unequal Loadings λ i = 0.62 and 0.10 λ i = and 0.35 C s r ij SD alpha beta ω h ECV 0.30 0.18 0.81 0.68 0.83 0.30 0.19 0.90 0.74 0.90 0.30 0.19 0.95 0.78 0.95 0.60 0.19 0.94 0.89 0.94 0.60 0.20 0.97 0.97 0.60 0.20 0.98 0.93 0.99 Bifactor Data For bifactor data sets, the degree of general factor saturation is the main factor which affects the values of indices for measuring internal consistency. As the degree 11

of general factor saturation decreases while keeping other variables the same, all the indices except SD become smaller. Alpha was least influenced by the degree of general factor saturation while most influenced by test length. Beta and ω h have the same values under all the conditions of bifactor data sets because the item loadings are equal in terms of the general factor, and each subscale has the same average within-subscale correlation. Beta and ω h are affected dramatically by the degree of general factor loadings. As in Table 2, when the general factor loading decreases from 0.25 to 0.05, beta and ω h drop by more than 70% of the values under the conditions with high general factor saturation. Although beta and ω h are also affected by item number, the effect is not remarkable as alpha, especially under the conditions with low degree of general factor saturation. All the coefficients, except ECV, increase as the degree of average interitem correlation increases. TABLE 2 Indices for measuring internal consistency for bi-factor data sets Subscale length 5 10 20 5 10 20 High GFS λ i = 0.25; α ik = 0.05 λ i = 0.50; α ik = 0.10 C s r ij SD alpha beta ω h ECV 0.27 0.02 0.79 0.72 0.72 0.27 0.02 0.88 0.81 0.81 0.27 0.02 0.94 0.85 0.85 0.54 0.05 0.85 0.85 0.55 0.05 0.96 0.88 0.88 0.55 0.05 0.98 0.89 0.89 Medium GFS λ i = 0.15; α ik = 0.15 λ i = 0.30; α ik = 0.30 C s r ij SD alpha beta ω h ECV 0.22 0.07 0.73 0.51 0.51 0.75 0.22 0.07 0.85 0.58 0.58 0.75 0.22 0.07 0.62 0.62 0.75 0.43 0.15 0.88 0.61 0.61 0.75 0.44 0.15 0.94 0.64 0.64 0.75 0.45 0.15 0.97 0.65 0.65 0.75 Low GFS λ i = 0.05; α ik = 0.25 λ i = 0.10; α ik = 0.50 C s r ij 0.16 0.17 0.17 0.32 0.34 0.34 SD alpha 0.12 0.66 0.12 0.80 0.12 0.89 0.25 0.83 0.25 0.91 0.25 0.95 12

beta 0.20 0.24 0.26 0.26 0.27 0.28 ω h ECV 0.20 0.58 0.24 0.58 0.26 0.58 0.26 0.58 0.27 0.58 0.28 0.58 Discussion and Recommendation To better evaluate which indices are appropriate for measuring internal consistency, we may review the factors which influence test internal consistency. 1) The degree of general factor saturation. Although all the coefficients are more or less influenced by that, only beta, ω h and ECV can clearly indicate the change in the degree of general factor saturation. 2) The degree of the average interitem correlation. Except ECV and SD, all the other coefficients can reflect the change in the degree of the average interitem correlation. 3) Independence of test length. Alpha, beta and ω h are all dependent of test length. They all increase with item number becoming larger while keeping other conditions the same. Alpha, as the estimate of reliability, is most influenced by test length among the three coefficients. Beta and ω h are independent of item number under the condition that the general factor saturation is zero, which means there is no general factor but orthogonal group factors. Under that condition, beta and ω h are always equal to zero. The indices independent of item number are Cronbach s r ij, SD and ECV. Nevertheless, these indices, if used alone, fail to depict the whole picture of test internal consistency. The results of the comparison of these indices show that measuring internal consistency with a statistic of a single value is not sufficient under the present stage, so a combination of measures for test internal consistency is recommended. To measure test internal consistency, we can first use ECV to assess whether a test is unidimensional or not, because it is the only index which can indicate the unidimensionality without the influence of other variables like test length or factor loading size. If ECV is one or close to one, then the test is unidimensional or very 13

close to unidimensionality, the average degree of the correlations between items (Cronbach s r ij) in the whole test and the standard deviation of these interitem correlations should be both reported for indicating test internal consistency. If ECV is rather less than 1(e.g., 0.75 or below in this study), then beta and ω h are recommended because they are more affected by the degree of general factor saturation than by the degree of average interitem correlations while Cronbach s r ij is the other way around; besides, beta and ω h are less affected by test length when ECV is 0.75 or below. Under the more stringent definition of internal consistency, only beta is recommended when ECV is rather less than 1 because beta decreases as discrepancy among the interitem correlations increases while ω h does not indicate the disagreement among the interitem correlations. To distinguish internal consistency from other terms in the fields of measurement and testing, explicit and less confusing names for different properties of a test are also needed. For example, the term internal consistency reliability is confusing and it can be replaced with reliability estimates under the condition of internal consistency. In short, more efforts are required for the clarification of different psychometrical terms so that appropriate measures can be developed and applied in the future. References Briggs, S. R., & Cheek, J. M. (1986). The role of factor analysis in the development and evaluation of personality scales. Journal of Personality, 54, 107-148. Chen, F. F., West, S. G. & Sousa, K. H. (2006). A comparison of bifactor and second-order models of quality-of-life. Multivariate Behavioral Research, 41, 189 225. Cortina, J. M. (1993). What is coefficient alpha? An examination of theory and applications. Journal of Applied Psychology, 78, 98-104. 14

Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, 16, 297-334. Green, S. B., Lissitz, R. W., & Mulaik, S. A. (1977). Limitations of coefficient alpha as an index of test unidimensionality. Educational and Psychological Measurement, 37, 827-838. Haertel, E. H. (2006). Reliability. In Brennan, R. L. (Ed.). Educational Measurement. (4th ed.). National Council on Measurement in Education. American Council on Education. Westport, CT : Praeger Publishers. Hattie, J. A. (1995). Methodology review: assessing unidimensionality. Applied Psychological Measurement, 9, 139-164. Horst, P. (1953). Correcting the Kuder-Richardson reliability for dispersion of item difficulties. Psychological Bulletin, 50, 371-374. Judd, C. M., Smith, E. R., & Kidder, L. H. (1991). Research Methods in social relations (6 th ed.). New York: Harcourt Brace Jovanovich. Kuder, G. F. & Richardson, M. W. (1937). The theory of estimation of test reliability. Psychometrika, 2, 151-161. McDonald, R. P. (1981). The dimensionality of tests and items. British Journal of Mathematical and Statistical Psychology, 34, 100-117. McDonald, R. P. (1999). Test theory: A unified treatment. Hillsdale: Erlbaum. Miller, M. B. (1995). Coefficient alpha: a basic introduction form the perspectives of classical test theory and structural equation modeling. Structural Equation Modeling, 2, 255-273. Novick, M. R., & Lewis, C. (1967). Coefficient alpha and the reliability of composite measurements. Psychometrika, 32, 1-13. Nunnally, J. C. (1978). Psychometric theory (2 nd ed.). New York: McGraw-Hill. Raykov, T. (1998). Coefficient alpha and composite reliability with interrelated nonhomogeneous items. Applied Psychological Measurement, 22, 375-385. Revelle, W. (1977). A program for hierarchical cluster analysis. Northwestern University. Computing Center Document 482, 1977. Revelle, W. (1979). Hierarchical cluster-analysis and the internal structure of tests. Multivariate Behavioral Research, 14 (1), 57-74. 15

Revelle, W. & Zinbarg, R. E. (2009). Coefficient alpha, beta, and the glb: comments on Sijtsma. Psychometrika, 74, 145-154. Rindskopf, D. & Rose, T. (1988). Some theory and applications of confirmatory second-order factor analysis. Multivariate Behavioral Research. 23, 51 67. Schmid, J. & Leiman, J. (1957). The development of hierarchical factor solutions. Psychometrika, 22, 53 61. Schmitt, N. (1996). Uses and Abuses of Coefficient Alpha. Psychological Assessment, 8, 350-353. Sijtsma, K. (2009). On the use, the misuse, and the very limited usefulness of Cronbach s alpha. Psychometrika, 74, 107-120. Spearman, C. (1904). The proof and measurement of association between two things. American Journal of Psychology, 15, 72-101. Streiner, D. L. (2003). Starting at the beginning: an introduction to coefficient alpha and internal consistency. Journal of Personality Assessment. 80, 99-103. Ten Berge, J. M. F., & Kiers, H. A. L. (1991). A numerical approach to the exact and the approximate minimum rank of a covariance matrix. Psychometrika, 56, 309-315. Ten Berge, J. M. F., & Sočan, G. (2004). The greatest lower bound to the reliability of a test and the hypothesis of unidimensionality. Psychometrika, 69, 613-625. Woodworth, R. S. (1920). Personal Data Sheet. Chicago: Stoelting. Yung, Y. F., Thissen, D. & McLeod, L. D. (1999). On the relationship between the higher-order factor model and the hierarchical factor model. Psychometrika, 64, 113 128. Zinbarg, R. E., Revelle, W., Yovel, I., & Li, W. (2005). Cronbach s α, Revelle s β, and McDonald s ω H : Their relations with each other and two alternative conceptualizations of reliability. Psychometrika, 70, 123-133. 16