The Application of a Cognitive Diagnosis Model via an. Analysis of a Large-Scale Assessment and a. Computerized Adaptive Testing Administration

Similar documents
Objectives. 6.3, 6.4 Quantifying the quality of hypothesis tests. Type I and II errors. Power of a test. Cautions about significance tests

Anchor Selection Strategies for DIF Analysis: Review, Assessment, and New Approaches

SIMULATIONS OF ERROR PROPAGATION FOR PRIORITIZING DATA ACCURACY IMPROVEMENT EFFORTS (Research-in-progress)

Author's personal copy

The vignette, task, requirement, and option (VITRO) analyses approach to operational concept development

Decision Analysis Rates, Proportions, and Odds Decision Table Statistics Receiver Operating Characteristic (ROC) Analysis

Presymptomatic Risk Assessment for Chronic Non- Communicable Diseases

Chapter 4: One Compartment Open Model: Intravenous Bolus Administration

Introducing Two-Way and Three-Way Interactions into the Cox Proportional Hazards Model Using SAS

Do People s First Names Match Their Faces?

Cognitive Load and Analogy-making in Children: Explaining an Unexpected Interaction

Bayesian design using adult data to augment pediatric trials

Merging of Experimental and Simulated Data Sets with a Bayesian Technique in the Context of POD Curves Determination

A Note on False Positives and Power in G 3 E Modelling of Twin Data

The meaning of Beta: background and applicability of the target reliability index for normal conditions to structural fire engineering

Sampling methods Simple random samples (used to avoid a bias in the sample)

Transitive Relations Cause Indirect Association Formation between Concepts. Yoav Bar-Anan and Brian A. Nosek. University of Virginia

Automatic System for Retinal Disease Screening

Remaining Useful Life Prediction of Rolling Element Bearings Based On Health State Assessment

Reinforcing Visual Grouping Cues to Communicate Complex Informational Structure

Patterns of Inheritance

Child attention to pain and pain tolerance are dependent upon anxiety and attention

Lecture 2: Linear vs. Branching time. Temporal Logics: CTL, CTL*. CTL model checking algorithm. Counter-example generation.

Linear Theory, Dimensional Theory, and the Face-Inversion Effect

Estimating shared copy number aberrations for array CGH data: the linear-median method

Risk and Rationality: Uncovering Heterogeneity in Probability Distortion

carinzz prophylactic regimens

R programs for splitting abridged fertility data into a fine grid of ages using the quadratic optimization method

Fundamental Concepts for Using Diagnostic Classification Models. Section #2 NCME 2016 Training Session. NCME 2016 Training Session: Section 2

This is an author-deposited version published in: Eprints ID: 15989

Regret theory and risk attitudes

Cocktail party listening in a dynamic multitalker environment

Min Kyung Hyun. 1. Introduction. 2. Methods

STAT 200. Guided Exercise 7

The Model and Analysis of Conformity in Opinion Formation

State-Trace Analysis of the Face Inversion Effect

GRUNDFOS DATA BOOKLET. Hydro Grundfos Hydro 2000 booster sets with 2 to 6 CR(E) pumps 50 Hz

Research Article ABSTRACT. Amanda Myhren-Bennett College of Nursing, University of South Carolina, Columbia, SC 29208, USA

Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories

Dental X-rays and Risk of Meningioma: Anatomy of a Case-Control Study

A model of HIV drug resistance driven by heterogeneities in host immunity and adherence patterns

Bayesian and Frequentist Approaches

MCAS Equating Research Report: An Investigation of FCIP-1, FCIP-2, and Stocking and. Lord Equating Methods 1,2

Randomized controlled trials: who fails run-in?

Title: Correlates of quality of life of overweight and obese patients: a pharmacy-based cross-sectional survey

Polymorbidity in diabetes in older people: consequences for care and vocational training

Research. Dental Hygienist Attitudes toward Providing Care for the Underserved Population. Introduction. Abstract. Lynn A.

Performance of the Vocal Source Related Features from the Linear Prediction Residual Signal in Speech Emotion Recognition

Identification and low-complexity regime-switching insulin control of type I diabetic patients

Contents. What is item analysis in general? Psy 427 Cal State Northridge Andrew Ainsworth, PhD

An Intuitive Approach to Understanding the Attributable Fraction of Disease Due to a Risk Factor: The Case of Smoking

André Cyr and Alexander Davies

An Algorithm for Probabilistic Least{Commitment Planning 3. Nicholas Kushmerick Steve Hanks Daniel Weld. University ofwashington Seattle, WA 98195

Origins of Hereditary Science

Running head: NESTED FACTOR ANALYTIC MODEL COMPARISON 1. John M. Clark III. Pearson. Author Note

University of Groningen. Variations in working memory capacity Gulbinaite, Rasa

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison

Prefrontal cortex fmri signal changes are correlated with working memory load

Comparison of Water Seal and Suction After Pulmonary Lobectomy: A Prospective, Randomized Trial

Treating Patients with HIV and Hepatitis B and C Infections: Croatian Dental Students Knowledge, Attitudes, and Risk Perceptions

CRIKEY - A Temporal Planner Looking at the Integration of Scheduling and Planning

Syncope in Children and Adolescents

Annie Quick and Saamah Abdallah, New Economics Foundation

Research Article Deep Learning Based Syndrome Diagnosis of Chronic Gastritis

Quadriceps muscles vastus medialis obliques, rectus femoris and vastus lateralis compared via electromyogram bicoherence analysis

Does Job Strain Increase the Risk for Coronary Heart Disease or Death in Men and Women?

The Mississippi Social Climate of Tobacco Control,

Brain regions supporting intentional and incidental memory: a PET study

Research Article Effects of Pectus Excavatum on the Spine of Pectus Excavatum Patients with Scoliosis

THE QUANTITATIVE GENETICS OF HETEROSIS. Kendall R. Lamkey and Jode W. Edwards INTRODUCTION

Numerous internal fixation techniques have been

The Mississippi Social Climate of Tobacco Control,

Migraine headache is one of the most debilitating RECONSTRUCTIVE

Connexion of Item Response Theory to Decision Making in Chess. Presented by Tamal Biswas Research Advised by Dr. Kenneth Regan

Chapter Forty-Six 46 Concordance Marjorie C. Weiss STUDY POINTS * What is meant and understood by the term concordance * How concordance differs from

King s Research Portal

Factors Affecting the Item Parameter Estimation and Classification Accuracy of the DINA Model

Comparative judgments of animal intelligence and pleasantness

Likelihood Ratio Based Computerized Classification Testing. Nathan A. Thompson. Assessment Systems Corporation & University of Cincinnati.

A modular neural-network model of the basal ganglia s role in learning and selecting motor behaviours

TO help 25.8 million Americans [1] with diabetes, a growing

University of Alberta

Chapter 11 Multiway Search Trees

Differences in the local and national prevalences of chronic kidney disease based on annual health check program data

Computerized semen analysis: objective measurement of semen characteristics is biased by subjective parameter setting

Item Response Theory (IRT): A Modern Statistical Theory for Solving Measurement Problem in 21st Century

Online publication date: 01 October 2010

Mendel Laid the Groundwork for Genetics Traits Genetics Gregor Mendel Mendel s Data Revealed Patterns of Inheritance Experimental Design purebred

Severe Psychiatric Disorders in Mid-Life and Risk of Dementia in Late- Life (Age Years): A Population Based Case-Control Study

A Brief Introduction to Bayesian Statistics

IN a recent article (Iwasa and Pomiankowski 1999),

Amino acids in retinitis pigmentosa

Does factor indeterminacy matter in multi-dimensional item response theory?

Relating mean blood glucose and glucose variability to the risk of multiple episodes of hypoglycaemia in type 1 diabetes

A Comparison of DIMTEST and Generalized Dimensionality Discrepancy. Approaches to Assessing Dimensionality in Item Response Theory. Ray E.

Getting to Goal: Managed Care Strategies for Children, Adolescents, and Adults With ADHD

A Comparison of Methods of Estimating Subscale Scores for Mixed-Format Tests

Chapter 1 Introduction. Measurement Theory. broadest sense and not, as it is sometimes used, as a proxy for deterministic models.

Influences of IRT Item Attributes on Angoff Rater Judgments

The dynamic evolution of the power exponent in a universal growth model of tumors

Transcription:

The Alication of a Cognitive Diagnosis Model via an Analysis of a Large-Scale Assessment and a Comuterized Adative Testing Administration by Meghan Kathleen McGlohen, B.S., M. A. Dissertation Presented to the Faculty of the Graduate School of The University of Texas at Austin In Partial Fulfillment Of the Requirements For the Degree of Doctor of Philosohy The University of Texas Austin May 2004 v

The Alication of a Cognitive Diagnosis Model via an Analysis of a Large-Scale Assessment and a Comuterized Adative Testing Administration Publication No. Meghan Kathleen McGlohen, Ph.D. The University of Texas at Austin, 2004 Suervisor: Hua Hua Chang Our society currently relies heavily on test scores to measure individual rogress, but tyical scores can only rovide a limited amount of information. For instance, a test score does not reveal which of the assessed toics were mastered and which were not well understood. According to the U.S. government, this is no longer sufficient. The No Child Left Behind Act of 2001 calls for diagnostic information to be rovided for each individual student, along with information for the arents, teachers, and rincials to use in addressing individual student needs. This oens the door for a new vi

area of sychometrics that focuses on the inclusion of diagnostic feedback in traditional standardized testing. This diagnostic assessment could even be combined with techniques already develoed in the arena of comuter adative testing to individualize the assessment rocess and rovide immediate feedback to individual students. This dissertation is comrised of two major comonents. First, a cognitive diagnosis-based model, namely the fusion model, is alied to two large-scale mandated tests administered by the Texas Education Agency; and secondly, comuter adative testing technology is incororated into the diagnostic assessment rocess as a way to develo a method of roviding interactive assessment and feedback for individual examinees mastery levels of the cognitive skills of interest. The first art requires attribute assignment of the standardized test items and the simultaneous IRT-based estimation of both the item arameters and the examinee variables under the fusion model. Examinees are classified dichotomously into mastery and non-mastery categories for the assigned attributes. Given this information, it is ossible to identify the attributes with which a given student needs additional hel. The second art focuses on alying CAT-based methodology, and in articular item selection, to the diagnostic testing rocess to form a dynamic test that is sensitive to individual resonse atterns while the examinee is being administered the test. This combination of comuter adative testing with diagnostic testing will contribute to the research field by enhancing the results that students and their arents and teachers receive from educational measurement. vii

CHAPTER ONE: INTRODUCTION Tyically, large-scale standardized assessments rovide a single summary score to reflect the overall erformance level of the examinee in a certain content area. The utility of large-scale standardized assessment would be enhanced if the assessment also rovided students and their teachers with useful diagnostic information in addition to the single overall score. Currently, smaller-scale assessments, such as teacher-made tests, are the means of roviding such helful feedback to students throughout the school year. Negligible concern is exressed about the considerable classroom time that is taken by the administration of these formative teacher-made tests because they are viewed as integral arts of instruction. Conversely, educators view standardized testing of any kind as lost instruction time (Linn, 1990). Some advantages of standardized tests over teacher-made tests are that they allow for the comarison of individuals across various educational settings, they are more reliable, and they are objective and equitable (Linn, 1990). The advantage of teacher-made tests, on the other hand, is that they rovide very secific information to the students regarding their strengths and weaknesses in the tested material. Large-scale standardized testing would be even more beneficial if it could also contribute to the educational rocess in a role beyond that of evaluation while maintaining these existing advantages, such as the reorting of diagnostic feedback. Then the students could use this information to target imrovement in areas where they are deficient. A new aroach to educational research has begun to effloresce in order to rovide the best of both worlds. This research area, dealing with the alication of 1

cognitive diagnosis in the assessment rocess, aims to rovide helful information to arents, teachers, and students, which can be used to direct additional instruction and study to the areas needed most by the individual student. This beneficial information rovided by diagnostic assessment deals with the fundamental elements or buildingblocks of the content area. The combination of these elements or attributes comrises the content domain of interest. This form of diagnostic assessment is an aroriate aroach to conducting formative assessment, because it rovides secific information regarding each measured attribute or content element to every examinee, rather than a single score result. An ideal assessment would not only be able to meet the meticulous sychometric standards of current large-scale assessments, but would also be able to rovide secific diagnostic information regarding the individual examinee s educational needs. In fact, the rovision of this additional diagnostic information by large-scale state assessments has recently become a requirement; the No Child Left Behind Act of 2001 mandates that such feedback be rovided to arents, teachers and students. Desite this requirement, constructing diagnostic assessment from scratch is exensive and imractical. A more affordable solution is to incororate diagnostic measurement into existing assessments that state and local governments are already administering to ublic school students. So, in order to incororate the benefits of diagnostic testing with the current assessment situation, cognitively diagnostic aroaches would need to be alied to an existing test. Diagnostic assessment is a very advantageous aroach to measurement. In traditional testing, different students may get the same score for different reasons 2

(Tatsuoka M. M. and Tatsuoka, K. K., 1989), but in diagnostic testing, these differences can be discovered and shared with the examinee and his/her teacher. Diagnostic assessment allows the testing rocess to serve an additional instructional urose in addition to the traditional uroses of assessment (Linn, 1990), and can be used to integrate instruction and assessment (Camione and Brown, 1990). Furthermore, diagnostic testing offers a means of selecting instructional material according to an individual s needs (Embretson, 1990). While traditional tests can accomlish assessment goals, such as a ranked comarison of examinees or grade assignments based on certain criteria, they do not rovide individualized information to teachers or test-takers regarding secific content in the domain of interest (Chiman, Nichols, and Brennan, 1995). Traditional assessment determines what an individual has learned, but not what s/he has the caacity to learn (Embretson, 1990). Diagnostic assessment can be alied to areas involving the identification of individuals who are likely to exerience difficulties in a given content domain, and it can hel rovide secific information regarding the kinds of hel an individual needs. Furthermore, the cognitive diagnosis can be used to gauge an individual s readiness to move on to higher levels of understanding and skill in the given content domain. (Gott, 1990,. 174). Current aroaches dealing with cognitive diagnosis focus solely on the estimation of the knowledge state, or attribute vector, of the examinees. This dissertation rooses the combination of the estimation of item resonse theory (IRT)-based individual ability levels (θˆ ) along with an emhasis on the diagnostic feedback rovided 3

by individual attribute vectors ( ˆ α ), thus bridging the current standard in testing technology with a new area of research aimed at heling students benefit from the testing rocess through diagnostic feedback. The goal of this research is to not only simultaneously measure individuals knowledge states and conventional unidimensional IRT ability levels, but to do so in an efficient way. This research will aly the advantages of comuterized adative testing to the new measurement area of cognitive diagnosis. The goal of comuterized adative testing is to tailor a test to each individual examinee by allowing the test to hone in on the examinees ability levels in an interactive manner. Accordingly, examinees are relieved from answering many items that are not reresentative of their abilities. To accomlish this goal, this research study relies on the technology of comuter adative testing. The aim of this research is to combine the advantages of comuterized adative testing with the helful feedback rovided by cognitively diagnostic assessment, to enhance the existing testing rocess. This dissertation rooses a customized diagnostic testing rocedure that rovides both conventional unidimesional ability estimates as well as a reort of attribute mastery status to the examinees, instead of just one or the other. The key idea of this work is to utilize the shadow test technique to otimize the estimation of the traditional IRT-based ability level, θˆ, and then select an item from this shadow test that is otimal for the cognitive attribute vector, ˆ α, for each examinee. But first, the founding concets of this research must be elucidated. 4

CHAPTER TWO: LITERATURE REVIEW Traditional IRT-Based Testing Item resonse theory (IRT) is a common foundation for wide-scale testing. IRT is based on the idea of test homogeneity (Loevinger, 1947) and logistic statistical modeling (Birnbaum, 1968), and uses these robabilistic models to describe the relationshi between item resonse atterns and underlying arameters. IRT uses the item as the unit of measure (rather than the entire test) to obtain ability scores that are on the same scale desite the differences in item administration across examinees (Wainer and Mislevy, 2000). As outlines in Rogers, Swaminathan and Hambleton s (1991) text, two main axioms are assumed when emloying IRT: (1) The erformance of an individual on a set of test items can be rationalized by an underlying construct, latent trait, or set thereof. The context of educational testing uses individual ability levels as the trait, which accounts for correct/incorrect resonse atterns to the test items. (2) The interconnection between this erformance and the intrinsic trait or set of traits can be reresented by a monotonically increasing function. That is to say, as the level of ability or trait of interest increases, the robability of a resonse reflecting this increase (secifically for this context, a correct resonse) also increases (Rogers, Swaminathan and Hambleton, 1991). A lethora of ossible IRT robability models are available, and a few of the most common will be briefly discussed. Each of these three models mas out a robabilistic 5

association between the items and the ability level of the examinee (j), denoted as θ j. These three models are resectively referred to as the one-, two-, and three-arameter logistic models. The one-arameter logistic model considers the difficulty level of the items on the test, denoted as b i for each item i. As the name suggests, item with a higher degree of difficult are harder to resond correctly to, and hence call on a higher level of the ability trait θ to do so. The robability of obtaining a correct resonse to item i given a secific ability level θ j is shown in Equation 1: ( ) = e θ j bi P ij Y ij = 1θ j ( ) ( 1+ e θ j b i ) (1) for i = 1, 2,, n, where n is the total number of items, and Y i denotes the resonse to item i (Rogers, Swaminathan and Hambleton, 1991). The one-arameter logistic model is also referred to as the Rasch model. Next, the two-arameter logistic model also involves this item difficulty arameter, b i, but includes another item arameter which deals with the item discrimination, denoted as a i. Item discrimination reflects an item s facility in discriminating between examinees of differing ability levels. The value of the item discrimination arameter is roortional to the sloe of the robability function at the location of b i on the ability axis (Rogers, Swaminathan and Hambleton, 1991). The robability of obtaining a correct resonse to item i given an ability level θ j, is shown in Equation 2, as described by the two-arameter logistic model: ( ) = edai θ j bi P i Y ij = 1θ j ( ) 1+ e Da i ( θ j b i ) (2) 6

for i = 1, 2,, n, and where D is a scaling constant, and n and Y i hold the same meaning as in the revious model (Rogers, Swaminathan and Hambleton, 1991). Finally, the three-arameter logistic model (3PL) includes both the difficulty level b i and discrimination arameter a i, but also adds a third item arameter, called the seudo-chance level, denoted as c i (Rogers, Swaminathan and Hambleton, 1991). The seudo-chance level allows for the instance where the lower asymtote of the robability function is greater than zero; that is to say, the examinees have some non-zero robability of resonding correctly to the item regardless of ability level. The 3PL robability of obtaining a correct resonse to item i given a θ j level of ability is shown in Equation 3: ( ) = c i + (1 c i ) P ij Y ij = 1θ j e Da i ( θ j b i ) 1 + e Da i ( θ j b i ) (3) for i = 1, 2,, n, and all variables are defined as reviously noted (Rogers, Swaminathan and Hambleton, 1991). Notice the similarities between these models, and in articular, how each is based on the revious. In fact, the latter two can simlify to the other one(s) by setting c i to zero (for the two-arameter logistic model) or by setting a i equal to one and c i to zero (for the one-arameter logistic model). These are just a few of the ossible model available for describing the relationshi between item resonse atterns and the latent individual ability levels and item arameters. The next ortion of this work will discuss a different aroach to testing called diagnostic assessment, and will eventually lead to a discussion about how to combine these long-established IRT aroaches with new diagnostic assessment techniques. 7

Diagnostic Assessment Models The focus of cognitive diagnosis is to rovide individual feedback to examinees regarding each of the attributes measured by the assessment. An attribute is defined as a task, subtask, cognitive rocess, or skill somehow involved in the measure (Tatsuoka, 1995,.330). But measuring individuals with resect to the attributes is not the only requirement of a good cognitive diagnosis model. A model must also allow the items to be examined in the context of the cognitive diagnosis, or else the results from the assessment cannot be standardized or understood in a larger testing framework (Hartz, 2002). In fact, Hartz, Roussos, and Stout (2002) describe three desirable characteristics of an ideal cognitive diagnosis model as the ability to (1) Allow the attributes to be araised with resect to individual examinees, (2) Allow the relationshi between the items and the attributes to be evaluated, (3) Statistically estimate the arameters involved in the model. There are at least fourteen models for diagnostic assessment, most of which fall under two major branches. The first deals with models based on Fischer s (1973) Linear Logistic Test Model (LLTM). The second branch of the cognitive diagnosis literature deals with models that emloy Tatsuoka s and Tatsuoka s (1982) Rule Sace methodology as a foundation. Each model has its strengths and weaknesses. A handful of the models will be described, followed by a justification of the model of choice for this research study. 8

Fischer s LLTM One of the first cognitive diagnosis models was the LLTM, develoed by Fischer in 1973. The LLTM is an exansion of the basic Rasch model to take into account the cognitive oerations required for correct item resonses. To do this, the Rasch item difficulty is artitioned into discrete cognitive attribute-based difficulties. Hence, the Rasch item difficulty (denoted as σ i ) equals the weighted sum of these attribute-based difficulties, as shown in Equation (4): σ = f η + c (4) i k where f ik is the weight of factor k in item i, η k is the difficulty arameter (or effect ) of factor k across the entire exam, and c is a normalizing constant (Fischer, 1973). The weight of factor k in item i, denoted as f ik, indicates the extent to which factor k is required by item i. (Please note that this concet is analogous to a Q-matrix entry, which is discussed in further detail subsequently.) Substituting this for the item difficulty arameter in the Rasch model yields the LLTM, as resented in Equation (5): P ( X = 1 ) ij 1+ e where X ij equals one when examinee j resonds correctly to item i and equals zero otherwise. The key idea in the LLTM that makes it alicable in the context of diagnostic assessment is that the item difficulty is comrised of the comosite of the influences of the basic cognitive oerations, or factors, necessary for correctly solving an item. These oerations can be thought of as cognitive building blocks or attributes. 9 ik k 1 θ = (5) j θ ( f ik ηk + c) k

The erson arameter, however, remains a single unidimensional ability arameter (θ j ), without any sort of attribute-secific estimate for individual examinees. The LLTM lays the foundation for exloring the cognitively diagnostic relationshi between items and their underlying attributes, but it does not incororate a measure to identify the resence or absence of such attributes in individual examinees. Rule Sace Methodology The rule sace methodology was develoed by Kikumi Tatsuoka and her associates (1982), and is comrised of two arts. The first art involves determining the relationshi between the items on a test and the attributes that they are measuring. Each examinee may or may not hold a mastery-level understanding of each attribute, and in fact, may hold a mastery-level understanding of any combination thereof. The combinations of attributes which are mastered and not mastered by an individual examinee are deicted in an attribute vector, which is also referred to as a knowledge state. The descrition of which items measure which attributes is illustrated in a Q- matrix. A Q-matrix is sometimes referred to as an incidence matrix, but not in this text. The Q-matrix is a [K x n] matrix of ones and zeros, where K is the number of attributers to be measured and n is the number of items on the exam. For a given element of the Q- matrix in the k th row and the i th column, a one indicates that item i does indeed measure attribute k and a zero indicates it does not. For examle, notice the following [3 x 4] Q- matrix: 10

i1 i2 i3 i4 A1 0 1 0 1 Q = A2 1 0 0 1 A3 1 0 1 0 The first item measures the second and third attributes, while the second item measures the first attribute only. The third item only measures the third attribute, while the fourth item measures the first two attributes. The consultation of exerts in the measured content domain is a good aroach to constructing the Q-matrix to determine if an item measures a articular attribute. Other ossible aroaches to constructing Q- matrices include borrowing from the test bluerint or intuitively evaluating each item to infer which attributes are being assessed. Each element in the Q-matrix is denoted as q ik where the subscrits i and k denote the item and attribute of interest, resectively. This q ik reresentation mirrors the f ik, arameter in Fischer s LLTM (symbolizing the weight of factor k in item I, as discussed on age 9 of this chater). They are not exactly the same, however, because Fischer s f ik, weight can take on values greater than unity, while the Q-matrix entry q ik cannot. They are synonymous, however, when all of the f ik weights are dichotomized. Next, the information rovided by the Q-matrix needs to be translated into a form that can be comared with individual observed item resonse atterns. An observed item resonse attern is a vector of ones and zeros that reresents how an individual erformed on a test. For examle, a resonse attern of [ 0 1 0 1 1 1 ] indicates that the individual resonded incorrectly to the first and the third items on the test, but answered each of the other four items correctly. This comarison between the Q-matrix and 11

individual observed item resonse atterns is accomlished by establishing a series of ideal resonse atterns. An ideal resonse attern is a resonse attern that is obtained through a articular hyothetical combination of mastery and non-mastery levels of the attributes (Tatsuoka, 1995). Notice that the word ideal does not indicate a erfect resonse attern, but rather suggests erfect fit with an underlying theory, (Tatsuoka, 1995,. 339). These ideal resonse atterns are ascertained systematically by using rules. A rule is defined as a descrition of a set of rocedures or oerations that one can use in solving a roblem in some well-defined rocedural domain, (Tatsuoka & Tatsuoka, 1987,.194). Rules are determined through logical task analysis (Tatsuoka, 1990) and deterministic methods used in artificial intelligence (Tatsuoka and Tatsuoka, 1987). A comuter rogram is used to generate the ossible ideal resonse atterns that would be obtained from the alication of a variety of rules. Indubitably, both correct rules and erroneous rules exist for any given assessment. The consistent alication of all of the correct rules would result in correct answers for the entire test. The consistent alication of some correct rules with other incorrect rules would result in a secific resonse attern including both ones and zeros. If a student consistently alies a secific combination of rules to all of the items in a test, then his/her resonse attern would match exactly the ideal resonse attern roduced by the comuter rogram for that exact combination of rules (Tatsuoka, 1995). This rovides a finite number of ideal resonse atterns with which the observed resonse atterns can be comared. 12

In order for an examinee to get an item right, s/he must ossess all of the attributes that the item measures. This is analogous to electrical circuitry: a current can only flow when all switches are closed. In this analogy, a closed switch symbolizes a mastered attribute and an electrical current reresents a correct resonse to an item. A correct resonse to an item can only be obtained by an examinee if all attributes involved in a give item are mastered (Tatsuoka and Tatsuoka, 1997; Tatsuoka, 1995). Boolean algebra can be used to exlain the cognitive requirements for item resonse atterns (Birenbaum and Tatsuoka, 1993). Boolean algebra is a mathematical disciline dealing with sets or collections of objects, such as attributes, and is commonly used in dealing with circuits. (For a detailed descrition of Boolean algebra, lease see Whitesitt, 1995). Furthermore, one secific feature of Boolean algebra, called Boolean descritive functions, can be used to ma the relationshi between the attributes and item resonse atterns (Tatsuoka, 1995). Consequently, the various attribute vectors, also known as a knowledge states, can be determined from the set of erroneous rules used to find the ideal resonse attern that corresonds with an individual s observed resonse attern. For instance, if an individual taking a math test knows and follows all of the alicable rules correctly excet that s/he does not know how to borrow in subtraction, then s/he would get every item right excet those that involve borrowing. Thus, his/her attribute vector would include a value of unity for all the attributes excet the one(s) that involve borrowing, which would be reresented in the attribute vector by a zero. Then diagnostic information would be rovided that exlains this individual needs more instruction in the area of borrowing in subtraction. 13

Unfortunately, this scenario does not adequately reflect the reality of a testing situation in its entirety. When an examinee alies an erroneous rule, s/he most likely does not aly the same erroneous rule consistently over the entire test (Birenbaum and Tatsuoka, 1993). Consequently, this inconsistency results in deviations of the observed resonse attern from the ideal resonse attern. Moreover, careless errors, uncertainty, fatigue, and/or a temorary lase in thinking can result in even more deviations of an observed resonse attern from the ideal resonse atterns. Tatsuoka (1990) states, Even if a student ossesses some systematic error [i.e. is following an erroneous rule], it is very rare to have the resonse attern erfectly matched with the atterns theoretically generated by its algorithm, (. 462). In fact, the total number of ossible resonse atterns is an exonential function of the number of items. For n items, 2 n ossible resonse atterns exist. Fortunately, Boolean algebra can be utilized to hel reduce the overwhelming number of ossible item resonse atterns to a quantity more comutationally manageable. These random errors made by examinees, referred to as slis, can be statistically thought of in terms of the robability of a sli occurring or even the robability of having any number of slis occur. Tatsuoka and Tatsuoka (1987) derive a theoretical distribution of the slis, and call it the bug distribution. The bug distribution calculates the robability of making u to s number of slis by multilying the robabilities of a sli occurring for all items where one did occur and the robabilities of a sli not occurring for all items where one did not occur, and summing across all s 14

slis for all ossible combinations of s slis. This robability distribution for a given rule R is resented in Equation (6): S n s= 0 Σ u = s i= 1 i ui i ui ( 1 i ) (6) with a mean of µ R = t i= 1 i + n q i i= t+ 1 and a variance of σ 2 R = n i= 1 q i i where the number of slis, denoted as s, ranges from one to S, and u i equals unity when a sli occurs on item i and is assigned a zero when it does not. Also, n is the number of items and t denotes the total score (Tatsuoka and Tatsuoka, 1987; Tatsuoka, 1995). Conveniently, resonse atterns that result from inconsistencies in rule alications cluster around the corresonding ideal resonse atterns; however, the resonse variance that makes u these clusters comlicates classification of the individual examinees because it can no longer be achieved by matching u observed resonse vectors exactly with the ideal resonse vectors. Due to the variability of ossible item resonse atterns, a method is needed for classifying individuals into knowledge states (i.e. identifying their attribute vector) when the resonse atterns do not reflect exactly an ideal resonse attern. To address this, the second art of the rule sace methodology deals with the construction of an organized sace for classifying the examinees resonse atterns into the knowledge state categories that are established in the first art of the methodology. 15

Once all ossible knowledge states are generated from the Q-matrix, item resonse theory is used to construct this classification sace for identifying each of the examinees into one of the redetermined knowledge states. Unfortunately, the attribute variables are unobservable constructs and cannot be reresented in such a sace directly. Instead, the item resonse theory functions are used in combination with a new arameter, ζ (zeta), develoed by Tatsuoka (1984) to measure the atyicality of the resonse atterns. Values of ζ are calculated as the standardized roduct of two residual matrices (where residuals are calculated as the difference between an observed and exected value and standardization is achieved by dividing the roduct by the standard deviation of its distribution). This arameter ζ can be deicted with the item resonse theory erson arameter θ, which symbolizes ability level, as axes in a Cartesian sace in which the knowledge states for each of the ideal resonse atterns can be grahically reresented (Tatsuoka, 1990), as demonstrated by Figure 1. ζ B A θ Figure 1: Two knowledge states in the two-dimensional rule sace. 16

Notice that knowledge state A is farther left on the θ (ability) scale than knowledge state B. This means that knowledge state B requires a higher level of ability to acquire than knowledge state A. Also notice that knowledge state A is much closer to the θ axis (where the value of ζ, the atyicality index, is zero), which means that this knowledge state occurs more frequently than B. Conversely, knowledge state B is farther away from the θ axis (i.e. the magnitude of ζ is greater), which means that it is a more atyical knowledge state. Likewise, the various observed resonse atterns can be reresented in terms of each attern s estimated ζ and θ levels. Furthermore, this Cartesian sace can be used to determine which ideal resonse attern an observed resonse attern is closest to (Birenbaum and Tatsuoka, 1993). Closeness between ideal and observed item resonse atterns can be aroximated by a distance measure, such as the Mahalanobis distance; this metric describes the distance between an observation and the centroid of all observations (Stevens, 1996). Also, Bayes decision rules may be used to minimize misclassifications (Birenbaum and Tatsuoka, 1993). Bayes decision rules simly use robabilities to determine which category is most likely when uncertainty is resent. The categories are then used to determine an individual s attribute mastery attern. Once this classification has been carried out, one can indicate with a secified robability level which attributes a given examinee is likely to have mastered, (Birenbaum and Tatsuoka, 1993,. 258). Attribute mastery atterns are reresented as a vector of zeros and ones, where a one signifies mastery of a given attribute and zeros signify non-mastery. An attribute 17

vector, containing k (the total number of attributes) elements, is estimated for each individual and denoted as α. For examle, if a test measures three attributes, the estimated attribute vector for a articular examinee may be α j = [ 0 1 1 ], meaning s/he has mastered the second and third attributes, but not the first. Attribute vectors can be used to rovide helful diagnostic information that secifies the area(s) each examinee needs hel in. For this examle examinee, the diagnostic information would include an indication that the student needs to work on the educational construct corresonding to the first attribute. The rule sace methodology allows for grahical reresentation of more concets than just the rule sace (as deicted in Figure 1). Other imortant features of this model can be illustrated visually, such as robability curves, influence diagrams, and knowledge state trees. An influence diagram dislays which items measure which attributes and which attributes influence other attributes. In Figure 2, the influence diagram reresents the item attribute relationshis as described by the Q-matrix as well as the relationshi between the attributes. 18

A2 A1 A4 A3 Figure 2: Nine items measuring four attributes. In this figure, an arrow indicates an item measures a certain attribute, with the arrowhead facing the item. In this examle, four items, numbers 3, 5, 6, and 9, measure two attributes each, while the remaining items measure one attribute each. Also, the influence of attribute 1 on attribute 3 is reresented by an arrow as well. This indicates that attribute 1 is a rerequisite for attribute 3. The rule sace methodology can also be used to construct a knowledge state tree. A knowledge state tree is a very imortant grahical reresentation because it ortrays the incremental relationshis between the knowledge states. This is articularly useful 19

because it draws out how to imrove from one knowledge state to a more mastered knowledge state (Tatsuoka and Tatsuoka, 1997). For examle, see Figure 3. θ = +3 Mastery State Cannot do A3 Cannot do A4 3.4 2.0 4.9 θ = 0 3.8 Cannot do A3 & A4 2.6 3.1 Cannot do A2 & A4 Cannot do A1, A3 & A4 Cannot do A2, A3 & A4 1.9 2.2 θ = -3 Cannot do A1, A2, A3, &A4 Figure 3: A tree reresenting eight knowledge states. 20

In traditional testing, higher scores would be assigned to individuals in the knowledge states that are higher in the figure. Knowledge states lower in the figure reresent lower traditional scores. Notice the knowledge states Cannot do A3 and Cannot do A4 are next to each other. In a traditional testing situation, individuals within both of these knowledge states would receive the same score. Alternatively, the rule sace method allowed these examinees to be rovided more secific information about their abilities. Obviously, it is ossible for an examinee to lack more than one attribute. This tree is useful because it shows the order in which the non-mastered attributes need to be learned. For instance, if a student cannot do attribute 1, 3, or 4 (i.e. his/her attribute vector is [0, 1, 0, 0]), then the diagram dictates that s/he should next move to the knowledge state Cannot do A3 & A4. Therefore, the student should learn attribute 1 next. Then there is a choice to move to either state Cannot do A3 or cannot do A4. In such a situation, the next aroriate knowledge state is the one with the shortest Mahalanobis distance (Tatsuoka and Tatsuoka, 1997). In this case, the next knowledge state to achieve would be cannot do A3 with a Mahalanobis distance 3.4; so the student needs to learn attribute 4. Then, once attribute 3 is mastered, the student will accomlish the knowledge state total mastery. Tatsuoka and Tatsuoka (1997) suggest rogramming this method of determining an aroriate ath for instruction in the form of a comuter adative tutor for remedial instruction. This is revolutionary because it embarks on the combination of the rule sace method of diagnostic assessment with comuter adative 21

assessment to rovide immediate feedback and immediate individualized instruction in the area(s) the examinee needs most. The rule sace methodology is revolutionary and advantageous in many ways, but there is always room for imrovement. One drawback is that the rule sace methodology does not take into account a way to evaluate the relationshis between the items and the attributes. The Unified Model DiBello, Stout, and Rousses (1995) based their new model, named the unified model, on the rule sace method. They attemted to imrove on one of the underlying ideas of the rule sace aroach. In the unified model, the source of random error is broken down into different four tyes of systematic error. In the rule sace model, all of these would be considered random slis. They examined the ossible sources of random errors and categorized them into four grous. Hence, while there is only one tye of random error in the rule sace model (slis) there are four sources of aberrant resonse variance in the unified model. First, they exlain that strategy selection is a source of resonse variation, because an examinee may answer an item using a different strategy than the one assigned in the Q-matrix. Second, comleteness of the Q-matrix is considered an imortant issue. An item may measure an attribute that is not listed in the Q-matrix. For examle, a worded math roblem includes a verbal comonent; if the Q-matrix does not contain a verbal attribute, then the Q-matrix would be considered incomlete. Third, the unified model takes 22

ositivity into account. Positivity addresses inconsistencies that arise when students who do not ossess a certain attribute haen to resond correctly to an item that measures the attribute, or when students who do ossess a certain attribute do not aly it correctly and resond erringly to an item measuring the ossessed attribute. Positivity takes on a high value when individuals who ossess an attribute use it correctly, while students who lack an attribute miss the items that measure it. The less this is the case, the lower the value of ositivity. Lastly, a category remains for random errors that are not caused by any of these other three issues. These are called slis and include mental glitches such as finding the correct solution to a roblem and then bubbling in a different multile-choice otion. Notice the term slis is used more generally for the rule sace aroach than the unified model. DiBello, Stout, and Rousses (1995) introduce a new arameter for dealing with the issues of strategy choice and incomleteness of the Q-matrix called the latent residual ability (confusingly, this is denoted as θ j, but is different than the IRT ability level θ j). The unified model is the first to include such a arameter. The latent ability sace consists of α Q, which is the art addressed in the Q-matrix, and α b, which is the remaining latent ability not included in α Q. This arameter θ j is intended to measure underlying construct α b, while α Q is measured by the arameter α j. One might ask, why not simly add more attributes to the Q-matrix to account for these issues? More attributes mean more attribute arameters. While additional arameters may allow for enhanced distinctions and alleviate certain classification roblems caused by strategy choice and incomleteness of the Q-matrix, these added 23

arameters would most likely comlicate the measurement rocess more than they would benefit it. More arameters require a greater degree of comlexity in the estimation rocedure. Also, an increase in the number of attribute arameters to be estimated requires an increase in the number of items on the test to obtain accetable reliability (DiBello, Stout, and Rousses, 1995). But this may not be ractical when so many educators feel test administration is already too long. For the sake of arsimony, including additional arameters is only advantageous when there is a real measurement interest in assessing the mastery/non-mastery of those attributes. Hence, the inclusion of these issues in a model without having to add more attribute arameters is otimal, and this is what DiBello, Stout, and Roussos (1995) have accomlished. The unified model is illustrated in Equation (7): P K α jk qik (1 α jk ) qik ( X 1α, θ ) = d π r P ( θ ) + (1 d ) P ( θ ) i j j = (7) i k= 1 ik ik where α jk denotes examinee j s mastery of attribute k, with a one indicating mastery and a zero denoting non-mastery. Also, q ik is the Q-matrix entry for item i and attribute k, andθ j is the latent residual ability and P(θ j ) is the Rasch model with the item difficulty arameter secified by the subscrit of P. The arameter π ik is the robability that erson j will correctly aly attribute k to item i given that erson j does indeed osses attribute k; mathematically, this is written as π P( = 1α = 1) with Y ijk equaling unity when ik = Y ijk jk correct alication of the attribute is resent. Lastly d i is the robability of selecting the Q-based strategy over all other ossible strategies. c i j i b i j 24

The unified model includes a large number of arameters to deal with a lethora of sychometric elements. Having such an amlified set of arameters in the model is both a blessing and a curse. It is a recarious balance that must be met between imroving accuracy through the inclusion of more arameters and the issue of statistical identifiability of the many arameters in the model. Jiang (1996) demonstrated that, in fact, the 2K i +3 item arameters contained in the unified model are just too many to be uniquely identified. This model is named the unified model because it uses a deterministic aroach to estimating knowledge state classification, and yet it also takes into account random errors. Hence, it unifies both deterministic and stochastic aroaches. The unified model is advantageous in that it takes in to account the necessity for assessing examinees with resect to underlying attributes, as well as the requirement for examining the relationshi between the items and the attributes rather than just one or the other. Some other advantages of the unified model include the innovative use of the latent residual ability θ j to hel avoid the roblems associated with too many latent classes being incororated into the assessment rocess (a suitable alternative to the addition of suerfluous attribute arameters) and the ability for the model to handle the use of multile solution strategies by examinees (DiBello, Stout, and Rousses, 1995). In order to aly the unified model in diagnostic assessment, the item arameters must be estimated, which is to say, the model must be calibrated. However, the model lacks racticality in this sense because the arameters involved are not uniquely statistically estimable (Jiang, 1996). 25

The Fusion Model The fusion model was based on the unified model (Hartz, Roussos, and Stout, 2002) which, in turn, was based on Tatsuoka s rule sace methodology (DiBello, Stout, and Roussos, 1995). The fusion model retains the advantages of the unified model while reducing the number of arameters involved so that they are statistically identifiable. The unified model has 2K i +3 arameters for each item, while the fusion model only has K+1 (where K is the number of attributes). Fischer s LLTM does not estimate individuals mastery level of each attribute. The Rule Sace model does not evaluate the relationshi between the items and the attributes. The Unified Model satisfies both of these needs, but does not have statistically estimable arameters. The fusion model simlifies the unified model so that the arameters may be estimated. The fusion model was selected as the cognitive analysis model of choice for this study due to the fact that it includes all three features described by Hartz, et al. (2002) as crucial for a successful cognitive diagnosis model, including (1) the estimation of examinees attribute mastery levels, (2) the ability to relate the items to the attributes, and (3) statistical identifiability of the model's arameters. The item resonse function for the fusion model is illustrated below in Equation (8), as described by Hartz, et al. (2002): where ( ) = π i * P X ij = 1α j,θ j K * r 1 α jk ) q ik P θ ik ( ) ci j (8) k=1 P ci ( θ j ) = The Rasch model with difficulty arameter i c. 26

K π * i = P( Yijk = 1α k= 1 jk = 1) q ik * r ik = P( Y ijk P( Y ijk = 1α = 1α jk jk = 0) = 1) Y ijk = 1 when examinee j correctly alies attribute k to item i, and 0 otherwise. α jk = 1 when examinee j has mastered attribute k, and 0 otherwise. c i = the amount the item resonse function relies on attribute assignments in the Q-matrix. Also, the attribute vector for individual j is denoted as α j, and θ j after accounting for the θ j is the residual ability arameter, which deals with content measured by the test that is not included in the Q- matrix. Parameters The fusion model introduces many new arameters not resent in any other model; therefore a brief exlanation of these arameters would be aroriate. First the three item arameters will be described, and then the two erson arameters will be exlained. Last, the attribute arameter will be discussed. The arameter * π i equals the robability of correctly alying all attributes required by item i given that the individual ossesses all of the required attributes for the item i. As a result, * π i reflects the difficulty of item i and is referred to as the Q-based item difficulty arameter. It affects a erson s caacity to answer the item correctly desite his/her ossession of all the involved attributes. Just as with the item difficulty of 27

classical test theory, the values of * π i must remain between zero and unity, and a high value indicates easiness rather than difficulty. Next, the arameter * r ik reresents the roortion of the robability of obtaining a correct resonse when the examinee does not have the required attribute with the robability of resonding correctly to the item when the required attribute is ossessed by the examinee. Hence, * r ik is considered the discrimination arameter of item i for attribute k. This arameter is described as the enalty for lacking attribute k. A high * r ik value signifies that attribute k is not imortant in roducing a correct resonse to item i (Hartz, 2002, Hartz, et al., 2002). This arameter is a ratio of robabilities, and its values also remain between zero and unity. Third, the arameter c i deals with the item resonse function s reliance on the residual ability arameter θ j, which measures the examinee on constructs that are beyond the scoe of the Q-matrix. Therefore, the c i arameter is the comleteness index for item i. This arameter s value remains between zero and three, with a high value meaning the Q- matrix is comlete and therefore a correct resonse to the item does not rely heavily on θ j. In sum, a good item would have a high one * π i arameter, one c i arameter, and a item measures. * π i, a high c i and a low * r ik. Each item has * r ik arameter for each attribute that the given The fusion model estimates two tyes of arameters for the examinees. The first is a vector denoted as α j which identifies the attributes which have been estimated as mastered by examinee j. The last arameter secific to the fusion model is the residual 28

ability arameter θ j. (Recall that though they are identically denoted, it is not the same as the theta arameter resent in tyical IRT models. Rather this arameter is comarable to the θ j arameter in the unified model.) This θ j arameter measures the left over ability construct required by the test but not included in the Q-matrix. This can be thought of as a measure of higher mental rocesses (Samejima, 1995,. 391), a measure for dealing with multile solution strategies, or merely a nuisance arameter, (DiBello, Stout, and Roussos, 1995,. 370). The inclusion of this arameter is imortant because it allows us to acknowledge the fact that Q-matrices are not comlete reresentations of what is required by an exam (Hartz, 2002; DiBello, Stout, and Roussos, 1995). The fusion model also consists of one arameter for each attribute measured in the assessment. This arameter is denoted as k and is the cutoff value for attribute k. These values are used to dichotomize the mastery/non-mastery assignments of each examinee, because the values in the attribute vector α j are continuous rather than dichotomous. The continuous values in the attribute vector that are greater than or equal to k are assigned the status of mastery for attribute k; the remaining values are assigned the status of non-mastery for that attribute. Parameter Estimation Like any model, the fusion model requires an estimation rocedure for the arameters involved. The fusion model has several different arameters that must be estimated simultaneously, and therefore requires a owerful rocedure for calculating 29

them. The Markov Chain Monte Carlo (MCMC) method is a versatile estimation rocedure that has become an imortant tool in alied statistics (Tierney, 1997). MCMC is used in arameter estimation in the fusion model because it can elegantly handle the otherwise arduous, erhas even imossible task of simultaneously estimating the K+1 item arameters as well as the K entries in the attribute vector of examinee arameters. MCMC Estimation. MCMC looks at the robability of the next observation assuming it deends solely on the current observation. Hence, this method is based on a Bayesian framework. The field of Bayesian statistics is based on Bayes theorem and uses a rior distribution, which involves known information about the arameters of interest, to estimate osterior distribution, which is the conditional robability distribution of the unobserved quantities of ultimate interest, given the observed data, (Gelman, et al., 1995,. 3). Bayesian techniques are an advantageous aroach to inferential statistical analyses in two imortant ways. First, the use of rior information utilizes known information about the relationshis between the variables involved while the data rovides information about unknown characteristics of the variables relationshis. For instance, ositive correlations exist between examinee arameters, so it would be aroriate not to waste this information during estimation (Hartz, 2002). A Bayesian framework allows for this sort of flexibility in the rior distribution of arameters (Hartz, 2002). And second, the rior distributions do not untenably influence the osterior distributions (Hartz, 2002). In sum, a Bayesian framework allows the 30

incororation of imortant information already known about the item arameter distributions, but is not adversely affected when such useful information is not available. (For more information regarding Bayesian analytic methods, lease see Gelman, et al., 1995.) To understand how MCMC estimates information about future observations or states, it is easiest to first consider a simle case involving a dichotomous random variable. At any given time, this variable could take on a value of zero or unity. A Markov Chain deals with such a variable over various mathematical states. Consider the robability of achieving a second state given the initial state. This robability can be reresented by a matrix with the number of rows and columns matching the number of ossible states. When only looking at two states, the transition matrix from state one to state two may be illustrated by the below robability matrix below: = 00 01 s 1 s 2 10 11 where 00 = the robability of the variable taking on a value of zero in the first state and zero in the second state. 01 = the robability of the variable taking on a value of zero in the first state and one in the second state. 10 = the robability of the variable taking on a value of one in the first state and zero in the second state. 31

32 11 = the robability of the variable taking on a value of one in both states. Regardless of whether the variable takes on a value of zero or one in either state, this matrix can be used to determine the robability of the transition to state two. A bit of matrix algebra can be used to determine the robability of obtaining a value of one or zero for the second state by multilying the transition matrix of robabilities and the vector of robabilities of the initial state as illustrated in Equation (9). = + + = = = = = = = = = = 1 0 1 11 0 10 1 01 0 00 1 0 11 10 01 00 2 2 1 1 1 1 1 1 2 s s s s s s s s s (9) Next, consider the robability of observing state 3. The robability of observing a zero or a one for the third state can be similarly determined, as outlined in Equation (10). = + + + + + + = = = = = = = = = = 1 0 1 2 11 01 10 0 10 11 00 10 1 11 01 01 00 0 10 01 2 00 1 0 2 11 10 01 00 3 3 1 1 1 1 1 1 3 ) ( ) ( ) ( ) ( s s s s s s s s s (10) Likewise, the robability vector for state 4 or even state 194 and so on can continue to be determined by raising the transition matrix to greater and greater exonents and then multilying by the vector of robabilities of the initial state. Consequently, as the number of iterations increases, the comlexity of the calculation also increases, but more imortantly, the reliance on the initial state decreases and the