Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit...

16
Q*Power User Manual April 2014 Beta Version Matthew Madison, University of Georgia Laine P. Bradshaw, University of Georgia Bill Hollingsworth, Hollingsworth Technologies, Inc. Ϙ 0 1 0 1 0 1 0 Ϟ

Transcript of Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit...

Page 1: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

Q*Power User Manual

April 2014

Beta Version

Matthew Madison, University of Georgia

Laine P. Bradshaw, University of Georgia

Bill Hollingsworth, Hollingsworth Technologies, Inc.

Ϙ 0 1

0 1 0

1 0 Ϟ

Page 2: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

Table of Contents

1 Introduction ...................................................................................................................................... 1

1.1.... Background ...................................................................................................................................... 1

1.2.... What is Q*Power? ........................................................................................................................... 5

2 How does Q*Power work? .............................................................................................................. 5

2.1.... User Input ........................................................................................................................................ 5

2.2.... Data Generation .............................................................................................................................. 8

2.3.... Estimate LCDM ............................................................................................................................... 8

2.4.... Analyze Output ............................................................................................................................... 10

2.5.... Report Results ................................................................................................................................ 11

3 References ...................................................................................................................................... 12

4 Contact Information ....................................................................................................................... 14

Page 3: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

1

1. Introduction

1.1 Background

Diagnostic classification models (DCMs; e.g., Rupp, Templin, & Henson, 2010) are

psychometric models that aim to classify respondents according to their mastery or non-mastery

of specified latent characteristics or skills. In contrast, when using item response theory (IRT)

models, the aim is to scale individuals on a continuum of general ability in a particular domain.

The demand for individualized diagnostic information in education (ETS, 2013), as well as other

settings, has fueled the recent wave of research on DCMs. These models are well-suited for

providing diagnostic feedback because of their practical efficiency and increased reliability when

compared to other multidimensional measurement models (Templin & Bradshaw, 2013). DCMs

have been applied in educational settings to classify examinees with respect to multiple attributes

(Lee, Park, & Taylan, 2011; Bradshaw, Izsák, Templin, & Jacobson, 2013) and in psychological

settings to diagnose mental disorders (Templin & Henson, 2006).

A core element of the design for a diagnostic assessment is a priori specifications of

which latent characteristics or attributes are measured by each item. This attribute-item

alignment is expressed in what is known as a Q-matrix (Tatsuoka, 1983). The Q-matrix is an

item-by-attribute matrix of 0’s and 1’s indicating which attributes are measured on each item. If

item i requires attribute j, then cell ij in the Q-matrix will be a 1, and 0 otherwise. Table 1

displays a hypothetical Q-matrix for an assessment with three items and three attributes. Here,

Item 1 measures Attribute 1, Item 2 measures Attribute 2, and Item 3 measures both Attribute 1

and 3 in conjunction.

Page 4: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

2

Table 1

Hypothetical Q-matrix

Item/Attribute Attribute 1 Attribute 2 Attribute 3

Item 1 1 0 0

Item 2 0 1 0

Item 3 1 0 1

Q-matrix designs may vary by many features. Basic features include the number of items

on the assessment and the number of attributes measured on the assessment. Other features

influence the complexity of the Q-matrix. Generally, the complexity of the Q-matrix increases as

the number of non-zero entries in the Q-matrix increases. The complexity may differ by the

number of items measuring each attribute, the number of attributes measured within each item,

and how often attributes are measured jointly with other attributes.

The specification of the Q-matrix precedes and supports any inference resulting from the

application of the DCM. In educational settings, the initial design of a Q-matrix typically

employs a combination of content-specific learning theory and expert insights provided by

teachers, content-experts, and psychometricians. Verifying the Q-matrix in practice is a tedious

process that involves extensive qualitative research investigating the specific mental processes

that examinees use to respond to items. This is a critical component of the assessment

development process because if the Q-matrix is not correct, the inferences resulting from the

application of the DCM will be fallacious. This verification is referred to as content validity

(Boorsboom & Mellenberg, 2007) and is arguably the most important characteristic of any

assessment.

Because the specification of the Q-matrix is such a demanding process, most prior

research on the Q-matrix has examined the impact of misspecification (Rupp & Templin, 2008;

Kunina-Habenicht, Rupp, & Wilhem, 2012). Using simulation studies, these articles show that

when the Q-matrix is not specified correctly, correct classification accuracy rates, both attribute-

Page 5: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

3

wise and overall profile, decrease to varying degrees depending on the degree of

misspecification. Rupp and Templin examine overfit (0’s changed to 1’s), underfit (1’s changed

to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among

attributes for a constrained non-compensatory DCM. They found that all four conditions resulted

in item parameter bias and misclassification of examinees with profiles corresponding to the Q-

matrix misspecifications. For example, when attribute i was deleted from the true Q-matrix,

many misclassifications occur for attribute profiles involving attribute i. Kunina-Habenicht et al.

investigate misspecified Q-matrices with random permutations of 30% of the entries and

incorrect dimensionality for a general diagnostic model. Similar to Rupp and Templin, they

found that classification accuracy decreases in both conditions.

Given that classification is the prime objective of DCMs, it is crucial that researchers and

practitioners are aware of all the assessment design variables that can possibly affect

classification accuracy. While correctly specifying the Q-matrix is vital, another factor that may

influence classification accuracy is the Q-matrix design and structure. DeCarlo (2011) points out

that in the deterministic-input, noisy-and-gate (DINA; e.g., Haertel, 1989; Junker & Sijtsma,

2001; de la Torre & Douglass, 2004) model and higher order models, the design of the Q-matrix

can affect classification accuracy, even if specified correctly. DeCarlo showed through an

analysis of the fraction subtraction data (Tatsuoka, 1990) and through an analytical argument,

that when attributes enter the DINA model only as interaction terms, i.e., attributes are not

measured in isolation, posterior probabilities are largely determined by prior probabilities. As a

result, examinee responses have little effect on their classifications, which is problematic, to say

the least. Madison and Bradshaw (2013) examined this phenomenon using a simulation study

Page 6: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

4

and found that even for a general diagnostic model, classification accuracy and reliability

decrease when the Q-matrix does not specify items that measure attributes in isolation.

There are also numerous other factors that could potentially affect classification accuracy.

Namely, examinee sample size, attribute correlations and mastery base rates, the length of test,

number of attributes, and the interaction among these factors all impact classification accuracy.

A simulation study that covers all possible combinations of the variables aforementioned is

practically and theoretically impossible. Madison and Bradshaw (2013) suggest that any

researcher interested in designing a test to be modeled with a DCM should perform what would

be analogous to a prospective power analysis for traditional statistical models. That is, perform a

simulation study to ensure that the model will meet the accuracy and reliability desires of the

researcher under hypothesized/expected test design specifications. However, it is unrealistic to

expect that many researchers outside of psychometrics will possess the time, statistical software

knowledge, and resources to perform such a task.

1.2 What is Q*Power?

In short, Q*Power is a free and easy-to-use program that performs the prospective DCM test

design analysis described above. Given a user-specified Q-matrix, and other hypothesized

sample and test characteristics, Q*Power supplies the user a report with predicted model

classification accuracy and reliability. Q*Power essentially works through three steps:

1. Read input specifications and generate data

2. Estimate the model

3. Analyze estimation output

The following section describes each step of Q*Power in detail.

Page 7: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

5

2. How does Q*Power work?

2.1 User Input

Similar to a prospective power analysis, the researcher must provide some details of the

research that are not known, but rather hypothesized or expected. These details can be obtained

from research literature, or from experience. The only information that Q*Power needs from the

user is the test design. Test design includes the Q-matrix, anticipated examinee sample size,

tetrachoric correlation among attributes, anticipated attribute effect size, and attribute mastery

base rates. To input these test design specifications, the user interacts with the web-based portion

of Q*Power at www.lainebradshaw.com/qpower. Figure 1 displays the Q*Power home page.

Input Q-matrix

On the Q*Power home page, the user will type their email address (used to send results) and

upload the Q-matrix as a .csv file with item numbers in the first column and no column headings.

An example Q-matrix is shown in Figure 2.

Input other test design specifications

After uploading the Q-matrix, the page will refresh with places to type in and/or select other test

design specifications. Examinee sample size can be any positive integer up to 10,000. The

tetrachoric correlation describes the relationship between the attributes. In educational setting,

a correlation between .5 and .8 is typical. The correlation must be between 0-.85 in increments of

.05. Attribute effect size refers to the quality of the items. For average attribute effects, masters

of the measured attributes have between [ ] probability of answering correctly. For large

attribute effects, masters of the measured attributes have between [ ] probability of

answering correctly. And for mixed attribute effects, masters of the measured attributes have

between [ ] probability of answering correctly. Many DCM simulations have used what

Page 8: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

6

Figure 1. Q*Power Home Page

Page 9: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

7

we are calling large parameters. However, in our experience, we have found that attribute effect

sizes tend to be closer to average. Since this is in the planning phase, we suggest being

conservative and entering average effects. For average effects, enter 1, for large effects, enter 2,

and for mixed, enter 3. Lastly, the user must enter mastery base rates for each attribute. Mastery

base rates must be real numbers in the range ( ). For example, an easier attribute could have

a base rate of .65, and a more difficult may have a mastery base rate of .35. If you are unsure,

enter a moderately difficult base rate of .5.

Figure 2. Example Q-Matrix .csv File

Page 10: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

8

2.2 Generation of data

After the user has entered test design specifications, Q*Power generates item response

data. The first step is to generate examinee attribute profiles according to the specified sample

size, attribute correlations, and base rates. To do this, Q*Power uses R, version 2.13.0.

Specifically, the bindata package (Leisch, Weingessel, & Hornik, 2012). This package contains

a function, rmvbin() that generates random multivariate binary data with given correlation.

After examinee generation, Q*Power generates item parameters according to the Q-

matrix and attribute effect size specified. Q*Power generates item parameters up to two-way

interactions. If an item measures more than three attributes, Q*Power generates only main

effects. This avoids parameters being so small that they become negligible.

Using the examinee attribute profiles and the item parameters, Q*Power generates item

response probabilities. These probabilities are then compared to a matrix of random uniform(0,1)

variates to obtain item responses. Now that data generation is complete, Q*Power moves to

estimation of the model.

2.3 Estimate the LCDM

The log-linear cognitive diagnosis model (LCDM; Henson, Templin, & Willse, 2009) is a

general diagnostic classification model that flexibly models attribute effects and interactions at

the item level. Other common DCMs (e.g., the DINA model) can be viewed as special cases of

the LCDM in which certain parameters are constrained across all items or all attributes. Using

the LCDM, the need for these constraints can be tested empirically to guide model specifications

that best represent the relationships among items and attributes exhibited by the data.

The LCDM item response function (IRF) is similar to multidimensional IRT model

specifications, with the distinguishing feature being that latent traits are not continuous, but

Page 11: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

9

rather, they are binary. These binary traits are referred to as attributes denoted by , where 0

for non-mastery and 1 for mastery. For a test that measures attributes, there are unique

patterns of attribute mastery or non-mastery an examinee may have. The LCDM item response

function assumes item independence conditional upon the examinee’s attribute pattern. These

patterns, or attribute profiles, are predetermined latent classes into which the LCDM

probabilistically classifies examinees. To demonstrate the LCDM IRF, consider item that

measures two attributes, Attribute 3 ( ) and Attribute 4 ( ).The LCDM models the

probability of a correct response for an examinee in class as:

( | ) ( ( )( ) ( )( ) ( )( ))

( ( )( ) ( )( ) ( )( ))

(1)

The person parameter in Equation 1 is the examinee’s profile , which is an length vector

denoting the individual attribute states within class . Namely, [ , , …, ] where =

1 for every attribute examinees in class have mastered and = 0 for every attribute

examinees in class have not mastered. The item parameters in Equation 1 include an intercept

( ), a simple main effect for Attribute 3 ( ( )), a simple main effect for Attribute 4 ( ( )),

and an interaction between the two attributes ( ( )) Similar to a reference-coded analysis of

variance (ANOVA) model, the non-masters of attribute(s) are the reference group for each

effect. For example, the simple main effect for Attribute 3 is only present in the IRF when an

examinee’s profile indicates they have mastered Attribute 3 (i.e., when ). Thus, likening

the LCDM to ANOVA methods helps interpret the parameters. For example, the intercept is the

log-odds of a correct response for examinees who have mastered neither Attribute 3 nor Attribute

4. The main effect for Attribute 3 is the increase in log-odds of a correct response for masters of

Attribute 3 and non-masters of Attribute 4, and the main effect for Attribute 4 is the increase in

Page 12: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

10

log-odds of a correct response for masters of Attribute 4 and non-masters of Attribute 3. Finally,

the interaction term is the change in log-odds for examinees who have mastered both Attribute 3

and 4.

Q*Power estimates the LCDM via an MCMC algorithm written in Fortran. The current

beta version of Q*Power sets the number of iterations and burn-in period based on the

complexity of the Q-matrix. After estimation, Q*Power moves to analyzing the output.

2.4 Analyze Output

Q*Power is interested in two outcome measures: classification accuracy and reliability.

Since Q*Power generated the examinees profiles, we can compare the true mastery state with the

estimated mastery state to obtain a point estimate for accuracy for each attribute. Reliability

measures the stability of classifications upon a hypothetical re-examination (Templin &

Bradshaw, 2013). From the estimated attribute probabilities, we can compute a point estimate

reliability for each attribute.

Now that we have point estimates for model accuracy and reliability, we need measures

of variability around these point estimates. Both the outcome statistics have distributions, with

variance. Using only a point estimate is dangerous in the planning phase because it is entirely

possible that the point estimate you obtain, say .9, is at the upper tail of this distribution that is

centered at, say .75. Since we do not know the distributions of accuracy and reliability, we use a

bootstrap procedure to obtain measures of variability. The bootstrap procedure involves taking m

samples from the data with replacement and calculating the value of the statistic in each of

these samples. Using Fortran, we take 5,000 bootstrap samples from the output and compute

model accuracy and reliability. The distribution of these statistics across the bootstrap samples is

Page 13: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

11

referred to as the bootstrap distribution. To obtain an interval, Q*Power takes the minimum and

maximum of the bootstrap distribution as the lower and upper bound of the interval.

2.5 Report Results

Once Q*Power has finished analyzing output, it will email the user with the results.

Figure 3 is an example of such a report. The first portion is a check for input specifications to

ensure that you are getting results for the correct input specifications.

Figure 3. Q*Power Emailed Report

The second and third sections contain the average, lower bound, and upper bounds of the

intervals for estimated classification accuracy and reliability.

Page 14: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

12

References

Boorsboom, D., & Mellenberg, G.D. (2007). Test validity in cognitive test. In J. P. Leighton &

M. J. Gierl (Eds.), Cognitive diagnostic test for education: Theory and applications (pp.

19-60). London: Cambridge University Press.

Bradshaw, L., Izsák, A., Templin, J., & Jacobson, E. (2013). Diagnosing teachers’

understandings of rational number: Building a multidimensional test within the diagnostic

classification framework. Manuscript under review.

Center for K-12 Assessment and Performance Management, ETS (2013). Coming together to

raise achievement: New assessments for the common core state standards.

DeCarlo, L. T. (2011). On the analysis of fraction subtraction data: The DINA model,

classification, latent class sizes, and the Q-matrix. Applied Psychological Measurement,

35, 8-26.

de la Torre, J., & Douglas, J. (2004). Higher-order latent trait models for cognitive diagnosis.

Psychometrika, 69, 333-353.

de la Torre, J. (2008). An empirically-based method of Q-matrix validation for the DINA

model: Development and applications. Journal of Educational Measurement, 45, 343–

362.

Haertel, E. H. (1989). Using restricted latent class models to map the skill structure of

achievement items. Journal of Educational Measurement, 26, 333-352.

Henson, R., Templin, J., & Willse, J. (2009). Defining a family of cognitive diagnosis models

using log linear models with latent variables. Psychometrika, 74, 191-210.

Page 15: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

13

Junker, B. W., & Sijtsma, K. (2001). Cognitive assessment models with few assumptions, and

connections with nonparametric item response theory. Applied Psychological

Measurement, 25, 258-272.

Kunina-Habenicht, O., Rupp, A. A., & Wilhem, O. (2012). The impact of model

misspecification on parameter estimation and item-fit assessment in log-linear diagnostic

classification models. Journal of Educational Measurement, 49(1), 59-81.

Lee, Y-S., Park, Y. S., & Taylan, D. (2011). A cognitive diagnostic modeling of attribute

mastery in Massachusetts, Minnesota, and the U.S. national sample using the TIMSS

2007. International Journal of Testing, 11, 144-177.

Leisch, F., Weingessel, A. & Hornik, K. (2012). bindata: generation of artificial

binary data. R package version 0.9-19. http://CRAN.R-project.org/package=bindata.

Madison, M. & Bradshaw, L. (2013). The effects of Q-matrix design on classification accuracy

in the LCDM. Paper presented at the annual meeting of the Northeastern Educational

Research Association in Rocky Hill, CT.

Rupp, A. A., & Templin, J. (2008). Effects of Q-matrix misspecification on parameter estimates

and misclassification rates in the DINA model. Educational and Psychological

Measurement, 68, 78-98.

Rupp, A. A., Templin, J., & Henson, R. (2010). Diagnostic measurement: Theory, methods, and

applications. New York, NY: Guilford.

Tatsuoka, K. K. (1983). Rule-space: An approach for dealing with misconceptions based on item

response theory. Journal of Educational Measurement, 20, 345-354.

Tatsuoka, K. K. (1990). Toward an integration of item-response theory and cognitive error

diagnosis. In N. Frederiksen, R. Glaser, A. Lesgold, & M. Safto (Eds.), Monitoring skills

Page 16: Î 0 1 0 1 0 1 0 Ôlainebradshaw.com/qpower/qpower_uguide.pdf · to 0’s), balanced misfit Q-matrices (exchanged 0’s and 1’s), and incorrect dependencies among attributes for

14

and knowledge acquisition (pp. 453–488). Hillsdale, NJ: Erlbaum.

Templin, J., & Bradshaw, L. (2013). The comparative reliability of diagnostic model

examinee estimates. Journal of Classification.

Templin, J. L., & Henson, R. A. (2006). Measurement of psychological disorders using cognitive

diagnosis models. Psychological Methods, 11, 287-305.

Contact Information

Matthew Madison

Graduate Research Assistant

Aderhold 125-A

University of Georgia

[email protected]

(803) 542-4309