Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

80
1 Psychometric Methods for Investigating DIF and Test Bias During Test Adaptation Across Languages and Cultures Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada Presented at the 5 th International Test Commission Conference July 6-8, 2006 Université catholique de Louvain (UCL) Brussels, Belgium

description

Psychometric Methods for Investigating DIF and Test Bias During Test Adaptation Across Languages and Cultures. Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada Presented at the 5 th International Test Commission Conference July 6-8, 2006 - PowerPoint PPT Presentation

Transcript of Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

Page 1: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

1

Psychometric Methods for Investigating DIF and Test Bias During Test Adaptation Across

Languages and Cultures Bruno D. Zumbo, Ph.D.

Professor

University of British Columbia

Vancouver, Canada

Presented at the 5th International Test Commission Conference

July 6-8, 2006

Université catholique de Louvain (UCL)

Brussels, Belgium

Page 2: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

2

Overview Methods for detecting differential item functioning (DIF) and

scale (or construct) equivalence typically are used in developing new measures, adapting existing measures, or validating test score inferences.

DIF methods allow the judgment of whether items (and ultimately the test they constitute) function in the same manner for various groups of examinees, essentially flagging problematic items or tasks. In broad terms, this is a matter of measurement invariance; that is, does the test perform in the same manner for each group of examinees?

You will be introduced to a variety of DIF methods, some developed by the presenter, for investigating item-level and scale-level (i.e., test-level) measurement invariance. The objective is to impart psychometric knowledge that will help enhance the fairness and equity of the inferences made from tests.

Page 3: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

3

Agenda1. What is measurement invariance, DIF, and

scale-level invariance?2. Construct versus item or scale equivalence3. Description of DIF methods4. Description of scale-level invariance, focus on

MG-EFA5. Recommendations 6. Appendix 1: Paper on DIF with rating scales,

using SPSS7. Appendix 2: Paper on construct comparability

from a factor analysis perspective.

Page 4: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

4

Why are we looking at item or scale-level bias?

Technological and theoretical changes over the past few decades have altered the way we think about test validity and the inferences we make from scores that arise from tests and measures.

If we want to use the measure in decision-making (or, in fact, simply use it in research) we need to conduct research to make sure that we do not have bias in our measures. Where our value statements come in here is that we need to have organizationally and socially relevant comparison groups (e.g., gender or minority status).

Page 5: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

5

Why are we looking at item or scale-level bias?

In recent years there has been a resurgence of thinking about validity in the field of testing and assessment. This resurgence has been partly motivated by the desire to expand the traditional views of validity to incorporate developments in qualitative methodology and in concerns about the consequences of decisions made as a result of the assessment process.

For a review of current issues in validity we recommend the recent papers by Zumbo (in press), Hubley and Zumbo (1996) and the edited volume by Zumbo (1998).

I will focus on item bias first and then turn to the scale-level approaches toward the end.

Page 6: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

6

Item and Test Bias

Methods for detecting differential item functioning (DIF) and item bias typically are used in the process of developing new measures, adapting existing measures, or validating test score inferences.

DIF methods allow one to judge whether items (and ultimately the test they constitute) are functioning in the same manner in various groups of examinees. In broad terms, this is a matter of measurement invariance; that is, is the test performing in the same manner for each group of examinees?

Page 7: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

7

Item and Test BiasConcerns about item bias emerged within the context of

test bias and high-stakes decision-making involving achievement, aptitude, certification, and licensure tests in which matters of fairness and equity were paramount.

Historically, concerns about test bias have centered around differential performance by groups based on gender or race. If the average test scores for such groups (e.g. men vs. women, Blacks vs. Whites) were found to be different, then the question arose as to whether the difference reflected bias in the test.

Given that a test is comprised of items, questions soon emerged about which specific items might be the source of such bias.

Page 8: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

8

Item and Test Bias

Given this context, many of the early item bias methods focused on (a) comparisons of only two groups of examinees, (b) terminology such as ‘focal’ and ‘reference’ groups to denote minority and majority groups, respectively, and (c) binary (rather than polytomous) scored items.

Due to the highly politicized environment in which item bias was being examined, two inter-related changes occurred. – First, the expression ‘item bias’ was replaced by the

more palatable term ‘differential item functioning’ or DIF in many descriptions.

Page 9: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

9

Item and Test Bias

– DIF was the statistical term that was used to simply describe the situation in which persons from one group answered an item correctly more often than equally knowledgeable persons from another group.

– Second, the introduction of the term ‘differential item functioning’ allowed one to distinguish item impact from item bias.

– Item impact described the situation in which DIF exists because there were true differences between the groups in the underlying ability of interest being measured by the item. Item bias described the situations in which there is DIF because of some characteristic of the test item or testing situation that is not relevant to the underlying ability of interest (and hence the test purpose).

Page 10: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

10

Item and Test Bias

Traditionally, consumers of DIF methodology and technology have been educational and psychological measurement specialists.

As a result, research has primarily focused on developing sophisticated statistical methods for detecting or ‘flagging’ DIF items rather than on refining methods to distinguish item bias from item impact and providing explanations for why DIF was occurring.

Page 11: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

11

Item and Test Bias

Although this is changing as increasing numbers of non-measurement specialists become interested in exploring DIF and item bias in tests, it has become apparent that much of the statistical terminology and software being used is not very accessible to many researchers.

In addition, I believe that we are in now in what I like to call the “Second Generation of DIF and Bias Detection Methods”. My presentation today will try and highlight this 2nd Generation.

Page 12: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

12

2nd Generation DIF and Bias1. Fairness and equity in testing.

• Policy and legislation define groups ahead of time.• Historically this is what DIF was about. To some extent it is still

the case today but at least three other classes of uses of DIF can be described.

2. Dealing with a possible “threat to internal validity.” • The composition and creation of groups is driven by an

investigators research questions & defined ahead of time.• DIF is investigated so that one can make group comparisons

and rule-out measurement artifact as an explanation for the group difference.

Page 13: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

13

2nd Generation DIF and Bias3. Trying to understand the (cognitive and/or psycho-

social) processes of item responding and test performance, and investigating whether these processes are the same for different groups of individuals. I think of this as “validity” (Zumbo, in press).

• Latent class or other such methods are used to “identify” or “create” groups and then these new “groups” are studied to see if one can learn about the process of responding to the items.

• Also can use qualitative methods like talk-aloud protocol analysis.

4. Empirical methods for investigating the interconnected ideas of: (a) “lack of invariance”(LOI) issues, (b) model-data fit, and (c) model appropriateness in model-based statistical measurement frameworks like IRT and other latent variable approaches (e.g., LOI is an assumption for CAT, linking, and many other IRT uses).

Page 14: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

14

Overview of DIF Methods

In this section we will: – discuss some of the conceptual features of

DIF;– discuss some of the statistical methods for

DIF detection;– consider an example with an eye toward

showing you how to conduct these analyses in SPSS.

Page 15: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

15

Conceptual Matters I

It is important to note that in this entry we have focused on ‘internal’ methods for studying potential item bias; i.e., within the test or measure itself. It is important for the reader to note that there is also a class of methods for studying potential item bias wherein we have a predictor and criterion relationship in the testing context.

For example, one may have a test that is meant to predict some criterion behavior. Item (or, in fact, test level) bias then focuses on whether the criterion and predictor relationship is the same for the various groups of interest.

Page 16: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

16

Conceptual Matters I

• DIF methods allow one to judge whether items (and ultimately the test they constitute) are functioning in the same manner in various groups of examinees.

• In broad terms, this is a matter of measurement invariance; that is, is the test performing in the same manner for each group of examinees?

Page 17: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

17

Conceptual Matters II: What is DIF?

DIF is the statistical term that is used to simply describe the situation in which persons from one group answered an item correctly more often than equally knowledgeable persons from another group.

In the context of social, attitude or personality measures, we are not talking about “knowledge”, per se, but different endorsement (or rating) after matching on the characteristic or attitude or interest.

Page 18: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

18

Conceptual Matters II b: What is DIF?

The key feature of DIF detection is that one is investigating differences in item scores after statistically matching the groups.

In our case the “groups” may represent different language or adaptations of the test or measure, groups that you want to compare on the measure of interest, or groups legislated for investigation for bias.

Page 19: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

19

Conceptual: DIF versus Item Bias

DIF is a necessary, but insufficient condition for item bias.

Item bias refers to unfairness of item toward one or more groups

Item bias is different than item impact which describes the situation in which DIF exists because there were true differences between the groups in the underlying ability of interest being measured by the item.

Page 20: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

20

Statistical frameworks for DIF

At least three frameworks for thinking about DIF have evolved in the literature: (1) modeling item responses via contingency tables and/or regression models, (2) item response theory, and (3) multidimensional models (see Zumbo & Hubley, 2003 for details).

We will focus on the first framework.

Page 21: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

21

One, of several, ways of classifying DIF methods

Matching variable

Observed score Latent variable

Binary

MH LogR

Conditional IRT Multidimensional SEM

Item

for

mat

Polytomous

Ordinal LogR Conditional IRT Multidimensional SEM

Page 22: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

22

Matching variable

Observed score Latent variable

Binary

MH LogR

Conditional IRT Multidimensional SEM

Item

for

mat

Polytomous

Ordinal LogR Conditional IRT Multidimensional SEM

Our focus will be here Our focus will be here and we will include and we will include

MH methods as well.MH methods as well.

Page 23: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

23

Statistical Definition I

A statistical implication of the definition of DIF (i.e., persons from one group answering an item correctly more often than equally knowledgeable persons from another group) is that one needs to match the groups on the ability of interest prior to examining whether there is a group effect.

Page 24: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

24

Statistical Definition II

That is, DIF exists when assuming that the same items are responded by two different groups, after conditioning on (i.e., statistically controlling for) the differences in item responses that are due to the ability being measured, the groups still differ.

Page 25: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

25

Statistical Definition III

Thus, within this framework, one is interested in stating a probability model that allows one to study the main effects of group differences (termed ‘uniform DIF’) and the interaction of group by ability (termed ‘non-uniform DIF’) after statistically matching on the test score.

Conceptually, this is akin to ANCOVA or ATI; therefore, all of the same caveats about causal claims in non-experimental designs apply. (DIF is a non-experimental design)

Page 26: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

26

Two classes of statistical methods, we focus on one.

Two broad classes of DIF detection methods: Mantel-Haenszel (MH) and logistic regression (LogR) approaches.

We will focus on LogR because it is the most general of the two methods allowing us to (a) investigate uniform and non-uniform DIF, (b) easily allow for additional matching variables, (c) easily extend our methods to rating or Likert scales, and (d) builds on the knowledge-base of regression modeling more generally defined .

Page 27: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

27

DIF Statistical Analysis

The LogR class of methods (Swaminathan & Rogers, 1990; Zumbo 1999) entails conducting a regression analysis (in the most common case, a logistic regression analysis as the scores are binary or ordinal LogR for Likert item scores) for each item wherein one tests the statistical effect of the grouping variable(s) and the interaction of the grouping variable and the total score after conditioning on the total score.

Page 28: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

Logistic Regression

Z T T T G 0 1 2 Z T T T G T G 0 1 2 3 T e s ts U n ifo rm D IF T e s ts N o n- un ifo rm D IF

P Ue

e

z

z( )

11

Page 29: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

29

Logistic Regression II

In the DIF detection regression methods, regression model building and testing has a natural order.

That is, – (i) one enters the conditioning variable (e.g., the total

score) in the first step and then – (ii) the variables representing the group and the

interaction are entered into the equation and – (ii) one tests whether step two variables that were

added statistically contribute to the model via looking at the 2-degree of freedom test and a measure of effect size.

Page 30: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

30

Example

A synthetic sample of military officers in Canada (532 monolingual English speaking; 226 monolingual French speaking).

This is a general aptitude test that measures verbal, spatial, and problem-solving skills, with 15, 15, and 30 items respectively.

The question at hand is the adequacy of the translation of the problem-solving items.

Page 31: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

31

Example (cont’d.)

What do we match on to compare item performance, item by item of the problem-solving scale?

• The standard answer is the total score of the problem-solving.

We are going to match on both the total score on the problem-solving scale and on the total of the spatial-abilities scale (hence matching on general intellectual functioning and not just problem-solving)– Because it is not a verbal measure, per se, the spatial-

abilities scale is not effected by language it is administered.

Page 32: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

32

Example (cont’d) Item #1

Lets start with the typical DIF analysis and condition only on the total scale of problem-solving (we are, after all, studying a problem-solving item)

Used indicator coding for language (English =1, French =0)

Chi-Squared, 2-df, 61.6, p<=.001.So, statistically we have signs of a DIF

item.

Page 33: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

33

Output from SPSS

Variables in the Equation

.218 .057 14.712 1 .000 1.244

2.317 .825 7.883 1 .005 10.148

-.001 .061 .000 1 .983 .999

-4.850 .718 45.576 1 .000 .008

problem

lang(1)

lang(1) by problem

Constant

Step1

a

B S.E. Wald df Sig. Exp(B)

Variable(s) entered on step 1: lang, lang * problem .a.

(Conditional) Odds ratio

Page 34: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

34

Plot of the “predicted probabilities of LogR analysis

French

Page 35: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

35

Plot of the “predicted probabilities of LogR analysis

French

Page 36: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

36

Example (cont’d) Item #1

Now lets do the multiple conditioning example with

Used indicator coding for language (English =1, French =0)

Chi-Squared, 2-df, 53.6, p<=.001.So, statistically we have signs of a DIF

item.

Page 37: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

37

Example (cont’d) Item #1

Variables in the Equation --- Note the results for "Lang"

15.489 1 .000 1.251

1.168 1 .280 .953

6.759 1 .009 8.704

.006 1 .939 1.005

31.198 1 .000 .012

PROBLEM

SPATIAL

LANG(1)

LANG(1) by Problem

Constant

Wald df Sig. Exp(B)

Page 38: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

38

Example (cont’d) Item #1

We conclude the item #1 displays DIF in this synthetic sample with an partial odds-ratio of over 8 times in favor of the English version – i.e. the French version of that item was over 8 times more difficult than the English version, while matching on general intellectual functioning.

This analysis would be conducted for each of the 30 problem-solving items

Page 39: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

39

Logistic Regression: Assessment on Balance

Strengths– Multivariate matching criteria (multidimensional

matching).– Can test for both uniform and non-uniform DIF– Significance test (like IRT likelihood ratio)

– Effect size (T2)

– Popular software (SAS, SPSS) … paper by Slocum, Gelin, & Zumbo (in press) describe SPSS in detail.

Limitations (relative to IRT)– Observed score metric– May need to purify matching score for scales with few items.

Page 40: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

40

DIF: MH Method (optional)

Mantel-Haenszel: another conditional procedure like LogR– Like standardization, typically uses total test score as

conditioning variable

Extension of chi-square test

3-way Chi-square:– 2(groups)-by-2(item score)-by-K(score levels)– The conditioning variable is categorized into k

levels (i.e., bins)

Page 41: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

Mantel Haenszel Formulas Item Score

Group 1 0 TotalReference Aj Bj Nrj

Focal Cj Dj NrjTotal M1j M0j T1

H

AC

BD

j

j

j

j

0 1: HA

CB

DAJ

J

J

J:

M H

A E A

V a r A

jj

jj

jj

2

2

5

( ) .

( )

Page 42: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

42

Lets look at the item #1 from LogR

Cannot easily do multiple matching variables so lets look at the conventional matching situation.

We need to divide up (bin) the total scale score so that we can do the MH. Because there is not a lot of range of scores lets do a fairly thick matching and divide it into two equal parts. {I tried other groupings and ended-up with missing cells in the 3-way table}.

Page 43: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

43

Histogram of the scale score

Cut offs for grouping; at

median

Statistics

problem758

0

10.0000

18.0000

25.0000

Valid

Missing

N

25

50

75

Percentiles

Page 44: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

44

SPSS result: 3-way table of DIF on item #1 from above

grouping * problem solving * language Crosstabulation

Count

68 93 161

27 344 371

95 437 532

205 15 220

3 3 6

208 18 226

1.00

2.00

grouping

Total

1.00

2.00

grouping

Total

languageEnglish

French

0 1

problem solving

Total

Still problematic

Page 45: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

45

SPSS result: MH test of DIF on item #1 from above

Tests of Conditional Independence

103.266 1 .000

100.613 1 .000

Cochran's

Mantel-Haenszel

Chi-Squared dfAsymp. Sig.

(2-sided)

Under the conditional independence assumption, Cochran'sstatistic is asymptotically distributed as a 1 df chi-squareddistribution, only if the number of strata is fixed, while theMantel-Haenszel statistic is always asymptotically distributedas a 1 df chi-squared distribution. Note that the continuitycorrection is removed from the Mantel-Haenszel statistic whenthe sum of the differences between the observed and theexpected is 0.

Yes DIF

Page 46: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

46

DIF: MH Method (optional)

Strengths:• Powerful• Has statistical test and effect size• Popular software (SAS, SPSS)

Limitations• Does not test for non-uniform DIF• Unlike LogR, you need to put the matching score

into bins (i.e., levels) and this is somewhat arbitrary and may affect the statistical decision regarding DIF.

Page 47: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

47

Important to keep in mind …You need to assume that the items are, in

essence, equivalent and that the grouping variable (or variable under study) is the reason that the items are performing differently.

Need to assume that the matching items are equivalent across languages– Screened items– Non-verbal items

No systematic bias because DIF will not detect this.

These assumptions must be defended!

Page 48: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

48

Measurement of the construct is equivalent across groups

Scale-level bias– Are there substantive differences across

groups in the operationalization or manifestation of the construct?

Scale-level equivalence does not guarantee item level equivalence (or vice versa)– This was shown by Zumbo (2003) in the

context of translation DIF for language tests.

Page 49: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

49

Methods for evaluating scale-level equivalence:

Factor analysis– Multi-group Confirmatory factor analysis– Multi-group Exploratory factor analysis

Multidimensional scaling

We will briefly describe the MG-EFA because the MG-CFA is described well by Barb Byrne and the MDS by Steve Sireci.

Page 50: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

50

MG-EFA

Construct comparability is established if the factor loadings of the items (i.e., the regressions of the items on to the latent variables) are invariant across groups. Therefore, if one is investigating the construct comparability across males and females of, for example, a large-scale test of science knowledge one needs to focus on establishing the similarity of the factor loadings across the groups of males and females.

Page 51: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

51

MG-EFA

Staying with our example of large-scale testing of science knowledge, it is important to note that the results of factor analyses based on the two groups combined may be quite different than if one were to analyze the same two groups separately.

That is, when conducting factor analyses if a set of factors are orthogonal in the combined population of men and women then the same factors are typically correlated when the groups are analyzed separately.

Page 52: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

52

MG-EFA

However, if one finds that the factors are correlated in the separate groups this does not imply that the same set of factors are orthogonal in the combined population of men and women. A practical implication of this principle is that one need not be surprised if the factor analysis results found in studies of combined groups does not replicate when the groups are examined separately with multi-group methods. Likewise, one must proceed cautiously in translating results from combined group factor analyses to separate group factor analyses.

Page 53: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

53

MG-EFA

The matter of construct comparability hence becomes one of comparing the factor solutions that have been conducted separately for the groups of interest. The comparison should be of a statistical nature involving some sort of index or test of similarity, rather than purely impressionistic, because factor loadings, like all statistics, are effected by sampling and measurement error.

Page 54: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

54

Purpose of this section

Review methodological issues that arise in the investigation of construct comparability across key comparison groups such as ethnic and gender groups, or adapted/translated versions of tests in the same or different cultures.

We describe a multiple-method approach to investigating construct comparability.

The multiple-method approach allows us to investigate how sensitive our results are to the method we have chosen and/or to some peculiarities of the data itself.

Page 55: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

55

Purpose of this section (cont’d.)

In particular, multi-group exploratory factor analysis (MG-EFA) is described, in the context of an example, as a complement to the standard multi-group confirmatory factor analysis (MG-CFA).

Our recommended methods are MG-CFA, MG-EFA, and MG-MDS. MG-EFA is described more fully in the paper and is the focus because it is the one that was been “bumped” off the list of options with the advent of MG-CFA in the 1970s.

Page 56: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

56

The empirical test of construct comparability is typically examined by a pair wise comparison of factors (i.e., latent variables) or dimensions across groups.

MG-CFA has become the standard and commonly recommended approach to investigating construct comparability.

Multiple methods (i.e., MDS and CFA) has been advocated and demonstrated by Sireci and Allalouf (2003), and Robin, Sireci, and Hambleton (in press).

The key, however, is in knowing the trade-offs between the methods. What benefits and baggage does each method bring with it?

Page 57: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

57

Factor Analysis Methods for Construct Comparability

Construct comparability is established if the factor loadings of the items (i.e., the regressions of the items on to the latent variables) are invariant across groups.

Important to note that the results of factor analyses based on the (two) groups combined may be quite different than if one were to analyze the same two groups separately.

Page 58: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

58

Factor Analysis Methods for Construct Comparability

An early statistical approach to comparing the factor solutions across groups is provided by exploratory factor analysis.

The strategy is to quantify or measure the similarity of the factor loadings across groups by rotating the two factor solutions to be maximally similar and then computing some sort of similarity index.

The emphasis in the previous sentence is on quantifying or measuring the similarity rather than computing a formal statistical hypothesis test.

Page 59: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

59

Factor Analysis Methods for Construct Comparability

At this juncture, some points are worth noting. – The exploratory factor analysis methods were

superseded by the MG-CFA because MG-CFA’s reliance on a formal hypothesis test made the decision of factorial comparability, as well as the whole factor analysis enterprise objective, at the very least it appears so on a surface analysis.

– The reliance on exploratory factor analysis’ similarity indices may capitalize on sample-specific subtleties (sometimes referred to as capitalizing on chance) and these similarity indices can be spuriously large and misleading in some situations …. Likewise MG-EFA (particularly the rotational aspect) is a powerful “wrench” to apply to the data, hence can be Procrustean.

Page 60: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

60

Factor Analysis Methods for Construct Comparability

There is an understanding now that the sole reliance on MG-CFA may be limiting for the following reasons. – CFA generally appears to perform optimally

when the factor pattern solution exhibits simple structure, typically, with each variable measuring one and only one factor in the domain. (simple structure is an ideal)

– Furthermore, many tests, items or tasks may measure multiple factors albeit some of them minor factors.

Page 61: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

61

Factor Analysis Methods for Construct Comparability

Many of the fit and modification indices need large sample sizes before they begin to evince their asymptotic properties. (not always a problem)

The statistical estimation and sampling theory of the commonly used estimators in CFA assume continuous observed variables but our measures are often skewed and/or binary or mixed item response data.

Some wonderful alternatives (Muthen’s work implemented in Mplus) but the computational size of these “weighted” methods are typically of order p-squared, p is the number of items, so we run into physical limitations. (sandwich estimators, Rao-Scott, two stage … with mopping up)

Page 62: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

62

Factor Analysis Methods for Construct Comparability

The above remarks regarding the MG-CFA and MG-EFA focus on the reliance of one method, in fact either method, to the exclusion of the other.

We are proposing that together MG-CFA and MG-EFA complement each other well.

Page 63: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

63

Demonstration of the Methodology with an Example Data Set

The School Achievement Indicators Program (SAIP) 1996 Science Assessment was developed by the Council of Ministers of Education, Canada (CMEC) to assess the performance of 13- and 16-year old students in science.

13-year-old students in two languages, namely, English and French.

5261 English-speaking and 2027 French-speaking students.

This study used the 66 binary items from the assessment data from Form B (13-year olds).

Page 64: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

64

MG-CFA

We tested the full and strong invariance hypotheses of CFA. (define full and strong)

Full invariance: the difference in chi-square values between the baseline model and the full invariance model is statistically significant, ² = 2554.55, df = 132, p < 0.0001, indicating that the hypothesis of full invariance is not tenable.

Strong invariance: ² = 612.12, df = 66, p < 0.0001, indicating

that the hypothesis of strong invariance is also not tenable. Therefore, neither the stricter test of full

invariance nor the weaker test of strong invariance are tenable in the SAIP data suggesting, on their own, that we do not have construct comparability.

Page 65: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

65

MG-EFA

Steps:– Separate factor analyses resulted in one major

factor and minor factors– We wrote an algorithm in SPSS to take the separate

varimax solutions and using methods from Cliff (1966) and Schönemann (1966), allows us to rotate two factor solutions to maximal congruence

– Compute the congruency measures– Because the items are binary also replicated this

with an EFA of the tetrachoric matrix.

Page 66: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

66

MG-EFA To check to see if we have spuriously inflated congruency

coefficients are we introduced a graphical method. Ideal case, large congruency (45 degree angle) going through (0,0). Here is factor 1. (method founded on Rupp & Zumbo, 2002)

Page 67: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

67

MG-EFA. 2nd minor factor

Page 68: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

68

MG-EFA

The primary factor shows a congruency coefficient of 0.98 (same with tetrachoric).

The secondary factor did not show invariance with a coefficient of .67 with ULS estimate and .23 with tetrachoric. Difference between ULS and tetrachoric suggests that secondary factor may be a non-linearity proxy due to binary data. (this may also effect the MG-CFA)

Also, the 95% confidence band for individual estimates showed one possible “deviant” item in Figure 2, on the second factor. We can investigate this as a follow-up.

Page 69: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

69

ConclusionIn our example the strict test of MG-CFA failed to

show construct comparability whereas the other method showed an invariant primary first factor with the minor factors to being comparable. The MG-CFA result may be effected by the non-linearity due to binary data.

We have a demonstration that the MG-EFA can be a useful complement to MG-CFA

We are still a bit dissatisfied because we want to investigate further the possibility of a more truly “exploratory” method because the current methods are all still “tests” of invariance.

Page 70: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

70

RECOMMENDATIONS

1. Select an acceptable DIF detection method.

2. Use more than one method for evaluating construct equivalence (e.g.,MG-CFA, MG-EFA, WMDS, SEM).

3. If you have complex item types, see an approach introduced by Gelin and Zumbo (2004) adapting MIMIC modeling.

Page 71: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

71

RECOMMENDATIONS (continued)

4. For both construct equivalence and DIF studies

Replicate

Physically Match (if possible)

5. For DIF studies—purify criterion (iterative), if necessary; latent variable matching is also

very productive (e.g., MIMIC).

Page 72: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

72

RECOMMENDATIONS (continued)

6. DIF analyses should be followed up by content analyses.

7. Translation DIF is usually explainable, others are bit more difficult (in translation items are homologues).

8. Use results from Content Analyses and DIF studies to inform future test development and adaptation efforts

Page 73: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

73

RECOMMENDATIONS (continued)

9. Realize that items flagged for DIF may not represent bias

--differences in cultural relevance

--true group proficiency differences

10. When subject-matter experts review items, they should evaluate both DIF and non-DIF items.

Page 74: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

74

RECOMMENDATIONS (continued)

11. If you have Likert responses you can use ordinal logistic regression (Zumbo, 1999; Slocum, Gelin, & Zumbo, in press).

– We have shown in recent collaborative research that the generalized Mantel is also a promising approach; it however, does not have the advantages of modeling (multiple conditioning variables, etc.).

Page 75: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

75

RECOMMENDATIONS (continued)

12.When sample sizes are small, use nonparametric IRT (e.g., TestGraf) and the cut-offs provided by Zumbo & Witarsa (2004) for sample sizes that 25 per group. http://www.educ.ubc.ca/faculty/zumbo/aera/papers/2004.html

Page 76: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

76

Remember:

“As with other aspects of test validation, DIF analysis is a process of collecting evidence. Weighting and interpreting that evidence will require careful judgment.

There is no single correct answer”

(Clauser & Mazor, 1998, p. 40).

Page 77: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

77

ReferencesBreithaupt, K., & Zumbo, B. D. (2002). Sample invariance of the structural equation model and the item

response model: a case study. Structural Equation Modeling, 9, 390-412. Clauser, B. E., & Mazor, K. M. (1998). Using statistical procedures to identify differentially functioning

test items. Educational Measurement: Issues and Practice, 17, 31-44. Hubley, A. M., & Zumbo, B. D. (1996). A dialectic on validity: Where we have been and where we are

going.  The Journal of General Psychology, 123, 207-215.Kristjansson, E., Aylesworth, R., McDowell, I, & Zumbo, B. D. (2005). A comparison of four methods for

detecting DIF in ordered response items. Educational and Psychological Measurement, , 65, 935-953.

Lu, I. R. R., Thomas, D. R., & Zumbo, B. D. (2005). Embedding IRT In Structural Equation Models: A Comparison With Regression Based On IRT Scores. Structural Equation Modeling, 12, 263-277.

Maller, S. J., French. B. F., & Zumbo, B. D. (in press). Item and Test Bias.  In Neil J. Salkind (Ed.), Encyclopedia of Measurement and Statistics (pp. xx-xx). Thousand Oaks, CA: Sage Press.

Rupp, A. A., & Zumbo, B. D. (2003). Which Model is Best? Robustness Properties to Justify Model Choice among Unidimensional IRT Models under Item Parameter Drift. {Theme issue in honor of Ross Traub} Alberta Journal of Educational Research, 49, 264-276.

Rupp, A. A., & Zumbo, B. D. (2004).. A Note on How to Quantify and Report Whether Invariance Holds for IRT Models: When Pearson Correlations Are Not Enough.  Educational and Psychological Measurement, 64, 588-599. {Errata, (2004) Educational and Psychological Measurement, 64, 991}

Rupp, A. A., & Zumbo, B. D. (2006). Understanding Parameter Invariance in Unidimensional IRT Models.  Educational and Psychological Measurement, 66, 63-84.

Slocum, S. L., Gelin, M. N., & Zumbo, B. D. (in press). Statistical and Graphical Modeling to Investigate Differential Item Functioning for Rating Scale and Likert Item Formats. In Bruno D. Zumbo (Ed.) Developments in the Theories and Applications of Measurement, Evaluation, and Research Methodology Across the Disciplines, Volume 1. Vancouver: Edgeworth Laboratory, University of British Columbia.

Page 78: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

78

ReferencesZumbo, B.D. (in press). Validity: Foundational Issues and Statistical Methodology.  In C.R. Rao

and S. Sinharay (Eds.) Handbook of Statistics,  Vol. 27: Psychometrics. Elsevier Science B.V.: The Netherlands.

Zumbo, B. D. (2005). Structural Equation Modeling and Test Validation. In Brian Everitt and David C. Howell, Encyclopedia of Behavioral Statistics, (pp. 1951-1958). Chichester, UK: John Wiley & Sons Ltd.   

Zumbo, B. D. (Ed.) (1998). Validity Theory and the Methods Used in Validation: Perspectives from the Social and Behavioral Sciences. Netherlands: Kluwer Academic Press.   {This is a special issue of the journal Social Indicators Research: An International and Interdisciplinary Journal for Quality-of-Life Measurement, Volume 45, No. 1-3, 509 pages}

Zumbo, B. D. (1999). A handbook on the theory and methods of differential item functioning (DIF): Logistic regression modeling as a unitary framework for binary and Likert-type (ordinal) item scores. Ottawa, ON: Directorate of Human Resources Research and Evaluation, Department of National Defense.

Zumbo, B. D. (2003). Does Item-Level DIF Manifest Itself in Scale-Level Analyses?: Implications for Translating Language Tests. Language Testing, 20, 136-147.

Zumbo, B. D., & Gelin, M.N. (in press). A Matter of Test Bias in Educational Policy Research: Bringing the Context into Picture by Investigating Sociological / Community Moderated Test and Item Bias. Educational Research and Policy Studies. 

Zumbo, B. D., & Hubley, A. M. (2003). Item Bias. In Rocío Fernández-Ballesteros (Ed.). Encyclopedia of Psychological Assessment, pp. 505-509.  Sage Press, Thousand Oaks, CA.

Zumbo, B. D., & Koh, K. H. (2005). Manifestation of Differences in Item-Level Characteristics in Scale-Level Measurement Invariance Tests of Multi-Group Confirmatory Factor Analyses. Journal of Modern Applied Statistical Methods, 4, 275-282. 

Page 79: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

79

ReferencesZumbo, B. D., & Rupp, A. A. (2004). Responsible Modeling of Measurement Data For Appropriate

Inferences: Important Advances in Reliability and Validity Theory. In David Kaplan (Ed.), The SAGE Handbook of Quantitative Methodology for the Social Sciences (pp. 73-92). Thousand Oaks, CA: Sage Press.           

Zumbo, B. D., Sireci, S. G., & Hambleton, R. K. (2003). Re-Visiting Exploratory Methods for Construct Comparability and Measurement Invariance: Is There Something to be Gained From the Ways of Old? Annual Meeting of the National Council for Measurement in Education (NCME), Chicago, Illinois.

Zumbo, B. D., & Witarsa, P.M. (2004). Nonparametric IRT Methodology For Detecting DIF In Moderate-To-Small Scale Measurement: Operating Characteristics And A Comparison With The Mantel Haenszel. Annual Meeting of the American Educational Research Association (AERA), San Diego, CA.

Page 80: Bruno D. Zumbo, Ph.D. Professor University of British Columbia Vancouver, Canada

80

Thank You For Your Attention!

Professor Bruno D. Zumbo

University of British Columbia

[email protected]

http://www.educ.ubc.ca/faculty/zumbo/zumbo.htm