 Research
 Open Access
 Published:
Detecting differential item functioning using generalized logistic regression in the context of largescale assessments
Largescale Assessments in Education volume 2, Article number: 4 (2014)
Abstract
Background
When studying student performance across different countries or cultures, an important aspect for comparisons is that of score comparability. In other words, it is imperative that the latent variable (i.e., construct of interest) is understood and measured equivalently across all participating groups or countries, if our inferences regarding performance can be regarded as valid. Relatively fewer studies examined an itemlevel approach to measurement equivalence, particularly in settings where a large number of groups is included.
Methods
This simulation study examines itemlevel differential item functioning (DIF) in the context of international largescale assessment (ILSA) using a generalized logistic regression approach. Manipulated factors included the number of groups (10 or 20), magnitude of DIF, percent of DIF items, the nature of DIF, as well as the percent of affected groups with DIF.
Results
Results suggested that the number of groups did not have an effect of the performance of the method (high power and low Type I error rates); however, other factors had impacted the accuracy. Specifically, Type I error rates were inflated in nonDIF conditions, while they were very conservative in all of the DIF conditions. Power was generally high, in particular in conditions where DIF magnitude was large, with one exception – in conditions where DIF was introduced in difficulty parameters and the percent of DIF items was 60.
Conclusions
Our findings presented a mixed picture with respect to the performance of the generalized logistic regression method in the context of large number of groups with large sample sizes. In the presence of DIF, the method was successful in distinguishing between DIF and nonDIF, as evidenced by low Type I error and high power rates. On the other hand, however, in the absence of DIF, the method yielded increased Type I errors.
Background
When studying student performance across different countries or cultures, an important aspect for comparisons is that of score comparability. In other words, it is imperative that the latent variable (i.e., construct of interest) is understood and measured equivalently across all participating groups or countries, if our inferences regarding performance can be regarded as valid. The psychometric property that typically must hold for scores to be comparable is known as measurement invariance (Meredith [1993]), absence of differential item functioning (Hambleton et al. [1991]; Mellenbergh [1994]; Swaminathan & Rogers [1990]), or lack of bias (Lord [1980]). Regardless of the term used, the literature on scale score equivalence in largescale achievement tests has received considerable attention (e.g., Ercikan [2002]; Hambleton [2002]; Oliveri et al. [2012]). Many of these investigations have focused on pairwise comparisons of countries (Oliveri et al. [2012]; Oliveri [2012]), the latter of which uses both empirical and simulated data. In the context of international attitude and behavior scale development, researchers have also used multiplegroup confirmatory factor analysis (MGCFA, Jöreskog [1971]; OECD [2010]). Under an MGCFA framework, researchers evaluate hypotheses of configural, metric, and scalar invariance and conduct difference tests to evaluate which level of invariance is supported by the responses.
Relatively fewer studies examined an itemlevel approach to measurement equivalence, particularly in settings where a large number of groups is included. One prominent exception includes the Programme for International Student Assessment (PISA), which uses an ANOVAlike approach to examine itembycountry interactions (OECD [2012]). This general paucity might stem from the fact that, until recently, methodologies that would allow for simultaneous examination of a large number of groups for comparisons were not available. As we note below, generalized versions of some methodologies used for the two group comparisons are now available to researchers for conducting comparisons across more than two groups; however, little information on the performance of these methods exists, in particular in the context of largescale studies such as PISA or the Trends in International Mathematics and Science Study (TIMSS). Also, to our knowledge, relatively little literature exists that questions the current practices of measurement invariance via MGCFA framework in such settings.
A recent study by Rutkowski and Svetina ([2014]) provides some evidence that typically recommended criteria for evaluating invariance in MGCFA (e.g., change in fit statistics; chisquare difference tests) may not always be appropriate when large numbers of groups are compared, as is typically the case in an international assessment or survey context. More specifically, Rutkowski and Svetina studied the performance of MGCFA and associated fit criteria when the number of groups are relatively large and found that as a measure of overall model fit, the chisquare statistic was not useful, as it suggested strong modeldata misfit across all studied conditions. Furthermore, the chisquare difference test (as typically applied) was also too conservative in the studied context. While results for the chisquare test were not surprising, fit indices results were unexpected to some extent. Namely, in examining the overall fit indices (RMSEA, CFI, TLI, and SRMR), the authors found that currently accepted cutoffs for the CFI, TLI, and SRMR as an overall fit indicators were generally suitable (although SRMR in isolation produced somewhat conservative results). However, when considering the relative fit (i.e., change in a fit index from configural to metric or metric to scalar) via changes in CFI and RMSEA, the authors found that the change in RMSEA associated with the metric invariance hypothesis to be increased or larger than a typically accepted difference of .010.
As mentioned above, there has been little research in DIF settings for multiple groups scenarios despite large interest in crosscultural and multilingual research (Fidalgo & Scalon [2010]). This may be partly due to the availability of methods that allow for more than two group comparisons simultaneously, given that a more common alternative approach is somewhat tedious and it requires several steps; namely, it involves pairwise comparisons among all group pairs. In practice, this pairwise comparisons method has been adopted by several researchers, and several variations on this approach exist. For example, as outlined in Ellis and Mead ([2000]), one way to conduct pairwise comparisons and multiple group DIF analysis is to consider in the first step fitting a separate item response theory (IRT) model to obtain item parameters for each group or country (of course, stringent IRT assumptions are assumed here to have been met prior to model fitting). Choice of software may depend on the analyst’s skill or software availability, but typical choices include BILOGMG (Zimowski et al. [1996]) or MULTILOG 7.03 (Thissen et al. [2003]). Then, an analyst would link the parameters using a linking software, for example, EQUATE 2.1 (Baker [1993]). After a successful linking procedure, item parameters would be compared among the groups.
Versions of this approach have been adopted in literature on crosscultural phenomena, including translation of surveys measuring personality (e.g., Ellis & Mead [2000]), global employee attitudes (e.g., Ryan et al. [2000]), dominance (e.g., Kulas et al. [2011]), emotional functioning (e.g., Petersen et al. [2003]), health/quality of life (e.g., Scott et al. [2006][2007]), and reading (e.g., Glas & Jehangir [2013]; Oliveri & von Davier [2011], [2014]), to name a few. Specifically, Oliveri and von Davier ([2011], [2014]) have provided empirical evidence that allowing a subset of item parameters to be uniquely estimated offers one way to improve modeltodata fit and reduces problems with comparability across heterogeneous populations and associated parameter estimate bias. Additionally, Glas and Jehangir ([2013]) proposed using a Lagrange multiplier (LM) test to identify poorly fitting (polytomous) background items where one country serves as the reference country and all other countries are pooled together to serve as the focal country. The single worst fitting item is identified and allowed to be freely estimated; the process is repeated until 95% of the residuals are sufficiently low or four items have been freely estimated for any given country. This approach is to some extent similar to the Oliveri and von Davier method in that some items are freely estimated, although Glas and Jehangir considered background questionnaire items and the process is iterative and, possibly, time intensive.
In the current study, emphasis is given to the investigation at an itemlevel, rather than a scale level but where achievement (rather than background) items are estimated simultaneously rather than using a multistep process adopted in the above mentioned studies. The emphasis here on itemlevel is given based on the following reasons. First, as noted above, typical cutoff criteria within the framework of MGCFA may not be appropriate for comparing large numbers of groups, hence the itemlevel analysis may prove more suitable. Second, motivated by invariance research in language testing and translations, Zumbo ([2003]) investigated whether itemlevel DIF manifested itself in the scalelevel analysis. Findings indicated that limiting investigations to the scalelevel only is often not sufficient since itemlevel DIF may be obscured in a scalelevel analyses (p. 146). In other words, itemlevel DIF should also be examined in conjunction with scalelevel investigations because additional insights may be gained. As such, the current paper investigates the performance of one method of DIF detection – generalized logistic regression – under conditions that closely follow international assessments.
The current research attempts to contribute to the literature on DIF when a large number of groups is considered and the analyses are conducted at an itemlevel. Specifically, using a Monte Carlo study, we are interested in answering the following research question: How does a generalized linear logistic regression method perform in identifying DIF when a large number of groups is considered? The remainder of the paper is organized as follows. In the next section, we introduce methods appropriate for itemlevel DIF analysis, with a particular focus on the approach used in the current study – the generalized logistic regression method. Next, we describe the methods and provide a rationale for the choices of the study design. Results, presented separately for 10 and 20group conditions, follow. Lastly, we conclude with a summary and a discussion of limitations.
Methods to examine itemlevel DIF
Several methods exist to investigate itemlevel DIF, including the MantelHaenszel approach (Holland & Thayer [1988]), logistic regression (Swaminathan & Rogers [1990]), which conditions on the observed scores, or methods rooted in item response theory (IRT), where item characteristic functions are estimated for each group and then compared to investigate DIF, such as Lord’s chisquare test (Lord [1980]) or Raju’s method (Raju [1988], [1990]).
Several factors likely influence any analysts choice of DIF method, including a preference for one of the aforementioned frameworks or methodological approaches (IRT vs. nonIRT), the type of DIF effect of interest (uniform, nonuniform, or both), underlying assumptions (parametric or nonparametric), the number of groups under comparison (two or more), and data characteristics (e.g., dichotomous or polytomous scoring; missing data). Over the last several decades, the literature has featured many DIF methods suitable for two groups (one reference and one focal) and reviews of the methods suitable for dichotomous or polytomous items (e.g., Camilli & Shepard [1994]; Millsap & Everson [1993]; Penfield & Lam [2000]; Potenza & Dorans [1995]). Furthermore, performance of different DIF methods have been examined (e.g., Li et al. [2012]; Penfield [2001]), including those that allow for multiple groups (i.e., > 2), such as generalized MantelHaenszel (e.g., Fidalgo & Madeira [2008]; Fidalgo & Scalon [2010]; Penfield [2001]), generalized Lord’s test (e.g., Kim et al. [1995]) and generalized logistic regression (Magis et al. [2013]; Magis et al. [2011]).
Given our study design, discussed subsequently, and our goal of studying itemlevel DIF when the number of groups compared is greater than two, we consider the performance of the generalized logistic regression method (Magis et al. [2011]). The method is an extension of the wellknown logistic regression approach to investigate DIF between two groups, as proposed by Swaminathan and Rogers ([1990]). Our choice of the generalized logistic regression as the method of choice to study DIF was based on the following. As suggested by Magis et al., the method allows for simultaneous estimation of group parameters, hence it avoids multistep, pairwise comparisons process (and it eliminates the necessity to use multiple software packages). In addition, its flexible framework allows for investigation of uniform, nonuniform, or both types of DIF within subgroups, which is an attractive feature since in international settings, it is likely that either type of DIF is plausible, given the inherent diversity and complexity of the studied context. And more so, it is an improvement over other generalized methods, such as the generalized MantelHaenszel method, which only tests for uniform DIF. Lastly, no merging of the focal groups is necessary.
As noted in the Methods section below, for our analysis, we used the generalized logistic regression approach implemented as a function in difR package (Magis et al. [2013]) in R (R Development Core Team [2012]). It is an appropriate method when data are dichotomously scored and where more than two groups are considered. Furthermore, this method allows for studying both uniform and nonuniform DIF and it investigates one item at the time, such that in the process, the remaining items not under consideration are assumed to be nonDIF^{a}. As the method of analysis in this study, we briefly highlight the main components of the generalized logistic regression. Interested readers are directed to the original work by Magis et al. ([2011]) for complete details.
The generalized logistic regression DIF model, as presented by Magis et al. ([2011]), has the following form:
where π_{ ig } is the probability of examinee i from group g correctly responding to an item, logit is the natural log of the odds of correctly answering an item, α and β are common intercept and slope parameters (i.e., for all groups), α_{ g } and β_{ g } are groupspecific slope and intercept parameters, and S_{ i } is the total test score for examinee i, which serves as a matching variable and a proxy for the ability level of the examinee. For model identification purposes, groupspecific parameters for the reference group (denoted as g = 0), α_{0} and β_{0}, are set to zero. In other words, if g = 0 (i.e., reference group), logit(π_{ ig }) = α + βS_{ i }, and if g ≠ 0 (i.e., focal groups, g = 1, 2, … , F), logit(π_{ ig }) = (α + α_{ g }) + (β + β_{ g })S_{ i }. An item is said to contain DIF if the probability π_{ ig } varies across the groups of examinees (i.e., there is an interaction between the group membership and the item response). This occurs when at least one of the group parameters, α_{ g } or β_{ g }, is different from zero. If all groupspecific parameters equal zero, an analyst would conclude that no DIF is present.
According to Magis et al. ([2011]), three types of DIF can be investigated using this framework: a) uniform DIF (UDIF), b) nonuniform DIF (NUDIF), and both types of DIF effects together (DIF). Tested null hypotheses for these three types include:
Statistically, DIF is assessed by an examination of the model parameters. Let τ represent a vector of model parameters, τ = (α, α_{1} … α_{ F }, β, β_{1} … β_{ F })^{T} for DIF and NUDIF and τ = (α, α_{1} … α_{ F }, β)^{T} for the UDIF, and$\hat{\mathit{\tau}}$be the maximum likelihood estimator of τ. Using maximum likelihood, we can test the null hypotheses in (2) to (4) by different methods, such as the Wald test or the likelihood ratio test (LRT). In the current study, we utilized LRT as a criterion to detect DIF, where the null (M_{0}) and alternative (M_{1}) models were given by:
and
DIF is then tested by the lambda statistic (Wilks [1938]), where$\mathbf{\Delta}=2log\left(\frac{{\mathit{L}}_{0}}{{\mathit{L}}_{1}}\right)$, L_{0} and L_{1} are the corresponding maximum of the likelihoods for the M_{0} and M_{1}, respectively, which follows an asymptotic chisquare distribution with degrees of freedom of the asymptotic null distribution of Δ equal to 2 F for DIF and F for both the UDIF and NUDIF. In other words, an item would be flagged as DIF when the lambda statistic is sufficiently large, or stated alternatively, when the alternative model is preferred.
Methods
Study design
As stated above, our study is situated within the context of ILSAs; hence, we designed our simulation study as follows. First, we follow the practice of test design and administration by adopting a rotated booklet design (see Table 1). In operational testing, such as in the TIMSS 2007 assessment in grade 4, 10 to 14 items per block were administered and each student received a total of four blocks (two for mathematics and two for science). In our booklet design, we aimed to achieve a similar set up for only one content assessment.
As Table 1 shows, for purposes of our study, each booklet contained three blocks, where each simulee received a single booklet. An approximately equal number of simulees received any one of the seven booklets within a group (with random assignment), and each group was administered all seven booklets. Each block contained 15 items, resulting in 45 items per simulee. The remaining items (those in blocks not administered to any one simulee at the time) were treated as missing by design in the analysis.
Our study design choices were motivated by what is typically observed in practice with ILSAs. To examine the impact of various factors on DIF detection, we manipulated several variables (see Manipulated factors). To generate the data, we used empirical approximations of item and person parameters from an existing dataset (see Data generation). In what follows, we offer rationale for the choices made in the study design, including manipulated factors (and respective levels) of the simulation study and the process of the selection/modification of item and person parameters.
Manipulated factors
Several factors were manipulated in the study, including

a)
number of groups (10 or 20);

b)
percent of groups affected by DIF (40% or 70%)

c)
percent of DIF items (20%, 40%, or 60%);

d)
nature of DIF (difficulty or discrimination); and

e)
magnitude of DIF (small or large).
The number of groups examined in the study was set at 10 or 20. One of our goals was to examine the performance of generalized logistic regression model as implemented in the difR package (Magis et al. [2013]) in a relatively large group setting; these group sizes approximate more closely the operational context of largescale surveys. For example, the Teaching and Learning International Survey typically has 20–30 participating educational systems, depending on the grade level (OECD [2010]). We do, however, recognize that other assessments, such as PISA or TIMSS, may have 60 participating educational systems or more. One of the primary reasons we focused on what might be considered lower bounds of the number of groups is the computing time required per analysis. In the preliminary analysis of conditions with 30 and 60 groups (with other design choices remaining the same), analysis of one replication required over 3 and 5.5 hours respectively on an IBM e1350 highperformance computing system^{b}. We further recognize that the computing time is not necessarily an issue in reality when only one dataset is considered. However, for the purposes of our simulation study, with multiple replications within any one condition, it was necessary to choose a smaller number of groups.
The number of groups affected by DIF was set to be 40% or 70% of total groups. This meant that in conditions with 10 groups, either 4 or 7 groups contained some items with DIF, and in conditions with 20 groups, 8 or 14 groups were affected by DIF. For all conditions, the same arbitrarily selected reference group was used, while the remaining groups were treated as focal. The assignment of groups affected by DIF was made randomly. The reference group had a moderate sample size (N = 6370 in 10group and N = 5341 in 20group conditions), was the same reference group across all conditions, and its nonDIF item parameters were similar to those of focal groups.
The percent of items that were modeled as DIFitems was set to be 20%, 40%, or 60%, which resulted in 21, 42, or 63 DIF items for selected groups (a total of 105 items were administered for any one group – see more detail below on how these parameters were selected and manipulated). The assignment of DIF items was also drawn at random, and was kept consistent across groups and levels of DIF (i.e., nature of DIF). Our choice for selecting a wide percent of items to be affected by DIF was primarily motivated by research on largescale assessments in language or crosscultural studies, where it is not uncommon to see a large number (percent) of items flagged as DIFitems. For example, Stubbe ([2011]) investigated DIF in the Germanlanguage versions of an international reading assessment, where he found an overall percentage of significant DIF items to be over 74%; percent of DIF items ranged from 71.4% to 77.4% for multiplechoice and constructed responses item types, respectively (2011). Further, Dorans and Middleton ([2012]) cogently argue, using wellestablished literature and empirical support, that language is a condition of measurement and as such, typical assumptions that identical items translated into testing languages are not necessarily equivalent. These “extreme assumptions” underpinning what the authors term “presumed” linking (in contrast to empirical linking) should be subjected to falsifiable tests as a means of building support for a presumed linking across potentially nonequivalent groups. The authors argue that invariance of relationships among test forms should be examined. Further, necessary conditions for scorescale comparability include a set of equivalent anchor items. As such, our design, even at the extreme, includes a set of noninvariant as well as invariant “anchor” items.
The nature of DIF was examined by changing the value of the difficulty or discrimination parameters for DIFitems in affected groups. This is akin to the investigation of uniform and nonuniform DIF, respectively. The generalized logistic method used in the analysis allows for investigation of both types of DIF; hence, in our analysis, we opted for DIF investigation of both types of DIF across all conditions. Focal groups that did not contain DIF items had difficulty and discrimination parameters equal to those of the reference group.
The magnitude of DIF was examined at two different levels. In conditions with small DIF, values for difficulty parameters for affected items differed from nonDIF items by .50 and discriminations varied by .40. In conditions with large DIF, differences were 1.00 and .80 for difficulty and discrimination parameters, respectively. Across all of the conditions, differences in value of item parameters was made at random, either in favor of the reference or focal group in question (i.e., differences in parameter values were either added or subtracted from the nonDIF item in question by a random draw). A note should be made regarding our choice of the difference values in introducing DIF. Within typical DIF investigations, some authors suggest that differences of .50 in difficulty, for example, might be considered large DIF (c.f., Goodman et al. [2011]). In our study, we used descriptors of small and large only to differentiate different levels of DIF, not necessarily the classification of the DIF magnitude. More importantly, we wanted to use values that can be found in examining difference in item parameters in largescale assessments. Specifically, in a study of 21 countries in PISA 2009, Rutkowski and Rutkowski ([2014]) found that differences in estimates of item difficulties between any two countries could be as large as 6.22 and as small as 1.95, with an absolute average of difficulty difference of .90 and absolute maximum mean difference across all items of 2.96^{c} (Rutkowski, L., & Rutkowski, D. One size does not fit all: The impact of item parameter differences and estimation methods on crosscultural achievement comparisons, submitted manuscript). Similarly, the authors found that the discrimination differences for the same data were found to be in range of .55 and 2.19, respectively, with an average maximum difference of 1.32.
Data generation
In order to simulate our data, we used existing released item parameters that could be found on typical ILSAs. Specifically, for nonDIF conditions, we estimated item parameters from the 2007 TIMSS grade 4 assessment. On 2007 TIMSS, a total of 179 items were released for grade 4 across all content areas, of which 174 were scored dichotomously (i.e., either multiplechoice or constructed response items with only two categories; Olson et al. [2008]). From this pool of 174 released items (i.e., their parameters), we randomly drew without replacement a total of 105 items. The 105 items represented the total number of items used in our study, where each of the seven blocks contained 15 items.
Once the 105 item parameters were selected at random, they were treated as nonDIF and were assigned to nonoverlapping blocks at random for the reference group and any focal group(s) selected to be nonDIF. Table 2 shows the average, minimum, and maximum value of the difficulty and discrimination item parameters for nonDIF items across the seven blocks. For example, as shown in Table 2, item 1 had an average difficulty of .27 across the seven blocks, and its difficulty parameter ranged from .23 to 1.06. As Table 2 suggests, item parameters drawn from the released 2007 TIMSS grade 4 assessment varied in both difficulty and discrimination, although these values would be considered typical in such assessments.
Person parameters for each group were drawn randomly from a normal distribution with some mean and standard deviation (see Table 3). The means and standard deviations varied across groups; the differences in population means (standard deviations) were purposefully introduced to resemble parameters likely to be found in largescale assessments. Values found in Table 3 were obtained by random sampling with replacement of the “standardized” plausible values from 44 participating countries from grade 4 mathematics on TIMSS 2007. Note that for 10group conditions, every other value for 20group conditions was used. Although these values are based on plausible values provided by TIMSS, they should not be used as estimate of countries’ abilities. Rather, we use them here only as an approximations in order to introduce group differences.
Table 3 also shows that the sample size varied across the groups. These sample sizes were chosen to resemble current largescale assessments, where group sample sizes are typically large and vary per participating country. Values for sample sizes were randomly drawn from ~ N(5000, 1000) for 20 groups; every other sample size from the 20group pool was assigned to the 10group. Within each group, approximately the same number of simulees received each of the booklets, also assigned at random.
We simulated binary responses to items according to a 2parameter logistic (2PL) model via the sim function implemented in irtoys package (Partchev [2012]) in R. The 2PL specifies the probability of person i endorsing an item j (X_{ ij } = 1) as:
where θ_{ i } is the person parameter for person i, b_{ j } is the location parameter (difficulty) for item j, and a_{ j } is the discrimination parameter for item j. A fully factorial design yielded a total of 50 conditions; 2 (number of groups) × 2 (percentage of DIFaffected groups) × 3 (percent of DIF items) × 2 (nature of DIF) × 2 (magnitude of DIF) + 2 (nonDIF conditions), each replicated 100 times.
Analysis
Generated data were analyzed within the framework of generalized logistic regression as implemented in the difR package under genDichoDif (method=“genLogistic”) function (Magis et al. [2013]) in R (R Development Core Team [2012]). The reference group (group 8) was used across all conditions in both 10 and 20group cases, while remaining groups were treated as focal^{d}. Default options were used for the analysis, including the test of both uniform and nonuniform DIF, LRT was used as the criterion to detect DIF items, and the method was not used iteratively to purify the set of anchor items. Items not administered per any one record were treated as missing in the analysis, although no missingness was introduced in the group membership (i.e., every simulee had a group membership designation).
Performance criteria
In order to evaluate the performance of the generalized logistic regression, as implemented in difR, we report Type I error and power rates. Type I error rates are reported as the proportion of items within a replication that were generated as nonDIF items but which are flagged as DIF in the analysis. Similarly, power rates are computed as the proportion of items that are simulated as DIF items that are also flagged as having DIF. Rates are averaged across 100 replications for each condition separately.
Results
For the purpose of organization, we present our results for the 10group conditions first, followed by the 20group conditions. Type I error rate for the NONDIF condition (where data for all 10 groups was modeled using the same set of item parameters) was inflated across the 100 replications; the mean (standard deviation) for the NONDIF condition was .20 (.02), which deviates from a typically used .05 or .01 levels. For the 10group DIF conditions, regardless of the level or nature of DIF introduced, the Type I error rates were all .00. Given the consistency across all studied conditions, we only mention the Type I error rates of zero and do not present these results in a table. Our finding, with respect to Type I error rates, suggests that the generalized logistic method was accurate in all DIF conditions by not flagging nonDIF items as having DIF.
Table 4 shows the average power rates (and standard deviations) for DIF conditions for the 10group cases. Rates were generally moderate to high across the studied conditions, with a few exceptions. The highest power rates were noted in conditions where the difference in item parameters for both the discrimination and difficulty were large. This is not unexpected, but it should also be noted that this difference was mostly noted in conditions where discrimination parameters varied (panel (a)); as panel (b) showed, uniform DIF was quite constant across different DIF levels. It was also noted that, generally, the power rates were higher in conditions where the percent of groups affected by DIF was at 40% compared to 70%. This seemed to be the case for DIF conditions where either discrimination or difficulty was invariant.
Several patterns were also noted. For example, in panel (a), where DIF was introduced in the discrimination parameter, the percent of DIF items within an affected group did not exhibit a clear pattern. For example, when discrimination DIF was .40 and 40% of groups were impacted by DIF, increasing the percent of DIF items did not necessarily result in higher power rates. For example, with 20% of DIF items, the estimated power rate was .70, while further increases in the percent of DIF items per group resulted in power rates of .65 and .71, respectively. In conditions with 70% of groups being impacted by DIF, more expected results were noted. Namely, power rates increased as the number of items with DIF increased, from .51, .57, and .63 for the 20%, 40%, and 60%, respectively. Interestingly, when discrimination DIF was .80, power rates slightly decreased as the percent of DIF items increased; this was the case for conditions with 40% and 70% of groups affected by DIF. It should be noted, however, that when discrimination differences were .80, all power rates were very high (ranged from .97 to 1.00).
Panel (b) shows the results for the conditions where DIF was introduced in the item difficulty parameter. It was noted that for all but four conditions, power rates were very high (ranged from .92 to 1.00). However, in conditions where the percent of DIF items was 60%, power rates are unacceptably low (~.20). This result was found regardless of the amount of DIF or the percent of groups that were affected by DIF items. Across all 10group conditions, standard deviations were very small, suggesting that across the 100 replications within any one condition, power rates were quite homogeneous.
Similar to the 10group conditions, Type I error rates were zero in all studied conditions in the 20group cases. Power results for the 20group conditions are presented in Table 4. Based on these results, there are clear parallels between the 10 and 20group cases. Similar to the 10group setting, the nonDIF condition had a Type I error rate much higher than would normally be expected. In particular, we estimated an operational error rate of 0.18. In other words, 18% of items, on average across replications, were identified as differentially functioning under the condition where items were invariant across groups. Further, and commensurate with the findings for the 10group case, the Type I error rates were zero to at least the third decimal place under all considered conditions, regardless of DIF location (difficulty or discrimination), magnitude of DIF, percent of items with DIF, or percept of groups with DIF items. Collectively, these findings suggest that the generalized logistic regression method has good specificity when some items have DIF and poorer specificity when no items have DIF^{e}.
Next, Table 4 indicates that, somewhat surprisingly, power to detect DIF items was similar to the 10group case, suggesting that the number of groups does not matter much for detecting DIF items. Rates of DIF detection were mostly moderate to high, with exceptions similar to the 10group case. In particular, when 60% of difficulty parameters vary across groups, the power to detect DIF is just .18. And this finding was consistent, regardless of DIF magnitude or the proportion of groups with DIF items. These results suggest that when high proportions of items function differently, the method considered does not do well at detecting these items. According to panel (a), we also found that when the DIF magnitude in the discrimination was small, detection rates were somewhat low (.55 to .72) and that detection rates were the lowest among these (.55 to .68) when a higher percentage of groups (70 compared to 40) had items with DIF in the discrimination parameters. In contrast, the magnitude of DIF in the discrimination was large, detection rates were very high, from .97 when 60% of discriminations had DIF to 1.00 when just 20% of discriminations had DIF. The findings were very consistent across both the 40% and 70% of DIFgroup conditions, suggesting that the percentage of groups with DIF items matters little.
With respect to the findings for the 20group case when DIF was located in the difficulty parameter, Table 4, panel (b) indicates largely very similar findings to the 10group case. That is, detection rates are high (.96 to 1.00), with the exception of the condition where 60% of items have DIF. And there are slight differences between the 40% and 70% of groups conditions when 40% of items have a small magnitude of DIF. Specifically, the proportion of DIF items detected as having DIF is slightly higher when a lower proportion of groups have DIF (.98 versus .96); however, this is a small difference. And there are no differences in detection rates between the 40% and 70% of groups conditions when 20% of items have DIF in the difficulty parameter, regardless of DIF magnitude. Similar to the 10group conditions, standard deviations were very small, again indicating relatively homogenous results for power rates across the 100 replications within any one condition.
Discussion and conclusions
As a whole, these findings present a mixed picture with respect to the performance of the generalized logistic regression method when conditions are similar to those found in many ILSA settings. On one hand, it seems that general logistic regression tended to detect DIF well in most cases (see remark on surprising results below). In the presence of DIF, the method was successful in distinguishing between DIF and nonDIF, as evidenced by low Type I error and high power rates. On the other hand, however, in the absence of DIF, the method yielded increased Type I errors.
More specifically, when there are some differentially functioning items in any of the considered conditions, generalized logistic regression does well at ignoring items that are invariant across groups, whether the parameter of interest is the discrimination or difficulty. Further, group size did not appear to be predictive of performance. That is, the findings across the two group sizes (10 and 20) were quite consistent. This finding is in contrast to previous research that used a multiplegroups confirmatory factor analytic approach to detect measurement invariance in large numbers of groups (Rutkowski & Svetina [2014]). Specifically, performance of the considered methods depended on the number of groups evaluated, with generally poorer performance connected to larger numbers of groups. Given that international surveys and assessments often feature dozens of systemlevel participants, a lack dependency on group size is a strength of the method considered here, at least where detection rates were high. And in many cases, power or detection rates were quite high, with more than 90% of DIF items being flagged in 28 of 48 considered conditions. In contrast, power rates were low in 20 of 48 conditions and surprisingly so in 8 of these conditions. Specifically, detecting DIF in discrimination was a challenge for this method when the DIF magnitude was small, with rates ranging from .55 to .72. But perhaps most surprising were the exceedingly low power rates when DIF was located in 60% of difficulty parameters, with rates that ranged from 18% to 20%. One possible explanation for this finding could be that with such a high proportion of DIF items, the method could not accurately identify a set of invariant items against which to compare the noninvariant items – in other words, the presence of DIF items masked the presence of other DIF items; however, it is puzzling that this finding was limited to the difficulty parameter. Nonetheless, uniform DIF is clearly not as easily distinguishable as nonuniform DIF in this particular condition. Lastly, it should be noted that in NONDIF conditions, regardless of the number of groups, Type I error rates were quite inflated. This result is consistent with previous research, which suggested that differences in group means (i.e., impact) may contribute to the inflated Type I error rates (DeMars [2010]; Jiang & Stout [1998]). As DeMars ([2010]) suggested, methods such as logistic regression that match groups based on their observed score may not result in well matched examinees on true proficiency. This may lead to false DIF detection due to inaccurate matching^{f}.
As with any study, our study has limitations due to the study design and the methodology used. With respect to the study design, the generalizability of our conclusions are limited by the selection of manipulated factors in the study. Although we aimed to design our study to closely resemble the context of ILSAs (e.g., rotated booklet design, large and varying sample size, released item parameters), due to space and time constraints, we did not investigate all potential scenarios that ILSAs may occupy in a real context. For example, our focus was only on dichotomously scored data, such that data were generated using a 2PL IRT model. It is often the case that on largescale assessments, several item types are present, including items scored dichotomously and polytomously. Further, in the current study, we did not allow for a lowerasymptote parameter, which may be justified for some item types, such as in multiplechoice items. Also, in cases where the effect is rather small (i.e., discrimination or difficulty parameters vary only by a very small amount), a question of the method’s performance remains unanswered. These limitations present an opportunity for further research, including a comparison between using the rotated booklet design with planned missing and conditions where no missingness is present. Additional research should consider this method and its suitability of its use in the context of ILSA studies, as high Type I error rates were observed in the absence of DIF. One avenue of exploration could be to invoke an iterative process called item purification to investigate whether Type I error rates stabilize^{g}. Further research should also examine the performance of the studied method in conditions where large amounts of uniform DIF is present, as our results pointed to a rather surprising phenomenon in conditions with 60% of DIF items and difficulty parameter as variant. Namely, power rates in those situations dropped considerably to unacceptable levels, thus further investigations regarding the “breaking point” of the method may be warranted.
From a methodological or practical view, a limitation is that only detection of DIF was investigated in the current study. In practical applications, once DIF is detected in an item or set of items, it is crucial to examine the underlying reasons as to why DIF occurred. As Ferne and Rupp ([2007]) suggested within the context of language testing, use of expert panels has historically been underutilized, citing reasons of cost and time. Even further, an important consideration by the analyst/testing developers should be made with regards to what is done with items that are flagged as DIF. Although not the focus of the current study, we advocate that detecting DIF in items on assessments is only the first, albeit important, step in testing. Once items are identified as functioning differentially (among the studied groups), substantive analysis should be implemented – for example, use of cultural or linguistic experts may be useful in shedding light on why DIF occurred. With that information, test developers could make more informed decisions to revise and/or remove the flagged items in order to ensure score comparability and appropriate inference regarding groups’ performance.
Endnotes
^{a}In this paper, we use the term nonDIF to suggest that the item is modeled as having measurement invariance or that no DIF is present in the item (sometimes referred to in the literature as DIFfree).
^{b}Preliminary analysis run in Mplus (Muthén, & Muthén [1998–2010]) for testing configural and metric invariance using similar design choices resulted in a single replication taking almost 24 hours.
^{c}The authors studied a total of 27 items; averages and maximums of item parameter differences here were based on 26 items because one item seemed to have provided estimates that were extremely rare (i.e., difficulty parameter estimates between two countries that differed up to 17 points). In order to provide a more general idea of how different item parameters between the countries could be, we eliminated this outlier in our reporting.
^{d}This group was chosen as a reference group because it was one of the groups not chosen at random to be impacted by DIF. The selection of this reference group was in that respect at random. This might be slightly different from ILSAs, where the host country of interest may serve as a reference group. Nonetheless, in DIF studies, the choice of reference and/or focal group(s) is often made by the analyst’s purpose and goals.
^{e}We calculated specificity as 1 – P(item is flagged as DIF  item is nonDIF) or 1 − Type I error rate.
^{f}Magis and De Boeck ([in press]) recently found inflated Type I error rates in MantelHaenszel approach, and similarly attribute the inflation to the item impact and difference in item discriminations. Although not the same approach as used in the current study, Magis and De Boeck study found potential sources of Type I error rate inflation as found in DeMars ([2010]) and the current study.
^{g}Item purification is an iterative process, which removes the items currently flagged as DIF from the test scores in order to obtain purified sets of items, unaffected by DIF. Process stops until two consecutive runs yield the same selection of the items. Currently, literature is unclear as to whether item purification controls the Type I error rate to some nominal value and/or maintains higher power (e.g., Magis & Facon [2013]).
References
Baker FB: EQUATE2: Computer Program for Equating two Metrics in Item Response Theory [Computer Program]. University of Wisconsin, Laboratory of Experimental Design, Madison; 1993.
Camilli G, Shepard LA: Methods for Identifying Biased Test Items. SAGE Publications, Thousand Oaks, CA; 1994.
DeMars CE: Type I error inflation for detecting DIF in the presence of impact. Educational and Psychological Measurement 2010, 70(6):961–972. doi:10.1177/0013164410366691 10.1177/0013164410366691
Dorans NJ, Middleton K: Addressing the extreme assumptions of presumed linkings. Journal of Educational Measurement 2012, 49(1):1–18. doi:10.1111/j.1745–3984.2011.00157.x 10.1111/j.17453984.2011.00157.x
Ellis BB, Mead AD: Assessment of the measurement equivalence of a Spanish translation of the 16PF questionnaire. Educational and Psychological Measurement 2000, 60(5):787–807. doi:10.1177/00131640021970781 10.1177/00131640021970781
Ercikan K: Disentangling sources of differential item functioning in Multilanguage assessments. International Journal of Testing 2002, 2(3–4):199–215. doi:10.1080/15305058.2002.9669493 10.1080/15305058.2002.9669493
Ferne T, Rupp AA: A synthesis of 15 years of research on DIF in language testing: methodological advances, challenges, and recommendations. Language Assessment Quarterly 2007, 4(2):113–148. doi:10.1080/15434300701375923 10.1080/15434300701375923
Fidalgo AM, Madeira JM: Generalized mantelhaenszel methods for differential item functioning detection. Educational and Psychological Measurement 2008, 68(6):940–958. doi:10.1177/0013164408315265 10.1177/0013164408315265
Fidalgo ÁM, Scalon JD: Using generalized mantelhaenszel statistics to assess DIF among multiple groups. Journal of Psychoeducational Assessment 2010, 28(1):60–69. doi:10.1177/0734282909337302 10.1177/0734282909337302
Glas C, Jehangir K: Modeling CountrySpecific Differential Item Functioning. In Handbook of International LargeScale Assessment: Background, Technical Issues, and Methods of Data Analysis. Edited by: Rutkowski L, Von Davier M, Rutkowski D. Chapman Hall/CRC Press, London; 2013:97–116.
Goodman J, Wilse J, Allen N, Klaric J: Identification of differential item functioning in assessment booklet designs with structurally missing data. Educational and Psychological Measurement 2011, 71(1):80–94. 10.1177/0013164410387341
Hambleton R: Adapting Achievement Tests into Multiple Languages for International Assessments. In Methodological Advances in CrossNational Surveys of Educational Achievement. Edited by: Porter AC, Gamoran A. National Academies Press, Washinton, DC; 2002.
Hambleton RK, Swaminathan H, Rogers HJ: Fundamentals of Item Response Theory. Sage, Newbury Park, CA; 1991.
Holland WP: Differential Item Performance and the MantelHaenszel Procedure. In Test Validity. Edited by: Wainer H, Braun H. LEA, Hillsdale, NJ; 1988:129–145.
Jiang H, Stout W: Improved type I error control and reduced estimation bias for DIF detection using SIBTEST. Journal of Educational and Behavioral Statistics 1998, 23(4):291–322. doi:10.2307/1165279 10.3102/10769986023004291
Jöreskog KG: Simultaneous factor analysis in several populations. Psychometrika 1971, 36(4):409–426. doi:10.1007/BF02291366 10.1007/BF02291366
Kim SH, Cohen AS, Park TH: Detection of differential item functioning in multiple groups. Journal of Educational Measurement 1995, 32(3):261–276. doi:10.2307/1435297 10.1111/j.17453984.1995.tb00466.x
Kulas JT, Thompson RC, Anderson MG: California psychological inventory dominance scale measurement equivalence: general population normative and Indian, U.K., and U.S. managerial samples. Educational and Psychological Measurement 2011, 71(1):245–257. doi:10.1177/0013164410391580 10.1177/0013164410391580
Li Y, Brooks GP, Johanson GA: Item discrimination and type I error in the detection of differential item functioning. Educational and Psychological Measurement 2012, 72(5):847–861. doi:10.1177/0013164411432333 10.1177/0013164411432333
Lord FM: Applications of Item Response Theory to Practical Testing Problems. Lawrence Erlbaum, New Jersey; 1980.
Magis D, Beland S, Raiche G: difR: Collection of Methods to Detect Dichotomous Differential Item Functioning (DIF) in Psychometrics. R Package Version 4.4. 2013.
Magis, D., & De Boeck, P. (in press). Type I error inflation in DIF identification with MantelHaenszel: An explanation and a solution. Educational and Psychological Measurement.
Magis D, Facon B: Item purification does Not always improve DIF detection a counterexample with Angoff’s delta plot. Educational and Psychological Measurement 2013, 73(2):293–311. doi:10.1177/0013164412451903 10.1177/0013164412451903
Magis D, Raîche G, Béland S, Gérard P: A generalized logistic regression procedure to detect differential item functioning among multiple groups. International Journal of Testing 2011, 11(4):365–386. doi:10.1080/15305058.2011.602810 10.1080/15305058.2011.602810
Mellenbergh GJ: Generalized linear item response theory. Psychological Bulletin 1994, 115: 300–307. 10.1037/00332909.115.2.300
Meredith W: Measurement invariance, factor analysis and factorial invariance. Psychometrika 1993, 58(4):525–543. doi:10.1007/BF02294825 10.1007/BF02294825
Millsap RE, Everson HT: Methodology review: statistical approaches for assessing measurement bias. Applied Psychological Measurement 1993, 17(4):297–334. doi:10.1177/014662169301700401 10.1177/014662169301700401
Muthén LK, Muthén BO: Mplus user’s Guide. Muthén & Muthén, Los Angeles, CA; 1998–2010.
TALIS Technical Report. OECD Publishing, Paris; 2010.
PISA 2009 Technical Report. OECD, Paris; 2012.
Oliveri ME: Investigation of Within Group Heterogeneity in Measurement Comparability Research (Unpublished PhD Dissertation). University of British Columbia, Vancouver, Canada; 2012.
Oliveri ME, Von Davier M: Investigation of model fit and score scale comparability in international assessments. Psychological Test and Assessment Modeling 2011, 53: 315–333.
Oliveri ME, Olson BF, Ercikan K, Zumbo BD: Methodologies for investigating item and testlevel measurement equivalence in international largescale assessments. International Journal of Testing 2012, 12(3):203–223. doi:10.1080/15305058.2011.617475 10.1080/15305058.2011.617475
Oliveri ME, Von Davier M: Toward increasing fairness in score scale calibrations employed in international largescale assessments. International Journal of Testing 2014, 14(1):1–21. doi:10.1080/15305058.2013.825265 10.1080/15305058.2013.825265
Olson J, Martin MO, Mullis IVS: TIMSS 2007 Technical Report. TIMSS & PIRLS International Study Center, Boston College, Chestnut Hill, MA; 2008.
Partchev I: Irtoys: Simple Interface to the Estimation and Plotting of IRT Models. R Package Version 0.1.6. 2012.
Penfield R: Assessing differential item functioning among multiple groups: a comparison of three mantelhaenszel procedures. Applied Measurement in Education 2001, 14(3):235–259. doi:10.1207/S15324818AME1403_3 10.1207/S15324818AME1403_3
Penfield RD, Lam TCM: Assessing differential item functioning in performance assessment: review and recommendations. Educational Measurement: Issues and Practice 2000, 19(3):5–15. doi:10.1111/j.1745–3992.2000.tb00033.x 10.1111/j.17453992.2000.tb00033.x
Petersen MA, Groenvold M, Bjorner JB, Aaronson N, Conroy T, Cull A, Fayers P, Hjermstad M, Sprangers M, Sullivan M: Use of differential item functioning analysis to assess the equivalence of translations of a questionnaire. Quality of Life Research 2003, 12(4):373–385. doi:10.1023/A:1023488915557 10.1023/A:1023488915557
Potenza MT, Dorans NJ: DIF assessment for polytomously scored items: a framework for classification and evaluation. Applied Psychological Measurement 1995, 19(1):23–37. doi:10.1177/014662169501900104 10.1177/014662169501900104
R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria; 2012.
Raju NS: The area between two item characteristic curves. Psychometrika 1988, 53(4):495–502. doi:10.1007/BF02294403 10.1007/BF02294403
Raju NS: Determining the significance of estimated signed and unsigned areas between Two item response functions. Applied Psychological Measurement 1990, 14(2):197–207. doi:10.1177/014662169001400208 10.1177/014662169001400208
Rutkowski L, Svetina D: Assessing the hypothesis of measurement invariance in the context of largescale international surveys. Educational and Psychological Measurement 2014, 74(1):31–57. doi:10.1177/0013164413498257 10.1177/0013164413498257
Ryan AM, Horvath M, Ployhart RE, Schmitt N, Slade LA: Hypothesizing differential item functioning in global employee opinion surveys. Personnel Psychology 2000, 53(3):531–562. doi:10.1111/j.1744–6570.2000.tb00213.x 10.1111/j.17446570.2000.tb00213.x
Scott NW, Fayers PM, Aaronson NK, Bottomley A, De Graeff A, Groenvold M, Koller M, Petersen MA, Sprangers MA: The use of differential item functioning analyses to identify cultural differences in responses to the EORTC QLQC30. Quality of Life Research 2007, 16(1):115–129. 10.1007/s1113600691201
Scott NW, Fayers PM, Bottomley A, Aaronson NK, De Graeff A, Groenvold M, Koller M, Petersen MA, Sprangers MA: Comparing translations of the EORTC QLQC30 using differential item functioning analyses. Quality of Life Research 2006, 15(6):1103–1115. 10.1007/s111360060040x
Stubbe TC: How do different versions of a test instrument function in a single language? A DIF analysis of the PIRLS 2006 German assessments. Educational Research and Evaluation 2011, 17(6):465–481. doi:10.1080/13803611.2011.630560 10.1080/13803611.2011.630560
Swaminathan H, Rogers HJ: Detecting differential item functioning using logistic regression procedures. Journal of Educational Measurement 1990, 27(4):361–370. doi:10.2307/1434855 10.1111/j.17453984.1990.tb00754.x
Thissen D, Chen WH, Bock RD: MULTILOG 7.03. Scientific Software International, Chicago; 2003.
Wilks SS: Weighting systems for linear functions of correlated variables when there is no dependent variable. Psychometrika 1938, 3(1):23–40. doi:10.1007/BF02287917 10.1007/BF02287917
Zimowski MF, Muraki E, Mislevy RJ, Bock RD: BilogMG: MultipleGroup IRT Analysis and Test Maintenance for Binary Items [Computer Software]. Scientific Software International, Chicago; 1996.
Zumbo BD: Does itemlevel DIF manifest itself in scalelevel analyses? Implications for translating language tests. Language Testing 2003, 20(2):136–147. doi:10.1191/0265532203lt248oa 10.1191/0265532203lt248oa
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
DS and LR reviewed the literature, designed and carried out the analyses, and prepared the manuscript. Both authors read and approved the final manuscript.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Svetina, D., Rutkowski, L. Detecting differential item functioning using generalized logistic regression in the context of largescale assessments. Largescale Assess Educ 2, 4 (2014). https://doi.org/10.1186/s4053601400045
Received:
Accepted:
Published:
Keywords
 Differential Item Functioning
 Item Response Theory
 Item Parameter
 Differential Item Functioning Item
 Difficulty Parameter