The symptom of DIF is that a "focal" group does better (or worse) on a particular item than a "reference" group, given those groups' performance on the rest of the test. When DIF exists, it prompts an investigation into the nature of the construct for the focal group (leading to test revision), and a re-estimation of respondents' measures.
Since any two samples will inevitably produce slightly different estimates of the difficulty of any item, an important aspect of DIF analysis is to screen out trivial or unreplicable item calibration differences across groups. There are three hurdles to establishing DIF:
1) Is the DIF statistically significant? A significance test between the two item calibrations is often the only screening applied. It is useful because it can save the researcher from taking seriously large calibration differences that have little supporting data. If there are only 10 members of a focal group, and a dichotomous item is reported as 1.0 logits harder for them, is there a statistical case for DIF? No, because the S.E. of 1.0 logits with only 10 replications is at least 0.6 logits, i.e., t<1.6. But DIF analysis is asymmetrical. Lack of statistical significance makes us skeptical that real DIF exists. Statistical significance does not assure us of it. "The most commonly occurring weakness in the application of [statistical] methods is undue emphasis on tests of significance." (Yates 1964).
2) Does the DIF have substantive implications? Are the different item calibrations, obtained from different groups, substantively the same, so that either measure leads to the same conclusions? It may be that dichotomous item 23 calibrates .1 logits easier for 10,000 girls than for 10,000 boys. Is there a case for a DIF adjustment? It might appear so, because the SE of 0.1 logits is statistically .03 logits. But the score impact of this DIF is, at most, .03 score points for an on-target item, .01 for an off- target item. Since such a small difference would not change our conclusions about group performance, it has no general repercussions. At worst, less than 3% of the focal group are penalized one point. Most of these would be at the ability level of the item's difficulty. If the item is far from a cut-point, then even real DIF would have no practical consequence.
But is this fair to everyone? If the DIF is taken as against individual boys, a .05 logit adjustment up in this item for boys may lift some boys above pass-fail. On the other hand, if the DIF is taken as in favor of individual girls, a .05 logit adjustment down in this item for girls may drop some girls below pass-fail. Since the calibration difference is so small, there is no way of knowing which, if either, is a "correct" adjustment. Further, once adjustment is embarked on, fairness mandates that the procedure be repeated for every other item. This leads to endless minuscule changes in measures and pass-fail decisions, and makes the final pass-fail decisions ever more arbitrary.
3) Is the DIF real or accidental? A single statistical test cannot differentiate between a real effect and an accident. A run of 10 coin-tosses producing 10 heads is unlikely, but does not prove the coin is biased. In advance of test administration, expert screening committees struggle in vain to identify and eliminate items that will manifest DIF. Once told that an item exhibits DIF against a focal group, however, it is easy for experts to imagine why this "must" be so. In the end, the "reality" of most statistically-detected DIF remains open to dispute.
The best test for DIF is "Does it replicate?" When the focal group is split in various ways, random or systematic, does the item continue to exhibit DIF in the same way for each subgroup? If so, the DIF shows evidence of being real. If not, it may be just an accident of sampling.
When DIF meets all three criteria there is reason to regard the DIF item as working like two "different" items. The data set may be restructured so that there is one representation of the item for the reference group and another for the focal group. Measures derived from this reconfigured data will be automatically adjusted for the DIF of that item.
Yates Frank (1964) Sir Ronald Fisher and the Design of Experiments. Biometrics 20:307-321
When to adjust for Differential Item Functioning. Du Y, Yates F. Rasch Measurement Transactions, 1995, 9:1 p.414
MOMS Program. Rasch Measurement Transactions, 1995, 9:1 p.413
Please help with Standard Dataset 4: Andrich Rating Scale Model
|Rasch Measurement Transactions (free, online)||Rasch Measurement research papers (free, online)||Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch||Applying the Rasch Model 3rd. Ed., Bond & Fox||Best Test Design, Wright & Stone|
|Rating Scale Analysis, Wright & Masters||Introduction to Rasch Measurement, E. Smith & R. Smith||Introduction to Many-Facet Rasch Measurement, Thomas Eckes||Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr.||Statistical Analyses for Language Testers, Rita Green|
|Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar||Journal of Applied Measurement||Rasch models for measurement, David Andrich||Constructing Measures, Mark Wilson||Rasch Analysis in the Human Sciences, Boone, Stave, Yale|
|in Spanish:||Análisis de Rasch para todos, Agustín Tristán||Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez|
|Forum||Rasch Measurement Forum to discuss any Rasch-related topic|
Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement
Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.
|Coming Rasch-related Events|
|Sept. 15-16, 2017, Fri.-Sat.||IOMC 2017: International Outcome Measurement Conference, Chicago, jampress.org/iomc2017.htm|
|Oct. 13 - Nov. 10, 2017, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|Oct. 25-27, 2017, Wed.-Fri.||In-person workshop: Applying the Rasch Model hands-on introductory workshop, Melbourne, Australia (T. Bond, B&FSteps), Announcement|
|Jan. 5 - Feb. 2, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|Jan. 10-16, 2018, Wed.-Tues.||In-person workshop: Advanced Course in Rasch Measurement Theory and the application of RUMM2030, Perth, Australia (D. Andrich), Announcement|
|Jan. 17-19, 2018, Wed.-Fri.||Rasch Conference: Seventh International Conference on Probabilistic Models for Measurement, Matilda Bay Club, Perth, Australia, Website|
|April 13-17, 2018, Fri.-Tues.||AERA, New York, NY, www.aera.net|
|May 25 - June 22, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|June 29 - July 27, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com|
|Aug. 10 - Sept. 7, 2018, Fri.-Fri.||On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com|
|Oct. 12 - Nov. 9, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|The HTML to add "Coming Rasch-related Events" to your webpage is:|
The URL of this page is www.rasch.org/rmt/rmt91e.htm