# Linking Constants with Common Items and Judges

Best Test Design (Wright & Stone, 1979 p. 96) gives formulae for evaluating the statistical quality of a linking constant between dichotomous tests with common items.

When Test A is given to a sample of NA persons, and Test B to a different sample of NB persons, then each item i of the K equally trustworthy common items is estimated to have two difficulties, DiA in Test A and DiB in Test B. DiA and DiB have standard errors of approximately 2.5/NA½ and 2.5/NB½ respectively.

If a cross-plot of the two sets of common items indicates that there is a single constant that adds to all difficulties and abilities in Test B to translate them onto the scale of Test A, then that constant is

with standard error

If NANBN, then

A test of the hypothesis that this linking constant explains the difference between difficulties of common items is

Links with Polytomous Items and Judges

These formulae can be generalized for tests containing polytomous items with both common items and common judges. Tests A and B have K common items and J common judges. Common items have difficulties DiA and DiB. Common judges have severities SjA and SjB. The standard errors of the item and judge measures are obtained from the reported results of analyses of Test A and Test B separately.

If cross-plots of the two sets of common item difficulties and of the two sets of judge severities exhibit approximately 45 trends, then the constants to add to difficulties (the Item Link), severities (the Judge Link) and abilities (GAB) in Test B to translate them onto the scale of Test A are

with standard errors of

A test of the hypothesis that each piece of the linking constant explains the difference between common measures in its facet has the form

When the standard errors of measures within the sets of linking items or judges differ noticeably and their precision is deemed to reflect the influence they should have on the linking constant, then the construction of information-weighted linking constants for the items or judges may be preferred. For instance, the Fisher information in each item difficulty shift is:

Then, the linking constant for the items is

with standard error of

A test of the hypothesis that this linking constant explains the difference between common items is

These item and judge links substitute directly in the earlier formulae for GAB and its standard error.

John Michael Linacre

Linacre J. M. (1998) Linking Constants with Common Items and Judges. Rasch Measurement Transactions 12:1 p. 621.

Linking Constants with Common Items and Judges. Linacre J. M. … Rasch Measurement Transactions, 1998, 12:1 p. 621.

Rasch Publications
Rasch Measurement Transactions (free, online) Rasch Measurement research papers (free, online) Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Applying the Rasch Model 3rd. Ed., Bond & Fox Best Test Design, Wright & Stone
Rating Scale Analysis, Wright & Masters Introduction to Rasch Measurement, E. Smith & R. Smith Introduction to Many-Facet Rasch Measurement, Thomas Eckes Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr. Statistical Analyses for Language Testers, Rita Green
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Journal of Applied Measurement Rasch models for measurement, David Andrich Constructing Measures, Mark Wilson Rasch Analysis in the Human Sciences, Boone, Stave, Yale
in Spanish: Análisis de Rasch para todos, Agustín Tristán Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez

 Forum Rasch Measurement Forum to discuss any Rasch-related topic

Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement

Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.

Coming Rasch-related Events
May 17 - June 21, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 12 - 14, 2024, Wed.-Fri. 1st Scandinavian Applied Measurement Conference, Kristianstad University, Kristianstad, Sweden http://www.hkr.se/samc2024
June 21 - July 19, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com
Aug. 5 - Aug. 6, 2024, Fri.-Fri. 2024 Inaugural Conference of the Society for the Study of Measurement (Berkeley, CA), Call for Proposals
Aug. 9 - Sept. 6, 2024, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com
Oct. 4 - Nov. 8, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 17 - Feb. 21, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
May 16 - June 20, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 20 - July 18, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Facets), www.statistics.com
Oct. 3 - Nov. 7, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com