The placement of items along a variable defines the construct. For dichotomies, the location of items by their measures is unambiguous. Construct development proceeds immediately, see Chapter 5 of Best Test Design (Wright & Stone, 1979). For rating scales, item location is more nebulous. Chapter 7 of Rating Scale Analysis (Wright & Masters, 1982) points out how the construct is redefined when different standpoints on item difficulty are adopted. When an instrument contains multiple item types, the challenge is to build a clear, comprehensive, useful and defensible construct definition.
When data are collected on a rating scale, each category threshold, "step", has its own difficulty. This can be parameterized as D_{ij}, the calibration of the jth step from category j-1 to category j for item i relative to the local origin of the measurement system. It can also be parameterized as D_{i} + F_{ij}, where D_{i} is the calibration of the item relative to the local origin and F_{ij} is the calibration of step j relative to D_{i}. Mathematically, the two forms are equivalent. Conceptually they are different.
Consider the SF-36 Health Survey instrument. It contains not only dichotomies but also rating scales with 3, 5 and 6 categories. In discussing an item on a 6 category rating scale ("All of the time" to "None of the Time"), such as "Did you feel tired", we would overwhelm ourselves, as well as our audience, were we to describe all 5 of its step calibrations. Rather, we want to locate that item at one point along the variable, so that it can be compared with a "Yes/No" dichotomy, such as "Accomplished less than you would like".
The mathematically convenient item location is the one at which F_{ij}=0 for the item. This is the point along the variable at which the highest and lowest categories for item i are equally probable, a type of mathematical "middle". The substantive meaning of this mathematical location, however, is usually obscure because it is an accident of how the categories have been labelled.
A more meaningful item location could be obtained by collapsing the rating scale into a dichotomy. The dichotomization of each item would be guided by one master criterion, such as "healthy vs. sick", with the actual pivot point for each item being decided by clinical judgement. Since the rating scale will not actually be collapsed into two categories, the pivot point may be in the middle of an ambivalent category, e.g., "Don't Know". Such uncertain categories, however, often contribute more noise than information and so are better treated as missing data.
Once the pivot point for each rating scale has been chosen by inspection of the category labels, then matching step calibrations can be imposed on the analysis. For instance, for the three category scale, "1. Yes, Limited A Lot", "2. Yes, Limited A Little", "3. No, Not Limited At All", the essential dichotomy between sickness and health contrasts categories 1 and 2 against category 3. The pivot point is between categories 2 and 3, where the step calibration is F_{i3}. This is the location relative to which the item difficulty is to be defined. Mathematically, D_{i} + F_{i3} D_{i}, so, to obtain this pivot, set (anchor) F_{i3}=0.
The effects of anchoring steps are:
1) the data do not change unless explicitly recoded,
2) item difficulties do change (see Figure 1), redefining the construct. At bottom of Figure 1, the new "easiest item" is "13", previously it was the "3" at extreme left.
3) but items on the same rating scale retain their relative difficulties, so that these subsets retain their sub-construct definitions (diagonals in Figure 1),
4) persons performances do not change nor do their relative measures (see Figure 2),
5) since the local origin of the measurement system is the mean item difficulty, all person measures may be changed, but by the same offsetting amount (diagonal offset in Figure 2).
Now each item difficulty is located according to a substantive criterion. Consequently, their ordering and relative location also have substantive meaning, aiding us in perceiving the criterion definition of the underlying construct.
Rita Bode
Rehabilitation Institute of Chicago
Pivot-Anchoring Items for Construct Definition. Bode R. … Rasch Measurement Transactions, 1997, 11:3 p. 576-7
Rasch Publications | ||||
---|---|---|---|---|
Rasch Measurement Transactions (free, online) | Rasch Measurement research papers (free, online) | Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch | Applying the Rasch Model 3rd. Ed., Bond & Fox | Best Test Design, Wright & Stone |
Rating Scale Analysis, Wright & Masters | Introduction to Rasch Measurement, E. Smith & R. Smith | Introduction to Many-Facet Rasch Measurement, Thomas Eckes | Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr. | Statistical Analyses for Language Testers, Rita Green |
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar | Journal of Applied Measurement | Rasch models for measurement, David Andrich | Constructing Measures, Mark Wilson | Rasch Analysis in the Human Sciences, Boone, Stave, Yale |
in Spanish: | Análisis de Rasch para todos, Agustín Tristán | Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez |
Forum | Rasch Measurement Forum to discuss any Rasch-related topic |
Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement
Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.
Coming Rasch-related Events | |
---|---|
Jan. 30-31, 2020, Thu.-Fri. | A Course on Rasch Measurement Theory - Part 1, Sydney, Australia, course flyer |
Feb. 3-7, 2020, Mon.-Fri. | A Course on Rasch Measurement Theory - Part 2, Sydney, Australia, course flyer |
Jan. 24 - Feb. 21, 2020, Fri.-Fri. | On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com |
Apr. 14-17, 2020, Tue.-Fri. | International Objective Measurement Workshop (IOMW), University of California, Berkeley, https://www.iomw.org/ |
May 22 - June 19, 2020, Fri.-Fri. | On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com |
June 26 - July 24, 2020, Fri.-Fri. | On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com |
June 29 - July 1, 2020, Mon.-Wed. | Measurement at the Crossroads 2020, Milan, Italy , https://convegni.unicatt.it/mac-home |
July 1 - July 3, 2020, Wed.-Fri. | International Measurement Confederation (IMEKO) Joint Symposium, Warsaw, Poland, http://www.imeko-warsaw-2020.org/ |
Aug. 7 - Sept. 4, 2020, Fri.-Fri. | On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com |
Oct. 9 - Nov. 6, 2020, Fri.-Fri. | On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com |
June 25 - July 23, 2021, Fri.-Fri. | On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com |
The URL of this page is www.rasch.org/rmt/rmt113e.htm
Website: www.rasch.org/rmt/contents.htm