We have problems in ordering students who have the same raw score.
We must do this to meet selection requirements. Here are 5 persons
who got the same raw score of 26 on the same items. Some students
answered correctly more difficult questions while some only
answered correctly easier questions. How can we make an ordering
among these students?
|5 Persons with Raw Score 26,|
Measure .73 logits, Model S.E. .35.
|Outfit Mean-square||Infit Mean-square||Point-biserial||Student|
The scientific principle on which measurement is based makes a point of specifying that for a fixed set of items all occurrences of the same raw scores must imply the same measures ("Sufficiency of raw scores on the same set of items"). So a scientist would not differentiate among the same raw scores on the same tests. However, a scientist would have already eliminated whatever responses were found to be irrelevant or contradictory to constructing the best measures.
Some psychometricians feel that a correct answer to a hard item merits a higher measure than a correct answer to an easy item. But then, does an incorrect answer to an easier item merit a lower measure? Remember, all these students achieved the same raw score, so success on an unusually hard item implies failure on at least one easier item. If you really think success on a hard item merits a higher measure, then you must disregard the concomitant failure on an easier item. If so, then you could make the response to that easier item, which you have decided to disregard, missing. Now that student has taken a shorter, more difficult test, but obtained the same raw score - the reported measure will be higher, resolving the ordering problem. Such students will have high outfit statistics in the original analysis, so student E is the likely winner! Student A is the loser.
In general, however, better fitting response strings are more reproducible and also easier to understand. We know what these students can do and what they cannot do. Consequently their performance on the next item is more predictable. Who is better, an erratic genius or a consistently high performer? The most consistent student, A, has the smallest mean-square fit statistics. This makes student A the winner and student E the loser!
Clearly, mathematics can take us only so far. In the end, we must also think again how to use measurement wisely as well as correctly.
Benjamin D. Wright
Note: If lucky guesses or careless mistakes are to be eliminated, then use a technique similar to "Guessing and Measurement" www.rasch.org/rmt/rmt62a.htm
Who is awarded First Prize when the raw scores are the same? Wright B.D. Rasch Measurement Transactions, 1998, 12:2 p. 629.
|Rasch Measurement Transactions (free, online)||Rasch Measurement research papers (free, online)||Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch||Applying the Rasch Model 3rd. Ed., Bond & Fox||Best Test Design, Wright & Stone|
|Rating Scale Analysis, Wright & Masters||Introduction to Rasch Measurement, E. Smith & R. Smith||Introduction to Many-Facet Rasch Measurement, Thomas Eckes||Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr.||Statistical Analyses for Language Testers, Rita Green|
|Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar||Journal of Applied Measurement||Rasch models for measurement, David Andrich||Constructing Measures, Mark Wilson||Rasch Analysis in the Human Sciences, Boone, Stave, Yale|
|in Spanish:||Análisis de Rasch para todos, Agustín Tristán||Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez|
|Forum||Rasch Measurement Forum to discuss any Rasch-related topic|
Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement
Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.
|Coming Rasch-related Events|
|Aug. 14 - 16, 2019. Wed.-Fri.||An Introduction to Rasch Measurement: Theory and Applications (workshop led by Richard M. Smith) https://www.hkr.se/pmhealth2019rs|
|August 25-30, 2019, Sun.-Fri.||Pacific Rim Objective Measurement Society (PROMS) 2019, Surabaya, Indonesia https://proms.promsociety.org/2019/|
|Oct. 11 - Nov. 8, 2019, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|Nov. 3 - Nov. 4, 2019, Sun.-Mon.||International Outcome Measurement Conference, Chicago, IL,http://jampress.org/iomc2019.htm|
|Jan. 24 - Feb. 21, 2020, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|May 22 - June 19, 2020, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|June 26 - July 24, 2020, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com|
|Aug. 7 - Sept. 4, 2020, Fri.-Fri.||On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com|
|Oct. 9 - Nov. 6, 2020, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|June 25 - July 23, 2021, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com|
The URL of this page is www.rasch.org/rmt/rmt122d.htm