In the 1991 Review of Research in Education yearbook, Wolf, Bixby, Glenn and Gardner perceive a need for a "new psychometrics" designed for student assessment. They claim this "new psychometrics" would have to satisfy four criteria: 1) be capable of measuring alternative pathways to mastery; 2) supersede the need for inter-judge agreement as a criterion for valid measurement; 3) measure and report many aspects of student performance; and 4) handle different units of analysis beyond the individual or the total. We are delighted to proclaim that these criteria have already been well-met by the psychometric theory, design and practice of many-faceted Rasch measurement (MFRM).
Examine the criteria:
1) MFRM is based on the reality that performance assessment involves judging complex behaviors. Many facets contribute to the measures earned by examinees. Facets frequently delineated are examinee ability, judge severity and task difficulty. But this does not mean that the analyst is constrained to think in some arbitrary way. Different observations of student performance can be conceptualized, for one analysis, as distinct elements of the same facet, but, for another analysis, as separate facets.
Consider an example in which students are required to perform two tasks: write an expository essay, and critique an essay. The tasks can be modeled together or separately, graded with the same or different grading schemes, and given identical or different weights. The contribution of the tasks to the final measure of the student depends upon the definition of the assessment. The assignment of weights and the combination of different tasks in separate models can be used to define alternative pathways to mastery. The student can demonstrate proficiency in areas that meet the requirements of one or more of several alternative models while failing to demonstrate proficiency in the areas represented by the remaining models and still receive a measure that justifies recognition of mastery. The ability to award partial credit also provides for differentiation in degrees of mastery. In this way, different pathways to mastery can be recognized and incorporated into the final student measure. Even further, a variety of different measurement instruments can be incorporated into a single analysis (Stahl & Lunz 1991) and substantively-determined weights assigned.
2) MFRM accepts and controls for differences in judge severities. Judges are not expected or required to grade identically just because they have expertise in the same field. Some judges may have higher standards and therefore grade more severely than others. Because the severity of the judge is an additive factor in the model and is taken into consideration before the student measures are calculated, the unique perception of any judge in the measurement process is acknowledged and accounted for.
3) In MFRM, interactions between facets can be specified and used as a diagnostic tool for identifying specific strengths and weaknesses of individual students or judges. Interactions between students and items or students and tasks, or judges and items, or judges and tasks can be explicitly specified and measured. Unusual performances can be flagged and students or judges informed about their unexpected behavior. Discussion may reveal a logical explanation for the performance or a need that can be remedied with additional instruction.
4) Groups with unique characteristics, whether the group consists of items, students or judges, can be analyzed as entities when MFRM strategies are employed. Group membership can be modelled as a separate facet and linear measures of the performance of different groups can be calculated. Groups of students or judges or selected groups of items may also be of interest in diagnosing student ability.
In summary, MFRM, splendidly implemented in the FACETS computer program, is an extremely flexible mode of analysis that yields objective, conjointly additive measures for all pertinent facets of a particular examination situation. The analysis models offer many options for studying the data so that the results may be diagnostic or evaluative for students, judges or items. MFRM provides the opportunity to control all aspects of an examination with more than two facets in a way that meets all the basic requirements in the "call for a new psychometrics". Rejoice oh Wolf, Bixby, Glenn and Gardner! Your "new psychometrics" is here.
Answering the "Call for a New Psychometrics", J Stahl & M Lunz Rasch Measurement Transactions, 1991, 5:1 p. 127-128
Please help with Standard Dataset 4: Andrich Rating Scale Model
|Rasch Measurement Transactions (free, online)||Rasch Measurement research papers (free, online)||Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch||Applying the Rasch Model 3rd. Ed., Bond & Fox||Best Test Design, Wright & Stone|
|Rating Scale Analysis, Wright & Masters||Introduction to Rasch Measurement, E. Smith & R. Smith||Introduction to Many-Facet Rasch Measurement, Thomas Eckes||Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr.||Statistical Analyses for Language Testers, Rita Green|
|Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar||Journal of Applied Measurement||Rasch models for measurement, David Andrich||Constructing Measures, Mark Wilson||Rasch Analysis in the Human Sciences, Boone, Stave, Yale|
|in Spanish:||Análisis de Rasch para todos, Agustín Tristán||Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez|
|Forum||Rasch Measurement Forum to discuss any Rasch-related topic|
Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement
Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.
|Coming Rasch-related Events|
|Sept. 15-16, 2017, Fri.-Sat.||IOMC 2017: International Outcome Measurement Conference, Chicago, jampress.org/iomc2017.htm|
|Oct. 13 - Nov. 10, 2017, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|Oct. 25-27, 2017, Wed.-Fri.||In-person workshop: Applying the Rasch Model hands-on introductory workshop, Melbourne, Australia (T. Bond, B&FSteps), Announcement|
|Jan. 5 - Feb. 2, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|Jan. 10-16, 2018, Wed.-Tues.||In-person workshop: Advanced Course in Rasch Measurement Theory and the application of RUMM2030, Perth, Australia (D. Andrich), Announcement|
|Jan. 17-19, 2018, Wed.-Fri.||Rasch Conference: Seventh International Conference on Probabilistic Models for Measurement, Matilda Bay Club, Perth, Australia, Website|
|April 13-17, 2018, Fri.-Tues.||AERA, New York, NY, www.aera.net|
|May 25 - June 22, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|June 29 - July 27, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com|
|Aug. 10 - Sept. 7, 2018, Fri.-Fri.||On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com|
|Oct. 12 - Nov. 9, 2018, Fri.-Fri.||On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com|
|The HTML to add "Coming Rasch-related Events" to your webpage is:|
The URL of this page is www.rasch.org/rmt/rmt51a.htm