MEASUREMENT RESEARCH ASSOCIATES
TEST INSIGHTS
June 2008
Greetings! 
 
I have worked with multiple-choice items and the committees who write them for many years. While computer-based testing enhances exam security, many people involved in certification exam programs still worry that candidates share information about what is on the exam. 
 
Ross Brown
Consistency of Common Item p-Values across Multiple Choice Exam Administrations

An often-voiced concern is that candidates memorize item content, and then share the information with colleagues. Candidates so forewarned of what is on a test may have an unfair advantage, and might be able to pass the test without having the required level of knowledge and skill. To determine if candidates are able to gain an unfair advantage in this way, a systematic study was conducted.
 
The study examined item p-values across exam administrations. The p-value is the percent of candidates who answer an item correctly. If the p-values of items that appear on more than one examination increase by less than 10% from one exam to the next, there is a low probability that individual candidates are benefiting from items being leaked.
 
The p-values of items used on consecutive exams were compared. The table below shows the results. For all five examinations, the majority of the common items did not show p-value increases of more than 10%. Furthermore, even though some of the tests included a large number of carryover items, the items that had higher p-values (easier) were a very small percentage of the total number of items on the second test.
 
Thus, the use of common items on certification examinations does not seem to increase the percent of candidates who answer the item correctly. This makes one wonder about the efficiency of the networks of individuals who supposedly memorize items during a test to share with their colleagues. Perhaps their memories are not terribly accurate, their colleagues are not helped by the information, or misinformation is actually transmitted. 


Exam

Number of Common Items

Number of items on test

Number of items with p-value increases of 10% or more on second test

Percent of Common items that got easier as a percentage of all items on the second test

Exam 1  

250

313

23 items

7%

Exam 2

53

144

6 items

4%

Exam 3

290

310

3 items

1%

Exam 4

52

150

3 items

2%

Exam 5

196

252

7 items

3%

 
Measurement Research Associates, Inc.
505 North Lake Shore Dr., Suite 1304
Chicago, IL  60611
Phone: (312) 822-9648     Fax: (312) 822-9650
 


Rasch-Related Resources: Rasch Measurement YouTube Channel
Rasch Measurement Transactions & Rasch Measurement research papers - free An Introduction to the Rasch Model with Examples in R (eRm, etc.), Debelak, Strobl, Zeigenfuse Rasch Measurement Theory Analysis in R, Wind, Hua Applying the Rasch Model in Social Sciences Using R, Lamprianou Journal of Applied Measurement
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Rasch Models for Measurement, David Andrich Constructing Measures, Mark Wilson Best Test Design - free, Wright & Stone
Rating Scale Analysis - free, Wright & Masters
Virtual Standard Setting: Setting Cut Scores, Charalambos Kollias Diseño de Mejores Pruebas - free, Spanish Best Test Design A Course in Rasch Measurement Theory, Andrich, Marais Rasch Models in Health, Christensen, Kreiner, Mesba Multivariate and Mixture Distribution Rasch Models, von Davier, Carstensen
Rasch Books and Publications: Winsteps and Facets
Applying the Rasch Model (Winsteps, Facets) 4th Ed., Bond, Yan, Heene Advances in Rasch Analyses in the Human Sciences (Winsteps, Facets) 1st Ed., Boone, Staver Advances in Applications of Rasch Measurement in Science Education, X. Liu & W. J. Boone Rasch Analysis in the Human Sciences (Winsteps) Boone, Staver, Yale Appliquer le modèle de Rasch: Défis et pistes de solution (Winsteps) E. Dionne, S. Béland
Introduction to Many-Facet Rasch Measurement (Facets), Thomas Eckes Rasch Models for Solving Measurement Problems (Facets), George Engelhard, Jr. & Jue Wang Statistical Analyses for Language Testers (Facets), Rita Green Invariant Measurement with Raters and Rating Scales: Rasch Models for Rater-Mediated Assessments (Facets), George Engelhard, Jr. & Stefanie Wind Aplicação do Modelo de Rasch (Português), de Bond, Trevor G., Fox, Christine M
Exploring Rating Scale Functioning for Survey Research (R, Facets), Stefanie Wind Rasch Measurement: Applications, Khine Winsteps Tutorials - free
Facets Tutorials - free
Many-Facet Rasch Measurement (Facets) - free, J.M. Linacre Fairness, Justice and Language Assessment (Winsteps, Facets), McNamara, Knoch, Fan

To be emailed about new material on www.rasch.org
please enter your email address here:

I want to Subscribe: & click below
I want to Unsubscribe: & click below

Please set your SPAM filter to accept emails from Rasch.org

www.rasch.org welcomes your comments:
Please email inquiries about Rasch books to books \at/ rasch.org

Your email address (if you want us to reply):

 

FORUMRasch Measurement Forum to discuss any Rasch-related topic

Coming Rasch-related Events
Oct. 6 - Nov. 3, 2023, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Facets), www.statistics.com
Oct. 12, 2023, Thursday 5 to 7 pm Colombian timeOn-line workshop: Deconstruyendo el concepto de validez y Discusiones sobre estimaciones de confiabilidad SICAPSI (J. Escobar, C.Pardo) www.colpsic.org.co
June 12 - 14, 2024, Wed.-Fri. 1st Scandinavian Applied Measurement Conference, Kristianstad University, Kristianstad, Sweden http://www.hkr.se/samc2024
Aug. 9 - Sept. 6, 2024, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com