Journal of Educational Measurement

Papers
(The TQCC of Journal of Educational Measurement is 1. The table below lists those papers that are above that threshold based on CrossRef citation counts [max. 250 papers]. The publications cover those that have been published in the past four years, i.e., from 2020-11-01 to 2024-11-01.)
ArticleCitations
Model‐Based Treatment of Rapid Guessing26
Optimizing Implementation of Artificial‐Intelligence‐Based Automated Scoring: An Evidence Centered Design Approach for Designing Assessments for AI‐based Scoring12
Random Responders in the TIMSS 2015 Student Questionnaire: A Threat to Validity?10
A Residual‐Based Differential Item Functioning Detection Framework in Item Response Theory8
Psychometric Methods to Evaluate Measurement and Algorithmic Bias in Automated Scoring8
Using Eye‐Tracking Data as Part of the Validity Argument for Multiple‐Choice Questions: A Demonstration7
Validity Arguments for AI‐Based Automated Scores: Essay Scoring as an Illustration7
Variation in Respondent Speed and its Implications: Evidence from an Adaptive Testing Scenario7
Generating Models for Item Preknowledge6
Linking and Comparability across Conditions of Measurement: Established Frameworks and Proposed Updates6
Score Comparability between Online Proctored and In‐Person Credentialing Exams6
Using Item Scores and Distractors in Person‐Fit Assessment6
Toward Argument‐Based Fairness with an Application to AI‐Enhanced Educational Assessments5
The Automated Test Assembly and Routing Rule for Multistage Adaptive Testing with Multidimensional Item Response Theory5
Examining the Impacts of Ignoring Rater Effects in Mixed‐Format Tests5
NCME Presidential Address 2022: Turning the Page to the Next Chapter of Educational Measurement5
Exploring the Impact of Random Guessing in Distractor Analysis4
Score Comparability Issues with At‐Home Testing and How to Address Them4
On Joining a Signal Detection Choice Model with Response Time Models4
Robust Estimation for Response Time Modeling4
Multiple‐Group Joint Modeling of Item Responses, Response Times, and Action Counts with the Conway‐Maxwell‐Poisson Distribution4
Validity Arguments Meet Artificial Intelligence in Innovative Educational Assessment4
Detecting Differential Item Functioning Using Posterior Predictive Model Checking: A Comparison of Discrepancy Statistics4
The Impact of Cheating on Score Comparability via Pool‐Based IRT Pre‐equating4
An Unsupervised‐Learning‐Based Approach to Compromised Items Detection4
On the Positive Correlation between DIF and Difficulty: A New Theory on the Correlation as Methodological Artifact3
A Statistical Test for the Detection of Item Compromise Combining Responses and Response Times3
Historical Perspectives on Score Comparability Issues Raised by Innovations in Testing3
A Unified Comparison of IRT‐Based Effect Sizes for DIF Investigations3
Latent Space Model for Process Data3
Specifying the Three Ws in Educational Measurement: Who Uses Which Scores for What Purpose?2
Assessing the Impact of Equating Error on Group Means and Group Mean Differences2
Controlling the Speededness of Assembled Test Forms: A Generalization to the Three‐Parameter Lognormal Response Time Model2
Anchoring Validity Evidence for Automated Essay Scoring2
Gender Bias in Test Item Formats: Evidence from PISA 2009, 2012, and 2015 Math and Reading Tests2
Fully Gibbs Sampling Algorithms for Bayesian Variable Selection in Latent Regression Models1
Explanatory Cognitive Diagnostic Modeling Incorporating Response Times1
Introduction to the Special Issue Maintaining Score Comparability: Recent Challenges and Some Possible Solutions1
Detecting Group Collaboration Using Multiple Correspondence Analysis1
Classical Item Analysis from a Signal Detection Perspective1
DIF Detection for Multiple Groups: Comparing Three‐Level GLMMs and Multiple‐Group IRT Models1
A Note on Latent Traits Estimates under IRT Models with Missingness1
A Factor Mixture Model for Item Responses and Certainty of Response Indices to Identify Student Knowledge Profiles1
Detecting Multidimensional DIF in Polytomous Items with IRT Methods and Estimation Approaches1
Using Linkage Sets to Improve Connectedness in Rater Response Model Estimation1
Validating Performance Standards via Latent Class Analysis1
Several Variations of Simple‐Structure MIRT Equating1
Cognitive Diagnostic Multistage Testing by Partitioning Hierarchically Structured Attributes1
Using Response Time in Multidimensional Computerized Adaptive Testing1
Sociocognitive Processes and Item Response Models: A Didactic Example1
Measuring the Impact of Peer Interaction in Group Oral Assessments with an Extended Many‐Facet Rasch Model1
Detecting Differential Item Functioning in CAT Using IRT Residual DIF Approach1
Pretest Item Calibration in Computerized Multistage Adaptive Testing1
Incorporating Test‐Taking Engagement into Multistage Adaptive Testing Design for Large‐Scale Assessments1
Measuring the Uncertainty of Imputed Scores1
Estimating Classification Accuracy and Consistency Indices for Multiple Measures with the Simple Structure MIRT Model1
1
An Exponentially Weighted Moving Average Procedure for Detecting Back Random Responding Behavior1
0.04629111289978