The TOEIC Speaking and Writing tests utilize a variety of test question types that require test takers to
construct responses, not simply to choose among prespecified options. Because these responses are
subjectively scored by human raters, there is a possibility that human error can reduce the accuracy
of test scores. The TOEIC® program employs multiple carefully developed procedures to monitor
rater performance in order to ensure that potential human error is kept to a minimum. Item level
scoring, calibration, benchmark responses, and topic notes help raters to understand the scoring
rubric accurately and to apply the same scoring criteria consistently over time. Back reading helps
scoring leaders monitor raters’ performance in time and improves scoring accuracy. Post-hoc rater
monitoring also provides useful information about an individual rater’s performance, which in turn
helps with rater training and monitoring and protects score accuracy and quality.