The use of student-response data in assessing classroom assessments has become increasingly important in recent years. Empirical techniques have been implemented to improve item quality on the basis of students’ responses, which mainly relies on numerical values. Although some educators find numerical values daunting, it is vital to understand how a commonly employed item-difficulty index can assist in assessing item quality.
The most commonly employed item-difficulty index, the p value, is calculated by dividing the number of students that responded correctly (R) to an item by the total number of students that responded to the item (T). The resulting value ranges from 0 to 1.00, with higher p values indicating that more students responded correctly to the item.
However, it is crucial to consider the probability of guessing the correct response when interpreting the p value. Items with high p values should not be viewed solely as “easy” items, and items with low p values should not be only considered “difficult.”
An item-discrimination index can be calculated by assessing the correlation coefficient between student’s total test scores and their performance on a particular item. This index measures how well an item discriminates between high and low performers in the class, thus indicating the quality of the item.
Overall, understanding these indices and incorporating student-response data can lead to significant improvements in assessment procedures in the classroom.