Item Response
Item Response
Item Response
Items can be classified as easy, moderate, or difficult based on their difficulty parameter.
1) Easy Items: Easy items are those that most people of all skill levels are likely to correctly
answer. They are situated near the bottom of the latent trait continuum and have a high
probability of giving the right answer. It tells us that easy items are useful only for testing
individuals low on the relevant trait as at higher levels, everyone answers correctly, and no
information is gained.
2) Difficult Items: Difficult items are challenging for most individuals, including those with
higher levels of latent traits. They have a low probability of a correct response and are located
towards the upper end of the latent trait continuum.
3) Moderate Items: Items in the moderate range: These are neither overly simple nor overly
complex. They are situated in the middle of the latent trait continuum and have a modest
likelihood of giving the right answer. It represents the point on the latent trait continuum where
the probability of a correct response is 0.5 (50%).
IRT equation using the single parameter of item difficulty level-:
b = Item Difficulty Level
Georg Rasch, a Danish mathematician, created this particular formula in 1960; as a result, this
IRT application is also known as a Rasch Model in his honour.
Second parameter: It uses both item difficulty and item discrimination. The item
discrimination parameter is a gauge of how well the item differentiates among individuals at a
specific level of the trait in question. Two items might be equally difficult overall like both are
answered correctly by 50 per cent of the examinees but still, we can’t tell the difference between
individuals possessing similar levels of the latent construct of interest (Gregory, 2004).
High Discrimination versus Low Discrimination: The slope of the curve determines the
discriminatory power of the item. Items with a steeper curve possess better discrimination,
meaning that it is better able to differentiate among individuals at this level of the trait. An item
with high discrimination will have different response patterns for individuals with high and
low trait levels, whereas an item with low discrimination will result in similar response patterns
across different trait levels. Items with high discrimination are desirable in assessments as they
provide more precise measurements of individual abilities. These variables have a greater
impact on how accurately individuals are ranked along the spectrum of latent traits. On the
other hand, items with low discrimination are less effective in discriminating between
individuals and may not contribute as much to the measurement precision. If an item is found
to exhibit negative discrimination, researchers should proceed with caution since as respondent
skill grows, the likelihood of selecting the right response shouldn't fall.
Third Parameter: The third parameter uses item difficulty, item discrimination, and the extent
to which the candidate can guess the correct answer (Baker, 2001).
Table 1
Showing differences in Classical Test Theory and Item Response Theory
The test is the unit of analysis. The item is the unit of analysis.
Measures with more items are deemed to be Measures with lesser items can be reliable too
more reliable. in comparison to their counterparts.
Position on the latent trait continuum is Position on the latent trait continuum is
derived by comparing the total test score from derived by comparing the distance between
a reference group. the items on the ability scale.
All items of a measure must have same scale All items of a measure can have different
points. response categories.
Used in a Variety of
Disciplines like market
research, health outcomes,
etc.
Various Approaches to
Test Construction in
Computerized Tailored IRT
Testing
REFERENCES
Columbia University Mailman School of Public Health (2016, August 5). Item Response
Theory. https://www.publichealth.columbia.edu/research/population-health-methods/item-
response-theory
Gregory, R. J. (2004). Psychological testing: History, principles, and applications. Pearson
Education India.
Kaplan, R. M., & Saccuzzo, D. P. (2017). Psychological testing: Principles, applications, and
issues. Cengage Learning.
Schneider, B., Hanges, P. J., Smith, D. B., & Salvaggio, A. N. (2003). Which comes first:
employee attitudes or organizational financial and market performance?. Journal of applied
psychology, 88(5), 836.
Singh, A.K. (2009). Tests, Measurements and Research Methods in Behavioural
Sciences. Bharati Bhawan, New Delhi.