Comparing estimates with other Rasch software

Top Up Down  A A

There are many Rasch-specific software packages and IRT packages which can be configured for Rasch models. Each implements particular estimation approaches and other assumptions or specifications about the estimates. Comparing or combining measures across packages can be awkward. There are three main considerations:

 

(a) choice of origin or zero-point

(b) choice of user-scaling multiplier.

(c) handling of extreme (zero, minimum possible and perfect, maximum possible) scores.

 

Here is one approach:

 

Produce person measures from Winsteps and the other computer program on the same data set. For Winsteps set USCALE=1 and UIMEAN=0.

Cross-plot the person measures with the Winsteps estimates on the x-axis. (This is preferential to comparing on item estimates, because these are more parameterization-dependent.)

Draw a best-fit line through the measures, ignoring the measures for extreme scores.

The slope is the user-scaling multiplier to apply. You can do this with USCALE= slope.

The intercept is the correction for origin to apply when comparing measures. You can do this with UIMEAN= y-axis intercept.

The departure of extreme scores from the best-fit line requires adjustment. You can do this with EXTRSCORE=. This may take multiple runs of Winsteps. If the measures for perfect, maximum  possible scores are above the best-fit line, and those for zero, minimum possible scores are below, then decrease EXTRSCORE= in 0.1 increments or less. If vice-versa, then increase EXTRSCORE= in 0.1 increments or less.

With suitable choices of UIMEAN=, USCALE= and EXTRSCORE=, the crossplotted person measures should approximate the identity line.

The item estimates are now as equivalent as they can be even if, due to different choice of parameterization or estimation procedure, they appear very different.

 

You may notice scatter of the person measures around the identity line or obvious curvature. These could reflect differential weighting of the items in a response string, the imposition of prior distributions, the choice of approximation to the logistic function, the choice of parameterization of the Rasch model or other reasons.These are generally specific to each software program and become an additional source of error when comparing measures.

 

There are technical details at Estimation.

 

Winsteps (JMLE) vs. ConQuest (MMLE), the estimate differences are primarily because:

1. ConQuest assumes a regular shape to the person-ability distribution. Winsteps does not.

2. ConQuest includes extreme person scores (zero and perfect) when estimating item difficulties. Winsteps does not.

There are also other technical differences, but these are usually inconsequential with large datasets.

 

Maximum Likelihood Estimates (MLE, any type) vs. "Warm" Mean Likelihood Estimates (WMLE):

Warm estimates are usually slightly more central than MLE estimates.