There is an interesting Law of the Hammer (see link *), which states the following truism: "If the only tool you have is a hammer, it is tempting to treat everything as if it were a nail." This Law may be working in some areas of Psychometrics, too.

If the only cognitive tools you have, are ordinary arithmetic and statistics on the real line (i.e., all negative and positive numbers), then you may be tempted to treat ALL measures as REAL numbers and apply the usual arithmetic and statistical operations (plus, times, arithmetic mean, etc.) on REAL numbers, to get whatever you want, e.g. estimates and predictions.

The Hammer of IRT/Rasch

This seems to be the case for IRT based on a naive concept of scores. Scores, in ITS's view, are just like real numbers, X for short, which can be neatly associated with (cumulative) probabilities via an elegant transformation (conversion) rule:

  • P(X) = exp(X) / (1 + exp(X) )

Once you have those probabilities, the whole machinery of probability theory and statistics can be invoked for descriptive or testing purposes. Adding one, two or three parameters (which have to be estimated) increases the explanatory power of the approach even if your responses are still binary 0 and 1's. Generalizing to polytomous and multidimensional IRT etc. is possible within the same adopted framework.

Still, all these developments rest on a highly questionable, implausible interpretation of X as the genuine measurement of an underlying (psychological) feature and P(X) as its associated probability. However, there is no empirical evidence or formal argument, that such an interpretation is possible, necessary, meaningful or even correct.

On the contrary: it is highly counter-intuitive. There are no (latent) psychological features or variables like X which are strictly unbounded to either side, i.e. could become - infinity or + infinity. If you happen to know one, please let me know.

An alternative view avoiding the concept of probability

The alternative is, to view the real number X as a transformed score f(S) from the double-bounded interval [A,B] between given A and B to the real numbers.

  • For instance, A could be 1 and B could be n, with equal-spaced anchor points 1, 2, 3, ..., n, so that [A,B] is the underlying continuous scale of a discrete n-point ruler or scale for the user/respondent.

Now let S be such a score on the n-point scale from 1 to n, and define X such that :

  • exp(X) := (S - A) / (B - S)

Then X is well-defined: it is the (natural) logarithm of (S - A) / (B - S). Now, after a few simple manipulations we get:

  • S = (1- P(X)) * A + P(X) * B so that P(X) / (1 - P(X)) = (S - A) / (B - S).

In other words: P(X) is just a weighting factor required to get the position (location) of the score S on the chosen n-point scale from A to B. Indeed, we could replace P(X) just be W(S), i.e. the weight of S on [A,B]) to get:

  • S = (1- W(S)) * A + W(S) * B so that W(S) / (1 - W(S)) = (S - A) / (B - S).

In other words, W(S) = P(X) is just a normalized version of S given that S is constrained to [A,B]. No probability interpretation is required!

What's more: we can define the quasi-addition (+) of scores S and T on any scale [A,B]. It is, in fact, relatively simple:

  • S (+) T := (S - A) * (T - A) divided by (S - A)*(T - A) + (B - S)*(B - T)

This quasi-sum (+) of S and T has (almost) ALL the properties of the usual sum of two real numbers, however, it is closed on the interval between A and B. Defining quasi-multiplication (*) of a score S with a real number is also quite simple. And once we have quasi-multiplication, the quasi-additive inverse is correctly defined as (-1) (*) S. Together, Module Theory provides the right framework for working out the algebra of bounded scales and scores.

Introducing the interval scale on [A,B]

The biggest surprise of all: it turns out that the n-point scale endowed with this quasi-arithmetic addition operation is indeed an interval scale. The same holds for the standard percentage scale [0,1], so often used in educational contexts, but also in Fuzzy Set Theory and Fuzzy Logic.

CONCLUSION:

Rasch, one of the founders of IRT, was an original mathematician and statistician, but he was rather focussed on classical statistics and measurement theory from the physical sciences, and probably he didn't know about the then-recent developments from what is known now as quasi-arithmetic calculus and module theory.

Also, in his time, Stevens was one of the hero's of a rather naive (=operationalist) scale theory for psychology and other social sciences (see e.g. the critiques of J. Michell, Luce, and other measurement theorists).

I am sure, that Rasch would be delighted to see that it is indeed possible to do serious measurement (theory) in the social sciences without the counter-intuitive assumptions that he had to make.

________________________________________________________________________

*) https://en.wikipedia.org/wiki/Law_of_the_instrument

More Paul Hubert Vossen's questions See All
Similar questions and discussions