Dear colleagues in the research community,
As we know, there are two approaches to hypothesis testing of cross-tables: testing for independence and testing for correlation between variables. In both cases, for exact probabilities, we ask the same question: what is the probability of getting "this table" and the "more extreme tables". For independence tests, the traditional exact test is the (dominant) Fisher-Freeman-Halton (FFH) statistic, and for correlation tests, the Mehta-Patel (MP) algorithm is a widely used solution. In some cases, especially when the table is sparse and ordinal, these algorithms give conflicting, if not opposite, inferences. I recently faced a table where the exact probability by FFH could be p = 1, while MP was p < 0.001 because of high correlation. In the attached note, I ponder this issue and compare their strategies. It seems that FFH's result is confusingly wrong, and the reason is the way the FFH algorithm treats tables with the same probability as the one of interest. This claim is strong, and it calls for a larger discussion within the research community about FFH: Should we change the logic of FFH to avoid confusing results? If we should, why? If we should not, why not?
Preprint Note on the deviating exact probabilities. Should we change ...