Conant's Rule (5 x the overall-state-set-size) gives us an effective minimum dataset size in order to be able to claim statistical significance for results based on analyzing the dataset. What I'm looking for is at the other end: is there a measure that will tell me when I'm using *too* much data, such that the effects I'm looking for are likely to be drowned in the data mass? What would I look for in the literature discussing this consideration?

More Doug Elias's questions See All
Similar questions and discussions