Researchers seem to often want to look for a "test" to show whether or not they should conclude the "truth" of a given 'hypothesis' (of some sort).  But often what is sought is not something that should be considered to have a "yes" or "no" answer.  Consider homoscedasticity versus heteroscedasticity.  Often it isn't a question of whether or not heteroscedasticity is present, it is a question of how much?  OLS says there is none.  Really?  It is actually a question of whether or not it is enough to matter in your case.  It is a matter of measuring it and/or other practical considerations. Further, when someone says "statistical test," don't they generally mean an hypothesis test, or even just a "significance" level without a type II error analysis?  Results then are very dubious.  You need to pay attention to effect size, and sample size, and what the 'test' is actually measuring, and interpretability. 

    

Accuracy depends upon standard deviation, sample size, and nonsampling error, and methodology.  But too often researchers want to take whatever they have and make a decision and they want to know if it is THE decision they should make, yes or no, based on a "test."  That test may only be concerned with a single aspect of a many faceted problem.  A 'test' cannot take the place of a reasoned sensitivity analysis based on sound statistical handling of data in the context of the subject matter, and theory, and a balanced approach to considering various risks involved. 

 

I have noticed on ResearchGate that often when a researcher is asking about a 'test,' they would probably, in my opinion, better expend their efforts on validation of results, using example data.  (I avoided the use of the term "test data" to distinguish this from the topic at hand ... but I really do mean "test data.") 

 

Often, though studying graphical results may seem more subjective than looking at some 'test statistic,' the graphics are far more useful and meaningful.  

 

So "How often does a "statistical test" have much value?"   

  

Anecdotes are not a scientific/statistical study, but may be a start to a discussion or lead to other considerations.  So, do you have an example of a project gone wrong because emphasis was placed on results from a 'statistical test'?  Any examples where it helped?  How? 

Thank you.  

More James R Knaub's questions See All
Similar questions and discussions