The purely statistical approach, i. e. statistical analysis in conjunction with Big Data, to the study of the human mind and the resulting behaviour is – once again – predominant in psychology, cognitive sciences, economics and related fields. The basic idea is that models that make use of sophisticated statistical analysis in conjunction with large databases allow for the increasingly precise predictions of future events.
The problem is that these models can function without any insight into the underlying principles that may explain the occurrence of these events. Given that a scientific explanation is given if and only if the complex visible phenomena are reduced to the simple invisible principles (call them “Laws” if you like), it seems that the statistical approach violates the idea of a scientific explanation in a fundamental way, because the do not give any insight into the simple invisibles – the underlying principles of nature.
On the other hand, the result of a scientific explanation is supposed to be a scientific theory. The sole criterion for the “value and justification” of a scientific theory, according to Albert Einstein, rests solely on the correspondence between the consequences from the theory and the facts as they are. It seems that the statistical approach – trivially – strives to achieves this goal. The models would, if successful, give the precise way in which events will occur. Alas, without any insight about why the events occur.
There is a simple (and uncharitable) explanation for the predominance of the statistical approach: Human behaviour is undetermined, i. e. there is no set of laws that allow one to deduce what a human will do next. This is a result that follows from basic economic analysis and can be confirmed by everyone. If you love bananas and hate apples and if you are given the choice between a banana and an apple, one has to conclude that you will chose the banana in all cases. Yet, we all know that we could simply change our minds and take the apple instead.
Economists have long recognized that the only way to predict future behaviour – which is the holy grail of economics and politics – can only be achieved by either extrapolate the probability of future behaviour from past behaviour or by straight up manipulation (or by any conceivable combination of these two ways). While the later is taken care of by advertising, public relation, education and so on, the former seems to be what sophisticated statistical analysis in conjunction with Big Data seems to be perfect for.
The question then is: Is there any saving grace for the statistical approach (without the incorporation of actual principles) or are these models simply an overly hyped way for scientists to serve economic and political interests?
I would like to stress that statistical models that include underlying principles are standard in the natural sciences. What makes the above mentioned approach stand out is the rejection of the principles except for the most basic assumption that somehow we extrapolate only from experience and nothing else.