I conduct 3 linear correlation tests (serial correlation LM test, runs test and VR test) on a stock market index full sample of 15 years and as well on each year separately for the same period of time (static windows) in order to capture a changing degree of efficiency. Strangely, the full sample results indicate a different state of efficiency than the average of the 15 subsamples results for the same period. In extreme cases the full sample result indicates inefficiency at a significants of 1% whereas the subsample are on average efficient. How can that be? Shouldn’t it be more or less the same?
Any help is very much appreciated...