In many probability-statistics textbooks and statistical contributions, the standard deviation of a random variable is proposed to be estimated by the square-root of the unbiased estimator of the variance, i.e. dividing the sum of square-deviations by n-1, being n the size of a random sample. Does anybody know why such an estimator is anchored in the statistical literature? Is there any technical reason supporting the use of this estimator?

More Juan Jose Egozcue's questions See All
Similar questions and discussions