Bootstrapping is where you sample a value from a population of data and then replace that value before drawing another value. If your data has rare extreme values, bootstrapping will undervalue these observations. If these extreme values are partly a result of mistakes, then this might be good. If these are a true part of the underlying distribution then this will be bad because it will underestimate the variability in the underlying population.
Randomization is where you withdraw values from a population of data without replacement. If one uses all of the data, then the rare original observations will be as common in each randomization as they were in the original data.
I partially agree witg Timothy A Ebert. The given definition of Bootstrapping isn't complete. In fact we distnguish two types of Bootstrap:
1 - The non parametric Bootstrap: when the data distribution is not known, so you have to perform a sampling with replacement as Timothy A Ebert said, and you'll have values that figures in original data.
2 - The Parametric Bootstrap: when you estimate the model or the data distribution, so you just simulate samples from the estimated model, so you can obtain values that doesn't figure in the original data.
In the second case, the extreme values behaves as well as in the original data, if the density (or the model) is correctly estimated.
I agree with Ouzzani, except that I would not use "bootstrap" for #2 (though a quick internet search turns up "parametric bootstrap"). I use the definitions set forth by Bryan Manly in his book "Randomization and Monte Carlo methods in Biology".
Using the definition of parametric bootstrap set forth in http://www.stat.umn.edu/geyer/5601/examp/parm.html. I do not see how this would differ from simulation (Monte Carlo simulation). I used the observed data to identify an underlying distribution. I then take that estimate as "true" and set up a random number generator to generate "similar" data sets. The concept of "replacement" does not apply. In both cases (bootstrap and simulation) the idea is that you are not changing the distribution of the resampled data. However, with bootstrap there is a problem with a rare observation that it is a unique value. There are no other observations that are near. I might have values like 2,4,3,6,8,4,27. In such a data set there is no possibility of getting a value of 26. In contrast, with simulation all values are possible. I could get 21.743 or I could get 57.3.
My apologies if my interpretation of the parametric bootstrap is in error. However, as described I do not see the "parametric bootstrap" method as being a bootstrap. The method is perfectly reasonable, the name is not.
In my knowledge, a rare observation is an observation with a tiny occurence probability , and the estimated density (which you use it to generate new samples) should take into account this smallness.
What I understood from your definition of rare observation is that it is impossible to reproduce a similar one. For example, even if you meet a person who weights 350kg, there is no chance to meet another one who weights 340kg, since both are very rare cases.
Also, If you generated your data in an independent way, then you have allowed the occurence of data which hasn't been observed.
Example: When I measured the weight of 10 persons I found the following results:
79 , 58 , 74 , 71 , 63 , 67 , 71 , 98 , 82 and 62
If I apply a non parametric bootstrap, it will be as saying "there is no chance to mme someone who weights 60, 70, 80 or 90".
This is why, in my opinion, the parametric bootstrap is more welcome.
For the name, you can refer to Efron in his paper:
EFRON, Bradley. Bayesian inference and the parametric bootstrap. The annals of applied statistics, 2012, vol. 6, no 4, p. 1971.
As I said, the problem is not with the method, it is with calling it a "parametric bootstrap." To me it does not seem to have any relation to non-parametric bootstrap except that it is a computer intensive approach to data analysis.
Rare observations: A rare observation is one that is many standard deviations distant from the mean. I have a rarer event if it is more standard deviations away from the mean than the other rare event.
In a data set, there will be rare events. Most of the time we are not able to go back and get more data. If the data are plotted with value on the x-axis and frequency on the y-axis, then there are gaps between observations as you move along the x-axis. The size of those gaps increases with increasing distance from the mean. If you estimate the probability of observing a specific value, that reliability of that estimate declines with increasing distance from the mean. What the simulation approach assumes is that you gathered enough data to accurately estimate the probability of the rare events.
The bootstrap method does not assume that a value of 96 is impossible, but it also does not factor in the possibility that such an event exists. It is more like it treats the value of 96 as missing data because there were no observations of 96 in your sample of ten values.
My objection is only with the name. I dislike the name "parametric bootstrap" because it includes the word "bootstrap" but has nothing at all to do with bootstrap in the sense of "non-parametric" bootstrap. The only thing in common is the use of a computer. The approach to data analysis is great, the name is the problem.
So if I have a parametric and a non-parametric bootstrap method, then it would seem like I could have a parametric and a non-parametric randomization test using the same logic that gave rise to the parametric bootstrap method. What is the difference between a parametric bootstrap and a parametric randomization test?
It is true that there is an ambiguity in the name since there is no difference between the parametric bootstrap and the parametric randomization, so I completely agree with you in all points