Principal Component Analysis, Factor Analysis and Linear Discriminant Analysis are all used for feature reduction. They all depend on using eigenvalues and eigenvectors to rotate and scale the vectors in order to project them to the new dimensions. They all assume the linearity of the observed data. They all use the formula WSW^-1 from linear algebra.
The question now is what is the difference between them conceptually? and when to use each of them? how each of them work?