Are there any survey papers on word embedding in NLP which covers the whole history of word embedding from simple topics like one-hot encoding to complex topics like w2v model?
I don't know of a single paper that covers all of what you call 'embeddings'. The term 'embeddings' normally refers to the approaches following Bengio or Collobert and Weston, but it doesn't normally refer to other distributional methods for creating semantic vectors which rely on counting co-occurence.
A good comparison of the two styles is Baroni's Don't Coun't Predict (http://www.aclweb.org/anthology/P14-1023). Sebastian Ruder also has a nice blog post on the history of embeddings (http://ruder.io/word-embeddings-1/index.html). I hope that helps a bit.
Mikolov, Tomas, Wen-tau Yih, and Geoffrey Zweig. "Linguistic regularities in continuous space word representations." Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2013.
Li, Jiwei, et al. "Visualizing and understanding neural models in NLP." arXiv preprint arXiv:1506.01066 (2015).
Levy, Omer, and Yoav Goldberg. "Dependency-based word embeddings." Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Vol. 2. 2014.