Hi! Preprocessing is not needed when using pre-trained language representation models like BERT. In particular, it uses all of the information in a sentence, even punctuation and stop-words, from a wide range of perspectives by leveraging a multi-head self attention mechanism.
You can find some well explained examples + GitHub code here: https://riccardo-cantini.netlify.app/post/