I am going to participate in a linguistic project which aims at shallow analysis of more than 100 languages but I can't find any comprehensive up-to-date list of tools for minor languages. At least POS taggers will be needed...
Here is a thread about Open Source tools for NLP https://mailman.uib.no/public/corpora/2017-February/026131.html with pointers to different tools http://stackoverflow.com/questions/22904025/java-or-python-for-natural-language-processing and https://stanbol.apache.org/docs/trunk/components/enhancer/nlp/
Here is additionally a thread about Interesting Corpus analysis tools & specific corpora https://mailman.uib.no/public/corpora/2018-July/028671.html
You maybe also interested in the site about NLP tools and corpora: http://clu.uni.no/icame/corpora/sites.html
The work done for "The Leipzig Corpora Collection", a multilingual corpus (more than a hundred languages), could be interesting for your project. This paper presents, for example, some relevant leads: "Building Large Resources for Text Mining: The Leipzig Corpora Collection" https://www.researchgate.net/publication/274250022_Building_Large_Resources_for_Text_Mining_The_Leipzig_Corpora_Collection
Check out the Linguistic Data Consortium (LDC), it has all kinds of linguistic corpora, though they may require purchase on some, and may be use-restricted.