Hello,

I want to construct a neighbour net on a dataset generated from GBS. (De novo assembly in ipyrad, 130 samples). Depending on how I filter, the dataset is quite large. When I filter more stringently (loci must be present in at least half the samples), the file loads fine in SplitsTree 4.17.0. However, I would ideally like to include more loci. At this point, loading the file becomes extremely slow. (After several hours, nothing seems to happen.)

Does someone have experience with constructing neighbor nets with large data? Is there a better implementation for this? E.g. I know the neighbor net algorithm is part of the R package phangorn (with the note that is experimental). Or would you suggest I decrease the size of my dataset e.g. by choosing one SNP per locus? So far, I am using the nexus files written by ipyrad (the full alignment).

Similar questions and discussions