Hi everyone!

I'm working on a multitemporal classification of the forest types from the Sentinel data. The reference data for each class is represented by multiple polygons scattered over the study area. After I overlay polygons and satellite data (88 S2 tiles from different period of year) my training dataset becomes really huge and I'm looking for a way to reduce its size.

I'm thinking of doing aggregation by polygon and date, and so far I have calculated mean value, max, min and some other quantiles. The cross-validation suggest that I'm not that far for when I'm using the complete dataset.

Does anyone know some better way or can provide me useful references how to perform aggregation per polygon?

Or should I completely abandon this approach and use some other method for reduction of data size?

Thanks in advance!

Similar questions and discussions