Compute time series for 70,000 locations (Speed up the processing)

Hi @Basile_Goussard, I have some additional remarks:

  • Do you consider all the time steps or do you filter out cloudy scenes beforehand using the eo:cloud_cover property? If not applied already, this could save you some time, reducing the number of Items you need to load.
  • Are you running the computation in a serial manner? If you have more than 1 CPU you could easily parallelize the whole computation using libraries like joblib, multiprocessing or dask
  • You could also think about grouping the geometries based on their location and to which Sentinel-2 tile(s) they correspond. This would allow to reduce the number of STAC API calls and reuse the same xarray object to get the info about multiple points at once.
  • Finally, really important, are you considering the baseline processing change that happened in 2022? Otherwise you might see strange differences in the indices you compute comparing years before 2022 and afterwards. References:
1 Like