Estimating the Area of Applicability (AOA) can be computationally intensive, depending on the amount of training data used for a model as well as the amount of new data the AOA has to be computed. This vignette goes over the possibility to (partly) compute the AOA in parallel. We will use the same data setup as the vignette “Area of applicability of spatial prediction models”. Please have a look there for a general introduction to the AOA and the details about the example data generation.
library(CAST)
library(virtualspecies)
library(caret)
library(raster)
library(sp)
library(sf)
library(viridis)
library(latticeExtra)
library(gridExtra)
<- stack(system.file("extdata","bioclim.grd",package="CAST"))
predictors <- generateSpFromPCA(predictors,
response means = c(3,1),sds = c(2,2), plot=F)$suitab.raster
<- predictors[[1]]
mask values(mask)[!is.na(values(mask))] <- 1
<- rasterToPolygons(mask)
mask
# Generate Clustered Training Samples
<- function(x,n,nclusters,maxdist,seed){
csample set.seed(seed)
<- sp::spsample(x, n = nclusters, type="random")
cpoints <- cpoints
result $clstrID <- 1:length(cpoints)
resultfor (i in 1:length(cpoints)){
<- rgeos::gBuffer(cpoints[i,], width = maxdist)
ext <- sp::spsample(ext, n = (n-nclusters)/nclusters,
newsamples type="random")
$clstrID <- rep(i,length(newsamples))
newsamples<- rbind(result,newsamples)
result
}$ID <- 1:nrow(result)
resultreturn(result)
}
<- csample(mask,75,15,maxdist=0.20,seed=15)
samplepoints
<- extract(predictors,samplepoints,df=TRUE)
trainDat $response <- extract (response,samplepoints)
trainDat<- merge(trainDat,samplepoints,by.x="ID",by.y="ID")
trainDat <- trainDat[complete.cases(trainDat),] trainDat
set.seed(10)
<- train(trainDat[,names(predictors)],
model_random $response,
trainDatmethod="rf",
importance=TRUE,
trControl = trainControl(method="cv"))
<- raster::predict(predictors,model_random) prediction_random
The simplest methods to compute the AOA in parallel is by providing a
cluster
object in the function call. This way the distance
calculation between training data and new data will run on multiple
cores (utilizing parallel::parApply
). This is recommended
if the training set is relatively small but new locations for which the
AOA should be computed are abundant.
library(doParallel)
library(parallel)
<- makeCluster(4)
cl registerDoParallel(cl)
<- aoa(studyArea,model,cl=cl) AOA
For even better performances, it is recommended to compute the AOA in
two steps. First, the DI of training data and the resulting DI threshold
is computed from the model or training data with the function
trainDI
. The result from trainDI is usually the first step
of the aoa
function, however it can be skipped by providing
the trainDI object in the function call. This makes it possible to
compute the AOA on multiple raster tiles at once (e.g. on different
cores). This is especially useful for very large prediction areas,
e.g. in global mapping.
= trainDI(model_random)
model_random_trainDI print(model_random_trainDI)
## DI of 74 observation
## Predictors: bio2 bio5 bio10 bio13 bio14 bio19
##
## AOA Threshold: 0.09237258
saveRDS(model_random_trainDI, "path/to/file")
If you have a large raster, you divide it into multiple smaller tiles and apply the trainDI object afterwards to each tile.
= crop(predictors, c(0,7,42.33333,54.83333))
r1 = crop(predictors, c(7,14,42.33333,54.83333))
r2 = crop(predictors, c(14,21,42.33333,54.83333))
r3
grid.arrange(spplot(r1[[1]], main = "Tile 1"),
spplot(r2[[1]], main = "Tile 2"),
spplot(r3[[1]], main = "Tile 3"), ncol = 3)
Use the trainDI
argument in the aoa
function to specify, that you want to use a previously computed trainDI
object.
= aoa(newdata = r1, trainDI = model_random_trainDI)
aoa_r1
grid.arrange(spplot(r1[[1]], main = "Tile 1: Predictors"),
spplot(aoa_r1$DI, main = "Tile 1: DI"),
spplot(aoa_r1$AOA, main = "Tile 1: AOA"), ncol = 3)
You can now run the aoa function in parallel on the different tiles!
Of course you can use for favorite parallel backend for this task, here
we use mclapply from the parallel
package.
library(parallel)
= mclapply(list(r1, r2, r3), function(tile){
tiles_aoa aoa(newdata = tile, trainDI = model_random_trainDI)
mc.cores = 3) },
grid.arrange(spplot(tiles_aoa[[1]]$AOA, main = "Tile 1"),
spplot(tiles_aoa[[2]]$AOA, main = "Tile 2"),
spplot(tiles_aoa[[3]]$AOA, main = "Tile 3"), ncol = 3)
For larger tasks it might be useful to save the tiles to you hard-drive and load them one by one to avoid filling up your RAM.
# Simple Example Code for raster tiles on the hard drive
= list.files("path/to/tiles", full.names = TRUE)
tiles
= mclapply(tiles, function(tile){
tiles_aoa = raster::stack(tile)
current aoa(newdata = current, trainDI = model_random_trainDI)
mc.cores = 3) },
It should be possible to combine both parallelization methods! For example on a High Performance Cluster, multiple raster tiles can be handled by different nodes and then, each node computes the distance between prediction and training data on multiple cores.