This package provides functions to simplify application of forecast
verification metrics to large datasets of ensemble forecasts. The design
goals of easyVerification
are:
The forecast metrics are imported from the
SpecsVerification
package. Additional verification metrics
not available through SpecsVerification
are implemented
directly. At the time of publication, the package offers functionality
to compute the following deterministic and probabilitistic scores and
skill scores:
EnsMe
), mean absolute
error(EnsMae
), mean squared error (EnsMse
),
and root mean squared error (EnsRmse
) of the ensemble mean
and their skill scores (e.g. EnsRmsess
)EnsCorr
)EnsSprErr
and
FairSprErr
)EnsRoca
) and its skill score
(EnsRocss
)FairRps
) and standard (EnsRps
) rank
probability scores and skill scores (e.g. FairRpss
)FairCrps
) and standard (EnsCrps
)
continuous ranked probability scores and skill scores
(e.g. FairCrpss
)DressIgn
, DressCrps
) and
their skill scores (DressIgnSs
, DressCrpss
)
with default ensemble dressing method (“silverman”)Ens2AFC
)Additional forecast verification metrics can be added by the user following the examples above.
You can get the latest version from CRAN
install.packages("easyVerification")
You can get the latest development version using
::install_github("MeteoSwiss/easyVerification") devtools
You can find out more about the package and its functionality in the vignette.
vignette("easyVerification")
The following example illustrates how to compute the continous ranked probability skill score of an ensemble forecast:
suppressPackageStartupMessages(library(easyVerification))
## check out what is included in easyVerification
ls(pos = "package:easyVerification")
#> [1] "climFairRpss" "convert2prob" "count2prob" "Ens2AFC" "EnsCorr"
#> [6] "EnsError" "EnsErrorss" "EnsIgn" "EnsIgnss" "EnsMae"
#> [11] "EnsMaess" "EnsMe" "EnsMse" "EnsMsess" "EnsRmse"
#> [16] "EnsRmsess" "EnsRoca" "EnsRocss" "EnsSprErr" "FairSprErr"
#> [21] "generateRef" "indRef" "size" "toyarray" "toymodel"
#> [26] "veriApply"
## set up the forecast and observation data structures
## assumption: we have 13 x 5 spatial instances, 15 forecast
## times and 51 ensemble members
<- toyarray(c(13, 5), N = 15, nens = 51)
tm <- veriApply("EnsCrpss", fcst = tm$fcst, obs = tm$obs)
fo.crpss
## if the data are organized differently such that forecast
## instance and ensemble members are NOT the last two array
## dimensions, this has to be indicated
## alternative setup:
## forecast instance, ensemble members, all forecast locations
## collated in one dimension
<- array(aperm(tm$fcst, c(3, 4, 1, 2)), c(15, 51, 13 * 5))
fcst2 <- array(aperm(tm$obs, c(3, 1, 2)), c(15, 13 * 5))
obs2 <- veriApply("EnsCrpss",
fo2.crpss fcst = fcst2, obs = obs2,
ensdim = 2, tdim = 1
)
## The forecast evaluation metrics are the same, but the
## data structure is different in the two cases
dim(fo.crpss$crpss)
#> NULL
dim(fo2.crpss$crpss)
#> NULL
range(fo.crpss$crpss - c(fo2.crpss$crpss))
#> Warning in min(x): no non-missing arguments to min; returning Inf
#> Warning in max(x): no non-missing arguments to max; returning -Inf
#> [1] Inf -Inf
As of easyVerification 0.1.7.0
, parallel processing is
supported under *NIX systems. The following minimal example illustrates
how to use the parallel processing capabilities of
easyVerification
.
## generate a toy-model forecast observation set of
## 10 x 10 forecast locations (e.g. lon x lat)
<- toyarray(c(10, 10))
tm
## run and time the ROC skill score for tercile forecasts without parallelization
system.time({
<- veriApply("EnsRocss", tm$fcst, tm$obs, prob = 1:2 / 3)
tm.rocss
})#> user system elapsed
#> 0.383 0.001 0.383
## run the ROC skill score with parallelization
system.time({
<- veriApply("EnsRocss", tm$fcst, tm$obs, prob = 1:2 / 3, parallel = TRUE)
tm.rocss.par
})#> [1] "Number of CPUs 3"
#> user system elapsed
#> 0.018 0.019 0.376
To get additional help and examples please see the vignette
{r, eval=FALSE} vignette('easyVerification')
or the help
pages of the functions in easyVerification
(e.g. {r, eval=FALSE} help(veriApply)
).