mcboost implements Multi-Calibration Boosting (Hebert-Johnson et al., 2018; Kim et al., 2019) for the multi-calibration of a machine learning model’s prediction. Multi-Calibration works best in scenarios where the underlying data & labels are unbiased but a bias is introduced within the algorithm’s fitting procedure. This is often the case, e.g. when an algorithm fits a majority population while ignoring or under-fitting minority populations.
For more information and example, see the package’s website.
More details with respect to usage and the procedures can be found in the package vignettes.
The current version can be downloaded from CRAN using:
install.packages("mcboost")
You can install the development version of mcboost from Github with:
::install_github("mlr-org/mcboost") remotes
Post-processing with mcboost
needs three components. We
start with an initial prediction model (1) and an auditing algorithm (2)
that may be customized by the user. The auditing algorithm then runs
Multi-Calibration-Boosting on a labeled auditing dataset (3). The
resulting model can be used for obtaining multi-calibrated
predictions.
In this simple example, our goal is to improve calibration for an
initial predictor
, e.g. a ML algorithm trained on an
initial task. Internally, mcboost
often makes use of
mlr3
and learners that come with
mlr3learners
.
library(mcboost)
library(mlr3)
First we set up an example dataset.
# Example Data: Sonar Task
= tsk("sonar")
tsk = sample(tsk$row_ids, 100) # 100 rows for training
tid = tsk$data(cols = tsk$feature_names, rows = tid)
train_data = tsk$data(cols = tsk$target_names, rows = tid)[[1]] train_labels
To provide an example, we assume that we have already a learner
l
which we train below. We can now wrap this initial
learner’s predict function for use with mcboost
, since
mcboost
expects the initial model to be specified as a
function
with data
as input.
= lrn("classif.rpart")
l $train(tsk$clone()$filter(tid))
l
= function(data) {
init_predictor # Get response prediction from Learner
= l$predict_newdata(data)$response
p # One-hot encode and take first column
one_hot(p)
}
We can now run Multi-Calibration Boosting by instantiating the object
and calling the multicalibrate
method. Note, that
typically, we would use Multi-Calibration on a separate validation set!
We furthermore select the auditor model, a
SubpopAuditorFitter
, in our case a
Decision Tree
:
= MCBoost$new(
mc init_predictor = init_predictor,
auditor_fitter = "TreeAuditorFitter")
$multicalibrate(train_data, train_labels) mc
Lastly, we predict on new data.
= setdiff(tsk$row_ids, tid) # held-out data
tstid = tsk$data(cols = tsk$feature_names, rows = tstid)
test_data $predict_probs(test_data) mc
While mcboost
in its defaults implements Multi-Accuracy
(Kim et al., 2019),
it can also multi-calibrate predictors (Hebert-Johnson
et al., 2018). In order to achieve this, we have to set the
following hyperparameters:
= MCBoost$new(
mc init_predictor = init_predictor,
auditor_fitter = "TreeAuditorFitter",
num_buckets = 10,
multiplicative = FALSE
)
mcboost
can also be used within a
mlr3pipeline
in order to use at the full end-to-end
pipeline (in the form of a GraphLearner
).
library(mlr3)
library(mlr3pipelines)
= ppl_mcboost(lrn("classif.rpart"))
gr = tsk("sonar")
tsk = sample(1:208, 108)
tid $train(tsk$clone()$filter(tid))
gr$predict(tsk$clone()$filter(setdiff(1:208, tid))) gr
The mcboost
vignettes Basics
and Extensions and Health
Survey Example demonstrate a lot of interesting showcases
for applying mcboost
.
This R package is licensed under the LGPL-3. If you encounter problems using this software (lack of documentation, misleading or wrong documentation, unexpected behaviour, bugs, …) or just want to suggest features, please open an issue in the issue tracker. Pull requests are welcome and will be included at the discretion of the maintainers.
As this project is developed with mlr3’s style guide in mind, the following resources can be helpful to individuals wishing to contribute: Please consult the wiki for a style guide, a roxygen guide and a pull request guide.
Please note that the mcboost project is released with a Contributor Code of Conduct. By contributing to this project, you agree to abide by its terms.
If you use mcboost
, please cite our package as well as
the two papers it is based on:
@article{pfisterer2021,
author = {Pfisterer, Florian and Kern, Christoph and Dandl, Susanne and Sun, Matthew and
Kim, Michael P. and Bischl, Bernd},
title = {mcboost: Multi-Calibration Boosting for R},
journal = {Journal of Open Source Software},
doi = {10.21105/joss.03453},
url = {https://doi.org/10.21105/joss.03453},
year = {2021},
publisher = {The Open Journal},
volume = {6},
number = {64},
pages = {3453}
}
# Multi-Calibration
@inproceedings{hebert-johnson2018,
title = {Multicalibration: Calibration for the ({C}omputationally-Identifiable) Masses},
author = {Hebert-Johnson, Ursula and Kim, Michael P. and Reingold, Omer and Rothblum, Guy},
booktitle = {Proceedings of the 35th International Conference on Machine Learning},
pages = {1939--1948},
year = {2018},
editor = {Jennifer Dy and Andreas Krause},
volume = {80},
series = {Proceedings of Machine Learning Research},
address = {Stockholmsmässan, Stockholm Sweden},
publisher = {PMLR}
}
# Multi-Accuracy
@inproceedings{kim2019,
author = {Kim, Michael P. and Ghorbani, Amirata and Zou, James},
title = {Multiaccuracy: Black-Box Post-Processing for Fairness in Classification},
year = {2019},
isbn = {9781450363242},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3306618.3314287},
doi = {10.1145/3306618.3314287},
booktitle = {Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society},
pages = {247--254},
location = {Honolulu, HI, USA},
series = {AIES '19}
}