1  Calibration

In this notebook, we present a Data Generative Process (DGP) that will be used to run simulations. We consider a binary outcome variable with a latent probability. We would like to build a classifier for that outcome variable, for which the estimated scores will reflect the true latent probability.

However, when building a classifier, if the model is not well calibrated, the estimated scores may not be good estimates of the latent probabilities. To have a better insight on the impact of having a non calibrated model, we will apply transformations to the true probabilities and examine how this affects the calibration metrics. These transformations will be termed as “decalibration” of the true probabilities.

Display the definitions of colors.
library(tidyverse)
── Attaching core tidyverse packages ──────────────────────── tidyverse 2.0.0 ──
✔ dplyr     1.1.4     ✔ readr     2.1.4
✔ forcats   1.0.0     ✔ stringr   1.5.0
✔ ggplot2   3.4.4     ✔ tibble    3.2.1
✔ lubridate 1.9.3     ✔ tidyr     1.3.0
✔ purrr     1.0.2     
── Conflicts ────────────────────────────────────────── tidyverse_conflicts() ──
✖ dplyr::filter() masks stats::filter()
✖ dplyr::lag()    masks stats::lag()
ℹ Use the conflicted package (<http://conflicted.r-lib.org/>) to force all conflicts to become errors
Display the definitions of colors.
wongBlack     <- "#000000"
wongGold      <- "#E69F00"
wongLightBlue <- "#56B4E9"
wongGreen     <- "#009E73"
wongYellow    <- "#F0E442"
wongBlue      <- "#0072B2"
wongOrange    <- "#D55E00"
wongPurple    <- "#CC79A7"

1.1 Data Generating Process

Let us consider a binary variable that is assumed to follow a Bernoulli distribution: \(D_i\sim B(p_i)\), where \(p_i\) is the probability of observing \(D_i = 1\). We define this probability \(p_i\) according to the following function \[p_i = \frac{1}{1+\exp(-\eta_i)}, \tag{1.1}\]

where \(\eta_i\) is defined as follows:

\[\eta_i = -0.1x_1 + 0.05x_2 + 0.2x_3 - 0.05x_4 + \varepsilon_i, \tag{1.2}\] where \(\varepsilon_i \sim \mathcal{N}(0, 0.5^2)\).

Let us simulate a toy dataset, following this DGP:

n <- 2000
x1 <- runif(n)
x2 <- runif(n)
x3 <- runif(n)
x4 <- runif(n)
epsilon_p <- rnorm(n, mean = 0, sd = .5)

The latent score:

eta <- -0.1*x1 + 0.05*x2 + 0.2*x3 - 0.05*x4  + epsilon_p

The true probabilities:

p <- (1 / (1 + exp(-eta)))

And the outcome binary variable:

d <- rbinom(n, size = 1, prob = p)

The distribution of the true probabilities are depicted in Figure 1.1

par(mar = c(4.1, 4.3, 2.1, 0.5))
hist(p, xlab = "p", main = "", xlim = c(0,1))
Figure 1.1: Distribution of the true probabilities

1.1.1 Poor Calibration

As previously mentioned, we would like to measure the calibration of a binary classifier. To do so, we will deliberately transform the probabilities. We consider two types of transformations:

  • one applied to the latent probability
  • another applied to the linear predictor.

In a more formal way, we will introduce two scaling parameters. The scaling parameter that modifies the latent probability changes Equation 1.1 which becomes: \[p_i^{u} = \left(\frac{1}{1+\exp(-\eta_i^u)}\right)^{\alpha} \tag{1.3}\] The scaling parameter that modifies the linear predictor changes Equation 1.4 which becomes: \[\eta_i^u = \gamma \times \left( (-0.1)x_1 + 0.05x_2 + 0.2x_3 - 0.05x_4 + \varepsilon_i\right), \tag{1.4}\]

We define a function to simulate data according to the GDP. This function allows to transform the probabilities (by default, it does not).

#' Simulates data
#'
#' @param n_obs number of desired observations
#' @param seed seed to use to generate the data
#' @param alpha scale parameter for the latent probability (if different 
#'   from 1, the probabilities are transformed and it may induce decalibration)
#' @param gamma scale parameter for the latent score (if different from 1, 
#'   the probabilities are transformed and it may induce decalibration)
sim_data <- function(n_obs = 2000, 
                     seed, 
                     alpha = 1, 
                     gamma = 1,
                     a = c(-0.1, 0.05, 0.2, -0.05)) {
  set.seed(seed)

  x1 <- runif(n_obs)
  x2 <- runif(n_obs)
  x3 <- runif(n_obs)
  x4 <- runif(n_obs)
  
  a1 <- a[1]
  a2 <- a[2]
  a3 <- a[3]
  a4 <- a[4]
  
  epsilon_p <- rnorm(n_obs, mean = 0, sd = .5)
  
  # True latent score
  eta <- a1*x1 + a2*x2 + a3*x3 + a4*x4  + epsilon_p
  # Transformed latent score
  eta_u <- gamma * eta
  
  # True probability
  p <- (1 / (1 + exp(-eta)))
  # Transformed probability
  p_u <- ((1 / (1 + exp(-eta_u))))^alpha

  # Observed event
  d <- rbinom(n_obs, size = 1, prob = p)

  tibble(
    # Event Probability
    p = p,
    p_u = p_u,
    # Binary outcome variable
    d = d,
    # Variables
    x1 = x1,
    x2 = x2,
    x3 = x3,
    x4 = x4
  )
}

We can then use that function to observe how it impacts the latent probability distribution.

1.1.1.1 Varying \(\alpha\)

Let us consider \(\alpha = \{1/3, 2/3, 1, 3/2, 3\}\).

alphas <- c(1/3, 2/3, 1, 3/2, 3)

We generate data using these values:

data_alphas <- 
  map(
  .x = alphas,
  .f = ~sim_data(
    n_obs = 2000, seed = 1, alpha = .x, gamma = 1
  )
)

Let us look at the transformed probabilities as a function as the true probabilities.

true_prob_alphas <- map(data_alphas, "p")
transformed_prob_alphas <- map(data_alphas, "p_u")
# Ordering by increasing values of the true proba
order_true_alphas <- map(true_prob_alphas, order)
true_prob_alphas <- map2(
  .x = true_prob_alphas, .y = order_true_alphas, .f = ~.x[.y])
transformed_prob_alphas <- map2(
  .x = transformed_prob_alphas, .y = order_true_alphas, .f = ~.x[.y])

colours <- RColorBrewer::brewer.pal(
  length(true_prob_alphas)+1, name = "Blues"
)
colours <- colours[-1]
colours[alphas == 1] <- "orange"

par(mar = c(4.1, 4.3, 2.1, 0.5))
plot(
  x = true_prob_alphas[[1]],
  y = transformed_prob_alphas[[1]], type = "l",
  xlab = latex2exp::TeX("$p$"),
  ylab = latex2exp::TeX("$p^u$"),
  col = colours[1],
  ylim = c(0, 1)
)
for (i in 2:length(true_prob_alphas)) {
  lines(
    x = true_prob_alphas[[i]],
    y = transformed_prob_alphas[[i]],
    col = colours[i]
  )
}
legend(
  "bottomright", col = colours, lty = 1,
  legend = latex2exp::TeX(str_c("$\\alpha = ",round(alphas, 2), "$"))
)
Figure 1.2: Transformed probabilities as a function of true probabilities

Let us also have a look at a histogram.

p_alphas <- map(data_alphas, ~sort(.x$p_u))
colours <- RColorBrewer::brewer.pal(
  length(p_alphas)+1, name = "Blues"
)
colours <- colours[-1]
colours[alphas == 1] <- "orange"

par(mar = c(4.1, 4.3, 2.1, 0.5), mfrow = c(3,2))
hist(
  p_alphas[[1]], breaks = seq(0, 1, by = .05),
  col = adjustcolor(colours[1], alpha.f = .4),
  xlab = "p", ylab = "Freq.",
  xlim = c(0, 1),
  main = latex2exp::TeX(str_c("$\\alpha = ", round(alphas[1],2), "$"))
)
for (i in 2:length(p_alphas)) {
  hist(
    p_alphas[[i]], breaks = seq(0, 1, by = .05),
    col = adjustcolor(colours[i], alpha.f = .4),
    xlab = "p", ylab = "Freq.",
    main = latex2exp::TeX(str_c("$\\alpha = ", round(alphas[i], 2), "$"))
  )
}
Figure 1.3: Effect of varying \(\alpha\) on the latent probabilities

1.1.1.2 Varying \(\gamma\)

Let us consider \(\gamma = \{c(1/3, 2/3, 1, 3/2, 3)\}\).

gammas <- c(c(1/3, 2/3, 1, 3/2, 3))

We generate data using these values:

data_gammas <- 
  map(
  .x = gammas,
  .f = ~sim_data(
    n_obs = 2000, seed = 1, alpha = 1, gamma = .x
  )
)

Let us look at the transformed probabilities as a function as the true probabilities.

true_prob_gammas <- map(data_gammas, "p")
transformed_prob_gammas <- map(data_gammas, "p_u")
# Ordering by increasing values of the true proba
order_true_gammas <- map(true_prob_gammas, order)
true_prob_gammas <- map2(
  .x = true_prob_gammas, .y = order_true_gammas, .f = ~.x[.y])
transformed_prob_gammas <- map2(
  .x = transformed_prob_gammas, .y = order_true_gammas, .f = ~.x[.y])

colours <- RColorBrewer::brewer.pal(
  length(true_prob_gammas)+1, name = "Blues"
)
colours <- colours[-1]
colours[gammas == 1] <- "orange"

par(mar = c(4.1, 4.3, 2.1, 0.5))
plot(
  x = true_prob_gammas[[1]],
  y = transformed_prob_gammas[[1]], type = "l",
  xlab = latex2exp::TeX("$p$"),
  ylab = latex2exp::TeX("$p^u$"),
  col = colours[1],
  ylim = c(0, 1)
)
for (i in 2:length(true_prob_gammas)) {
  lines(
    x = true_prob_gammas[[i]],
    y = transformed_prob_gammas[[i]],
    col = colours[i]
  )
}
legend(
  "bottomright", col = colours, lty = 1,
  legend = latex2exp::TeX(str_c("$\\gamma = ",round(gammas, 2), "$"))
)
Figure 1.4: Transformed probabilities as a function of true probabilities

We look at the distribution of the values with histograms as well.

p_gammas <- map(data_gammas, ~sort(.x$p_u))
colours <- RColorBrewer::brewer.pal(
  length(p_gammas)+1, name = "Blues"
)
colours <- colours[-1]
colours[gammas == 1] <- "orange"

par(mar = c(4.1, 4.3, 2.1, 0.5), mfrow = c(3,2))
hist(
  p_gammas[[1]], breaks = seq(0, 1, by = .05),
  col = adjustcolor(colours[1], alpha.f = .4),
  xlab = "p", ylab = "Freq.",
  xlim = c(0, 1),
  main = latex2exp::TeX(str_c("$\\gamma = ", round(gammas[1],2), "$"))
)
for (i in 2:length(p_gammas)) {
  hist(
    p_gammas[[i]], breaks = seq(0, 1, by = .05),
    col = adjustcolor(colours[i], alpha.f = .4),
    xlab = "p", ylab = "Freq.",
    main = latex2exp::TeX(str_c("$\\gamma = ", round(gammas[i], 2), "$"))
  )
}
Figure 1.5: Effect of varying \(\gamma\) on the latent probabilities

1.2 Measuring Calibration

A model is calibrated if:

\[\mathbb{P}(D = d \mid s(x) = p) = p,\]

for all values \(p \in[0,1]\) and values \(d = \{0,1\}\), where \(s(x)\) is the estimated score that is output by a model.

Warning

In our case, when we vary \(\alpha\) or \(\gamma\), \(s(x)\) corresponds to \(p^u\), the transformed probabilities.

To have an idea of the calibration of a model, we can use two ways:

  1. a metric-based technique
  2. a visualization-based technique.

1.2.1 Metrics

Let us define functions to compute the following metrics:

  • MSE based on true probabilities
  • Brier Score
  • Expected Calibration Error
  • MSE on quantile-based bins (QMSE)
  • Weighted MSE on the calibration curve defined on a continuum of values.
Note

Each metric is defined below.

1.2.1.1 MSE Based on true probabilities

As we observe the true probabilities in our simulations, we can compare the transformed probabilities \(p^u\) to the true probabilities \(p\). The MSE writes: \[\text{MSE} = \frac{1}{n}\sum_{i=1}^{n} \left(p_i - p_i^u\right)^2. \tag{1.5}\]

1.2.1.2 Brier Score

The target variable \(D\) is binary and takes values in \(\{0,1\}\). In that case, the Brier Score writes: \[\text{BS} = \frac{1}{n}\sum_{i=1}^{n} (s(x_i) - d_i)^{2}, \tag{1.6}\] where \(s(x_i)\) is the predicted score (we use here either the true value \(p\) or its “uncalibrated” version \(p^u\)) for observation \(i\) and \(d_i\) is the outcome variable.

brier_score <- function(obs, scores) mean((scores - obs)^2)

1.2.1.3 Expected Calibration Error

To measure the calibration of a model, one of the metrics, Expected Calibration Error (or ECE), consists in splitting the predicted scores \(s(x)\) into bins and computing in each bins two metrics: the accuracy (average of the empirical probabilities in the sample, or fractions of correctly predicted classes) and the confidence (average of the predicted scores in the bin). The Expected Calibration Error is then computed as the average over the bins

\[\text{ECE} = \sum_{b=1}^{B} \frac{n_b}{n} \mid \text{acc}(b) - \text{conf}(b) \mid, \tag{1.7}\]

where \(n_b\) is the number of observations in bin \(b\), \(n\) is the number of observations, \(\text{acc}(b)\) is the accuracy of bin \(b\), and \(\text{conf}(b)\) is the the confidence in bin \(b\).

The accuracy of a bin is defined as the average of the correctly predicted classes in the bin: \[\text{acc}(b) = \frac{1}{n_b} \sum_{i \in b} \mathrm{1}_{\hat{d}_i = d_i}, \tag{1.8}\] where \(\hat{d}_i\) is the predicted class (0 or 1) for observation \(i\): \[\hat{d}_i = \begin{cases} 1, & s(x) \geq \tau\\ 0, & s(x) < \tau\\ \end{cases},\]

where \(\tau\) is the classification threshold, with \(\tau = .5\) by default and where \(s(x)\) is the propensity score.

The confidence of a bin is defined as the average of the predicted scores in the bin: \[\text{conf}(b) = \frac{1}{n_b} \sum_{i \in b} s(x) \tag{1.9}\]

We define a function to compute the bins. We will consider here the deciles to create the bins (and therefore set k=10 in our function).

#' Computes summary statistics for binomial observed data and predicted scores
#' returned by a model
#'
#' @param obs vector of observed events
#' @param scores vector of predicted probabilities
#' @param k number of classes to create (quantiles, default to `10`)
#' @param threshold classification threshold (default to `.5`)
#' @return a tibble where each row correspond to a bin, and each columns are:
#' - `score_class`: level of the decile that the bin represents
#' - `nb`: number of observation
#' - `mean_obs`: average of obs (proportion of positive events)
#' - `mean_score`: average predicted score (confidence)
#' - `sum_obs`: number of positive events (number of positive events)
#' - `accuracy`: accuracy (share of correctly predicted, using the
#'    threshold)
get_summary_bins <- function(obs,
                             scores,
                             k = 10, 
                             threshold = .5) {
  breaks <- quantile(scores, probs = (0:k) / k)
  tb_breaks <- tibble(breaks = breaks, labels = 0:k) |>
    group_by(breaks) |>
    slice_tail(n = 1) |>
    ungroup()
  
  x_with_class <- tibble(
    obs = obs,
    score = scores,
  ) |>
    mutate(
      score_class = cut(
        score,
        breaks = tb_breaks$breaks,
        labels = tb_breaks$labels[-1],
        include.lowest = TRUE
      ),
      pred_class = ifelse(score > threshold, 1, 0),
      correct_pred = obs == pred_class
    )
  
  x_with_class |>
    group_by(score_class) |>
    summarise(
      nb = n(),
      mean_obs = mean(obs),
      mean_score = mean(score), # confidence
      sum_obs = sum(obs),
      accuracy = mean(correct_pred)
    ) |>
    ungroup() |>
    mutate(
      score_class = as.character(score_class) |> as.numeric()
    ) |>
    arrange(score_class)
}

Let us consider, to illustrate how this function works, the case where the probabilities are transformed using \(\alpha=\) 0.3333333. We can consider that these values would be the estimated score returned by an uncalibrated model. We would then obtain the following summary statistics for each bin:

get_summary_bins(
  obs = data_alphas[[1]]$d, 
  scores = data_alphas[[1]]$p_u,
  k = 10, threshold = .5
)
# A tibble: 10 × 6
   score_class    nb mean_obs mean_score sum_obs accuracy
         <dbl> <int>    <dbl>      <dbl>   <int>    <dbl>
 1           1   200    0.315      0.673      63    0.315
 2           2   200    0.35       0.726      70    0.35 
 3           3   200    0.39       0.753      78    0.39 
 4           4   200    0.445      0.773      89    0.445
 5           5   200    0.585      0.790     117    0.585
 6           6   200    0.495      0.808      99    0.495
 7           7   200    0.535      0.823     107    0.535
 8           8   200    0.63       0.840     126    0.63 
 9           9   200    0.68       0.862     136    0.68 
10          10   200    0.675      0.893     135    0.675

We then define the e_calib_error() function, which computes the ECE using the bins defined thanks to the get_summary_bins() function.

#' Expected Calibration Error
#'
#' @param obs vector of observed events
#' @param scores vector of predicted probabilities
#' @param k number of classes to create (quantiles, default to `10`)
#' @param threshold classification threshold (default to `.5`)
e_calib_error <- function(obs,
                          scores, 
                          k = 10, 
                          threshold = .5) {
  summary_bins <- get_summary_bins(
    obs = obs, scores = scores, k = k, threshold = .5
  )
  summary_bins |>
    mutate(ece_bin = nb * abs(accuracy - mean_score)) |>
    summarise(ece = 1 / sum(nb) * sum(ece_bin)) |>
    pull(ece)
}

For example, the ECE when \(s(x)\) is obtained using transformed probabilites with \(\alpha=\) 0.3333333 equals:

e_calib_error(
  obs = data_alphas[[1]]$d, 
  scores = data_alphas[[1]]$p_u, 
  k = 10, threshold = .5
)
[1] 0.2842172

While when we use the true probabilites instead of the transformed ones, the ECE equals:

e_calib_error(
  obs = data_alphas[[which(alphas == 1)]]$d,
  scores = data_alphas[[which(alphas == 1)]]$p,
  k = 10, threshold = .5
)
[1] 0.1101167

Note that due to the random noise introduced in the DGP, the ECE is not equal to 0.

1.2.1.4 QMSE

Let us compute the MSE on based on the quantile-defined bins (as explained in Section 1.2.1.3). To do so, we first calculate the quantiles of the predicted scores: \(p_\tau\). This allows us to define bins \(b=1,\ldots, B\) based on the quantiles.

For each bin, we compute the average of the observed event: \[\bar{d}_b = \frac{1}{n_b} \sum_{i\in b} d_i,\] where \(n_b\) is the number of observation in bin \(b\).

Then, we can compute the confidence in each bin, i.e, the average of the predicted score \(s(x)\): \[\text{conf}(b) = \frac{1}{n_b} \sum_{i\in b} s(x),\]

where \(s(x)\) are the estimated probabilities.

The Quantile-based MSE is then defined as:

\[\text{QMSE} = \frac{1}{n}\sum_{b=1}^{B} n_b \left[\bar{d}_b - \text{conf}(b)\right]^2.\]

#' Quantile-Based MSE
#'
#' @param obs vector of observed events
#' @param scores vector of predicted probabilities
#' @param k number of classes to create (quantiles, default to `10`)
#' @param threshold classification threshold (default to `.5`)
qmse_error <- function(obs,
                       scores, 
                       k = 10, 
                       threshold = .5) {
  summary_bins <- get_summary_bins(
    obs = obs, scores = scores, k = k, threshold = .5
  )
  summary_bins |>
    mutate(qmse_bin = nb * (mean_obs - mean_score)^2) |>
    summarise(qmse = 1/sum(nb) * sum(qmse_bin)) |>
    pull(qmse)
}

1.2.1.5 WMSE

In a similar fashion to the QMSE, we define the Weighted MSE (WMSE). We define a metric to measure calibration based on how far from the perfect calibration is the model. More specifically, for each of the values at which we computed \(E(D \mid s(x) = p)\), with \(p \in [0,1]\) we compute the squared difference between \(E(D \mid s(x) = p)\) and \(D\). We then aggregate the results using a weighted mean, where the weights are the estimated density of the propensity scores \(s(x)\) at the corresponding values at which \(E(D \mid s(x) = p)\) was estimated.

Let us define a function, local_ci_scores(), that identifies the nearest neighbors of a certain predicted score and then calculates the mean scores in that neighborhood accompanied with its confidence interval. This functions requires the binom.confint() function from {binom}.

library(binom)
#' @param obs vector of observed events
#' @param scores vector of predicted probabilities
#' @param tau value at which to compute the confidence interval
#' @param nn fraction of nearest neighbors
#' @prob level of the confidence interval (default to `.95`)
#' @param method Which method to use to construct the interval. Any combination
#'  of c("exact", "ac", "asymptotic", "wilson", "prop.test", "bayes", "logit",
#'  "cloglog", "probit") is allowed. Default is "all".
#' @return a tibble with a single row that corresponds to estimations made in
#'   the neighborhood of a probability $p=\tau$`, using the fraction `nn` of
#'   neighbors, where the columns are:
#'  - `score`: score tau in the neighborhood of which statistics are computed
#'  - `mean`: estimation of $E(d | s(x) = \tau)$
#'  - `lower`: lower bound of the confidence interval
#'  - `upper`: upper bound of the confidence interval
local_ci_scores <- function(obs,
                            scores,
                            tau,
                            nn,
                            prob = .95,
                            method = "probit") {
  
  # Identify the k nearest neighbors based on hat{p}
  k <- round(length(scores) * nn)
  rgs <- rank(abs(scores - tau), ties.method = "first")
  idx <- which(rgs <= k)
  
  binom.confint(
    x = sum(obs[idx]),
    n = length(idx),
    conf.level = prob,
    methods = method
  )[, c("mean", "lower", "upper")] |>
    tibble() |>
    mutate(xlim = tau) |>
    relocate(xlim, .before = mean)
}
local_ci_scores(
  obs = data_alphas[[which(alphas == 1)]]$d,
  scores = data_alphas[[which(alphas == 1)]]$p_u,
  tau = .1, nn = .15, prob = .5, method = "probit"
)
# A tibble: 1 × 4
   xlim  mean lower upper
  <dbl> <dbl> <dbl> <dbl>
1   0.1 0.327 0.309 0.345

Later on, we can compute the mean observed events in the neighborhood of multiple values between 0 and 1:

linspace <- seq(0, 1, length.out = 100)
map(
  .x = linspace,
  .f = ~local_ci_scores(
    obs = data_alphas[[which(alphas == 1)]]$d,
    scores = data_alphas[[which(alphas == 1)]]$p_u,
    tau = .x, 
    nn = .15, prob = .5, method = "probit")
) |> 
  bind_rows()
# A tibble: 100 × 4
     xlim  mean lower upper
    <dbl> <dbl> <dbl> <dbl>
 1 0      0.327 0.309 0.345
 2 0.0101 0.327 0.309 0.345
 3 0.0202 0.327 0.309 0.345
 4 0.0303 0.327 0.309 0.345
 5 0.0404 0.327 0.309 0.345
 6 0.0505 0.327 0.309 0.345
 7 0.0606 0.327 0.309 0.345
 8 0.0707 0.327 0.309 0.345
 9 0.0808 0.327 0.309 0.345
10 0.0909 0.327 0.309 0.345
# ℹ 90 more rows

We define the weighted_mse() function that will then rely results obtained with the local_ci_scores() and compute the WMSE.

#' Compute the Weighted Mean Squared Error to assess the calibration of a model
#'
#' @param local_scores tibble with expected scores obtained with the 
#'   `local_ci_scores()` function
#' @param scores vector of raw predicted probabilities
weighted_mse <- function(local_scores, scores) {
  # To account for border bias (support is [0,1])
  scores_reflected <- c(-scores, scores, 2 - scores)
  dens <- density(
    x = scores_reflected, from = 0, to = 1, 
    n = length(local_scores$xlim)
  )
  # The weights
  weights <- dens$y
  local_scores |>
    mutate(
      wmse_p = (xlim - mean)^2,
      weight = !!weights
    ) |>
    summarise(wmse = sum(weight * wmse_p) / sum(weight)) |>
    pull(wmse)
}

1.2.1.6 Local Calibration Score

library(locfit)
locfit 1.5-9.8   2023-06-11

Attaching package: 'locfit'
The following object is masked from 'package:purrr':

    none
#' Calibration score using Local Regression
#' 
#' @param obs vector of observed events
#' @param scores vector of predicted probabilities
local_calib_score <- function(obs, 
                              scores) {
  
  # Add a little noise to the scores, to avoir crashing R
  scores <- scores + rnorm(length(scores), 0, .001)
  locfit_0 <- locfit(
    formula = d ~ lp(scores, nn = 0.15, deg = 0), 
    kern = "rect", maxk = 200, 
    data = tibble(
      d = obs,
      scores = scores
    )
  )
  # Predictions on [0,1]
  linspace_raw <- seq(0, 1, length.out = 100)
  # Restricting this space to the range of observed scores
  keep_linspace <- which(linspace_raw >= min(scores) & linspace_raw <= max(scores))
  linspace <- linspace_raw[keep_linspace]
  
  locfit_0_linspace <- predict(locfit_0, newdata = linspace)
  locfit_0_linspace[locfit_0_linspace > 1] <- 1
  locfit_0_linspace[locfit_0_linspace < 0] <- 0
  
  # Squared difference between predicted value and the bissector, weighted by the density of values
  scores_reflected <- c(-scores, scores, 2 - scores)
  dens <- density(
    x = scores_reflected, from = 0, to = 1, 
    n = length(linspace_raw)
  )
  # The weights
  weights <- dens$y[keep_linspace]
  
  weighted.mean((linspace - locfit_0_linspace)^2, weights)
}

1.3 Settings for the Simulations

We will now run some simulations on uncalibrated models. To do so, we will consider more or less distorted probabilities \(p^u\) (see Section 1.1.1). For each simulation, we first split the dataset into two parts: a calibration and a test set. While this step is not necessary here, it will be of crucial importance in Chapter 2. On the test set, we will compute standard metrics (such as MSE, accuracy, and so on, in Section 1.5) and then we will consider the previously defined calibration metrics (Section 1.4).

For each value of \(\alpha\) or \(\gamma\) that transform the initial probabilities, we replicate the simulations 200 times.

Let us define a function, get_samples() which generates data from the DGP described in Section 1.1.

#' Get calibration/test samples from the DGP
#'
#' @param seed seed to use to generate the data
#' @param n_obs number of desired observations
#' @param alpha scale parameter for the latent probability (if different 
#'   from 1, the probabilities are transformed and it may induce decalibration)
#' @param gamma scale parameter for the latent score (if different from 1, 
#'   the probabilities are transformed and it may induce decalibration)
get_samples <- function(seed,
                        n_obs = 2000,
                        alpha = 1,
                        gamma = 1) {
  set.seed(seed)
  data_all <- sim_data(
    n_obs = n_obs, seed = seed, alpha = alpha, gamma = gamma
  )
  
  # Calibration/test sets----
  data <- data_all |> select(d, x1:x4)
  probas <- data_all |> select(p)

  calib_index <- sample(1:nrow(data), size = .6 * nrow(data), replace = FALSE)
  tb_calib <- data |> slice(calib_index)
  tb_test <- data |> slice(-calib_index)
  probas_calib <- probas |> slice(calib_index)
  probas_test <- probas |> slice(-calib_index)

  list(
    data_all = data_all,
    data = data,
    tb_calib = tb_calib,
    tb_test = tb_test,
    probas_calib = probas_calib,
    probas_test = probas_test,
    calib_index = calib_index,
    seed = seed,
    n_obs = n_obs,
    alpha = alpha,
    gamma = gamma
  )
}

Let us consider 200 replications for each value of \(\alpha = \left\{\frac{1}{3}, \frac{2}{3}, 1, \frac{3}{2}, 3\right\}\) and then \(\gamma = \left\{\frac{1}{3}, \frac{2}{3}, 1, \frac{3}{2}, 3\right\}\).

n_repl <- 200 # number of replications
n_obs <- 2000 # number of observations to draw
grid_alpha <- expand_grid(alpha = c(1/3, 2/3, 1, 3/2, 3), seed = 1:n_repl)
grid_gamma <- expand_grid(gamma = c(1/3, 2/3, 1, 3/2, 3), seed = 1:n_repl)

1.4 Standard Metrics on Simulations

Before turning to calibration, let us have a look at some standard metrics computed on data drawn from the PGD. We will consider 200 datasets for each value for \(\alpha\) or \(\gamma\), and compute various standard metrics in each replication.

In our predictive modeling framework, denoted as \(h\), we consider a binary variable \(D\) with observed values represented by \(d\), and corresponding observed features denoted as \(\boldsymbol X\) with a realization \(\boldsymbol x\). The predictive model \(h\) takes these inputs and generates a score \(s(\boldsymbol x)\). This score serves as an estimation of the probability that the binary variable \(D\) equals 1. Effectively, \(h\) maps the observed features to a numerical score, reflecting the likelihood of observing \(D=1\). The relationship can be succinctly expressed as \(\mathbb{P}(D=1 | \boldsymbol x) \approx s(\boldsymbol x)\).

Based on the drawn probabilities (that we are able to get here only because we know the data generating process) \(p\) and the “predicted” scores \(s(\boldsymbol x)\), we can compute the Mean Squared Error (MSE): \[ MSE(h) = \frac{1}{n} \sum_{i=1}{n}\left(s(\boldsymbol x) - p\right)^2 \] By defining a probability threshold \(\tau\), we can transform the score into a binary variable \(\hat{D}\) such that: \[ \hat{D} = \begin{cases} 1, & \text{if } s(\boldsymbol x) \geq \tau,\\ 0, & \text{if } s(\boldsymbol x) < \tau,\\ \end{cases} \] We will consider different values of \(\tau\) for each replication of \(h\).

For specific values of \(\tau\), we can compute multiple goodness of fit metrics, based on the confusion table shown in Table 1.1.

Table 1.1: Confusion matrix.
Predicted Positive (P) Predicted Negative (N)
Actual Positive (A) True Positive (TP) False Negative (FN)
Actual Negative (B) False Positive (FP) True Negative (TN)

The cells in this table give the following metrics:

  • True Positive (TP): Instances where the model correctly predicts the positive class.
  • False Negative (FN): Instances where the model predicts the negative class, but the true class is positive.
  • False Positive (FP): Instances where the model predicts the positive class, but the true class is negative.
  • True Negative (TN): Instances where the model correctly predicts the negative class.

Based on these metrics, we compute the following metrics:

  • Accuracy: Accuracy is a measure of the overall correctness of the model. It is calculated as the ratio of correctly predicted instances (both true positives and true negatives) to the total number of instances. The formula is given by: \[ \text{Accuracy} = \frac{\text{True Positives} + \text{True Negatives}}{\text{Total Instances}}\]

  • Misclassification Rate: The misclassification rate, also known as the error rate, represents the proportion of incorrectly classified instances among all instances. It is calculated as the ratio of the sum of false positives and false negatives to the total number of instances. The formula is given by: \[\text{Misclassification Rate} = \frac{\text{False Positives} + \text{False Negatives}}{\text{Total Instances}}\]

  • Sensitivity (True Positive Rate or Recall): Sensitivity measures the ability of the model to correctly identify instances of the positive class. It is the ratio of true positives to the total number of actual positive instances. The formula is given by: \[\text{Sensitivity} = \frac{\text{True Positives}}{\text{True Positives} + \text{False Negatives}}\]

  • Specificity (True Negative Rate): Specificity measures the ability of the model to correctly identify instances of the negative class. It is the ratio of true negatives to the total number of actual negative instances. The formula is given by: \[ \text{Specificity} = \frac{\text{True Negatives}}{\text{True Negatives} + \text{False Positives}}\]

Let us define function compute_gof() that will compute those metrics for some true probabilities \(p\), observed values \(d\), predicted scores \(s(\boldsymbol x)\), and a specific probability threshold \(\tau\).

#' Computes goodness of fit metrics
#' 
#' @param true_prob true probabilities
#' @param obs observed values (binary outcome)
#' @param pred predicted scores
#' @param threshold classification threshold (default to `.5`)
compute_gof <- function(true_prob,
                        obs, 
                        pred, 
                        threshold = .5) {
  
  # MSE
  mse <- mean((true_prob - pred)^2)
  
  pred_class <- as.numeric(pred > threshold)
  confusion_tb <- tibble(
    obs = obs,
    pred = pred_class
  ) |> 
    count(obs, pred)
  
  TN <- confusion_tb |> filter(obs == 0, pred == 0) |> pull(n)
  TP <- confusion_tb |> filter(obs == 1, pred == 1) |> pull(n)
  FP <- confusion_tb |> filter(obs == 0, pred == 1) |> pull(n)
  FN <- confusion_tb |> filter(obs == 1, pred == 0) |> pull(n)
  
  if (length(TN) == 0) TN <- 0
  if (length(TP) == 0) TP <- 0
  if (length(FP) == 0) FP <- 0
  if (length(FN) == 0) FN <- 0
  
  n_pos <- sum(obs == 1)
  n_neg <- sum(obs == 0)
  
  # Accuracy
  acc <- (TP + TN) / (n_pos + n_neg)
  # Missclassification rate
  missclass_rate <- 1 - acc
  # Sensitivity (True positive rate)
  # proportion of actual positives that are correctly identified as such
  TPR <- TP / n_pos
  # Specificity (True negative rate)
  # proportion of actual negatives that are correctly identified as such
  TNR <- TN / n_neg
  # False positive Rate
  FPR <- FP / n_neg
  
  tibble(
    mse = mse,
    accuracy = acc,
    missclass_rate = missclass_rate,
    sensitivity = TPR,
    specificity = TNR,
    threshold = threshold,
    FPR = FPR
  )
}

We encompass these functions into a second one that will:

  1. draw data from the DGP
  2. transform the true probabilities accordingly with either \(\alpha\) or \(\gamma\), to obtain \(p^u\) which will be considered as the score return by some model \(s(\boldsymbol x) := p^u\).
  3. compute the goodness of fit metrics using those values.

Recall that the R objects grid_alpha and grid_gamma contain the value of \(\alpha\) and \(\gamma\), respectively, to use to transform the probabilities. Each value is repeated 200 times to replicate the generation of the data from the same PGD, using the same type of probability transformation at each replication.

#' Computes goodness of fit metrics for a replication
#'
#' @param i row number of the grid to use for the simulation
#' @param grid grid tibble with the seed number (column `seed`) and the deformations value (either `alpha` or `gamma`)
#' @param n_obs desired number of observation
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
compute_gof_simul <- function(i,
                              grid,
                              n_obs,
                              type = c("alpha", "gamma")) {
  current_seed <- grid$seed[i]
  if (type == "alpha") {
    transform_scale <- grid$alpha[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = transform_scale, gamma = 1
    )
  } else if (type == "gamma") {
    transform_scale <- grid$gamma[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = 1, gamma = transform_scale
    )
  } else {
    stop("Transform type should be either alpha or gamma.")
  }
  
  
  # Get the calib/test datasets with true probabilities
  data_all_calib <- current_data$data_all |>
    slice(current_data$calib_index)
  
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)
  
  # Calibration set
  true_prob_calib <- data_all_calib$p_u
  obs_calib <- data_all_calib$d
  pred_calib <- data_all_calib$p
  # Test set
  true_prob_test <- data_all_test$p_u
  obs_test <- data_all_test$d
  pred_test <- data_all_test$p
  
  metrics_simul_calib <- map(
    .x = seq(0, 1, by = .01),
    .f = ~compute_gof(
      true_prob = true_prob_calib,
      obs = obs_calib,
      pred = pred_calib, 
      threshold = .x
    )
  ) |> 
    list_rbind()
  
  metrics_simul_test <- map(
    .x = seq(0, 1, by = .01),
    .f = ~compute_gof(
      true_prob = true_prob_test,
      obs = obs_test,
      pred = pred_test, 
      threshold = .x
    )
  ) |> 
    list_rbind()
  
  roc_calib <- pROC::roc(obs_calib, pred_calib)
  auc_calib <- as.numeric(pROC::auc(roc_calib))
  roc_test <- pROC::roc(obs_test, pred_test)
  auc_test <- as.numeric(pROC::auc(roc_test))
  
  metrics_simul_calib |> 
    mutate(
      auc = auc_calib,
      seed = current_seed,
      scale_parameter = transform_scale,
      type = type,
      sample = "calibration"
    ) |> 
    bind_rows(
      metrics_simul_test |> 
        mutate(
          auc = auc_test,
          seed = current_seed,
          scale_parameter = transform_scale,
          type = type,
          sample = "test"
        )
    )
}

Let us apply this function to the different simulations. We begin with probabilities transformed according to the variation of the parameter \(\alpha\).

library(future)
nb_cores <- future::availableCores()-1
plan(multisession, workers = nb_cores)
progressr::with_progress({
  p <- progressr::progressor(steps = nrow(grid_alpha))
  metrics_alpha <- furrr::future_map(
    .x = 1:nrow(grid_alpha),
    .f = ~{
      p()
      compute_gof_simul(
        i = .x, 
        grid = grid_alpha, 
        n_obs = n_obs, 
        type = "alpha"
      )
    },
    .options = furrr::furrr_options(seed = FALSE)
  )
})

metrics_alpha <- list_rbind(metrics_alpha)

We do the same for \(\gamma\):

library(future)
nb_cores <- future::availableCores()-1
plan(multisession, workers = nb_cores)
progressr::with_progress({
  p <- progressr::progressor(steps = nrow(grid_gamma))
  metrics_gamma <- furrr::future_map(
    .x = 1:nrow(grid_gamma),
    .f = ~{
      p()
      compute_gof_simul(
        i = .x, 
        grid = grid_gamma, 
        n_obs = n_obs, 
        type = "gamma"
      )
    },
    .options = furrr::furrr_options(seed = FALSE)
  )
})

metrics_gamma <- list_rbind(metrics_gamma)

We define function boxplot_simuls_metrics() to plot the results of the simulations. This function will produce a panel of boxplots. Each row of the panel will correspond to a metric whereas each column will correspond to a value for either \(\alpha\) or \(\gamma\). On each figure, the x-axis will correspond to the value used for the probability threshold \(\tau\), and the y-axis will correspond to the values of the metric.

#' Boxplots for the simulations to visualize the distribution of some 
#' traditional metrics as a function of the probability threshold.
#' And, ROC curves
#' The resulting figure is a panel of graphs, with vayring values for the 
#' transformation applied to the probabilities (in columns) and different 
#' metrics (in rows).
#' 
#' @param tb_metrics tibble with computed metrics for the simulations
#' @param type type of transformation: `"alpha"` or `"gamma"`
#' @param metrics names of the metrics computed
boxplot_simuls_metrics <- function(tb_metrics,
                                   type = c("alpha", "gamma"),
                                   metrics) {
  scale_parameters <- unique(tb_metrics$scale_parameter)
  
  par(mfrow = c(length(metrics), length(scale_parameters)))
  for (i_metric in 1:length(metrics)) {
    metric <- metrics[i_metric]
    for (i_scale_parameter in 1:length(scale_parameters)) {
      scale_parameter <- scale_parameters[i_scale_parameter]
      
      tb_metrics_current <- tb_metrics |> 
        filter(scale_parameter == !!scale_parameter)
      
      if (metric == "roc") {
        seeds <- unique(tb_metrics_current$seed)
        if (i_metric == 1) {
          # first row
          title <- latex2exp::TeX(
            str_c("$\\", type, " = ", round(scale_parameter, 2), "$")
          )
          size_top <- 2.1
        } else if (i_metric == length(metrics)) {
          # Last row
          title <- ""
          size_top <- 1.1
        } else {
          title <- ""
          size_top <- 1.1
        }
        
        if (i_scale_parameter == 1) {
          # first column
          y_lab <- str_c(metric, "\n True Positive Rate") 
          size_left <- 5.1
        } else {
          y_lab <- ""
          size_left <- 4.1
        }
        
        par(mar = c(4.5, size_left, size_top, 2.1))
        plot(
          0:1, 0:1,
          type = "l", col = NULL,
          xlim = 0:1, ylim = 0:1,
          xlab = "False Positive Rate", 
          ylab = y_lab,
          main = ""
        )
        for (i_seed in 1:length(seeds)) {
          tb_metrics_current_seed <- 
            tb_metrics_current |> 
            filter(seed == seeds[i_seed])
          lines(
            x = tb_metrics_current_seed$FPR, 
            y = tb_metrics_current_seed$sensitivity,
            lwd = 2, col = adjustcolor("black", alpha.f = .04)
          )
        }
        segments(0, 0, 1, 1, col = "black", lty = 2)
        
      } else {
        # not ROC
        tb_metrics_current <- 
          tb_metrics_current |> 
          filter(threshold %in% seq(0, 1, by = .1))
        form <- str_c(metric, "~threshold")
        if (i_metric == 1) {
          # first row
          title <- latex2exp::TeX(
            str_c("$\\", type, " = ", round(scale_parameter, 2), "$")
          )
          size_top <- 2.1
        } else if (i_metric == length(metrics)) {
          # Last row
          title <- ""
          size_top <- 1.1
        } else {
          title <- ""
          size_top <- 1.1
        }
        
        if (i_scale_parameter == 1) {
          # first column
          y_lab <- metric
        } else {
          y_lab <- ""
        }
        
        par(mar = c(4.5, 4.1, size_top, 2.1))
        boxplot(
          formula(form), data = tb_metrics_current,
          xlab = "Threshold", ylab = y_lab,
          main = title
        )
      }
    }
  }
}

We aim to create a set of boxplots to visually assess the influence of probability transformations using \(\alpha\) or \(\gamma\) on standard metrics. Whenever \(\alpha \neq 1\) or \(\gamma \neq 1\), the resulting scores \(p^u\) represent values akin to those obtained from an uncalibrated model. Our focus is on investigating whether these transformations have any discernible impact on the standard metrics. The key question is to understand if adjusting the probabilities through these transformations introduces noticeable changes in the model’s performance as evaluated by the standard metrics. The results are shown in Figure 1.6 for vayring values of \(\alpha\), and in Figure 1.7 for vayring values of \(\gamma\).

metrics <- c("mse", "accuracy", "sensitivity", "specificity", "roc", "auc")
Display the R codes to produce the Figure.
boxplot_simuls_metrics(
  tb_metrics = metrics_alpha |> filter(sample == "test"), 
  type = "alpha", metrics = metrics
)
Figure 1.6: Calibration transformations made by varying \(\alpha\) and impact on standard metrics. The model is calibrated when \(\alpha=1\).

Display the R codes to produce the Figure.
boxplot_simuls_metrics(tb_metrics = metrics_gamma |> filter(sample == "test"),
                       type = "gamma", metrics = metrics)
Figure 1.7: Calibration transformations made by varying \(\gamma\) and impact on standard metrics. The model is calibrated when \(\beta=1\).

Let us also visualize some standard metrics with a fix threshold of \(\tau=.5\).

plot_boxplot_metric_2 <- function(tb_metrics, type, metrics, metrics_labs) {
  
  df_plot <- tb_metrics |> filter(threshold == 0.5)
  colours <- RColorBrewer::brewer.pal(
    5+1, name = "Blues"
  )
  colours <- colours[-1]
  colours[3] <- "orange"
  if (length(unique(df_plot$scale_parameter)) == 3) {
    colours <- colours[c(1, 3, 5)]
  }
  
  
  
  for (i_metric in 1:length(metrics)) {
    y_lab <- str_c("$", type, "=", sort(round(unique(df_plot$scale_parameter), 2)), "$")
    metric <- metrics[i_metric]
    metric_lab <- metrics_labs[i_metric]
    boxplot(
      formula(str_c(metric, "~ scale_parameter")),
      data = df_plot,
      las = 1,
      xlab = "", ylab = "",
      main = metric_lab,
      col = rev(colours),
      horizontal = TRUE,
      yaxt = "n"
    )
    axis(
      side = 2, at = 1:length(y_lab), 
      labels = latex2exp::TeX(y_lab), las = 2
    )
  }
}

We focus on the folllwing metrics:

metrics <- c("accuracy", "sensitivity", "specificity", "auc")
metrics_labs <- c("Accuracy", "Sensitivity", "Specificity", "AUC")
Display the R codes used to produce the Figure.
par(mar = c(4.1, 4.1, 2.1, 2.1), mfrow = c(2,4))
plot_boxplot_metric_2(
  tb_metrics = metrics_alpha |> filter(scale_parameter %in% c(1/3, 1, 3)) |> 
    filter(sample == "test"), 
  type = "alpha", 
  metrics = metrics, metrics_labs = metrics_labs
)
plot_boxplot_metric_2(
  tb_metrics = metrics_gamma |> filter(scale_parameter %in% c(1/3, 1, 3)) |> 
    filter(sample == "test"), 
  type = "gamma", 
  metrics = metrics, metrics_labs = metrics_labs
)
Figure 1.8: Standard Goodness of Fit Metrics on 200 Simulations for each Value of \(\alpha\) (top) or \(\gamma\) (bottom. The probability threshold is set to \(\tau=0.5\)

1.5 Calibration Metrics on Simulations

We define the f_simul() function to run a single simulation.

#' Performs one replication for a simulation
#' 
#' @param i row number of the grid to use for the simulation
#' @param grid grid tibble with the seed number (column `seed`) and the deformations value (either `alpha` or `gamma`)
#' @param n_obs desired number of observation
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
#' @param linspace values at which to compute the mean observed event when computing the WMSE
f_simul <- function(i, 
                    grid, 
                    n_obs, 
                    type = c("alpha", "gamma"),
                    linspace = NULL) {
  
  if (is.null(linspace)) linspace <- seq(0, 1, length.out = 100)
  
  current_seed <- grid$seed[i]
  if (type == "alpha") {
    transform_scale <- grid$alpha[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = transform_scale, gamma = 1
    )
  } else if (type == "gamma") {
    transform_scale <- grid$gamma[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = 1, gamma = transform_scale
    )
  } else {
    stop("Transform type should be either alpha or gamma.")
  }
  
  # Get the calib/test datasets with true probabilities
  data_all_calib <- current_data$data_all |>
    slice(current_data$calib_index)
  
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)

  # Transformed probabilities
  p_u_calib <- data_all_calib$p_u
  p_u_test <- data_all_test$p_u
  # Observed events
  d_calib <- data_all_calib$d
  d_test <- data_all_test$d
  
  # Mean observed events
  expected_events_calib <- map(
  .x = linspace,
  .f = ~local_ci_scores(
    obs = data_all_calib$d,
    scores = data_all_calib$p_u, 
    tau = .x, 
    nn = .15, prob = .5, method = "probit")
) |> 
  bind_rows()
  
  expected_events_test <- map(
  .x = linspace,
  .f = ~local_ci_scores(
    obs = data_all_test$d,
    scores = data_all_test$p_u, 
    tau = .x, 
    nn = .15, prob = .5, method = "probit")
) |> 
  bind_rows()
  
  # Compute Metrics
  ## Calibration set
  mse_calib <- mean((data_all_calib$p - data_all_calib$p_u)^2)
  brier_calib <- brier_score(obs = d_calib, score = p_u_calib)
  ece_calib <- e_calib_error(
    obs = d_calib, scores = p_u_calib, k = 10, threshold = .5
  )
  qmse_calib <- qmse_error(
    obs = d_calib, score = p_u_calib, k = 10, threshold = .5
  )
  wmse_calib <- weighted_mse(
    local_scores = expected_events_calib, scores = p_u_calib
  )
  lcs_calib <- local_calib_score(obs = d_calib, scores = p_u_calib)
  
  ## Test Set
  mse_test <- mean((data_all_test$p - data_all_test$p_u)^2)
  brier_test <- brier_score(obs = d_test, score = p_u_test)
  ece_test <- e_calib_error(
    obs = d_test, scores = p_u_test, k = 10, threshold = .5
  )
  qmse_test <- qmse_error(
    obs = d_test, score = p_u_test, k = 10, threshold = .5
  )
  wmse_test <- weighted_mse(
    local_scores = expected_events_test, scores = p_u_test
  )
  lcs_test <- local_calib_score(obs = d_test, scores = p_u_test)
  
  tibble(
    seed = grid$seed[i],
    scale_parameter = transform_scale,
    type = type,
    sample = "calibration",
    mse = mse_test,
    brier = brier_test,
    ece = ece_test,
    qmse = qmse_test,
    wmse = wmse_test,
    lcs = lcs_test
  ) |> 
    bind_rows(
      tibble(
        seed = grid$seed[i],
        scale_parameter = transform_scale,
        type = type,
        sample = "test",
        mse = mse_test,
        brier = brier_test,
        ece = ece_test,
        qmse = qmse_test,
        wmse = wmse_test,
        lcs = lcs_test
      )
    )
}

We run the replications of the simulations for each value of \(\alpha\):

library(future)
nb_cores <- future::availableCores()-1
plan(multisession, workers = nb_cores)
progressr::with_progress({
  p <- progressr::progressor(steps = nrow(grid_alpha))
  simul_alpha <- furrr::future_map(
    .x = 1:nrow(grid_alpha),
    .f = ~{
      p()
      f_simul(
        i = .x, 
        grid = grid_alpha, 
        n_obs = n_obs, 
        type = "alpha", 
        linspace = NULL
      )
    },
    .options = furrr::furrr_options(seed = FALSE)
  )
})

simul_alpha <- list_rbind(simul_alpha)

And we do the same for each value of \(\gamma\):

library(future)
nb_cores <- future::availableCores()-1
plan(multisession, workers = nb_cores)
progressr::with_progress({
  p <- progressr::progressor(steps = nrow(grid_gamma))
  simul_gamma <- furrr::future_map(
    .x = 1:nrow(grid_gamma),
    .f = ~{
      p()
      f_simul(
        i = .x, 
        grid = grid_gamma, 
        n_obs = n_obs, 
        type = "gamma", 
        linspace = NULL
      )
    },
    .options = furrr::furrr_options(seed = FALSE)
  )
})

simul_gamma <- list_rbind(simul_gamma)

1.5.1 Results

calib_metrics <- 
  simul_alpha |>
  bind_rows(simul_gamma) |> 
  pivot_longer(
    cols = c("mse", "brier", "ece", "qmse", "wmse", "lcs"),
    names_to = "metric", values_to = "value"
  ) |> 
  mutate(
    metric = factor(
      metric,
      levels = c("mse", "brier", "ece", "qmse", "wmse", "lcs"),
      labels = c("MSE", "Brier Score", "ECE", "QMSE", "WMSE", "LCS")
    )
  )
load("calib_metrics.rda")
#' @param current_metric name of the metric to display (MSE, Brier Score, 
#'   ECE, QMSE, or WMSE)
#' @param calib_metrics tibble with the metrics computed on the test set
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
plot_boxplot_metric <- function(current_metric,
                                calib_metrics, 
                                type = c("alpha", "gamma"),
                                sample = "test") {
  
  data_plot <- calib_metrics |>
    filter(sample == !!sample) |> 
    filter(metric == current_metric, type == !!type) |>
    mutate(
      label = str_c("$\\", type, "$=", scale_parameter, "$")
    )
  
  labels_y <- rep(unique(data_plot$label))
  par(mar = c(4.1, 4.1, 2.1, 2.1))
  boxplot(
    value ~ scale_parameter,
    data = data_plot,
    xlab = latex2exp::TeX(str_c("$\\", type, "$")),
    ylab = current_metric
  )
}

Let us also display these results in a different way.

#' @param current_metric name of the metric to display (MSE, Brier Score, 
#'   ECE, QMSE, or WMSE)
#' @param calib_metrics tibble with the metrics computed on the test set
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
plot_boxplot_metric_2 <- function(current_metric,
                                  calib_metrics, 
                                  type = c("alpha", "gamma"),
                                  sample = "test") {
  
  data_plot <- calib_metrics |>
    filter(metric == current_metric, type == !!type) |>
    filter(sample == !!sample) |> 
    mutate(
      label = str_c("$\\", type, "$=", round(scale_parameter, 2), "$")
    ) |> 
    mutate(scale_parameter = round(scale_parameter, 2))
  
  
  if (as.character(current_metric) == "QMSE") {
    title <- "QBS"
  } else if (as.character(current_metric) == "MSE") {
    title <- "True MSE"
  } else {
    title <- current_metric
  }
  
  labels_y <- rep(unique(data_plot$label))
  colours <- RColorBrewer::brewer.pal(
    5+1, name = "Blues"
  )
  colours <- colours[-1]
  colours[3] <- "orange"
  if (length(labels_y) == 3) {
    colours <- colours[c(1, 3, 5)]
  }
  boxplot(
    value ~ scale_parameter,
    data = data_plot,
    ylab = "",
    xlab = "",
    las = 1,
    main = title,
    col = colours,
    horizontal = TRUE,
    yaxt = "n"
  )
  axis(
    side = 2, at = 1:length(labels_y), 
    labels = latex2exp::TeX(labels_y), las = 2
  )
}
Display the R codes used to produce the Figure.
tb_lab <- expand_grid(
  metric = c("MSE", "Brier Score", "ECE", "LCS"),
  type = c("alpha", "gamma")
) |> 
  mutate(
    metric = factor(metric, levels = c("MSE", "Brier Score", "ECE", "LCS")),
    type = factor(type, levels = c("alpha", "gamma"))
  ) |> 
  arrange(type, metric) |> 
  mutate(
    metric = as.character(metric)
  )
par(mfrow = c(2,4), mar = c(4.1, 2.5, 2.1, 2.1))
for (i in 1:nrow(tb_lab)) {
  plot_boxplot_metric_2(
    current_metric = tb_lab$metric[i], 
    calib_metrics = calib_metrics,
    type = tb_lab$type[i]
  )
}
Figure 1.9: Calibration Metrics on 200 Simulations for each Value of \(\alpha\) (top) or \(\gamma\) (bottom).

1.6 Visualizations

Let us visualize the calibration for the different decalibrated models. We can either use a quantile-based approach, or a method based on moving averages.

1.6.1 Quantile-Based Bins

We visualize the calibration curve similarly to what is done with the calibration_curve() method from sci-kit learn.

The x-axis of the calibration plot reports the mean predicted probabilities computed on different bins, where the bins are defined using the deciles of the predicted scores (using the get_summary_bins() function previously defined). On the y-axis, the corresponding fraction of positive events (\(D=1\)) are reported.

As we simulated data on multiple samples, we can show the calibration curve for all the replications. We can also focus on a single replication and use the confidence interval computed in the get_summary_bins() function.

We define a function, calibration_curve_quant_simul(), to get the summary for each bin, for a single replication.

#' Get the calibration curve for one simulation
#' using the quantile-based approach
#' 
#' @param i row number of the grid to use for the simulation
#' @param grid grid tibble with the seed number (column `seed`) and the deformations value (either `alpha` or `gamma`)
#' @param n_obs desired number of observation
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
#' @param linspace values at which to compute the mean observed event when computing the WMSE
calibration_curve_quant_simul <- function(i, 
                                          grid, 
                                          n_obs, 
                                          type = c("alpha", "gamma"),
                                          linspace = NULL) {

  if (is.null(linspace)) linspace <- seq(0, 1, length.out = 100)
  
  current_seed <- grid$seed[i]
  if (type == "alpha") {
    transform_scale <- grid$alpha[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = transform_scale, gamma = 1
    )
  } else if (type == "gamma") {
    transform_scale <- grid$gamma[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = 1, gamma = transform_scale
    )
  } else {
    stop("Transform type should be either alpha or gamma.")
  }
  
  # Get the test dataset with true probabilities
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)
  
  summary_bins <- get_summary_bins(
    obs = data_all_test$d,
    scores = data_all_test$p_u, 
    k = 10, threshold = .5)
  summary_bins |> 
    select(score_class, mean_score, mean_obs) |> 
    mutate(
      seed = grid$seed[i],
      scale_parameter = transform_scale,
      type = type
    )
}

1.6.1.1 All the Simulations

Let us retrieve the calibration curve values for each value of \(\alpha\), anb then for \(\gamma\), for all replications.

For \(\alpha\):

tb_calibration_curve_quant_alphas <- map(
  .x = 1:nrow(grid_alpha),
  .f = ~calibration_curve_quant_simul(
    i = .x, 
    grid = grid_alpha, 
    n_obs = n_obs, 
    type = "alpha"
  )
) |> 
  list_rbind()

For \(\gamma\):

tb_calibration_curve_quant_gammas <- map(
  .x = 1:nrow(grid_gamma),
  .f = ~calibration_curve_quant_simul(
    i = .x, 
    grid = grid_gamma, 
    n_obs = n_obs, 
    type = "gamma"
  )
) |> 
  list_rbind()

We bind those two tibbles into one:

tb_calibration_curve_quant <- 
  tb_calibration_curve_quant_alphas |> 
  bind_rows(tb_calibration_curve_quant_gammas)

Next, let us define a plot function to display the calibration curves for each value of one of the scaling parameter (\(\alpha\) or \(\gamma\)), for all the replications.

If we want to add barplots showing the distributions of the true probabilities on top of the graphs, we need to define bins and compute the average number of observation in each bin over the 200 replications.

get_count <- function(seed, alpha = 1, gamma = 1) {
  breaks <- seq(0, 1, by = .05)
  current_data <- get_samples(seed = seed, n_obs = 2000, alpha = alpha, gamma = gamma)
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)
 
  nb_0_bins <- table(cut(data_all_test |> filter(d == 0) |> pull(p_u), breaks = breaks))
  nb_1_bins <- table(cut(data_all_test |> filter(d == 1) |> pull(p_u), breaks = breaks))
  tibble(
    bins = names(nb_0_bins),
    nb_0_bins = as.vector(nb_0_bins),
    nb_1_bins = as.vector(nb_1_bins),
    seed = seed,
    alpha = alpha,
    gamma = gamma
  )
}

We define a grid with all the combinations of values for \(\alpha\) and \(\gamma\) used in the simulations.

grid <- expand_grid(seed = 1:n_repl, alpha = alphas) |> 
  mutate(gamma = 1) |> 
  bind_rows(
    expand_grid(seed = 1:n_repl, gamma = gammas) |> 
      mutate(alpha = 1)
  )

Then, we can apply the get_count() function to get the average number of observation in each bin.

counts_samples <- map(
  .x = 1:nrow(grid), 
  ~get_count(
    seed = grid$seed[.x], 
    alpha = grid$alpha[.x],
    gamma = grid$gamma[.x]
  ),
  .progress = TRUE
) |> 
  list_rbind() |> 
  group_by(bins, alpha, gamma) |> 
  summarise(
    nb_0_bins = mean(nb_0_bins),
    nb_1_bins = mean(nb_1_bins),
  )

Let us save for use in Chapter 2.

save(counts_samples, file = "simul-calib-counts_samples.rda")

For example, in the 200 simulations run with \(\alpha=1\) and \(\gamma=3\):

counts_samples |> 
  filter(alpha == 1, gamma == 3)
# A tibble: 20 × 5
# Groups:   bins, alpha [20]
   bins       alpha gamma nb_0_bins nb_1_bins
   <chr>      <dbl> <dbl>     <dbl>     <dbl>
 1 (0,0.05]       1     3     12.6       3.94
 2 (0.05,0.1]     1     3     22.8       9.54
 3 (0.1,0.15]     1     3     23.4      12.4 
 4 (0.15,0.2]     1     3     24.7      14.2 
 5 (0.2,0.25]     1     3     23.6      15.6 
 6 (0.25,0.3]     1     3     23.4      17.0 
 7 (0.3,0.35]     1     3     21.7      17.7 
 8 (0.35,0.4]     1     3     22.2      19.1 
 9 (0.4,0.45]     1     3     22.2      19.6 
10 (0.45,0.5]     1     3     21.5      20.6 
11 (0.5,0.55]     1     3     20.6      22.0 
12 (0.55,0.6]     1     3     20.5      22.6 
13 (0.6,0.65]     1     3     20.1      24.0 
14 (0.65,0.7]     1     3     19.4      24.9 
15 (0.7,0.75]     1     3     18.8      26.5 
16 (0.75,0.8]     1     3     18.7      28.0 
17 (0.8,0.85]     1     3     17.7      29.6 
18 (0.85,0.9]     1     3     16.4      31.1 
19 (0.9,0.95]     1     3     13.4      31.4 
20 (0.95,1]       1     3      6.20     20.1 

Let us now define the function that will plot the calibration curves.

#' Plot the calibration curve for a set of simulations, using quantile-based
#' bins to obtain the estimations
#'
#' @param tb_calibration_curve
#' @param counts_samples tibble with the average number of observation with
#'   true probability in each bin defined along the [0,1] segment.
#' @param type deformation probability type (either `alpha` or `gamma`)
plot_calibration_quant_simul <- function(tb_calibration_curve,
                                         counts_samples,
                                         type = c("alpha", "gamma")) {
  tb_calibration_curve_curr <-  tb_calibration_curve |> 
    filter(type == !!type)
  
  scale_params <- unique(tb_calibration_curve_curr$scale_parameter)
  seeds <- unique(tb_calibration_curve_curr$seed)
  
  for (scale_param in scale_params) {
    title <- str_c("$\\", type, " = $", round(scale_param, 2))
    
    # Histogram
    ## Calculate the heights for stacking
    if (type == "alpha") {
      counts_samples_current <- 
        counts_samples |> 
        filter(gamma == 1, alpha == !!scale_param)
    } else {
      counts_samples_current <- 
        counts_samples |> 
        filter(alpha == 1, gamma == !!scale_param)
    }
    heights <- rbind(
      counts_samples_current$nb_0_bins, 
      counts_samples_current$nb_1_bins
    )
    col_bars <- c("#CC79A7", "#E69F00")
    par(mar = c(0.5, 4.5, 3.0, 0.5))
    barplot(
      heights, 
      col = col_bars, 
      border = "white", 
      space = 0,
      xlab = "", ylab = "", main = latex2exp::TeX(title),
      axes = FALSE,
    )
    par(mar = c(4.1, 4.5, 0.5, 0.5))
    colour <- ifelse(scale_param == 1, wongOrange, wongBlue)
    plot(
      0:1, 0:1,
      type = "l", col = NULL,
      xlim = 0:1, ylim = 0:1,
      xlab = latex2exp::TeX("$p^u$"), 
      ylab = latex2exp::TeX("$\\hat{E}(D | p^u = p^c)$"),
      main = ""
    )
    for (i_simul in seeds) {
      tb_current <- tb_calibration_curve_curr |> 
        filter(
          scale_parameter == scale_param,
          seed == i_simul
        )
      lines(
        tb_current$mean_score, tb_current$mean_obs,
        lwd = 1, cex = .1, 
        col = adjustcolor(colour, alpha.f = 0.1), t = "b",
      )
      segments(0, 0, 1, 1, col = "black", lty = 2)
    }
  }
}

Lastly, let us visualize the calibration curves, using the plot_calibration_quant_simul() function.

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3))
plot_calibration_quant_simul(
  tb_calibration_curve = tb_calibration_curve_quant, 
  type = "alpha",
  counts_samples = counts_samples
)
Figure 1.10: Calibration curve for the 200 replications for varying values of \(\alpha\)

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3))
plot_calibration_quant_simul(
  tb_calibration_curve = tb_calibration_curve_quant, 
  type = "gamma", counts_samples = counts_samples
)
Figure 1.11: Calibration curve for the 200 replications for varying values of \(\gamma\)

1.6.2 Calibration Curve with Local Regression

We would like to introduce another way of showing the calibration curve, not based on bins. Instead, to have a smoother curve, we will fit a local regression model with degree 0. Using a degree 0 allows us to calculate the local mean of the observed events in the neighborhood of a predicted probability. We will regress the observed events on the predicted probabilities. Then, in a second step, we will use the trained model to make predictions on a line space with values in \([0,1]\).

The {locfit} library is required:

library(locfit)

Let us define a function, calibration_curve_locfit_simul() that will perform the local regression for a single replication of our simulations:

#' Get the calibration curve for one simulation
#' using the local regression-based approach
#' 
#' @param i row number of the grid to use for the simulation
#' @param grid grid tibble with the seed number (column `seed`) and the
#'   deformations value (either `alpha` or `gamma`)
#' @param n_obs desired number of observation
#' @param type deformation probability type (either `alpha` or `gamma`); the 
#' name should match with the `grid` tibble
calibration_curve_locfit_simul <- function(i,
                                           grid,
                                           n_obs,
                                           type = c("alpha", "gamma")) {
  
  
  current_seed <- grid$seed[i]
  
  if (type == "alpha") {
    transform_scale <- grid$alpha[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = transform_scale, gamma = 1
    )
  } else if (type == "gamma") {
    transform_scale <- grid$gamma[i]
    current_data <- get_samples(
      seed = current_seed, n_obs = n_obs, alpha = 1, gamma = transform_scale
    )
  } else {
    stop("Transform type should be either alpha or gamma.")
  }
  
  # Get the test dataset with true probabilities
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)
  
  locfit_0 <- locfit(
    formula = d ~ lp(p_u, nn = 0.15, deg = 0), 
    kern = "rect", maxk = 200, data = data_all_test
  )
  
  scores <- data_all_test$p_u
  # Predictions on [0,1]
  linspace_raw <- seq(0, 1, length.out = 100)
  # Restricting this space to the range of observed scores
  keep_linspace <- which(linspace_raw >= min(scores) & linspace_raw <= max(scores))
  linspace <- linspace_raw[keep_linspace]
  
  score_c_locfit_0 <- predict(locfit_0, newdata = linspace)
  score_c_locfit_0[score_c_locfit_0 > 1] <- 1
  score_c_locfit_0[score_c_locfit_0 < 0] <- 0
  
  tibble(
    seed = grid$seed[i],
    scale_parameter = transform_scale,
    type = type,
    xlim = linspace,
    locfit_pred = score_c_locfit_0
  )
}

The curves for all the values of \(\alpha\):

tb_calibration_curve_locfit_alphas <- map(
  .x = 1:nrow(grid_alpha),
  .f = ~calibration_curve_locfit_simul(
    i = .x, 
    grid = grid_alpha, 
    n_obs = n_obs, 
    type = "alpha"
  )
) |> 
  list_rbind()

And for \(\gamma\):

tb_calibration_curve_locfit_gammas <- map(
  .x = 1:nrow(grid_gamma),
  .f = ~calibration_curve_locfit_simul(
    i = .x, 
    grid = grid_gamma, 
    n_obs = n_obs, 
    type = "gamma"
  )
) |> 
  list_rbind()

We bind those two tibbles into one:

tb_calibration_curve_locfit <- 
  tb_calibration_curve_locfit_alphas |> 
  bind_rows(tb_calibration_curve_locfit_gammas)

Let us define a function to plot the calibration curves for all simulation and values of one of the scaling parameter (\(\alpha\) or \(\gamma\)).

#' Plot the calibration curve for a set of simulations, using local regression
#' to obtain the estimations
#'
#' @param tb_calibration_curve
#' @param type deformation probability type (either `alpha` or `gamma`)
#' @param counts_samples tibble with the average number of observation with
#'   true probability in each bin defined along the [0,1] segment.
plot_calibration_locfit_simuls <- function(tb_calibration_curve,
                                           type = c("alpha", "gamma"),
                                           counts_samples = counts_samples) {
  
  tb_calibration_curve_curr <-  tb_calibration_curve |> 
    filter(type == !!type)
  
  scale_params <- unique(tb_calibration_curve_curr$scale_parameter)
  
  
  df_plot <- tb_calibration_curve_curr |> 
    group_by(type, scale_parameter, xlim) |> 
    summarise(
      mean = mean(locfit_pred),
      lower = quantile(locfit_pred, probs = .025),
      upper = quantile(locfit_pred, probs = .975),
      .groups = "drop"
    )
  
  for (scale_param in scale_params) {
    title <- str_c("$\\", type, " = $", round(scale_param, 2))
    # Histogram
    ## Calculate the heights for stacking
    if (type == "alpha") {
      counts_samples_current <- 
        counts_samples |> 
        filter(gamma == 1, alpha == !!scale_param)
    } else {
      counts_samples_current <- 
        counts_samples |> 
        filter(alpha == 1, gamma == !!scale_param)
    }
    heights <- rbind(
      counts_samples_current$nb_0_bins, 
      counts_samples_current$nb_1_bins
    )
    col_bars <- c("#CC79A7", "#E69F00")
    par(mar = c(0.5, 4.5, 3.0, 0.5))
    barplot(
      heights, 
      col = col_bars, 
      border = "white", 
      space = 0,
      xlab = "", ylab = "", main = latex2exp::TeX(title),
      axes = FALSE,
    )
    
    par(mar = c(4.1, 4.5, 0.5, 0.5))
    df_plot_current <- df_plot |> filter(scale_parameter == scale_param)
    colour <- ifelse(scale_param == 1, wongOrange, wongBlue)
    plot(
      df_plot_current$xlim, df_plot_current$mean,
      type = "l", col = colour,
      xlim = 0:1, ylim = 0:1,
      xlab = latex2exp::TeX("$p^u$"), 
      ylab = latex2exp::TeX("$\\hat{E}(D | p^u = p^c)$"),
      main = ""
    )
    polygon(
      c(df_plot_current$xlim, rev(df_plot_current$xlim)),
      c(df_plot_current$lower, rev(df_plot_current$upper)),
      col = adjustcolor(col = colour, alpha.f = .4),
      border = NA
    )
    segments(0, 0, 1, 1, col = "black", lty = 2)
  }
}

Lastly, we can plot the graphs:

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3, 1,3))
plot_calibration_locfit_simuls(
  tb_calibration_curve = tb_calibration_curve_locfit,
  type = "alpha", counts_samples = counts_samples
)
Figure 1.12: Calibration curve for different values of \(\gamma\), using local regression

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3, 1,3))
plot_calibration_locfit_simuls(
  tb_calibration_curve = tb_calibration_curve_locfit,
  type = "gamma", counts_samples = counts_samples
)
Figure 1.13: Calibration curve for different values of \(\gamma\), using local regression

1.6.2.1 Looking at a Single Replication

We would like to plot the calibration curve for a single replication obtained with local regression, also showing the distribution of the observed events.

We define a function that simulates data, estimates the calibration curve with local regression, and then returns the simulated data and the calibration curve.

#' Simulates data and estimate the calibration curve with local regression
#' 
#' @param seed desired seed
#' @param n_obs number of desired observations
#' @param seed seed to use to generate the data
#' @param type deformation probability type (either `alpha` or `gamma`)
#' @param transform_scale scale value (either `alpha` or `gamma`)
#' @param ci_level level for confidence intervals. 
#'   For 95% confidence interval: .95 (if `NULL`, confidence  intervals are 
#'   not computed)
#' @param nn fraction of nearest neighbors
#' @returns list with two elements:
#'   - `data_all_test`: simulated data
#'   - `tb_calibration_curve_locfit`: calibration curve 
calibration_curve_locfit_param <- function(seed,
                                           n_obs,
                                           transform_scale,
                                           type = c("alpha", "gamma"),
                                           ci_level = NULL,
                                           nn = 10) {
  
  
  if (type == "alpha") {
    current_data <- get_samples(
      seed = seed, n_obs = n_obs, alpha = transform_scale, gamma = 1
    )
  } else if (type == "gamma") {
    current_data <- get_samples(
      seed = seed, n_obs = n_obs, alpha = 1, gamma = transform_scale
    )
  } else {
    stop("Transform type should be either alpha or gamma.")
  }
  
  # Get the test dataset with true probabilities
  data_all_test <- current_data$data_all |>
    slice(-current_data$calib_index)
  
  locfit_0 <- locfit(
    formula = d ~ lp(p_u, nn = 0.15, deg = 0), 
    kern = "rect", maxk = 200, data = data_all_test
  )
  
  scores <- data_all_test$p_u
  
  # Predictions on [0,1]
  linspace_raw <- seq(0, 1, length.out = 100)
  # Restricting this space to the range of observed scores
  keep_linspace <- which(linspace_raw >= min(scores) & linspace_raw <= max(scores))
  linspace <- linspace_raw[keep_linspace]
  
  score_c_locfit_0 <- predict(locfit_0, newdata = linspace)
  score_c_locfit_0[score_c_locfit_0 < 0] <- 0
  score_c_locfit_0[score_c_locfit_0 > 1] <- 1
  
 
  tb_calibration_curve_locfit <- tibble(
    seed = seed,
    scale_parameter = transform_scale,
    type = type,
    xlim = linspace,
    locfit_pred = score_c_locfit_0
  )
  
   # Confidence intervals
  if (!is.null(ci_level)) {
    tb_ci <- map(
      .x = linspace,
      .f = ~local_ci_scores(
        obs = data_all_test$d, 
        scores = data_all_test$p_u,
        tau = .x, 
        nn = nn, 
        prob = ci_level, 
        method = "probit"
      )
    ) |> 
      bind_rows()
    tb_calibration_curve_locfit <- 
      tb_calibration_curve_locfit |> 
      left_join(
        tb_ci, by = "xlim"
      )
  }
  
  list(
    data_all_test = data_all_test,
    tb_calibration_curve_locfit = tb_calibration_curve_locfit
  )
  
}

Let us simulate data for varying values of \(\alpha\) and estimate the calibration curve with local regression of degree 0:

res_calibration_curve_locfit_alphas <- map(
  .x = alphas,
  .f = ~calibration_curve_locfit_param(
    seed = 1, 
    n_obs = n_obs,
    transform_scale = .x,
    type = "alpha",
    ci_level = .95, 
    nn = .15
  )
)

We do the same for varying values of \(\gamma\):

res_calibration_curve_locfit_gammas <- map(
  .x = gammas,
  .f = ~calibration_curve_locfit_param(
    seed = 1, 
    n_obs = n_obs,
    transform_scale = .x,
    type = "gamma",
    ci_level = .95, 
    nn = .15
  )
)

Then, we define a function to plot the calibration curves for varying values of one of the scaling parameter (\(\alpha\) or \(\gamma\)):

#' Plot the calibration curve for a set of simulations, using local regression
#' to obtain the estimations
#'
#' @param tb_calibration_curve calibration data from 
#'   `calibration_curve_locfit_param()`
#' @param type deformation probability type (either `alpha` or `gamma`)
#' @param mfrow mfrow for plot
plot_calibration_locfit_simul <- function(tb_calibration_curve,
                                          type = c("alpha", "gamma"),
                                          mfrow = NULL) {
  if (!is.null(mfrow)) par(mfrow = mfrow)
  scale_params <- map(tb_calibration_curve, "tb_calibration_curve_locfit") |> 
    map_dbl(~unique(.x$scale_parameter))

  for (i in 1:length(tb_calibration_curve)) {
    scale_param <- scale_params[i]
    df_plot_current <- tb_calibration_curve[[i]]$tb_calibration_curve_locfit
    df_data_current <- tb_calibration_curve[[i]]$data_all_test
    title <- str_c("$\\", type, " = $", round(scale_param, 2))
    colour <- ifelse(scale_param == 1, wongOrange, wongBlue)
    
    # Histogram with values
    par(mar = c(0.5, 4.3, 3.0, 0.5))
    # Event == 0
    hist(
      df_data_current |> filter(d == 0) |> pull(p_u),
      breaks = seq(0, 1, by = .05),
      col = adjustcolor("#CC79A7", alpha.f = .4), border = "white",
      axes = FALSE,
      xlab = "", ylab = "",
      main = latex2exp::TeX(title)
    )
    # Event == 1
    hist(
      df_data_current |> filter(d == 1) |> pull(p_u),
      breaks = seq(0, 1, by = .05),
      col = adjustcolor("#E69F00", alpha.f = .4), border = "white",
      add = TRUE,
      axes = FALSE,
      xlab = "", ylab = ""
    )
    par(mar = c(4.1, 4.3, 0.5, 0.5))
    plot(
      df_plot_current$xlim, df_plot_current$locfit_pred,
      type = "l", col = colour,
      xlim = 0:1, ylim = 0:1,
      xlab = "Predicted probability", ylab = "Mean predicted probability",
      main = ""
    )
    polygon(
      c(df_plot_current$xlim, rev(df_plot_current$xlim)),
      c(df_plot_current$lower, rev(df_plot_current$upper)),
      col = adjustcolor(col = colour, alpha.f = .4),
      border = NA
    )
    segments(0, 0, 1, 1, col = "black", lty = 2)
  }
}

We can now plot the curves:

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3))
plot_calibration_locfit_simul(
  tb_calibration_curve = res_calibration_curve_locfit_alphas, 
  type = "alpha"
)
Figure 1.14: Calibration curve for different values of \(\alpha\), using local regression (for a single replication)

mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3))
plot_calibration_locfit_simul(
  tb_calibration_curve = res_calibration_curve_locfit_gammas, 
  type = "gamma"
)
Figure 1.15: Calibration curve for different values of \(\alpha\), , using local regression (for a single replication

1.6.3 Calibration Curve with Moving Average and Confidence Intervals

Instead of running 200 replications, we could also use the computed confidence interval returned by our local_ci_scores() function.

First, we compute the confidence intervals for each value of \(\alpha\).

calib_curve_alpha_ci <- map(
  .x = alphas,
  .f = function(alpha) {
    linspace_raw <- seq(0, 1, length.out = 100)
    scores <- data_alphas[[which(alphas == alpha)]]$p_u
    keep_linspace <- which(
      linspace_raw >= min(scores) & linspace_raw <= max(scores)
    )
    linspace <- linspace_raw[keep_linspace]
    map(
      .x = linspace,
      .f = ~local_ci_scores(
        obs = data_alphas[[which(alphas == alpha)]]$d,
        scores = data_alphas[[which(alphas == alpha)]]$p_u,
        tau = .x, 
        nn = .15, prob = .5, method = "probit")
    ) |> 
      bind_rows() |> 
      mutate(alpha = alpha)
  }
)

Then, we can plot the calibration curve:

Display the R codes used to create the Figure.
mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3, 1,3))

for (i in 1:length(calib_curve_alpha_ci)) {
  calib_curve_alpha_ci_curr <- calib_curve_alpha_ci[[i]]
  alpha <- unique(calib_curve_alpha_ci_curr$alpha)
  title <- str_c("$\\alpha = $", round(alpha, 2))
  
  # Histogram
  ## Calculate the heights for stacking
  counts_samples_current <- 
    counts_samples |> 
    filter(gamma == 1, alpha == !!alpha)
  heights <- rbind(
    counts_samples_current$nb_0_bins, 
    counts_samples_current$nb_1_bins
  )
  col_bars <- c("#CC79A7", "#E69F00")
  par(mar = c(0.5, 4.3, 3.0, 0.5))
  barplot(
    heights, 
    col = col_bars, 
    border = "white", 
    space = 0,
    xlab = "", ylab = "", main = latex2exp::TeX(title),
    axes = FALSE,
  )
  
  par(mar = c(4.1, 4.3, 0.5, 0.5))
  
  plot(
    calib_curve_alpha_ci_curr$xlim, calib_curve_alpha_ci_curr$mean, 
    type = "l", main = "",
    xlim = c(0, 1), ylim = c(0, 1),
    xlab = "Predicted probability", ylab = "Mean predicted probability"
  )
  col_ic <- ifelse(alpha == 1, wongOrange, wongBlue)
  polygon(
    c(calib_curve_alpha_ci_curr$xlim, rev(calib_curve_alpha_ci_curr$xlim)),
    c(calib_curve_alpha_ci_curr$lower, rev(calib_curve_alpha_ci_curr$upper)),
    col = adjustcolor(col = col_ic, alpha.f = .4),
    border = NA
  )
  segments(0, 0, 1, 1, col = "black", lty = 2)
}
Figure 1.16: Calibration curve for different values of \(\alpha\)

First, we compute the confidence intervals for each value of \(\gamma\).

calib_curve_gamma_ci <- map(
  .x = gammas,
  .f = function(gamma) {
    map(
      .x = seq(0, 1, length.out = 100),
      .f = ~local_ci_scores(
        obs = data_gammas[[which(gammas == gamma)]]$d,
        scores = data_gammas[[which(gammas == gamma)]]$p_u,
        tau = .x, 
        nn = .15, prob = .5, method = "probit")
    ) |> 
      bind_rows() |> 
      mutate(gamma = gamma)
  }
)

Then, we can plot the calibration curve:

Display the R codes used to create the Figure.
mat <- matrix(1:10, nrow = 2)
layout(mat, heights = c(1,3, 1,3))

for (i in 1:length(calib_curve_gamma_ci)) {
  calib_curve_gamma_ci_curr <- calib_curve_gamma_ci[[i]]
  gamma <- unique(calib_curve_gamma_ci_curr$gamma)
  title <- str_c("$\\gamma = $", round(gamma, 2))
  
  # Histogram
  ## Calculate the heights for stacking
  counts_samples_current <- 
    counts_samples |> 
    filter(alpha == 1, gamma == !!gamma)
  heights <- rbind(
    counts_samples_current$nb_0_bins, 
    counts_samples_current$nb_1_bins
  )
  col_bars <- c("#CC79A7", "#E69F00")
  par(mar = c(0.5, 4.3, 3.0, 0.5))
  barplot(
    heights, 
    col = col_bars, 
    border = "white", 
    space = 0,
    xlab = "", ylab = "", main = latex2exp::TeX(title),
    axes = FALSE,
  )
  
  par(mar = c(4.1, 4.3, 0.5, 0.5))
  
  plot(
    calib_curve_gamma_ci_curr$xlim, calib_curve_gamma_ci_curr$mean, 
    type = "l", main = "",
    xlim = c(0, 1), ylim = c(0, 1),
    xlab = "Predicted probability", ylab = "Mean predicted probability"
  )
  col_ic <- ifelse(gamma == 1, wongOrange, wongBlue)
  polygon(
    c(calib_curve_gamma_ci_curr$xlim, rev(calib_curve_gamma_ci_curr$xlim)),
    c(calib_curve_gamma_ci_curr$lower, rev(calib_curve_gamma_ci_curr$upper)),
    col = adjustcolor(col = col_ic, alpha.f = .4),
    border = NA
  )
  segments(0, 0, 1, 1, col = "black", lty = 2)
}
Figure 1.17: Calibration curve for different values of \(\gamma\)