# Influenza Forecasting Framework based on Gaussian Processes

## Contents

## Abstract

This paper presents a novel framework for seasonal epidemic forecasting using Gaussian process regression. Resulting retrospective forecasts, trained on a subset of the publicly available CDC influenza-like-illness (ILI) data-set, outperformed four state-of-the-art models when compared using the official CDC scoring rule (log-score).

## Background

Each year, the seasonal influenza epidemic affects public health at a massive scale, resulting in 38 million cases, 400 000 hospitalizations, and 22 000 deaths in the United States in 2019/20 alone ([1]). Given this, reliable forecasts of future influenza development are invaluable, because they allow for improved public health policies and informed resource development and allocation.

## Related Work

Given the value of epidemic forecasts, the CDC regularly publishes ILI data and has funded a seasonal ILI forecasting challenge. This challenge has lead to four state of the art models in the field; MSS, a physical susceptible-infected-recovered model with assumed linear noise [4]; SARIMA, a framework based on seasonal auto-regressive moving average models [2]; and LinEns, an ensemble of three linear regression models.

## Motivation

It has been shown that LinEns forecasts outperform the other frameworks on the ILI data-set. However, this framework assumes a deterministic relationship between the epidemic week and its case count, which does not reflect the stochastic nature of the trend. Therefore, it is natural to ask whether a similar framework that assumes a stochastic relationship between these variables would provide better performance. This motivated the development of the proposed Gaussian process regression framework, and the subsequent performance comparison to the benchmark models.

## Gaussian Process Regression

Consider the following set up: let [math]X = [\mathbf{x}_1,\ldots,\mathbf{x}_n][/math] [math](d\times n)[/math] be your training data, [math]\mathbf{y} = [y_1,y_2,\ldots,y_n]^T[/math] be your noisy observations where [math]y_i = f(\mathbf{x}_i) + \epsilon_i[/math], [math](\epsilon_i:i = 1,\ldots,n)[/math] i.i.d. [math]\sim \mathcal{N}(0,{\sigma}^2)[/math], and [math]f[/math] is the trend we are trying to model (by [math]\hat{f}[/math]). Let [math]\mathbf{x}^*[/math] [math](d\times 1)[/math] be your test data point, and [math]\hat{y} = \hat{f}(\mathbf{x}^*)[/math] be your predicted outcome.

Instead of assuming a deterministic form of [math]f[/math], and thus of [math]\mathbf{y}[/math] and [math]\hat{y}[/math] (as classical linear regression would, for example), Gaussian process regression assumes [math]f[/math] is stochastic. More precisely, [math]\mathbf{y}[/math] and [math]\hat{y}[/math] are assumed to have a joint prior distribution. Indeed, we have

$$ (\mathbf{y},\hat{y}) \sim \mathcal{N}(0,\Sigma(X,\mathbf{x}^*)) $$

where [math]\Sigma(X,\mathbf{x}^*)[/math] is a matrix of covariances dependent on some kernel function [math]k[/math]. In this paper, the kernel function is assumed to be Gaussian and takes the form

$$ k(\mathbf{x}_i,\mathbf{x}_j) = \sigma^2\exp(-\frac{1}{2}(\mathbf{x}_i-\mathbf{x}^j)^T\Sigma(\mathbf{x}_i-\mathbf{x}_j)). $$

It is important to note that this gives a joint prior distribution of **functions** (**Fig. 1** left, grey curves).

By restricting this distribution to contain only those functions (**Fig. 1** right, grey curves) that agree with the observed data points [math]\mathbf{x}[/math] (**Fig. 1** right, solid black) we obtain the posterior distribution for [math]\hat{y}[/math] which has the form

$$ p(\hat{y} | \mathbf{x}^*, X, \mathbf{y}) \sim \mathcal{N}(\mu(\mathbf{x}^*,X,\mathbf{y}),\sigma(\mathbf{x}^*,X)) $$

**Figure 1. Gaussian process regression**: Select the functions from your joint prior distribution (left, grey curves) with mean [math]0[/math] (left, bold line) that agree with the observed data points (right, black bullets). These form your posterior distribution (right, grey curves) with mean [math]\mu(\mathbf{x})[/math] (right, bold line). Red triangle helps compare the two images (location marker) [3].

## Data-set Description

Let [math]d_j^i[/math] denote the number of epidemic cases recorded in week [math]j[/math] of season [math]i[/math], and let [math]j^*[/math] and [math]i^*[/math] denote the current week and season, respectively. The ILI data-set contains $d_j^i$ for all previous weeks and seasons, up to the current season with a 1-3 week publishing delay. Note that a season refers to the time of year when the epidemic is prevalent (e.g. an influenza season lasts 30 weeks and contains the last 10 weeks of year k, and the first 20 weeks of year k+1). The goal is to predict [math]\hat{y}_T = \hat{f}_T(x^*) = d^{i^*}_{j* + T}[/math] where [math]T, \;(T = 1,\ldots,K)[/math] is the target week (how many weeks into the future that you want to predict).

To do this, a design matrix [math]X[/math] is constructed where each element [math]X_{ji} = d_j^i[/math] corresponds to the number of cases in week (row) j of season (column) i. The training outcomes [math]y_{i,T}, i = 1,\ldots,n[/math] correspond to the number of cases that were observed in target week [math]T,\; (T = 1,\ldots,K)[/math] of season [math]i, (i = 1,\ldots,n)[/math].

## Proposed Framework

To compute [math]\hat{y}[/math], the following algorithm is executed.

- Let [math]J \subseteq \{j^*-4 \leq j \leq j^*\}[/math] (subset of possible weeks).
- Assemble the Training Set [math]\{X_J, \mathbf{y}_{T,J}\}[/math]
- Train the Gaussian process
- Calculate the
**distribution**of [math]\hat{y}_{T,J}[/math] using [math]p(\hat{y}_{T,J} | \mathbf{x}^*, X_J, \mathbf{y}_{T,J}) \sim \mathcal{N}(\mu(\mathbf{x}^*,X,\mathbf{y}_{T,J}),\sigma(\mathbf{x}^*,X_J))[/math] - Set [math]\hat{y}_{T,J} =\mu(x^*,X_J,\mathbf{y}_{T,J})[/math]
- Repeat steps 2-5 for all sets of weeks [math]J[/math]
- Determine the best 3 performing sets J (on the 2010/11 and 2011/12 validation sets)
- Calculate the ensemble forecast by averaging the 3 best performing predictive distribution densities i.e. [math]\hat{y}_T = \frac{1}{3}\sum_{k=1}^3 \hat{y}_{T,J_{best}}[/math]

## Results

To demonstrate the accuracy of their results, retrospective forecasting was done on the ILI data-set. In other words, the Gaussian process model was trained assuming a previous season (2012/13) was the current season. In this fashion, the forecast could be compared to the already observed true outcome.

To produce a forecast for the entire 2012/13 season, 30 Gaussian processes were trained (each influenza season has 30 test points [math]\mathbf{x^*}[/math]) and a curve connecting the predicted outputs [math]y_T = \hat{f}(\mathbf{x^*)}[/math] was plotted (**Fig.2**, blue line). As shown in **Fig.2**, this forecast (blue line) was reliable for both 1 (left) and 3 (right) week targets, given that the 95% prediction interval (**Fig.2**, purple shaded) contained the true values (**Fig.2**, red x's) 95% of the time.

**Figure 2. Retrospective forecasts and their uncertainty**: One week retrospective influenza forecasting for two targets (T = 1, 3). Red x’s are the true observed values, and blue lines and purple shaded areas represent point forecasts and 95% prediction intervals, respectively.

Moreover, as shown in **Fig.3**, the novel Gaussian process regression framework outperformed all benchmarks for four different targets [math](T = 1,\ldots, 4)[/math] , including LinEns, when compared using the official CDC scoring criterion *log-score*. Log-score describes the logarithmic probability of the forecast being within in an interval around the true value.

**Figure 3. Average log-score gain of proposed framework**: Each bar shows the mean seasonal log-score gain of the proposed framework vs. the given state-of-the-art model, and each panel corresponds to a different target week [math] T = 1,...4 [/math].

## Conclusion

This paper presented a novel framework for forecasting seasonal epidemics using Gaussian process regression that outperformed multiple state-of-the-art forecasting methods on the CDC's ILI data-set. Hence, this work may play a key role in future influenza forecasting and as result, the improvement of public health policies and resource allocation.

## Critique

The proposed framework provides a computationally efficient method to forecast any seasonal epidemic count data that is easily extendable to multiple target types. In particular; one can compute key parameters such as the peak infection incidence ([math]\hat{y} = max_{0 \leq j \leq 52} d^i_j [/math]), the timing of the peak infection incidence ([math]\hat{y} = argmax_{0 \leq j \leq 52} d^i_j[/math]) and the final epidemic size of a season ([math]\hat{y} = \sum_{j=1}^{52} d^i_j[/math]). However, given it is not a physical model, it cannot provide insights on parameters describing the disease spread. Moreover, the framework requires training data and hence, is not applicable for non-seasonal epidemics.

## References

[1] CDC website https://www.cdc.gov/flu/about/burden/2019-2020.html#flu-burden-infographic

[2] Ray, E. L., Sakrejda, K., Lauer, S. A., Johansson, M. A.,and Reich, N. G. (2017).Infectious disease prediction with kernel conditional densityestimation.Statistics in Medicine, 36(30):4908–4929.

[3] Schulz, E., Speekenbrink, M., and Krause, A. (2017).A tutorial on gaussian process regression with a focus onexploration-exploitation scenarios.bioRxiv.

[4] Zimmer, C., Leuba, S. I., Cohen, T., and Yaesoubi, R.(2019).Accurate quantification of uncertainty in epidemicparameter estimates and predictions using stochasticcompartmental models.Statistical Methods in Medical Research,28(12):3591–3608.PMID: 30428780.