Presented By

Qianlin Song, William Loh, Junyue Bai, Phoebe Choi

# Confusing Supervised Learning

Confusing supervised learning (CSL) offers a solution to the issue at hand. A major area of improvement can be seen in the choice of risk measure. In traditional supervised learning, assuming the risk measure is mean squared error (MSE), the expected risk functional is

$$R(g) = \int_x (f(x) - g(x))^2 p(x) \; \mathrm{d}x$$

where $p(x)$ is the prior distribution of the input variable $x$. In practice, model optimizations are performed using the empirical risk

$$R_e(g) = \sum_{i=1}^n (y_i - g(x_i))^2$$

When the problem involves different tasks, the model should optimize for each data point depending on the given task. Let $f_j(x)$ be the true ground-truth function for each task $j$. Therefore, for some input variable $x_i$, an ideal model $g$ would predict $g(x_i) = f_j(x_i)$. With this, the risk functional can be modified to fit this new task for traditional supervised learning methods.

$$R(g) = \int_x \sum_{j=1}^n (f_j(x) - g(x))^2 p(f_j) p(x) \; \mathrm{d}x$$

We call $(f_j(x) - g(x))^2 p(f_j)$ the confusing multiple mappings. Then the optimal solution $g^*(x)$ to the mapping is $\bar{f}(x) = \sum_{j=1}^n p(f_j) f_j(x)$ under this risk functional. However, the optimal solution is not conditional on the specific task at hand but rather on the entire ground-truth functions. Therefore, for every non-trivial set of tasks where $f_k(x) \neq f_\ell(x)$ for some input $x$, $R(g^*) \gt 0$ which implies that there is an unavoidable confusion risk.

To overcome this issue, the authors introduce two types of learning functions:

• Deconfusing function — allocation of which samples come from the same task
• Mapping function — mapping relation from input to output of every learned task