View on GitHub

SeminarLearningTheory

Website for the Seminar on Learning Theory, taught WS18/19 by Michael Kamp and Pascal Welke

Introduction & a Gentle Start

The main goal of the book is to provide statistical learning theory for supervised, passive batch learning from random training data. For that, the following formal notation is used.

Empirical Risk Minimization

Generalization Bound for ERM in a Restricted Case

For the following result, three assumptions are made:

With this, the following generalization bound can be given:

Corollary:

finite $\mathcal{H},\delta\in (0,1), \epsilon >0, m\geq\frac{\log\frac{|\mathcal{H}|}{\delta}}{\epsilon}$

$\Rightarrow \forall \mathcal{D}$,$f$ realizable, with probability $1-\delta$ for iid sample $S$ of size $m$, $h_S=ERM_\mathcal{H}(S)$ it holds that

$\mathcal{L}_{\mathcal{D},f}(h_S)\leq\epsilon$

Questions:

Q: When calculating the error of an hypothesis why do we compare the error of the obtained hypothesis with $h^{*}$ (in hypothesis space $\mathcal{H}$ ) and not with $f$?

A: In general we compare results with $h^{*}$, because given the hypothesis space $\mathcal{H}$ it is the result closest to the real function one could possibly obtain.

Q: Is it possible to calculate $h^{*}$?

A: In general this is impossible, since we cannot compute $\mathcal{L}_{\mathcal{D},f}$. However, in Chapter 4 we will see that under certain conditions, $h_S\rightarrow h^*$ for $|S|\rightarrow\infty$.

Q: In the sample set $S$ are duplicates of instances allowed? And what is the actual influence of duplicats in sample $S$?

A: Yes, duplicates are allowed because the instances are i.i.d. and therefore it is possible. In practical apoaches duplicates in the sample set are very rare. Nevertheless, the algorithm should be prepared for this case since some methods could break, for example an inversion of a matrix because of singularity.

Q: In the proof of the Corollary where do we use the assumption on realizability?

A: We utilize the assumption when assuming there exists an hypothesis with $\mathcal{L}_{S}(h)=0$.

Q: Why $\varepsilon>0$ and not $\varepsilon \in [0,1]$ since it is a probability?

A: Later on epsilon could be the value of a loss function and considering $\varepsilon>0$ one can apply the corollary.

Q: Would anyone ever calculate m utilizing the corollary?

A: For legal bounds for example in medicine being able to calculate a probability of an error is very important. Of cause the realizability assumption is very strong. Later on we are going to see a similar corollary without the assumption.

Q: Is the upper bound in the corollary a sharp bound?

A: Sometimes the considered bounds are right but generally one can only proof uper and lower bounds.

Q: The goal utilizing ERM is always to minimize the risk oder the input space. Rare cases (with respect to distribution $\mathcal{D}$) are not considered because they have a very low probabilty but especially those rare cases (for example the corner case in autonomous driving) are the ones that are most interesting because of their strong impacts. How can the algorithms be good on those rare cases?

A: ERM sadly doesn’t work very well for those cases. But there are practical solutions for example

Q: Having a machine that is very conservative in risk taking and tends to overestimate errors could be problematic. So a quite philosophical question is what is better: overestimate or underestimate the error?

A: From a learning perspective, it is better to overestimate the error. For example, if the error measure (i.e., loss function) is non-continuous, non-convex, and/or non-differentiable, a standard way of dealing with it is by using a surrogate convex upper bound (e.g., the absolute error is a convex upper bound to the 0-1-loss).

In practice, a hypothesis is not necessarily evaluated with the same error measure it has been trained. E.g., a classification model can be trained using the hinge-loss, but the application requires a high precision with little importance to the recall. Then the model can be evaluated using these measures. It is an interesting practical problem, how to tweak the training process so that the resulting model is good given the application driven metric, especially in cases where the model cannot be directly optimized using this metric (because the metric is non-continuous, non-convex, because optimization with that metric has other undesireable side-effects, etc.).

Q: What is overfitting?

A: the Oxford dictionary defines overfitting as: “The production of an analysis which corresponds too closely or exactly to a particular set of data, and may therefore fail to fit additional data or predict future observations reliably.”

In ML, this refers to a model that has a low training error (empirical risk) $\mathcal{L}_{S}$ but a high risk $\mathcal{L}_{\mathcal{H},f}$.