Supervised here essentially means: "doing things the old fashioned way". Old fashioned in the context of deep learning (DL), of course, so it's still fairly new. Also, "old fashioned" doesn't always mean bad - it's just that later on we'll discuss ways to train networks that clearly outperform approaches using supervised training.
Nonetheless, "supervised training" is a starting point for all projects one would encounter in the context of DL, and hence it is worth studying. Also, while it typically yields inferior results to approaches that more tightly couple with physics, it can be the only choice in certain application scenarios where no good model equations exist.
For supervised training, we're faced with an
unknown function $f^(x)=y^
The
$$ \text{arg min}_{\theta} \sum_i \Big(f(x_i ; \theta)-y^*_i \Big)^2 . $$ (supervised-training)
This will give us
Irrespective of our choice of metric, this formulation gives the actual "learning" process for a supervised approach.
The training data typically needs to be of substantial size, and hence it is attractive
to use numerical simulations solving a physical model
On the other hand, this approach inherits the common challenges of replacing experiments with simulations: first, we need to ensure the chosen model has enough power to predict the behavior of real-world phenomena that we're interested in. In addition, the numerical approximations have numerical errors which need to be kept small enough for a chosen application. As these topics are studied in depth for classical simulations, and the existing knowledge can likewise be leveraged to set up DL training tasks.
---
height: 220px
name: supervised-training
---
A visual overview of supervised training. Quite simple overall, but it's good to keep this
in mind in comparison to the more complex variants we'll encounter later on.
One of the central advantages of the supervised approach above is that
we obtain a surrogate model, i.e., a new function that mimics the behavior of the original
Despite this, it's important to be careful:
NNs can quickly generate huge numbers of in between results. Consider a CNN layer with
Nonetheless, replacing complex and expensive solvers with fast, learned approximations is a very attractive and interesting direction.
Let's finally look at a code example that trains a neural network:
we'll replace a full solver for turbulent flows around airfoils with a surrogate model from {cite}thuerey2020dfp
.