🇬🇧 Maintenance 17:00 - 21:00 today. GitLab may be temporarily unavailable. 🇩🇪 Wartungsarbeiten heute 17:00 - 21:00. GitLab ist möglicherweise vorübergehend nicht verfügbar.
@@ -16,7 +16,8 @@ This formulation however, has strong limitations and is prone to over-fitting an
...
@@ -16,7 +16,8 @@ This formulation however, has strong limitations and is prone to over-fitting an
```math
```math
E_{data} = y_{true} - y_p(x, w)
E_{data} = y_{true} - y_p(x, w)
```
```math
E_{total} = E_{data} + λE_{regularization}
E_{total} = E_{data} + λE_{regularization}
```
```
where Lambda defines the relative effect of the regularization term. $`E_{regularization}`$ is typically defined as a function of the weight vector (w) and the variations in this dependency lead to alternative regularization methods. The underlying idea is to enforce the optimizer to decay the weight values towards zero, unless the opposite is enforces by the data. In statistics, this is called [parameter shrinkage method](https://en.wikipedia.org/wiki/Shrinkage_(statistics)).
where Lambda defines the relative effect of the regularization term. $`E_{regularization}`$ is typically defined as a function of the weight vector (w) and the variations in this dependency lead to alternative regularization methods. The underlying idea is to enforce the optimizer to decay the weight values towards zero, unless the opposite is enforces by the data. In statistics, this is called [parameter shrinkage method](https://en.wikipedia.org/wiki/Shrinkage_(statistics)).
...
@@ -33,6 +34,7 @@ For mathematical convenience, let’s go in to the number domain, rather than st
...
@@ -33,6 +34,7 @@ For mathematical convenience, let’s go in to the number domain, rather than st