🇬🇧 Maintenance 17:00 - 21:00 today. GitLab may be temporarily unavailable. 🇩🇪 Wartungsarbeiten heute 17:00 - 21:00. GitLab ist möglicherweise vorübergehend nicht verfügbar.
@@ -80,6 +80,12 @@ If you are more curious about its derivation and the underlying mathematics of t
...
@@ -80,6 +80,12 @@ If you are more curious about its derivation and the underlying mathematics of t
* Christopher M. Bishop, Pattern Recognition and Machine Learning, Ch. 7.
* Christopher M. Bishop, Pattern Recognition and Machine Learning, Ch. 7.
* Nello Cristianini and John Shawe-Taylor, [An Introduction to Support Vector Machines and other Kernel-Based Learning Methods](https://www.cambridge.org/core/books/an-introduction-to-support-vector-machines-and-other-kernelbased-learning-methods/A6A6F4084056A4B23F88648DDBFDD6FC)
* Nello Cristianini and John Shawe-Taylor, [An Introduction to Support Vector Machines and other Kernel-Based Learning Methods](https://www.cambridge.org/core/books/an-introduction-to-support-vector-machines-and-other-kernelbased-learning-methods/A6A6F4084056A4B23F88648DDBFDD6FC)
## Bayesian Linear Regression
In the “regression problem” (see lecture notes for “ode to learning” and “regression”), we have discussed that selection of model complexity is needed to be compatible with the data (dimensions, volume) in order to minimize the over-fitting problem. We have also seen that we can force regularization on the lost function to give additional penalty for the over-fitting. Nonetheless, its impact is limited as the nature of the base function (i.e. our scientific hypothesis) is still there, affecting the overall behaviour of the ML model deployed.
As an alternative path, we can follow the probabilistic learning to alleviate the over-fitting in the regression analysis.