author: niplav, created: 2021-04-19, modified: 2022-02-02, language: english, status: on hold, importance: 2, confidence: likely
Solutions to the textbook “Pattern Recognition and Machine Learning” by Christopher M. Bishop.
(*) Consider the sum-of-squares error function given by (1.2) in which the function
$y(x, \textbf{w})$
is given by the polynomial (1.1). Show that the coefficients$\textbf{w}=\{w_i\}$
that minimizes this error function are given by the solution to the following set of linear equations
where
Here a suffix
$i$
or$j$
denotes the index of a component, whereas$(x)^i$
denotes$x$
raised to the power of$i$
.
Recap: formula 1.1 is
and formula 1.2 (the error function) is
Substituting 1.1 into 1.2 gives
Differentiating after $\textbf{w}$
then returns
I really should learn multivariable calculus.
(*) Using the definition (1.38) show that
$\text{var}[f(x)]$
satisfies (1.39).