C. Donovan
11 April 2018
NB: If it's not in the lecture or lab, it's not in the exam
We'll fit a basic NN to some image data for classification and see how we did
[R: You'll do similar in the lab this week]
Simple in principle:
It's a search over multiple dimensions (dictated by number of parameters/weights).
Nasty ones (like NNs)
Note:
Simple in principle:
This is a gradient search, iterating over multiple dimensions (dictated by number of parameters/weights).
Refer H, T & F sections 11.3 & 11.4. Simplified version follows.
Consider the following simple NN \[ y = \beta_0 + \beta_1z_1 + \beta_2z_2 \] where
\[ \begin{align*} z_1 &= \frac{1}{1+e^{-(\alpha_0 + \alpha_1x_1 + \alpha_2x_2)}}\\ z_2 &= \frac{1}{1+e^{(\alpha_3 + \alpha_4x_1 + \alpha_5x_2)}}\\ \end{align*} \]
We're seeking to optmise the weights (the \( \alpha \) and \( \beta \)).