\relax 
\@writefile{toc}{\contentsline {paragraph}{Required Library}{1}}
\@writefile{toc}{\contentsline {paragraph}{Required R Code}{1}}
\@writefile{toc}{\contentsline {section}{\numberline {1}Classification and Regression}{1}}
\newlabel{section classification and regression}{{1}{1}}
\@writefile{toc}{\contentsline {paragraph}{Classification}{1}}
\@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces Example of linear separability.}}{1}}
\newlabel{fig linear sep}{{1}{1}}
\@writefile{toc}{\contentsline {paragraph}{Regression}{2}}
\@writefile{toc}{\contentsline {section}{\numberline {2}Perceptron}{2}}
\newlabel{section perceptron}{{2}{2}}
\@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces Perceptron (from \url {http://en.wikipedia.org/wiki/File:Perceptron.svg}).}}{2}}
\newlabel{fig perceptron}{{2}{2}}
\@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces Threshold binary activation function.}}{2}}
\newlabel{fig threshold bin}{{3}{2}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.1}Prediction}{2}}
\citation{novikoff1962convergence}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.2}Training}{3}}
\newlabel{equation update w}{{2}{3}}
\@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces Example of weight update. The chosen example is enclosed in the green box. The update pulls the plane to a direction that making prediction on the example is `less' incorrect.}}{3}}
\newlabel{fig update w}{{4}{3}}
\citation{hornik1989multilayer}
\@writefile{toc}{\contentsline {section}{\numberline {3}Multi-layer Perceptron}{5}}
\newlabel{section mlp}{{3}{5}}
\@writefile{lof}{\contentsline {figure}{\numberline {5}{\ignorespaces A 3-layer perceptron with 2 hidden layers (from \url {http://www.statistics4u.com/fundstat_eng/cc_ann_bp_function.html}.)}}{5}}
\newlabel{fig mlp}{{5}{5}}
\@writefile{lof}{\contentsline {figure}{\numberline {6}{\ignorespaces Sigmoid activation function.}}{6}}
\newlabel{fig sigmoid}{{6}{6}}
\newlabel{equation yli}{{3}{6}}
\@writefile{toc}{\contentsline {subsection}{\numberline {3.1}Prediction}{6}}
\@writefile{lof}{\contentsline {figure}{\numberline {7}{\ignorespaces Feedforward (left) and backpropagation (right).}}{6}}
\newlabel{fig ff bp}{{7}{6}}
\@writefile{toc}{\contentsline {subsection}{\numberline {3.2}Training}{6}}
\@writefile{lof}{\contentsline {figure}{\numberline {8}{\ignorespaces Illustration for the gradient descent method. The blue line is the tangent at the current value of the parameter $w$. If we update $w$ by subtracting an amount proportional to the gradient at that point, the value of $E$ will be pushed along the arrow and hence decrease. However, this method only guarantees to converge to a local minimum.}}{7}}
\newlabel{fig gradient descent}{{8}{7}}
\newlabel{equation dE/dzLi}{{4}{7}}
\newlabel{equation dE/dzli}{{5}{7}}
\newlabel{equation dE/dw}{{6}{7}}
\bibstyle{apalike}
\bibdata{ref}
\bibcite{hornik1989multilayer}{{1}{1989}{{Hornik et~al.}}{{}}}
\bibcite{novikoff1962convergence}{{2}{1962}{{Novikoff}}{{}}}
\@writefile{lof}{\contentsline {figure}{\numberline {9}{\ignorespaces MLP for Exercise 3\hbox {}.2. All the biases are 0.5. All the neurons have the sigmoid activation function.}}{8}}
\newlabel{fig mlp exercise}{{9}{8}}
\@writefile{toc}{\contentsline {section}{\numberline {4}Submission}{8}}
