\relax 
\@writefile{toc}{\contentsline {section}{\numberline {1}The perceptron}{1}}
\@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces transforming $y=-1$ datapoints into $y=1$ datapoints}}{1}}
\newlabel{all_positive}{{1}{1}}
\@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces perceptron update: the plane normal $w$ moves in the direction of misclassified $x$ until the $x$ is on the correct side.}}{2}}
\newlabel{perceptron_update}{{2}{2}}
\@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces bias unit}}{3}}
\newlabel{bias}{{3}{3}}
\@writefile{toc}{\contentsline {section}{\numberline {2}Multilayer perceptrons}{3}}
\@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces multilayer perceptron}}{4}}
\newlabel{multilayer_perceptron}{{4}{4}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.1}More than linear functions, example: XOR}{4}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.2} Construction and structure of NNets}{4}}
\@writefile{lof}{\contentsline {figure}{\numberline {5}{\ignorespaces XOR NNet}}{5}}
\newlabel{XOR_nnet}{{5}{5}}
\@writefile{lof}{\contentsline {figure}{\numberline {6}{\ignorespaces NNet fully connected}}{6}}
\newlabel{nnet}{{6}{6}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.3} Kolmogorov theorem, expressive power of NNet}{6}}
\@writefile{toc}{\contentsline {section}{\numberline {3}Training, Error backpropagation}{6}}
\@writefile{toc}{\contentsline {section}{\numberline {4}Toy Example}{8}}
\@writefile{lof}{\contentsline {figure}{\numberline {7}{\ignorespaces NN Example}}{8}}
\newlabel{toy}{{7}{8}}
\bibstyle{abbrv}
\bibdata{../../bibtex/ir,../../bibtex/aslam,../../bibtex/query_local,../../bibtex/learning,../../bibtex/other,../../bibtex/math,bibtex/local,bibtex/1mq}
\@writefile{toc}{\contentsline {subsection}{\numberline {4.1}Feed Forward}{9}}
\@writefile{toc}{\contentsline {subsection}{\numberline {4.2}Back Propagation}{9}}