\relax \@writefile{toc}{\contentsline {section}{\numberline {1}The perceptron}{1}} \@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces transforming $y=-1$ datapoints into $y=1$ datapoints}}{1}} \newlabel{all_positive}{{1}{1}} \@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces perceptron update: the plane normal $w$ moves in the direction of misclassified $x$ until the $x$ is on the correct side.}}{2}} \newlabel{perceptron_update}{{2}{2}} \@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces bias unit}}{3}} \newlabel{bias}{{3}{3}} \@writefile{toc}{\contentsline {section}{\numberline {2}Multilayer perceptrons}{3}} \@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces multilayer perceptron}}{4}} \newlabel{multilayer_perceptron}{{4}{4}} \@writefile{toc}{\contentsline {subsection}{\numberline {2.1}More than linear functions, example: XOR}{4}} \@writefile{toc}{\contentsline {subsection}{\numberline {2.2} Construction and structure of NNets}{4}} \@writefile{lof}{\contentsline {figure}{\numberline {5}{\ignorespaces XOR NNet}}{5}} \newlabel{XOR_nnet}{{5}{5}} \@writefile{lof}{\contentsline {figure}{\numberline {6}{\ignorespaces NNet fully connected}}{6}} \newlabel{nnet}{{6}{6}} \@writefile{toc}{\contentsline {subsection}{\numberline {2.3} Kolmogorov theorem, expressive power of NNet}{6}} \@writefile{toc}{\contentsline {section}{\numberline {3}Training, Error backpropagation}{6}} \@writefile{toc}{\contentsline {section}{\numberline {4}Toy Example}{8}} \@writefile{lof}{\contentsline {figure}{\numberline {7}{\ignorespaces NN Example}}{8}} \newlabel{toy}{{7}{8}} \bibstyle{abbrv} \bibdata{../../bibtex/ir,../../bibtex/aslam,../../bibtex/query_local,../../bibtex/learning,../../bibtex/other,../../bibtex/math,bibtex/local,bibtex/1mq} \@writefile{toc}{\contentsline {subsection}{\numberline {4.1}Feed Forward}{9}} \@writefile{toc}{\contentsline {subsection}{\numberline {4.2}Back Propagation}{9}}