\relax 
\providecommand\hyper@newdestlabel[2]{}
\providecommand\HyperFirstAtBeginDocument{\AtBeginDocument}
\HyperFirstAtBeginDocument{\ifx\hyper@anchor\@undefined
\global\let\oldcontentsline\contentsline
\gdef\contentsline#1#2#3#4{\oldcontentsline{#1}{#2}{#3}}
\global\let\oldnewlabel\newlabel
\gdef\newlabel#1#2{\newlabelxx{#1}#2}
\gdef\newlabelxx#1#2#3#4#5#6{\oldnewlabel{#1}{{#2}{#3}}}
\AtEndDocument{\ifx\hyper@anchor\@undefined
\let\contentsline\oldcontentsline
\let\newlabel\oldnewlabel
\fi}
\fi}
\global\let\hyper@last\relax 
\gdef\HyperFirstAtBeginDocument#1{#1}
\providecommand*\HyPL@Entry[1]{}
\citation{05-this_notes}
\citation{05-this_slides}
\citation{05-this_slides2}
\citation{05-tj-qk}
\HyPL@Entry{0<</S/D>>}
\@writefile{toc}{\contentsline {section}{\numberline {1}语言模型 \\ Language Models}{1}{section.1}\protected@file@percent }
\@writefile{toc}{\contentsline {subsection}{\numberline {1.1}介绍 \\ Introduction}{1}{subsection.1.1}\protected@file@percent }
\newlabel{gram-raw}{{1.1}{1}{介绍 \\ Introduction}{subsection.1.1}{}}
\@writefile{toc}{\contentsline {subsubsection}{\numberline {1.1.1}N元语法 \\ N-Gram Language Models}{2}{subsubsection.1.1.1}\protected@file@percent }
\newlabel{n-window}{{1}{2}{N元语法 \\ N-Gram Language Models}{equation.1.1}{}}
\@writefile{toc}{\contentsline {subsection}{\numberline {1.2}基于N元语法计算条件共现频率 \\ Calculation Prediction Probability in N-Gram Based Language Models}{2}{subsection.1.2}\protected@file@percent }
\newlabel{n-gram}{{1.2}{2}{基于N元语法计算条件共现频率 \\ Calculation Prediction Probability in N-Gram Based Language Models}{subsection.1.2}{}}
\@writefile{toc}{\contentsline {subsubsection}{\numberline {1.2.1}基于N元语法计算条件共现频率}{2}{subsubsection.1.2.1}\protected@file@percent }
\citation{05-tj-qk}
\newlabel{bi-gram}{{2}{3}{基于N元语法计算条件共现频率}{equation.1.2}{}}
\newlabel{tri-gram}{{3}{3}{基于N元语法计算条件共现频率}{equation.1.3}{}}
\@writefile{toc}{\contentsline {subsubsection}{\numberline {1.2.2}预设参数n对n元语法模型的影响}{3}{subsubsection.1.2.2}\protected@file@percent }
\citation{05-stat-fail}
\citation{05-bengio03}
\citation{05-w2v}
\citation{05-w2v-model}
\citation{05-w2v-opti}
\citation{05-tj}
\@writefile{toc}{\contentsline {subsubsection}{\numberline {1.2.3}N元语法模型的评价}{4}{subsubsection.1.2.3}\protected@file@percent }
\@writefile{toc}{\contentsline {subsection}{\numberline {1.3}基于窗口的语言模型 \\ Window-based Neural Language Model}{4}{subsection.1.3}\protected@file@percent }
\@writefile{toc}{\contentsline {section}{\numberline {2}循环神经网络（RNN）的应用 \\ Application of RNN}{4}{section.2}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces RNN单元的输入输出结构}}{5}{figure.1}\protected@file@percent }
\newlabel{05-RNN}{{1}{5}{RNN单元的输入输出结构}{figure.1}{}}
\@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces RNN可以用于文本生成。}}{5}{figure.2}\protected@file@percent }
\newlabel{05-prediction}{{2}{5}{RNN可以用于文本生成。}{figure.2}{}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.1}RNN语言模型的特点 \\ Advantages, Disadvantages and Applications of RNNs}{6}{subsection.2.1}\protected@file@percent }
\@writefile{toc}{\contentsline {subsection}{\numberline {2.2}梯度爆炸和梯度消失的对策 \\ Solution to the Exploding and Vanishing Gradients}{6}{subsection.2.2}\protected@file@percent }
\newlabel{05-solution-vanish}{{2.2}{6}{梯度爆炸和梯度消失的对策 \\ Solution to the Exploding and Vanishing Gradients}{subsection.2.2}{}}
\@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces 梯度修剪缓解梯度爆炸问题的一个实证。}}{7}{figure.3}\protected@file@percent }
\newlabel{05-wall}{{3}{7}{梯度修剪缓解梯度爆炸问题的一个实证。}{figure.3}{}}
\@writefile{toc}{\contentsline {subsection}{\numberline {2.3}基于RNN的机器翻译 \\ Application: RNN Translation Model}{7}{subsection.2.3}\protected@file@percent }
\bibcite{05-this_notes}{notes05}
\bibcite{05-this_slides}{lecture06}
\bibcite{05-this_slides2}{lecture07}
\bibcite{05-stat-fail}{On Chomsky and the Two Cultures of Statistical Learning}
\bibcite{05-bengio03}{A Neural Probabilistic Language Model}
\bibcite{05-w2v}{Word2Vec in gensim}
\@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces 最简单的基于RNN的机器翻译模型}}{8}{figure.4}\protected@file@percent }
\newlabel{05-rnn-nmt}{{4}{8}{最简单的基于RNN的机器翻译模型}{figure.4}{}}
\@writefile{toc}{\contentsline {section}{\numberline {3}章节附录}{8}{section.3}\protected@file@percent }
\@writefile{toc}{\contentsline {subsection}{\numberline {3.1}维数灾难 \\ Curse of Dimensionality}{8}{subsection.3.1}\protected@file@percent }
\newlabel{curse}{{3.1}{8}{维数灾难 \\ Curse of Dimensionality}{subsection.3.1}{}}
\bibcite{05-w2v-model}{Word2Vec CBOW and Skip-Gram}
\bibcite{05-w2v-opti}{Word2Vec Optimization}
\bibcite{05-tj}{统计自然语言处理}
\bibcite{05-tj-qk}{统计自然语言处理}
\bibcite{05-js}{计算语言学导论}
\ttl@finishall
