Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
T
Toy-MT-Introduction
概览
Overview
Details
Activity
Cycle Analytics
版本库
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
问题
0
Issues
0
列表
Board
标记
里程碑
合并请求
0
Merge Requests
0
CI / CD
CI / CD
流水线
作业
日程表
图表
维基
Wiki
代码片段
Snippets
成员
Collapse sidebar
Close sidebar
活动
图像
聊天
创建新问题
作业
提交
Issue Boards
Open sidebar
NiuTrans
Toy-MT-Introduction
Commits
207e0648
Commit
207e0648
authored
May 04, 2020
by
xiaotong
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
updates of the content figure
parent
3d106c96
隐藏空白字符变更
内嵌
并排
正在显示
5 个修改的文件
包含
116 行增加
和
797 行删除
+116
-797
Book/ChapterPreface/ChapterPreface.tex
+6
-5
Book/ChapterPreface/Figures/figure-preface.tex
+41
-0
Book/mt-book-xelatex.idx
+0
-149
Book/mt-book-xelatex.ptc
+68
-642
Book/mt-book-xelatex.tex
+1
-1
没有找到文件。
Book/ChapterPreface/ChapterPreface.tex
查看文件 @
207e0648
...
@@ -14,11 +14,11 @@
...
@@ -14,11 +14,11 @@
\end{center}
\end{center}
\vspace
{
2em
}
\vspace
{
2em
}
\begin{spacing}
{
1.
4
}
\begin{spacing}
{
1.
2
}
让计算机进行自然语言的翻译是人类长期的梦想,也是人工智能的终极目标之一。自上世纪九十年代起,机器翻译也迈入了基于统计建模的时代,而发展到今天,深度学习等机器学习方法已经在机器翻译中得到了大量的应用,取得了令人瞩目的进步。
让计算机进行自然语言的翻译是人类长期的梦想,也是人工智能的终极目标之一。自上世纪九十年代起,机器翻译也迈入了基于统计建模的时代,而发展到今天,深度学习等机器学习方法已经在机器翻译中得到了大量的应用,取得了令人瞩目的进步。
在这个时代背景下,对机器翻译的模型、方法、实现技术进行深入了解是自然语言处理领域研究者和实践者所渴望的。本书全面的回顾了近三十年内机器翻译的技术发展历程,并围绕统计建模和深度学习两个主题对机器翻译的技术方法进行了全面介绍。在写作中,笔者力求用朴实的语言和
图画
阐述机器翻译的基本思想,同时对相关的技术前沿进行讨论。本书可以供计算机相关专业高年级本科生及研究生学习之用,也可以作为自然语言处理,特别是机器翻译相关研究人员的参考资料。
在这个时代背景下,对机器翻译的模型、方法、实现技术进行深入了解是自然语言处理领域研究者和实践者所渴望的。本书全面的回顾了近三十年内机器翻译的技术发展历程,并围绕统计建模和深度学习两个主题对机器翻译的技术方法进行了全面介绍。在写作中,笔者力求用朴实的语言和
实例
阐述机器翻译的基本思想,同时对相关的技术前沿进行讨论。本书可以供计算机相关专业高年级本科生及研究生学习之用,也可以作为自然语言处理,特别是机器翻译相关研究人员的参考资料。
本书共分为七个章节。章节的顺序参考了机器翻译技术发展的时间脉络,同时兼顾形成机器翻译知识体系的内在逻辑。各章节的主要内容包括:
本书共分为七个章节。章节的顺序参考了机器翻译技术发展的时间脉络,同时兼顾形成机器翻译知识体系的内在逻辑。各章节的主要内容包括:
...
@@ -40,13 +40,14 @@
...
@@ -40,13 +40,14 @@
\vspace
{
0.4em
}
\vspace
{
0.4em
}
\end{itemize}
\end{itemize}
其中,第一章是对机器翻译的整体介绍。第二章和第五章是对统计建模和深度学习方法的介绍,分别建立了两个机器翻译范式的基础知识体系
\ \dash
\
统计机器翻译和神经机器翻译。统计机器翻译部分(第三、四章)涉及早期的基于单词的翻译模型,以及本世纪初流行的基于短语和句法的翻译模型;神经机器翻译(第六、七章)代表了
机器翻译的技术前沿,内容主要涉及了基于端到端表示学习的机器翻译建模方法。特别在第七章,本书对构建先进神经机器翻译的技术进行了较为全面的介绍
,为相关相关科学问题的研究和实用系统的开发提供了可落地的思路。图
\ref
{
fig:preface
}
展示了本书各个章节及核心概念之间的关系。
其中,第一章是对机器翻译的整体介绍。第二章和第五章是对统计建模和深度学习方法的介绍,分别建立了两个机器翻译范式的基础知识体系
\ \dash
\
统计机器翻译和神经机器翻译。统计机器翻译部分(第三、四章)涉及早期的基于单词的翻译模型,以及本世纪初流行的基于短语和句法的翻译模型;神经机器翻译(第六、七章)代表了
当今机器翻译的前沿,内容主要涉及了基于端到端表示学习的机器翻译建模方法。特别是,第七章对一些先进的神经机器翻译方法进行了讨论
,为相关相关科学问题的研究和实用系统的开发提供了可落地的思路。图
\ref
{
fig:preface
}
展示了本书各个章节及核心概念之间的关系。
{
\red
用最简单的方式阐述机器翻译的基本思想
}
是笔者所期望达到的
。但是,书中难免会涉及一些概念形式化定义和算法的抽象描述,笔者也都尽所能通过图形进行解释
。不过,本书所包含的内容较为广泛,难免有疏漏之处,望读者海涵,并指出不当之处。
{
\red
用最简单的方式阐述机器翻译的基本思想
}
是笔者所期望达到的
目标。但是,书中难免会涉及一些概念形式化定义和算法的抽象描述,笔者也都尽所能通过图例进行解释(本书共320张插图)
。不过,本书所包含的内容较为广泛,难免有疏漏之处,望读者海涵,并指出不当之处。
\begin{figure}
[htp]
\begin{figure}
[htp]
\centering
\centering
%\includegraphics[scale=0.36]{./Chapter6/Figures/mt-history.png}
\centering
\input
{
./ChapterPreface/Figures/figure-preface
}
\caption
{
本书各章节及核心概念关系图
}
\caption
{
本书各章节及核心概念关系图
}
\label
{
fig:preface
}
\label
{
fig:preface
}
\end{figure}
\end{figure}
...
...
Book/ChapterPreface/Figures/figure-preface.tex
0 → 100644
查看文件 @
207e0648
\begin{tikzpicture}
\tikzstyle
{
secnode
}
=[font=
\scriptsize
,minimum height=4.0em,minimum width=22em,draw,thick,fill=white,drop shadow]
\tikzstyle
{
conceptnode
}
=[font=
\scriptsize
,minimum height=1.5em,minimum width=5em]
\tikzstyle
{
conceptnodesmall
}
=[font=
\scriptsize
,minimum height=1.0em,minimum width=4.4em]
% section 1
\node
[secnode,anchor=south west,minimum width=22.5em,red,fill=white] (sec1) at (0,0)
{}
;
\node
[anchor=north] (sec1label) at ([yshift=-0.2em]sec1.north)
{
\small
{
机器翻译简介
}}
;
\node
[conceptnode,anchor=south west,fill=red!15,thin] (sec1box1) at ([xshift=0.5em,yshift=0.5em]sec1.south west)
{
\footnotesize
{
发展历史
}}
;
\node
[conceptnode,anchor=west,fill=red!15,thin] (sec1box2) at ([xshift=0.5em]sec1box1.east)
{
\footnotesize
{
评价方法
}}
;
\node
[conceptnode,anchor=west,fill=red!15,thin] (sec1box3) at ([xshift=0.5em]sec1box2.east)
{
\footnotesize
{
应用情况
}}
;
\node
[conceptnode,anchor=west,fill=red!15,thin] (sec1box4) at ([xshift=0.5em]sec1box3.east)
{
\footnotesize
{
系统
\&
数据
}}
;
\node
[anchor=north west,draw,thick,fill=white,red,rounded corners] (sec1title) at ([xshift=-0.3em,yshift=0.3em]sec1.north west)
{{
\footnotesize\bfnew
{
第一章
}}}
;
% section 2
\node
[secnode,anchor=south,blue,fill=white] (sec2) at ([xshift=-6.5em,yshift=3em]sec1.north)
{}
;
\node
[anchor=north] (sec2label) at (sec2.north)
{
\small
{
词法、语法及统计建模基础
}}
;
\node
[conceptnode,anchor=south west,fill=blue!15,thin,minimum width=4em,align=left] (sec2box1) at ([xshift=0.5em,yshift=0.4em]sec2.south west)
{
\tiny
{
概率论与
}
\\\tiny
{
统计建模基础
}}
;
\node
[anchor=west,draw,dotted,thick,minimum height=2em,minimum width=16.6em,align=left] (sec2box2) at ([xshift=0.3em]sec2box1.east)
{}
;
\node
[conceptnodesmall,minimum width=5em,anchor=south west,fill=blue!15,thin] (sec2box3) at ([xshift=0.4em,yshift=0.3em]sec2box2.south west)
{
\scriptsize
{
中文分词
}}
;
\node
[conceptnodesmall,minimum width=5em,anchor=west,fill=blue!15,thin] (sec2box4) at ([xshift=0.4em]sec2box3.east)
{
\scriptsize
{$
n
$
元语法模型
}}
;
\node
[conceptnodesmall,minimum width=5em,anchor=west,fill=blue!15,thin] (sec2box5) at ([xshift=0.4em]sec2box4.east)
{
\scriptsize
{
句法分析
}}
;
\node
[anchor=north west,draw,thick,fill=white,blue,rounded corners] (sec2title) at ([xshift=-0.3em,yshift=0.3em]sec2.north west)
{{
\footnotesize\bfnew
{
第二章
}}}
;
\draw
[->,very thick] ([xshift=2em,yshift=0.2em]sec1.north) .. controls +(north:2.5em) and +(south:2.5em) .. ([xshift=-0em,yshift=-0.2em]sec2.south);
% section 5
\node
[secnode,anchor=south,orange,fill=white] (sec5) at ([xshift=7em,yshift=10em]sec1.north)
{}
;
\node
[anchor=north] (sec5label) at (sec5.north)
{
\small
{
人工神经网络和神经语言建模
}}
;
\node
[conceptnode,minimum width=4em,anchor=south west,fill=orange!15,thin,minimum width=4em,align=left] (sec5box1) at ([xshift=0.5em,yshift=0.4em]sec5.south west)
{
\tiny
{
线性代数基础
}
\\\tiny
{
与感知机
}}
;
\node
[conceptnode,minimum width=4em,anchor=west,fill=orange!15,thin,minimum width=4em,align=left] (sec5box2) at ([xshift=0.2em]sec5box1.east)
{
\tiny
{
多层神经网络
}
\\\tiny
{
与实现方法
}}
;
\node
[conceptnode,minimum width=4em,anchor=west,fill=orange!15,thin,minimum width=4em,align=left] (sec5box3) at ([xshift=0.2em]sec5box2.east)
{
\tiny
{
模型训练
}
\\\tiny
{
(反向传播)
}}
;
\node
[conceptnode,minimum width=4em,anchor=west,fill=ublue!15,thin,minimum width=4em,align=left] (sec5box4) at ([xshift=0.2em]sec5box3.east)
{
\tiny
{
神经语言模型
}
\\\tiny
{
(FNN等)
}}
;
\node
[conceptnode,minimum width=4em,anchor=west,fill=ublue!15,thin,minimum width=4em,align=left] (sec5box5) at ([xshift=0.2em]sec5box4.east)
{
\tiny
{
表示学习与
}
\\\tiny
{
预训练模型
}}
;
\node
[anchor=north west,draw,thick,fill=white,orange,rounded corners] (sec5title) at ([xshift=-0.3em,yshift=0.3em]sec5.north west)
{{
\footnotesize\bfnew
{
第五章
}}}
;
\draw
[->,very thick] ([yshift=-9.8em]sec5.south) -- ([yshift=-0.2em]sec5.south);
\end{tikzpicture}
\ No newline at end of file
Book/mt-book-xelatex.idx
查看文件 @
207e0648
\indexentry{未登录词|hyperpage}{13}
\indexentry{Out of Vocabulary Word,OOV Word|hyperpage}{13}
\indexentry{子词切分|hyperpage}{13}
\indexentry{Sub-word Segmentation|hyperpage}{13}
\indexentry{标准化|hyperpage}{13}
\indexentry{Normalization|hyperpage}{13}
\indexentry{数据清洗|hyperpage}{13}
\indexentry{Dada Cleaning|hyperpage}{13}
\indexentry{数据选择|hyperpage}{15}
\indexentry{Data Selection|hyperpage}{15}
\indexentry{数据过滤|hyperpage}{15}
\indexentry{Data Filtering|hyperpage}{15}
\indexentry{开放词表|hyperpage}{18}
\indexentry{Open-Vocabulary|hyperpage}{18}
\indexentry{子词|hyperpage}{19}
\indexentry{Sub-word|hyperpage}{19}
\indexentry{字节对编码|hyperpage}{19}
\indexentry{双字节编码|hyperpage}{19}
\indexentry{Byte Pair Encoding,BPE|hyperpage}{19}
\indexentry{正则化|hyperpage}{22}
\indexentry{Regularization|hyperpage}{22}
\indexentry{过拟合问题|hyperpage}{22}
\indexentry{Overfitting Problem|hyperpage}{22}
\indexentry{反问题|hyperpage}{22}
\indexentry{Inverse Problem|hyperpage}{22}
\indexentry{适定的|hyperpage}{22}
\indexentry{Well-posed|hyperpage}{22}
\indexentry{不适定问题|hyperpage}{22}
\indexentry{Ill-posed Problem|hyperpage}{22}
\indexentry{降噪|hyperpage}{23}
\indexentry{Denoising|hyperpage}{23}
\indexentry{泛化|hyperpage}{23}
\indexentry{Generalization|hyperpage}{23}
\indexentry{标签平滑|hyperpage}{25}
\indexentry{Label Smoothing|hyperpage}{25}
\indexentry{相互适应|hyperpage}{26}
\indexentry{Co-Adaptation|hyperpage}{26}
\indexentry{集成学习|hyperpage}{27}
\indexentry{Ensemble Learning|hyperpage}{27}
\indexentry{容量|hyperpage}{28}
\indexentry{Capacity|hyperpage}{28}
\indexentry{宽残差网络|hyperpage}{29}
\indexentry{Wide Residual Network|hyperpage}{29}
\indexentry{探测任务|hyperpage}{30}
\indexentry{Probing Task|hyperpage}{30}
\indexentry{表面信息|hyperpage}{30}
\indexentry{Surface Information|hyperpage}{30}
\indexentry{语法信息|hyperpage}{30}
\indexentry{Syntactic Information|hyperpage}{30}
\indexentry{语义信息|hyperpage}{30}
\indexentry{Semantic Information|hyperpage}{30}
\indexentry{词嵌入|hyperpage}{31}
\indexentry{Embedding|hyperpage}{31}
\indexentry{数据并行|hyperpage}{31}
\indexentry{Data Parallelism|hyperpage}{31}
\indexentry{模型并行|hyperpage}{31}
\indexentry{Model Parallelism|hyperpage}{31}
\indexentry{小批量训练|hyperpage}{31}
\indexentry{Mini-batch Training|hyperpage}{31}
\indexentry{课程学习|hyperpage}{33}
\indexentry{Curriculum Learning|hyperpage}{33}
\indexentry{推断|hyperpage}{34}
\indexentry{Inference|hyperpage}{34}
\indexentry{解码|hyperpage}{34}
\indexentry{Decoding|hyperpage}{34}
\indexentry{搜索错误|hyperpage}{34}
\indexentry{Search Error|hyperpage}{34}
\indexentry{模型错误|hyperpage}{34}
\indexentry{Modeling Error|hyperpage}{34}
\indexentry{重排序|hyperpage}{36}
\indexentry{Re-ranking|hyperpage}{36}
\indexentry{双向推断|hyperpage}{36}
\indexentry{Bidirectional Inference|hyperpage}{36}
\indexentry{批量推断|hyperpage}{40}
\indexentry{Batch Inference|hyperpage}{40}
\indexentry{批量处理|hyperpage}{40}
\indexentry{Batching|hyperpage}{40}
\indexentry{二值网络|hyperpage}{41}
\indexentry{Binarized Neural Networks|hyperpage}{41}
\indexentry{自回归翻译|hyperpage}{42}
\indexentry{Autoregressive Translation|hyperpage}{42}
\indexentry{非自回归翻译|hyperpage}{42}
\indexentry{Regressive Translation|hyperpage}{42}
\indexentry{繁衍率|hyperpage}{42}
\indexentry{Fertility|hyperpage}{42}
\indexentry{偏置|hyperpage}{43}
\indexentry{Bias|hyperpage}{43}
\indexentry{退化|hyperpage}{44}
\indexentry{Degenerate|hyperpage}{44}
\indexentry{过翻译|hyperpage}{45}
\indexentry{Over Translation|hyperpage}{45}
\indexentry{欠翻译|hyperpage}{45}
\indexentry{Under Translation|hyperpage}{45}
\indexentry{充分性|hyperpage}{46}
\indexentry{Adequacy|hyperpage}{46}
\indexentry{系统融合|hyperpage}{46}
\indexentry{System Combination|hyperpage}{46}
\indexentry{假设选择|hyperpage}{47}
\indexentry{Hypothesis Selection|hyperpage}{47}
\indexentry{多样性|hyperpage}{47}
\indexentry{Diversity|hyperpage}{47}
\indexentry{重排序|hyperpage}{48}
\indexentry{Re-ranking|hyperpage}{48}
\indexentry{混淆网络|hyperpage}{49}
\indexentry{Confusion Network|hyperpage}{49}
\indexentry{动态线性层聚合方法|hyperpage}{53}
\indexentry{Dynamic Linear Combination of Layers,DLCL|hyperpage}{53}
\indexentry{相互适应|hyperpage}{57}
\indexentry{Co-adaptation|hyperpage}{57}
\indexentry{数据增强|hyperpage}{59}
\indexentry{Data Augmentation|hyperpage}{59}
\indexentry{回译|hyperpage}{59}
\indexentry{Back Translation|hyperpage}{59}
\indexentry{迭代式回译|hyperpage}{60}
\indexentry{Iterative Back Translation|hyperpage}{60}
\indexentry{前向翻译|hyperpage}{61}
\indexentry{Forward Translation|hyperpage}{61}
\indexentry{分布式表示|hyperpage}{61}
\indexentry{Distributed Representation|hyperpage}{61}
\indexentry{预训练|hyperpage}{61}
\indexentry{Pre-training|hyperpage}{61}
\indexentry{微调|hyperpage}{61}
\indexentry{Fine-tuning|hyperpage}{61}
\indexentry{多任务学习|hyperpage}{63}
\indexentry{Multitask Learning|hyperpage}{63}
\indexentry{知识精炼|hyperpage}{64}
\indexentry{Knowledge Distillation|hyperpage}{64}
\indexentry{模型压缩|hyperpage}{64}
\indexentry{Model Compression|hyperpage}{64}
\indexentry{学习难度|hyperpage}{64}
\indexentry{Learning Difficulty|hyperpage}{64}
\indexentry{教师模型|hyperpage}{65}
\indexentry{Teacher Model|hyperpage}{65}
\indexentry{学生模型|hyperpage}{65}
\indexentry{Student Model|hyperpage}{65}
\indexentry{基于单词的知识精炼|hyperpage}{65}
\indexentry{Word-level Knowledge Distillation|hyperpage}{65}
\indexentry{基于序列的知识精炼|hyperpage}{65}
\indexentry{Sequence-level Knowledge Distillation|hyperpage}{65}
\indexentry{中间层输出|hyperpage}{66}
\indexentry{Hint-based Knowledge Transfer|hyperpage}{66}
\indexentry{注意力分布|hyperpage}{66}
\indexentry{Attention To Attention Transfer|hyperpage}{67}
\indexentry{循环一致性|hyperpage}{69}
\indexentry{Circle Consistency|hyperpage}{69}
\indexentry{翻译中回译|hyperpage}{70}
\indexentry{On-the-fly Back-translation|hyperpage}{70}
\indexentry{神经架构搜索|hyperpage}{73}
\indexentry{Neural Architecture Search|hyperpage}{73}
Book/mt-book-xelatex.ptc
查看文件 @
207e0648
...
@@ -2,714 +2,140 @@
...
@@ -2,714 +2,140 @@
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\select@language {english}
\select@language {english}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {part}{\@mypartnumtocformat {I}{
机器翻译基础}}{13
}{part.1}
\contentsline {part}{\@mypartnumtocformat {I}{
神经机器翻译}}{9
}{part.1}
\ttl@starttoc {default@1}
\ttl@starttoc {default@1}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {1}
机器翻译简介}{15
}{chapter.1}
\contentsline {chapter}{\numberline {1}
人工神经网络和神经语言建模}{11
}{chapter.1}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {1.1}
机器翻译的概念}{15
}{section.1.1}
\contentsline {section}{\numberline {1.1}
深度学习与人工神经网络}{12
}{section.1.1}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.2}机器翻译简史}{18}{section.1.2
}
\contentsline {s
ubsection}{\numberline {1.1.1}发展简史}{12}{subsection.1.1.1
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.2.1}人工翻译}{18}{subsection.1.2.1
}
\contentsline {subs
ubsection}{早期的人工神经网络和第一次寒冬}{12}{section*.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.2.2}机器翻译的萌芽}{19}{subsection.1.2.2
}
\contentsline {subs
ubsection}{神经网络的第二次高潮和第二次寒冬}{13}{section*.4
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.2.3}机器翻译的受挫}{20}{subsection.1.2.3
}
\contentsline {subs
ubsection}{深度学习和神经网络方法的崛起}{14}{section*.5
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {1.
2.4}机器翻译的快速成长}{21}{subsection.1.2.4
}
\contentsline {subsection}{\numberline {1.
1.2}为什么需要深度学习}{15}{subsection.1.1.2
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.2.5}机器翻译的爆发}{22}{subsection.1.2.5
}
\contentsline {subs
ubsection}{端到端学习和表示学习}{15}{section*.7
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.3}机器翻译现状}{23}{section.1.3
}
\contentsline {s
ubsubsection}{深度学习的效果}{16}{section*.9
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {1.
4}机器翻译方法}{24}{section.1.4
}
\contentsline {section}{\numberline {1.
2}神经网络基础}{16}{section.1.2
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {1.
4.1}基于规则的机器翻译}{26}{subsection.1.4
.1}
\contentsline {subsection}{\numberline {1.
2.1}线性代数基础}{16}{subsection.1.2
.1}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.4.2}基于实例的机器翻译}{26}{subsection.1.4.2
}
\contentsline {subs
ubsection}{标量、向量和矩阵}{17}{section*.11
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.4.3}统计机器翻译}{27}{subsection.1.4.3
}
\contentsline {subs
ubsection}{矩阵的转置}{18}{section*.12
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.4.4}神经机器翻译}{28}{subsection.1.4.4
}
\contentsline {subs
ubsection}{矩阵加法和数乘}{18}{section*.13
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.4.5}对比分析}{29}{subsection.1.4.5
}
\contentsline {subs
ubsection}{矩阵乘法和矩阵点乘}{19}{section*.14
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.5}翻译质量评价}{30}{section.1.
5}
\contentsline {s
ubsubsection}{线性映射}{20}{section*.1
5}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.5.1}人工评价}{30}{subsection.1.5.1
}
\contentsline {subs
ubsection}{范数}{21}{section*.16
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {1.
5.2}自动评价}{31}{subsection.1.5
.2}
\contentsline {subsection}{\numberline {1.
2.2}人工神经元和感知机}{22}{subsection.1.2
.2}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
BLEU}{31}{section*.15
}
\contentsline {subsubsection}{
感知机\ \raisebox {0.5mm}{------}\ 最简单的人工神经元模型}{23}{section*.19
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
TER}{33}{section*.16
}
\contentsline {subsubsection}{
神经元内部权重}{24}{section*.22
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
基于检测点的评价}{33}{section*.17
}
\contentsline {subsubsection}{
神经元的输入\ \raisebox {0.5mm}{------}\ 离散 vs 连续}{25}{section*.24
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.6}机器翻译应用}{34}{section.1.
6}
\contentsline {s
ubsubsection}{神经元内部的参数学习}{25}{section*.2
6}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.7}开源项目与评测}{36}{section.1.7
}
\contentsline {s
ubsection}{\numberline {1.2.3}多层神经网络}{26}{subsection.1.2.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {1.7.1}开源机器翻译系统}{36}{subsection.1.7.1
}
\contentsline {subs
ubsection}{线性变换和激活函数}{26}{section*.28
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
统计机器翻译开源系统}{37}{section*.19
}
\contentsline {subsubsection}{
单层神经网络$\rightarrow $多层神经网络}{28}{section*.35
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ubsection}{神经机器翻译开源系统}{38}{section*.20
}
\contentsline {subs
ection}{\numberline {1.2.4}函数拟合能力}{29}{subsection.1.2.4
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ubsection}{\numberline {1.7.2}常用数据集及公开评测任务}{40}{subsection.1.7.2
}
\contentsline {s
ection}{\numberline {1.3}神经网络的张量实现}{33}{section.1.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {1.8}推荐学习资源}{42}{section.1.8
}
\contentsline {s
ubsection}{\numberline {1.3.1} 张量及其计算}{34}{subsection.1.3.1
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {
chapter}{\numberline {2}词法、语法及统计建模基础}{47}{chapter.2
}
\contentsline {
subsubsection}{张量}{34}{section*.45
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {2.1}问题概述 }{48}{section.2.1
}
\contentsline {s
ubsubsection}{张量的矩阵乘法}{36}{section*.48
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {2.2}概率论基础}{49}{section.2.2
}
\contentsline {s
ubsubsection}{张量的单元操作}{37}{section*.50
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.2.1}随机变量和概率}{50}{subsection.2.2.1
}
\contentsline {subsection}{\numberline {
1.3.2}张量的物理存储形式}{38}{subsection.1.3.2
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.2.2}联合概率、条件概率和边缘概率}{51}{subsection.2.2.2
}
\contentsline {subsection}{\numberline {
1.3.3}使用开源框架实现张量计算}{38}{subsection.1.3.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.2.3}链式法则}{52}{subsection.2.2.3
}
\contentsline {subsection}{\numberline {
1.3.4}前向传播与计算图}{40}{subsection.1.3.4
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.2.4}贝叶斯法则}{53}{subsection.2.2.4
}
\contentsline {subsection}{\numberline {
1.3.5}神经网络实例}{43}{subsection.1.3.5
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ubsection}{\numberline {2.2.5}KL距离和熵}{55}{subsection.2.2.5
}
\contentsline {s
ection}{\numberline {1.4}神经网络的参数训练}{44}{section.1.4
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ubsection}{信息熵}{55}{section*.27
}
\contentsline {subs
ection}{\numberline {1.4.1}损失函数}{45}{subsection.1.4.1
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ubsection}{KL距离}{56}{section*.29
}
\contentsline {subs
ection}{\numberline {1.4.2}基于梯度的参数优化}{45}{subsection.1.4.2
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
交叉熵}{56}{section*.30
}
\contentsline {subsubsection}{
梯度下降}{46}{section*.68
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {2.3}中文分词}{57}{section.2.3
}
\contentsline {s
ubsubsection}{梯度获取}{48}{section*.70
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {2.3.1}基于词典的分词方法}{58}{subsection.2.3.1
}
\contentsline {subs
ubsection}{基于梯度的方法的变种和改进}{51}{section*.74
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.3.2}基于统计的分词方法}{59}{subsection.2.3.2
}
\contentsline {subsection}{\numberline {
1.4.3}参数更新的并行化策略}{54}{subsection.1.4.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ubsection}{统计模型的学习与推断}{59}{section*.3
4}
\contentsline {subs
ection}{\numberline {1.4.4}梯度消失、梯度爆炸和稳定性训练}{56}{subsection.1.4.
4}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
掷骰子游戏}{60}{section*.36
}
\contentsline {subsubsection}{
易于优化的激活函数}{56}{section*.77
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
全概率分词方法}{62}{section*.40
}
\contentsline {subsubsection}{
梯度裁剪}{57}{section*.81
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {2.4}$n$-gram语言模型 }{64}{section.2.4
}
\contentsline {s
ubsubsection}{稳定性训练}{58}{section*.82
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.4.1}建模}{65}{subsection.2.4.1
}
\contentsline {subsection}{\numberline {
1.4.5}过拟合}{59}{subsection.1.4.5
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.4.2}未登录词和平滑算法}{67}{subsection.2.4.2
}
\contentsline {subsection}{\numberline {
1.4.6}反向传播}{60}{subsection.1.4.6
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
加法平滑方法}{68}{section*.46
}
\contentsline {subsubsection}{
输出层的反向传播}{61}{section*.85
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
古德-图灵估计法}{69}{section*.48
}
\contentsline {subsubsection}{
隐藏层的反向传播}{63}{section*.89
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
Kneser-Ney平滑方法}{70}{section*.50
}
\contentsline {subsubsection}{
程序实现}{64}{section*.92
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {
2.5}句法分析(短语结构分析)}{72}{section.2
.5}
\contentsline {section}{\numberline {
1.5}神经语言模型}{66}{section.1
.5}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
2.5.1}句子的句法树表示}{72}{subsection.2
.5.1}
\contentsline {subsection}{\numberline {
1.5.1}基于神经网络的语言建模}{66}{subsection.1
.5.1}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {2.5.2}上下文无关文法}{74}{subsection.2.5.2
}
\contentsline {subs
ubsection}{基于前馈神经网络的语言模型}{67}{section*.95
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {2.5.3}规则和推导的概率}{78}{subsection.2.5.3
}
\contentsline {subs
ubsection}{基于循环神经网络的语言模型}{69}{section*.98
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {2.6}小结及深入阅读}{80}{section.2.6
}
\contentsline {s
ubsubsection}{基于自注意力机制的语言模型}{70}{section*.100
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {part}{\@mypartnumtocformat {II}{统计机器翻译}}{83}{part.2}
\contentsline {subsubsection}{语言模型的评价}{71}{section*.102}
\ttl@stoptoc {default@1}
\ttl@starttoc {default@2}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {
chapter}{\numberline {3}基于词的机器翻译模型}{85}{chapter.3
}
\contentsline {
subsection}{\numberline {1.5.2}单词表示模型}{72}{subsection.1.5.2
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {3.1}什么是基于词的翻译模型}{85}{section.3.1
}
\contentsline {s
ubsubsection}{One-hot编码}{72}{section*.103
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {s
ection}{\numberline {3.2}构建一个简单的机器翻译系统}{87}{section.3.2
}
\contentsline {s
ubsubsection}{分布式表示}{72}{section*.105
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {
3.2.1}如何进行翻译?}{87}{subsection.3.2.1
}
\contentsline {subsection}{\numberline {
1.5.3}句子表示模型及预训练}{74}{subsection.1.5.3
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
机器翻译流程}{88}{section*.63
}
\contentsline {subsubsection}{
简单的上下文表示模型}{74}{section*.109
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
人工翻译 vs. 机器翻译}{89}{section*.65
}
\contentsline {subsubsection}{
ELMO模型}{76}{section*.112
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {3.2.2}基本框架}{89}{subsection.3.2.2
}
\contentsline {subs
ubsection}{GPT模型}{76}{section*.114
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subs
ection}{\numberline {3.2.3}单词翻译概率}{90}{subsection.3.2.3
}
\contentsline {subs
ubsection}{BERT模型}{77}{section*.116
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{
什么是单词翻译概率?}{90}{section*.67
}
\contentsline {subsubsection}{
为什么要预训练?}{78}{section*.118
}
\defcounter {refsection}{0}\relax
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{如何从一个双语平行数据中学习?}{90}{section*.69}
\contentsline {section}{\numberline {1.6}小结及深入阅读}{79}{section.1.6}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{如何从大量的双语平行数据中学习?}{92}{section*.70}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.2.4}句子级翻译模型}{93}{subsection.3.2.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基础模型}{93}{section*.72}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{生成流畅的译文}{95}{section*.74}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.2.5}解码}{97}{subsection.3.2.5}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {3.3}基于词的翻译建模}{100}{section.3.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.3.1}噪声信道模型}{100}{subsection.3.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.3.2}统计机器翻译的三个基本问题}{102}{subsection.3.3.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{词对齐}{103}{section*.83}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于词对齐的翻译模型}{103}{section*.86}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于词对齐的翻译实例}{105}{section*.88}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {3.4}IBM模型1-2}{106}{section.3.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.4.1}IBM模型1}{106}{subsection.3.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.4.2}IBM模型2}{108}{subsection.3.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.4.3}解码及计算优化}{109}{subsection.3.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.4.4}训练}{110}{subsection.3.4.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{目标函数}{110}{section*.93}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{优化}{111}{section*.95}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {3.5}IBM模型3-5及隐马尔可夫模型}{117}{section.3.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.1}基于产出率的翻译模型}{117}{subsection.3.5.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.2}IBM 模型3}{120}{subsection.3.5.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.3}IBM 模型4}{121}{subsection.3.5.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.4} IBM 模型5}{123}{subsection.3.5.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.5}隐马尔可夫模型}{124}{subsection.3.5.5}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{隐马尔可夫模型}{125}{section*.107}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{词对齐模型}{126}{section*.109}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.5.6}解码和训练}{127}{subsection.3.5.6}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {3.6}问题分析}{127}{section.3.6}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.6.1}词对齐及对称化}{127}{subsection.3.6.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.6.2}Deficiency}{128}{subsection.3.6.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.6.3}句子长度}{129}{subsection.3.6.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {3.6.4}其他问题}{130}{subsection.3.6.4}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {3.7}小结及深入阅读}{130}{section.3.7}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {4}基于短语和句法的机器翻译模型}{133}{chapter.4}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {4.1}翻译中的结构信息}{133}{section.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.1.1}更大粒度的翻译单元}{134}{subsection.4.1.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.1.2}句子的结构信息}{136}{subsection.4.1.2}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {4.2}基于短语的翻译模型}{138}{section.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.1}机器翻译中的短语}{138}{subsection.4.2.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.2}数学建模及判别式模型}{141}{subsection.4.2.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于翻译推导的建模}{141}{section*.121}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{对数线性模型}{142}{section*.122}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{搭建模型的基本流程}{143}{section*.123}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.3}短语抽取}{144}{subsection.4.2.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{与词对齐一致的短语}{145}{section*.126}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{获取词对齐}{146}{section*.130}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{度量双语短语质量}{147}{section*.132}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.4}调序}{148}{subsection.4.2.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于距离的调序}{148}{section*.136}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于方向的调序}{149}{section*.138}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于分类的调序}{151}{section*.141}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.5}特征}{151}{subsection.4.2.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.6}最小错误率训练}{152}{subsection.4.2.6}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.2.7}栈解码}{155}{subsection.4.2.7}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{翻译候选匹配}{156}{section*.146}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{翻译假设扩展}{156}{section*.148}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{剪枝}{157}{section*.150}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{解码中的栈结构}{159}{section*.152}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {4.3}基于层次短语的模型}{160}{section.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.3.1}同步上下文无关文法}{163}{subsection.4.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{文法定义}{163}{section*.157}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{推导}{164}{section*.158}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{胶水规则}{165}{section*.159}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{处理流程}{166}{section*.160}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.3.2}层次短语规则抽取}{166}{subsection.4.3.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.3.3}翻译模型及特征}{168}{subsection.4.3.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.3.4}CYK解码}{169}{subsection.4.3.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.3.5}立方剪枝}{172}{subsection.4.3.5}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {4.4}基于语言学句法的模型}{175}{section.4.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.1}基于句法的翻译模型分类}{177}{subsection.4.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.2}基于树结构的文法}{177}{subsection.4.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{树到树翻译规则}{179}{section*.176}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于树结构的翻译推导}{181}{section*.178}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{树到串翻译规则}{183}{section*.181}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.3}树到串翻译规则抽取}{184}{subsection.4.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{树的切割与最小规则}{185}{section*.183}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{空对齐处理}{188}{section*.189}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{组合规则}{189}{section*.191}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{SPMT规则}{190}{section*.193}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{句法树二叉化}{191}{section*.195}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.4}树到树翻译规则抽取}{192}{subsection.4.4.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于节点对齐的规则抽取}{193}{section*.199}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于对齐矩阵的规则抽取}{194}{section*.202}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.5}句法翻译模型的特征}{196}{subsection.4.4.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.6}基于超图的推导空间表示}{197}{subsection.4.4.6}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {4.4.7}基于树的解码 vs 基于串的解码}{199}{subsection.4.4.7}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于树的解码}{201}{section*.209}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于串的解码}{202}{section*.212}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {4.5}小结及深入阅读}{204}{section.4.5}
\defcounter {refsection}{0}\relax
\contentsline {part}{\@mypartnumtocformat {III}{神经机器翻译}}{207}{part.3}
\ttl@stoptoc {default@2}
\ttl@starttoc {default@3}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {5}人工神经网络和神经语言建模}{209}{chapter.5}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.1}深度学习与人工神经网络}{210}{section.5.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.1.1}发展简史}{210}{subsection.5.1.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{早期的人工神经网络和第一次寒冬}{210}{section*.214}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{神经网络的第二次高潮和第二次寒冬}{211}{section*.215}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{深度学习和神经网络方法的崛起}{212}{section*.216}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.1.2}为什么需要深度学习}{213}{subsection.5.1.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{端到端学习和表示学习}{213}{section*.218}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{深度学习的效果}{214}{section*.220}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.2}神经网络基础}{214}{section.5.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.2.1}线性代数基础}{214}{subsection.5.2.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{标量、向量和矩阵}{215}{section*.222}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{矩阵的转置}{216}{section*.223}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{矩阵加法和数乘}{216}{section*.224}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{矩阵乘法和矩阵点乘}{217}{section*.225}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{线性映射}{218}{section*.226}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{范数}{219}{section*.227}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.2.2}人工神经元和感知机}{220}{subsection.5.2.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{感知机\ \raisebox {0.5mm}{------}\ 最简单的人工神经元模型}{221}{section*.230}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{神经元内部权重}{222}{section*.233}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{神经元的输入\ \raisebox {0.5mm}{------}\ 离散 vs 连续}{223}{section*.235}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{神经元内部的参数学习}{223}{section*.237}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.2.3}多层神经网络}{224}{subsection.5.2.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{线性变换和激活函数}{224}{section*.239}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{单层神经网络$\rightarrow $多层神经网络}{226}{section*.246}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.2.4}函数拟合能力}{227}{subsection.5.2.4}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.3}神经网络的张量实现}{231}{section.5.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.3.1} 张量及其计算}{232}{subsection.5.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{张量}{232}{section*.256}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{张量的矩阵乘法}{234}{section*.259}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{张量的单元操作}{235}{section*.261}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.3.2}张量的物理存储形式}{236}{subsection.5.3.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.3.3}使用开源框架实现张量计算}{236}{subsection.5.3.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.3.4}前向传播与计算图}{238}{subsection.5.3.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.3.5}神经网络实例}{241}{subsection.5.3.5}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.4}神经网络的参数训练}{242}{section.5.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.1}损失函数}{243}{subsection.5.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.2}基于梯度的参数优化}{243}{subsection.5.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{梯度下降}{244}{section*.279}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{梯度获取}{246}{section*.281}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于梯度的方法的变种和改进}{249}{section*.285}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.3}参数更新的并行化策略}{252}{subsection.5.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.4}梯度消失、梯度爆炸和稳定性训练}{254}{subsection.5.4.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{易于优化的激活函数}{254}{section*.288}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{梯度裁剪}{255}{section*.292}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{稳定性训练}{256}{section*.293}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.5}过拟合}{257}{subsection.5.4.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.4.6}反向传播}{258}{subsection.5.4.6}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{输出层的反向传播}{259}{section*.296}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{隐藏层的反向传播}{261}{section*.300}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{程序实现}{262}{section*.303}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.5}神经语言模型}{264}{section.5.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.5.1}基于神经网络的语言建模}{264}{subsection.5.5.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于前馈神经网络的语言模型}{265}{section*.306}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于循环神经网络的语言模型}{267}{section*.309}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{基于自注意力机制的语言模型}{268}{section*.311}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{语言模型的评价}{269}{section*.313}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.5.2}单词表示模型}{270}{subsection.5.5.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{One-hot编码}{270}{section*.314}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{分布式表示}{270}{section*.316}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {5.5.3}句子表示模型及预训练}{272}{subsection.5.5.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{简单的上下文表示模型}{272}{section*.320}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{ELMO模型}{274}{section*.323}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{GPT模型}{274}{section*.325}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{BERT模型}{275}{section*.327}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{为什么要预训练?}{276}{section*.329}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {5.6}小结及深入阅读}{277}{section.5.6}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {6}神经机器翻译模型}{279}{chapter.6}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.1}神经机器翻译的发展简史}{279}{section.6.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.1.1}神经机器翻译的起源}{281}{subsection.6.1.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.1.2}神经机器翻译的品质 }{283}{subsection.6.1.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.1.3}神经机器翻译的优势 }{286}{subsection.6.1.3}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.2}编码器-解码器框架}{288}{section.6.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.2.1}框架结构}{288}{subsection.6.2.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.2.2}表示学习}{289}{subsection.6.2.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.2.3}简单的运行实例}{290}{subsection.6.2.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.2.4}机器翻译范式的对比}{291}{subsection.6.2.4}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.3}基于循环神经网络的翻译模型及注意力机制}{292}{section.6.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.1}建模}{292}{subsection.6.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.2}输入(词嵌入)及输出(Softmax)}{296}{subsection.6.3.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.3}循环神经网络结构}{300}{subsection.6.3.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{循环神经单元(RNN)}{300}{section*.351}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{长短时记忆网络(LSTM)}{300}{section*.352}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{门控循环单元(GRU)}{302}{section*.355}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{双向模型}{304}{section*.357}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{多层循环神经网络}{304}{section*.359}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.4}注意力机制}{305}{subsection.6.3.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{翻译中的注意力机制}{306}{section*.362}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{上下文向量的计算}{307}{section*.365}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{注意力机制的解读}{310}{section*.370}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.5}训练}{312}{subsection.6.3.5}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{损失函数}{312}{section*.373}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{长参数初始化}{313}{section*.374}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{优化策略}{314}{section*.375}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{梯度裁剪}{314}{section*.377}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{学习率策略}{314}{section*.378}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{并行训练}{316}{section*.381}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.6}推断}{317}{subsection.6.3.6}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{贪婪搜索}{319}{section*.385}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{束搜索}{320}{section*.388}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{长度惩罚}{321}{section*.390}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.3.7}实例-GNMT}{322}{subsection.6.3.7}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.4}Transformer}{323}{section.6.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.1}自注意力模型}{325}{subsection.6.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.2}Transformer架构}{326}{subsection.6.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.3}位置编码}{328}{subsection.6.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.4}基于点乘的注意力机制}{331}{subsection.6.4.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.5}掩码操作}{333}{subsection.6.4.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.6}多头注意力}{334}{subsection.6.4.6}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.7}残差网络和层正则化}{335}{subsection.6.4.7}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.8}前馈全连接网络子层}{336}{subsection.6.4.8}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.9}训练}{337}{subsection.6.4.9}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.4.10}推断}{340}{subsection.6.4.10}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.5}序列到序列问题及应用}{340}{section.6.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.5.1}自动问答}{341}{subsection.6.5.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.5.2}自动文摘}{341}{subsection.6.5.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.5.3}文言文翻译}{342}{subsection.6.5.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.5.4}对联生成}{342}{subsection.6.5.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {6.5.5}古诗生成}{343}{subsection.6.5.5}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {6.6}小结及深入阅读}{344}{section.6.6}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {7}神经机器翻译实战 \ \raisebox {0.5mm}{------}\ 参加一次比赛}{347}{chapter.7}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.1}神经机器翻译并不简单}{347}{section.7.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.1.1}影响神经机器翻译性能的因素}{348}{subsection.7.1.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.1.2}搭建神经机器翻译系统的步骤 }{349}{subsection.7.1.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.1.3}架构选择 }{350}{subsection.7.1.3}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.2}数据处理}{350}{section.7.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.2.1}分词}{351}{subsection.7.2.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.2.2}标准化}{352}{subsection.7.2.2}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.2.3}数据清洗}{353}{subsection.7.2.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.2.4}子词切分}{355}{subsection.7.2.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{大词表和OOV问题}{356}{section*.428}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{子词}{356}{section*.430}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{双字节编码(BPE)}{357}{section*.432}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{其他方法}{360}{section*.435}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.3}建模与训练}{360}{section.7.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.3.1}正则化}{360}{subsection.7.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{L1/L2正则化}{362}{section*.437}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{标签平滑}{363}{section*.438}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{Dropout}{364}{section*.440}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{Layer Dropout}{365}{section*.443}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.3.2}增大模型容量}{366}{subsection.7.3.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{宽网络}{366}{section*.445}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{深网络}{367}{section*.447}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{增大输入层和输出层表示能力}{369}{section*.449}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{大模型的分布式计算}{369}{section*.450}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.3.3}大批量训练}{369}{subsection.7.3.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{为什么需要大批量训练}{370}{section*.451}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{如何构建批次}{371}{section*.454}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.4}推断}{372}{section.7.4}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.4.1}推断优化}{372}{subsection.7.4.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{推断系统的架构}{372}{section*.456}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{自左向右推断 vs 自右向左推断}{373}{section*.458}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{推断加速}{374}{section*.459}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.4.2}译文长度控制}{381}{subsection.7.4.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{长度惩罚因子}{382}{section*.465}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{译文长度范围约束}{383}{section*.467}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{覆盖度模型}{383}{section*.468}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.4.3}多模型集成}{384}{subsection.7.4.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{假设选择}{385}{section*.469}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{局部预测融合}{386}{section*.471}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{译文重组}{387}{section*.473}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.5}进阶技术}{388}{section.7.5}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.5.1}深层模型}{388}{subsection.7.5.1}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{Post-Norm vs Pre-Norm}{388}{section*.476}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{层聚合}{391}{section*.479}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{深层模型的训练加速}{392}{section*.481}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{渐进式训练}{392}{section*.482}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{分组稠密连接}{392}{section*.484}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{学习率重置策略}{393}{section*.486}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{深层模型的鲁棒性训练}{395}{section*.488}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.5.2}单语数据的使用}{396}{subsection.7.5.2}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{伪数据}{397}{section*.491}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{预训练}{399}{section*.494}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{联合训练}{401}{section*.497}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.5.3}知识精炼}{401}{subsection.7.5.3}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{什么是知识精炼}{402}{section*.499}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{知识精炼的基本方法}{403}{section*.500}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{机器翻译中的知识精炼}{404}{section*.502}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {7.5.4}双向训练}{406}{subsection.7.5.4}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{有监督对偶学习}{406}{section*.504}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{无监督对偶学习}{407}{section*.505}
\defcounter {refsection}{0}\relax
\contentsline {subsubsection}{翻译中回译}{408}{section*.507}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {7.6}小结及深入阅读}{408}{section.7.6}
\defcounter {refsection}{0}\relax
\contentsline {part}{\@mypartnumtocformat {IV}{附录}}{413}{part.4}
\ttl@stoptoc {default@3}
\ttl@starttoc {default@4}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {A}附录A}{415}{Appendix.1.A}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {A.1}基准数据集}{415}{section.1.A.1}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {A.2}平行语料}{416}{section.1.A.2}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {A.3}相关工具}{417}{section.1.A.3}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {A.3.1}数据预处理工具}{417}{subsection.1.A.3.1}
\defcounter {refsection}{0}\relax
\contentsline {subsection}{\numberline {A.3.2}评价工具}{418}{subsection.1.A.3.2}
\defcounter {refsection}{0}\relax
\contentsline {chapter}{\numberline {B}附录B}{419}{Appendix.2.B}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {B.1}IBM模型3训练方法}{419}{section.2.B.1}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {B.2}IBM模型4训练方法}{421}{section.2.B.2}
\defcounter {refsection}{0}\relax
\contentsline {section}{\numberline {B.3}IBM模型5训练方法}{423}{section.2.B.3}
\contentsfinish
\contentsfinish
Book/mt-book-xelatex.tex
查看文件 @
207e0648
...
@@ -127,7 +127,7 @@
...
@@ -127,7 +127,7 @@
%\include{Chapter4/chapter4}
%\include{Chapter4/chapter4}
%\include{Chapter5/chapter5}
%\include{Chapter5/chapter5}
%\include{Chapter6/chapter6}
%\include{Chapter6/chapter6}
\include
{
Chapter7/chapter7
}
%
\include{Chapter7/chapter7}
%\include{ChapterAppend/chapterappend}
%\include{ChapterAppend/chapterappend}
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论