{
"cells": [
{
"cell_type": "markdown",
"id": "ba91d77a",
"metadata": {
"origin_pos": 0
},
"source": [
"# 双向循环神经网络\n",
":label:`sec_bi_rnn`\n",
"\n",
"在序列学习中,我们以往假设的目标是:\n",
"在给定观测的情况下\n",
"(例如,在时间序列的上下文中或在语言模型的上下文中),\n",
"对下一个输出进行建模。\n",
"虽然这是一个典型情景,但不是唯一的。\n",
"还可能发生什么其它的情况呢?\n",
"我们考虑以下三个在文本序列中填空的任务。\n",
"\n",
"* 我`___`。\n",
"* 我`___`饿了。\n",
"* 我`___`饿了,我可以吃半头猪。\n",
"\n",
"根据可获得的信息量,我们可以用不同的词填空,\n",
"如“很高兴”(\"happy\")、“不”(\"not\")和“非常”(\"very\")。\n",
"很明显,每个短语的“下文”传达了重要信息(如果有的话),\n",
"而这些信息关乎到选择哪个词来填空,\n",
"所以无法利用这一点的序列模型将在相关任务上表现不佳。\n",
"例如,如果要做好命名实体识别\n",
"(例如,识别“Green”指的是“格林先生”还是绿色),\n",
"不同长度的上下文范围重要性是相同的。\n",
"为了获得一些解决问题的灵感,让我们先迂回到概率图模型。\n",
"\n",
"## 隐马尔可夫模型中的动态规划\n",
"\n",
"这一小节是用来说明动态规划问题的,\n",
"具体的技术细节对于理解深度学习模型并不重要,\n",
"但它有助于我们思考为什么要使用深度学习,\n",
"以及为什么要选择特定的架构。\n",
"\n",
"如果我们想用概率图模型来解决这个问题,\n",
"可以设计一个隐变量模型:\n",
"在任意时间步$t$,假设存在某个隐变量$h_t$,\n",
"通过概率$P(x_t \\mid h_t)$控制我们观测到的$x_t$。\n",
"此外,任何$h_t \\to h_{t+1}$转移\n",
"都是由一些状态转移概率$P(h_{t+1} \\mid h_{t})$给出。\n",
"这个概率图模型就是一个*隐马尔可夫模型*(hidden Markov model,HMM),\n",
"如 :numref:`fig_hmm`所示。\n",
"\n",
"\n",
":label:`fig_hmm`\n",
"\n",
"因此,对于有$T$个观测值的序列,\n",
"我们在观测状态和隐状态上具有以下联合概率分布:\n",
"\n",
"$$P(x_1, \\ldots, x_T, h_1, \\ldots, h_T) = \\prod_{t=1}^T P(h_t \\mid h_{t-1}) P(x_t \\mid h_t), \\text{ where } P(h_1 \\mid h_0) = P(h_1).$$\n",
":eqlabel:`eq_hmm_jointP`\n",
"\n",
"现在,假设我们观测到所有的$x_i$,除了$x_j$,\n",
"并且我们的目标是计算$P(x_j \\mid x_{-j})$,\n",
"其中$x_{-j} = (x_1, \\ldots, x_{j-1}, x_{j+1}, \\ldots, x_{T})$。\n",
"由于$P(x_j \\mid x_{-j})$中没有隐变量,\n",
"因此我们考虑对$h_1, \\ldots, h_T$选择构成的\n",
"所有可能的组合进行求和。\n",
"如果任何$h_i$可以接受$k$个不同的值(有限的状态数),\n",
"这意味着我们需要对$k^T$个项求和,\n",
"这个任务显然难于登天。\n",
"幸运的是,有个巧妙的解决方案:*动态规划*(dynamic programming)。\n",
"\n",
"要了解动态规划的工作方式,\n",
"我们考虑对隐变量$h_1, \\ldots, h_T$的依次求和。\n",
"根据 :eqref:`eq_hmm_jointP`,将得出:\n",
"\n",
"$$\\begin{aligned}\n",
" &P(x_1, \\ldots, x_T) \\\\\n",
" =& \\sum_{h_1, \\ldots, h_T} P(x_1, \\ldots, x_T, h_1, \\ldots, h_T) \\\\\n",
" =& \\sum_{h_1, \\ldots, h_T} \\prod_{t=1}^T P(h_t \\mid h_{t-1}) P(x_t \\mid h_t) \\\\\n",
" =& \\sum_{h_2, \\ldots, h_T} \\underbrace{\\left[\\sum_{h_1} P(h_1) P(x_1 \\mid h_1) P(h_2 \\mid h_1)\\right]}_{\\pi_2(h_2) \\stackrel{\\mathrm{def}}{=}}\n",
" P(x_2 \\mid h_2) \\prod_{t=3}^T P(h_t \\mid h_{t-1}) P(x_t \\mid h_t) \\\\\n",
" =& \\sum_{h_3, \\ldots, h_T} \\underbrace{\\left[\\sum_{h_2} \\pi_2(h_2) P(x_2 \\mid h_2) P(h_3 \\mid h_2)\\right]}_{\\pi_3(h_3)\\stackrel{\\mathrm{def}}{=}}\n",
" P(x_3 \\mid h_3) \\prod_{t=4}^T P(h_t \\mid h_{t-1}) P(x_t \\mid h_t)\\\\\n",
" =& \\dots \\\\\n",
" =& \\sum_{h_T} \\pi_T(h_T) P(x_T \\mid h_T).\n",
"\\end{aligned}$$\n",
"\n",
"通常,我们将*前向递归*(forward recursion)写为:\n",
"\n",
"$$\\pi_{t+1}(h_{t+1}) = \\sum_{h_t} \\pi_t(h_t) P(x_t \\mid h_t) P(h_{t+1} \\mid h_t).$$\n",
"\n",
"递归被初始化为$\\pi_1(h_1) = P(h_1)$。\n",
"符号简化,也可以写成$\\pi_{t+1} = f(\\pi_t, x_t)$,\n",
"其中$f$是一些可学习的函数。\n",
"这看起来就像我们在循环神经网络中讨论的隐变量模型中的更新方程。\n",
"\n",
"与前向递归一样,我们也可以使用后向递归对同一组隐变量求和。这将得到:\n",
"\n",
"$$\\begin{aligned}\n",
" & P(x_1, \\ldots, x_T) \\\\\n",
" =& \\sum_{h_1, \\ldots, h_T} P(x_1, \\ldots, x_T, h_1, \\ldots, h_T) \\\\\n",
" =& \\sum_{h_1, \\ldots, h_T} \\prod_{t=1}^{T-1} P(h_t \\mid h_{t-1}) P(x_t \\mid h_t) \\cdot P(h_T \\mid h_{T-1}) P(x_T \\mid h_T) \\\\\n",
" =& \\sum_{h_1, \\ldots, h_{T-1}} \\prod_{t=1}^{T-1} P(h_t \\mid h_{t-1}) P(x_t \\mid h_t) \\cdot\n",
" \\underbrace{\\left[\\sum_{h_T} P(h_T \\mid h_{T-1}) P(x_T \\mid h_T)\\right]}_{\\rho_{T-1}(h_{T-1})\\stackrel{\\mathrm{def}}{=}} \\\\\n",
" =& \\sum_{h_1, \\ldots, h_{T-2}} \\prod_{t=1}^{T-2} P(h_t \\mid h_{t-1}) P(x_t \\mid h_t) \\cdot\n",
" \\underbrace{\\left[\\sum_{h_{T-1}} P(h_{T-1} \\mid h_{T-2}) P(x_{T-1} \\mid h_{T-1}) \\rho_{T-1}(h_{T-1}) \\right]}_{\\rho_{T-2}(h_{T-2})\\stackrel{\\mathrm{def}}{=}} \\\\\n",
" =& \\ldots \\\\\n",
" =& \\sum_{h_1} P(h_1) P(x_1 \\mid h_1)\\rho_{1}(h_{1}).\n",
"\\end{aligned}$$\n",
"\n",
"因此,我们可以将*后向递归*(backward recursion)写为:\n",
"\n",
"$$\\rho_{t-1}(h_{t-1})= \\sum_{h_{t}} P(h_{t} \\mid h_{t-1}) P(x_{t} \\mid h_{t}) \\rho_{t}(h_{t}),$$\n",
"\n",
"初始化$\\rho_T(h_T) = 1$。\n",
"前向和后向递归都允许我们对$T$个隐变量在$\\mathcal{O}(kT)$\n",
"(线性而不是指数)时间内对$(h_1, \\ldots, h_T)$的所有值求和。\n",
"这是使用图模型进行概率推理的巨大好处之一。\n",
"它也是通用消息传递算法 :cite:`Aji.McEliece.2000`的一个非常特殊的例子。\n",
"结合前向和后向递归,我们能够计算\n",
"\n",
"$$P(x_j \\mid x_{-j}) \\propto \\sum_{h_j} \\pi_j(h_j) \\rho_j(h_j) P(x_j \\mid h_j).$$\n",
"\n",
"因为符号简化的需要,后向递归也可以写为$\\rho_{t-1} = g(\\rho_t, x_t)$,\n",
"其中$g$是一个可以学习的函数。\n",
"同样,这看起来非常像一个更新方程,\n",
"只是不像我们在循环神经网络中看到的那样前向运算,而是后向计算。\n",
"事实上,知道未来数据何时可用对隐马尔可夫模型是有益的。\n",
"信号处理学家将是否知道未来观测这两种情况区分为内插和外推,\n",
"有关更多详细信息,请参阅 :cite:`Doucet.De-Freitas.Gordon.2001`。\n",
"\n",
"## 双向模型\n",
"\n",
"如果我们希望在循环神经网络中拥有一种机制,\n",
"使之能够提供与隐马尔可夫模型类似的前瞻能力,\n",
"我们就需要修改循环神经网络的设计。\n",
"幸运的是,这在概念上很容易,\n",
"只需要增加一个“从最后一个词元开始从后向前运行”的循环神经网络,\n",
"而不是只有一个在前向模式下“从第一个词元开始运行”的循环神经网络。\n",
"*双向循环神经网络*(bidirectional RNNs)\n",
"添加了反向传递信息的隐藏层,以便更灵活地处理此类信息。\n",
" :numref:`fig_birnn`描述了具有单个隐藏层的双向循环神经网络的架构。\n",
"\n",
"\n",
":label:`fig_birnn`\n",
"\n",
"事实上,这与隐马尔可夫模型中的动态规划的前向和后向递归没有太大区别。\n",
"其主要区别是,在隐马尔可夫模型中的方程具有特定的统计意义。\n",
"双向循环神经网络没有这样容易理解的解释,\n",
"我们只能把它们当作通用的、可学习的函数。\n",
"这种转变集中体现了现代深度网络的设计原则:\n",
"首先使用经典统计模型的函数依赖类型,然后将其参数化为通用形式。\n",
"\n",
"### 定义\n",
"\n",
"双向循环神经网络是由 :cite:`Schuster.Paliwal.1997`提出的,\n",
"关于各种架构的详细讨论请参阅 :cite:`Graves.Schmidhuber.2005`。\n",
"让我们看看这样一个网络的细节。\n",
"\n",
"对于任意时间步$t$,给定一个小批量的输入数据\n",
"$\\mathbf{X}_t \\in \\mathbb{R}^{n \\times d}$\n",
"(样本数$n$,每个示例中的输入数$d$),\n",
"并且令隐藏层激活函数为$\\phi$。\n",
"在双向架构中,我们设该时间步的前向和反向隐状态分别为\n",
"$\\overrightarrow{\\mathbf{H}}_t \\in \\mathbb{R}^{n \\times h}$和\n",
"$\\overleftarrow{\\mathbf{H}}_t \\in \\mathbb{R}^{n \\times h}$,\n",
"其中$h$是隐藏单元的数目。\n",
"前向和反向隐状态的更新如下:\n",
"\n",
"$$\n",
"\\begin{aligned}\n",
"\\overrightarrow{\\mathbf{H}}_t &= \\phi(\\mathbf{X}_t \\mathbf{W}_{xh}^{(f)} + \\overrightarrow{\\mathbf{H}}_{t-1} \\mathbf{W}_{hh}^{(f)} + \\mathbf{b}_h^{(f)}),\\\\\n",
"\\overleftarrow{\\mathbf{H}}_t &= \\phi(\\mathbf{X}_t \\mathbf{W}_{xh}^{(b)} + \\overleftarrow{\\mathbf{H}}_{t+1} \\mathbf{W}_{hh}^{(b)} + \\mathbf{b}_h^{(b)}),\n",
"\\end{aligned}\n",
"$$\n",
"\n",
"其中,权重$\\mathbf{W}_{xh}^{(f)} \\in \\mathbb{R}^{d \\times h}, \\mathbf{W}_{hh}^{(f)} \\in \\mathbb{R}^{h \\times h}, \\mathbf{W}_{xh}^{(b)} \\in \\mathbb{R}^{d \\times h}, \\mathbf{W}_{hh}^{(b)} \\in \\mathbb{R}^{h \\times h}$\n",
"和偏置$\\mathbf{b}_h^{(f)} \\in \\mathbb{R}^{1 \\times h}, \\mathbf{b}_h^{(b)} \\in \\mathbb{R}^{1 \\times h}$都是模型参数。\n",
"\n",
"接下来,将前向隐状态$\\overrightarrow{\\mathbf{H}}_t$\n",
"和反向隐状态$\\overleftarrow{\\mathbf{H}}_t$连接起来,\n",
"获得需要送入输出层的隐状态$\\mathbf{H}_t \\in \\mathbb{R}^{n \\times 2h}$。\n",
"在具有多个隐藏层的深度双向循环神经网络中,\n",
"该信息作为输入传递到下一个双向层。\n",
"最后,输出层计算得到的输出为\n",
"$\\mathbf{O}_t \\in \\mathbb{R}^{n \\times q}$($q$是输出单元的数目):\n",
"\n",
"$$\\mathbf{O}_t = \\mathbf{H}_t \\mathbf{W}_{hq} + \\mathbf{b}_q.$$\n",
"\n",
"这里,权重矩阵$\\mathbf{W}_{hq} \\in \\mathbb{R}^{2h \\times q}$\n",
"和偏置$\\mathbf{b}_q \\in \\mathbb{R}^{1 \\times q}$\n",
"是输出层的模型参数。\n",
"事实上,这两个方向可以拥有不同数量的隐藏单元。\n",
"\n",
"### 模型的计算代价及其应用\n",
"\n",
"双向循环神经网络的一个关键特性是:使用来自序列两端的信息来估计输出。\n",
"也就是说,我们使用来自过去和未来的观测信息来预测当前的观测。\n",
"但是在对下一个词元进行预测的情况中,这样的模型并不是我们所需的。\n",
"因为在预测下一个词元时,我们终究无法知道下一个词元的下文是什么,\n",
"所以将不会得到很好的精度。\n",
"具体地说,在训练期间,我们能够利用过去和未来的数据来估计现在空缺的词;\n",
"而在测试期间,我们只有过去的数据,因此精度将会很差。\n",
"下面的实验将说明这一点。\n",
"\n",
"另一个严重问题是,双向循环神经网络的计算速度非常慢。\n",
"其主要原因是网络的前向传播需要在双向层中进行前向和后向递归,\n",
"并且网络的反向传播还依赖于前向传播的结果。\n",
"因此,梯度求解将有一个非常长的链。\n",
"\n",
"双向层的使用在实践中非常少,并且仅仅应用于部分场合。\n",
"例如,填充缺失的单词、词元注释(例如,用于命名实体识别)\n",
"以及作为序列处理流水线中的一个步骤对序列进行编码(例如,用于机器翻译)。\n",
"在 :numref:`sec_bert`和 :numref:`sec_sentiment_rnn`中,\n",
"我们将介绍如何使用双向循环神经网络编码文本序列。\n",
"\n",
"## (**双向循环神经网络的错误应用**)\n",
"\n",
"由于双向循环神经网络使用了过去的和未来的数据,\n",
"所以我们不能盲目地将这一语言模型应用于任何预测任务。\n",
"尽管模型产出的困惑度是合理的,\n",
"该模型预测未来词元的能力却可能存在严重缺陷。\n",
"我们用下面的示例代码引以为戒,以防在错误的环境中使用它们。\n"
]
},
{
"cell_type": "code",
"execution_count": 1,
"id": "40b9f7d4",
"metadata": {
"execution": {
"iopub.execute_input": "2023-08-18T07:18:55.992113Z",
"iopub.status.busy": "2023-08-18T07:18:55.991547Z",
"iopub.status.idle": "2023-08-18T07:19:45.366029Z",
"shell.execute_reply": "2023-08-18T07:19:45.365163Z"
},
"origin_pos": 2,
"tab": [
"pytorch"
]
},
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"perplexity 1.1, 131129.2 tokens/sec on cuda:0\n",
"time travellerererererererererererererererererererererererererer\n",
"travellerererererererererererererererererererererererererer\n"
]
},
{
"data": {
"image/svg+xml": [
"\n",
"\n",
"\n"
],
"text/plain": [
""
]
},
"metadata": {
"needs_background": "light"
},
"output_type": "display_data"
}
],
"source": [
"import torch\n",
"from torch import nn\n",
"from d2l import torch as d2l\n",
"\n",
"# 加载数据\n",
"batch_size, num_steps, device = 32, 35, d2l.try_gpu()\n",
"train_iter, vocab = d2l.load_data_time_machine(batch_size, num_steps)\n",
"# 通过设置“bidirective=True”来定义双向LSTM模型\n",
"vocab_size, num_hiddens, num_layers = len(vocab), 256, 2\n",
"num_inputs = vocab_size\n",
"lstm_layer = nn.LSTM(num_inputs, num_hiddens, num_layers, bidirectional=True)\n",
"model = d2l.RNNModel(lstm_layer, len(vocab))\n",
"model = model.to(device)\n",
"# 训练模型\n",
"num_epochs, lr = 500, 1\n",
"d2l.train_ch8(model, train_iter, vocab, lr, num_epochs, device)"
]
},
{
"cell_type": "markdown",
"id": "0ffafe68",
"metadata": {
"origin_pos": 4
},
"source": [
"上述结果显然令人瞠目结舌。\n",
"关于如何更有效地使用双向循环神经网络的讨论,\n",
"请参阅 :numref:`sec_sentiment_rnn`中的情感分类应用。\n",
"\n",
"## 小结\n",
"\n",
"* 在双向循环神经网络中,每个时间步的隐状态由当前时间步的前后数据同时决定。\n",
"* 双向循环神经网络与概率图模型中的“前向-后向”算法具有相似性。\n",
"* 双向循环神经网络主要用于序列编码和给定双向上下文的观测估计。\n",
"* 由于梯度链更长,因此双向循环神经网络的训练代价非常高。\n",
"\n",
"## 练习\n",
"\n",
"1. 如果不同方向使用不同数量的隐藏单位,$\\mathbf{H_t}$的形状会发生怎样的变化?\n",
"1. 设计一个具有多个隐藏层的双向循环神经网络。\n",
"1. 在自然语言中一词多义很常见。例如,“bank”一词在不同的上下文“i went to the bank to deposit cash”和“i went to the bank to sit down”中有不同的含义。如何设计一个神经网络模型,使其在给定上下文序列和单词的情况下,返回该单词在此上下文中的向量表示?哪种类型的神经网络架构更适合处理一词多义?\n"
]
},
{
"cell_type": "markdown",
"id": "f30b84f3",
"metadata": {
"origin_pos": 6,
"tab": [
"pytorch"
]
},
"source": [
"[Discussions](https://discuss.d2l.ai/t/2773)\n"
]
}
],
"metadata": {
"language_info": {
"name": "python"
},
"required_libs": []
},
"nbformat": 4,
"nbformat_minor": 5
}