论文标题
转移熵瓶颈:序列信息传递的学习顺序
Transfer Entropy Bottleneck: Learning Sequence to Sequence Information Transfer
论文作者
论文摘要
当呈现两个统计依赖性变量的数据流时,预测其中一个变量(目标流)的未来可以从有关其历史记录和其他变量历史的信息(源流)(源流)中受益。例如,可以使用温度和气压读数来预测气象站温度的波动。但是,建模此类数据时的挑战是,神经网络很容易依靠目标流中最大的关节相关性,这可能会忽略从源到源到目标流的至关重要但小的信息传输。同样,在某些情况下,目标流可能以前是独立建模的,并且使用该模型为新的联合模型提供了信息。在这里,我们开发了一种信息瓶颈方法,用于在两个依赖数据流上有条件学习。我们称之为转移熵瓶颈(TEB)的方法允许人们学习一个模型,该模型可以将从源变量传输到目标变量传递的有向信息,同时量化模型中的此信息传输。因此,TEB提供了一种有用的新信息瓶颈方法,用于建模两个统计依赖的数据流,以便对其中一个进行预测。
When presented with a data stream of two statistically dependent variables, predicting the future of one of the variables (the target stream) can benefit from information about both its history and the history of the other variable (the source stream). For example, fluctuations in temperature at a weather station can be predicted using both temperatures and barometric readings. However, a challenge when modelling such data is that it is easy for a neural network to rely on the greatest joint correlations within the target stream, which may ignore a crucial but small information transfer from the source to the target stream. As well, there are often situations where the target stream may have previously been modelled independently and it would be useful to use that model to inform a new joint model. Here, we develop an information bottleneck approach for conditional learning on two dependent streams of data. Our method, which we call Transfer Entropy Bottleneck (TEB), allows one to learn a model that bottlenecks the directed information transferred from the source variable to the target variable, while quantifying this information transfer within the model. As such, TEB provides a useful new information bottleneck approach for modelling two statistically dependent streams of data in order to make predictions about one of them.