论文标题
汇总学习:一种学习神经网络分类器的矢量定量方法
Aggregated Learning: A Vector-Quantization Approach to Learning Neural Network Classifiers
论文作者
论文摘要
我们考虑学习神经网络分类器的问题。在信息瓶颈(IB)原则下,我们与此分类问题相关联是一个表示学习问题,我们称之为“ IB学习”。我们表明,IB学习实际上等同于量化问题的特殊类别。然后,速率延伸理论的经典结果表明,IB学习可以从“向量量化”方法中受益,即同时学习多个输入对象的表示。这种方法有助于一些变分技术,从而导致一个新颖的学习框架“汇总学习”,用于与神经网络模型进行分类。在此框架中,几个对象由单个神经网络共同分类。通过对标准图像识别和文本分类任务进行广泛的实验来验证该框架的有效性。
We consider the problem of learning a neural network classifier. Under the information bottleneck (IB) principle, we associate with this classification problem a representation learning problem, which we call "IB learning". We show that IB learning is, in fact, equivalent to a special class of the quantization problem. The classical results in rate-distortion theory then suggest that IB learning can benefit from a "vector quantization" approach, namely, simultaneously learning the representations of multiple input objects. Such an approach assisted with some variational techniques, result in a novel learning framework, "Aggregated Learning", for classification with neural network models. In this framework, several objects are jointly classified by a single neural network. The effectiveness of this framework is verified through extensive experiments on standard image recognition and text classification tasks.