论文标题

两位塔语言模型的自发新兴偏好

Spontaneous Emerging Preference in Two-tower Language Model

论文作者

He, Zhengqi, Toyoizumi, Taro

论文摘要

基础语言模型的不断增长的规模在各种下游任务中带来了显着的性能增长。由于存在大型基础语言模型(例如部署成本,可用性问题和环境成本)所带来的副作用,因此有一些兴趣探索其他可能的方向,例如分裂和构成方案。在本文中,我们提出了一个基本问题:语言过程自然可以分配吗?我们通过简单的两个较高语言模型设置研究了这个问题,其中两个具有相同配置的语言模型是并排训练的。在这种环境中,我们发现了自发的新兴偏好现象,其中一些令牌始终被一座塔更好地预测,而另一些则由另一个塔楼。无论模型配置和类型如何,这种现象在质量上都是稳定的,这表明这是自然语言的内在特性。这项研究表明,自然语言的有趣特性仍在等待被发现,这可能有助于自然语言处理技术的未来发展。

The ever-growing size of the foundation language model has brought significant performance gains in various types of downstream tasks. With the existence of side-effects brought about by the large size of the foundation language model such as deployment cost, availability issues, and environmental cost, there is some interest in exploring other possible directions, such as a divide-and-conquer scheme. In this paper, we are asking a basic question: are language processes naturally dividable? We study this problem with a simple two-tower language model setting, where two language models with identical configurations are trained side-by-side cooperatively. With this setting, we discover the spontaneous emerging preference phenomenon, where some of the tokens are consistently better predicted by one tower while others by another tower. This phenomenon is qualitatively stable, regardless of model configuration and type, suggesting this as an intrinsic property of natural language. This study suggests that interesting properties of natural language are still waiting to be discovered, which may aid the future development of natural language processing techniques.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源