论文标题

深度神经网络在食用更多数据时总是表现更好吗?

Do Deep Neural Networks Always Perform Better When Eating More Data?

论文作者

Yang, Jiachen, Zhang, Zhuo, Gong, Yicheng, Ma, Shukun, Guo, Xiaolan, Yang, Yue, Xiao, Shuai, Wen, Jiabao, Li, Yang, Gao, Xinbo, Lu, Wen, Meng, Qinggang

论文摘要

现在,数据已成为深度学习的缺点。自己领域的研究人员分享了“深层神经网络在吃更多的数据时可能并不总是更好”,这仍然缺乏实验验证和令人信服的指导理论。为了填补这一缺乏,我们从相同独立的分布(IID)和无分布(OOD)设计实验,这给出了有力的答案。为了指导的目的,根据结果的讨论,提出了两种理论:在IID条件下,信息量决定了每个样本的有效性,样本的贡献和班级之间的差异决定了样本信息的数量和类信息的量;在OOD条件下,样品的跨域程度决定了贡献,而由不相关的元素引起的偏置是跨域的重要因素。以上理论从数据的角度提供了指导,这可以促进人工智能的广泛实用应用。

Data has now become a shortcoming of deep learning. Researchers in their own fields share the thinking that "deep neural networks might not always perform better when they eat more data," which still lacks experimental validation and a convincing guiding theory. Here to fill this lack, we design experiments from Identically Independent Distribution(IID) and Out of Distribution(OOD), which give powerful answers. For the purpose of guidance, based on the discussion of results, two theories are proposed: under IID condition, the amount of information determines the effectivity of each sample, the contribution of samples and difference between classes determine the amount of sample information and the amount of class information; under OOD condition, the cross-domain degree of samples determine the contributions, and the bias-fitting caused by irrelevant elements is a significant factor of cross-domain. The above theories provide guidance from the perspective of data, which can promote a wide range of practical applications of artificial intelligence.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源