论文标题
神经添加剂模型:神经网的可解释的机器学习
Neural Additive Models: Interpretable Machine Learning with Neural Nets
论文作者
论文摘要
深神经网络(DNN)是强大的黑盒预测指标,在各种任务上都取得了令人印象深刻的表现。但是,它们的准确性是以可理解为代价的:通常不清楚他们是如何做出决定的。这阻碍了它们适用于医疗保健等高利益决策领域的适用性。我们提出了神经添加剂模型(NAM),将DNN的某些表达性与广义添加剂模型的固有清晰度结合在一起。 NAMS学习了每个参与单个输入功能的神经网络的线性组合。这些网络是共同训练的,可以学习其输入功能与输出之间任意复杂的关系。我们关于回归和分类数据集的实验表明,NAM比广泛使用的可理解模型(例如逻辑回归和浅决策树)更准确。它们的性能类似于准确的现有最新的广义添加剂模型,但更灵活,因为它们基于神经网,而不是增强的树木。为了证明这一点,我们展示了如何将NAM用于合成数据以及由于其合成性而在综合数据和Compas累犯数据上学习,并证明了NAM的不同性能使他们能够为Covid-19训练更复杂的可解释模型。
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks. However, their accuracy comes at the cost of intelligibility: it is usually unclear how they make their decisions. This hinders their applicability to high stakes decision-making domains such as healthcare. We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models. NAMs learn a linear combination of neural networks that each attend to a single input feature. These networks are trained jointly and can learn arbitrarily complex relationships between their input feature and the output. Our experiments on regression and classification datasets show that NAMs are more accurate than widely used intelligible models such as logistic regression and shallow decision trees. They perform similarly to existing state-of-the-art generalized additive models in accuracy, but are more flexible because they are based on neural nets instead of boosted trees. To demonstrate this, we show how NAMs can be used for multitask learning on synthetic data and on the COMPAS recidivism data due to their composability, and demonstrate that the differentiability of NAMs allows them to train more complex interpretable models for COVID-19.