论文标题
使用量子退火器使用特征选择的优势
An Advantage Using Feature Selection with a Quantum Annealer
论文作者
论文摘要
特征选择是统计预测建模中的一种技术,它标识了与目标变量具有牢固统计连接的记录中的特征。在训练中,与训练中的统计变量较弱的特征不仅降低了数据的尺寸,从而降低了算法的时间复杂性,还可以降低数据中噪声,从而有助于避免过度拟合。总的来说,特征选择有助于培训良好的统计模型,该模型的性能良好并且稳定。鉴于经典计算缺乏可扩展性,当前技术仅考虑功能的预测能力,而不是功能本身之间的冗余。利用量子退火(QA)的功能选择的最新进步提供了一种可扩展的技术,旨在最大程度地提高特征的预测能力,同时最大程度地减少冗余。结果,预计该算法将有助于偏见/差异权衡,从而为培训统计模型提供更好的功能。本文通过利用开源数据集并评估每个训练有素的统计模型众所周知的预测算法的功效来测试这种直觉,并评估了这种直觉。数值结果利用利用QA的算法选择的功能显示了优势。
Feature selection is a technique in statistical prediction modeling that identifies features in a record with a strong statistical connection to the target variable. Excluding features with a weak statistical connection to the target variable in training not only drops the dimension of the data, which decreases the time complexity of the algorithm, it also decreases noise within the data which assists in avoiding overfitting. In all, feature selection assists in training a robust statistical model that performs well and is stable. Given the lack of scalability in classical computation, current techniques only consider the predictive power of the feature and not redundancy between the features themselves. Recent advancements in feature selection that leverages quantum annealing (QA) gives a scalable technique that aims to maximize the predictive power of the features while minimizing redundancy. As a consequence, it is expected that this algorithm would assist in the bias/variance trade-off yielding better features for training a statistical model. This paper tests this intuition against classical methods by utilizing open-source data sets and evaluate the efficacy of each trained statistical model well-known prediction algorithms. The numerical results display an advantage utilizing the features selected from the algorithm that leveraged QA.