论文标题
任意大型标记为机器学习培训的随机满足公式
Arbitrarily Large Labelled Random Satisfiability Formulas for Machine Learning Training
论文作者
论文摘要
应用深度学习来解决硬组合问题的现实实例具有巨大的潜力。朝这个方向进行的研究集中在布尔满意度(SAT)问题上,这是由于其理论中心性和实际重要性。但是,面对的主要障碍是,训练集仅限于大小的随机公式,比实际兴趣的公式小几个数量级,这引起了人们对概括的严重关注。这是因为标记增加大小的随机公式会迅速棘手。通过以基本方式利用概率方法,我们完全删除了此障碍:我们展示了如何正确标记任何所需大小的随机公式,而无需解决基本的决策问题。此外,通过改变简单的标量参数,可以调整分类任务的分类任务对我们发电机产生的公式的难度。这为机器学习方法开辟了一个全新的复杂水平,可以使人满意。使用我们的发电机,我们训练现有的最新模型,以预测10,000个变量的公式。我们发现它们没有比随机猜测更好。作为新发电机可以实现的目标,我们提出了一个新颖的分类器,该分类器的性能要比随机猜测在同一数据集上的99%要好得多。至关重要的是,与基于公式的句法特征学习的过去方法不同,我们的分类器在求解器计算的简短前缀上进行学习,这是我们期望具有独立兴趣的方法。
Applying deep learning to solve real-life instances of hard combinatorial problems has tremendous potential. Research in this direction has focused on the Boolean satisfiability (SAT) problem, both because of its theoretical centrality and practical importance. A major roadblock faced, though, is that training sets are restricted to random formulas of size several orders of magnitude smaller than formulas of practical interest, raising serious concerns about generalization. This is because labeling random formulas of increasing size rapidly becomes intractable. By exploiting the probabilistic method in a fundamental way, we remove this roadblock entirely: we show how to generate correctly labeled random formulas of any desired size, without having to solve the underlying decision problem. Moreover, the difficulty of the classification task for the formulas produced by our generator is tunable by varying a simple scalar parameter. This opens up an entirely new level of sophistication for the machine learning methods that can be brought to bear on Satisfiability. Using our generator, we train existing state-of-the-art models for the task of predicting satisfiability on formulas with 10,000 variables. We find that they do no better than random guessing. As a first indication of what can be achieved with the new generator, we present a novel classifier that performs significantly better than random guessing 99% on the same datasets, for most difficulty levels. Crucially, unlike past approaches that learn based on syntactic features of a formula, our classifier performs its learning on a short prefix of a solver's computation, an approach that we expect to be of independent interest.