论文标题
用于几个射击模拟电路建模和设计的图表神经网络
Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling and Design
论文作者
论文摘要
能够在不运行昂贵模拟的情况下预测电路的性能是可以催化自动设计的所需功能。在本文中,我们提出了一种有监督的预处理方法,以学习可以适应新电路拓扑或看不见的预测任务的电路表示。我们假设,如果我们训练一个可以预测广泛电路实例的输出直流电压的神经网络(NN),它将被迫学习有关每个电路元素的作用以及它们如何相互作用的可推广知识。由于获得地面真相标签所需的直流模拟相对便宜,因此可以轻松地按大规模收集此监督学习目标的数据集。然后,该表示形式将有助于几次概括,以使观察到的电路指标需要更多耗时的模拟来获得地面真相标签。为了应对不同电路的可变拓扑结构,我们将每个电路描述为图形,并使用图形神经网络(GNN)学习节点嵌入。我们表明,与随机初始初始化的模型相比,对输出节点电压预测预测的GNN可以鼓励可以适应新的看不见的拓扑或预测新电路级属性的新电路水平效率,这些效率高出10倍。我们进一步表明,我们可以通过2倍(几乎与使用Oracle模型一样好)将基于SOTA模型的优化方法提高样本效率,这是通过预算的GNN作为学习模型的特征提取器的样本效率。
Being able to predict the performance of circuits without running expensive simulations is a desired capability that can catalyze automated design. In this paper, we present a supervised pretraining approach to learn circuit representations that can be adapted to new circuit topologies or unseen prediction tasks. We hypothesize that if we train a neural network (NN) that can predict the output DC voltages of a wide range of circuit instances it will be forced to learn generalizable knowledge about the role of each circuit element and how they interact with each other. The dataset for this supervised learning objective can be easily collected at scale since the required DC simulation to get ground truth labels is relatively cheap. This representation would then be helpful for few-shot generalization to unseen circuit metrics that require more time consuming simulations for obtaining the ground-truth labels. To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings. We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties with up to 10x more sample efficiency compared to a randomly initialized model. We further show that we can improve sample efficiency of prior SoTA model-based optimization methods by 2x (almost as good as using an oracle model) via fintuning pretrained GNNs as the feature extractor of the learned models.