论文标题
Harl:基于神经网络的基于层次的自适应增强学习计划
HARL: Hierarchical Adaptive Reinforcement Learning Based Auto Scheduler for Neural Networks
论文作者
论文摘要
为了有效地使用神经网络进行推断,基础张量程序需要足够的调整工作,然后才能部署到生产环境中。通常,需要充分探索巨大的张量候选者,以找到表现最好的候选者。这对于使神经网络产品满足现实世界应用的高需求,例如自然语言处理,自动驾驶等。但是,由于较大的搜索空间和缺乏智能搜索指南,当前的自动安排人员需要数小时到数天的调整时间,以找到整个神经网络的表现最佳的张量程序。 在本文中,我们提出了基于加固学习(RL)的自动安装仪Harl专为有效张量程序探索而设计的。 Harl使用层次RL体系结构,其中在各个不同级别的搜索粒度上做出基于学习的决策。它还可以自动实时调整勘探配置,以更快地融合性能。结果,与最先进的自动安排机相比,Harl将张量操作员的性能提高了22%,搜索速度提高了4.3倍。推理性能和搜索速度在端到端神经网络上也得到了显着提高。
To efficiently perform inference with neural networks, the underlying tensor programs require sufficient tuning efforts before being deployed into production environments. Usually, enormous tensor program candidates need to be sufficiently explored to find the one with the best performance. This is necessary to make the neural network products meet the high demand of real-world applications such as natural language processing, auto-driving, etc. Auto-schedulers are being developed to avoid the need for human intervention. However, due to the gigantic search space and lack of intelligent search guidance, current auto-schedulers require hours to days of tuning time to find the best-performing tensor program for the entire neural network. In this paper, we propose HARL, a reinforcement learning (RL) based auto-scheduler specifically designed for efficient tensor program exploration. HARL uses a hierarchical RL architecture in which learning-based decisions are made at all different levels of search granularity. It also automatically adjusts exploration configurations in real-time for faster performance convergence. As a result, HARL improves the tensor operator performance by 22% and the search speed by 4.3x compared to the state-of-the-art auto-scheduler. Inference performance and search speed are also significantly improved on end-to-end neural networks.