强化学习
计算机科学
启发式
水准点(测量)
序列(生物学)
图形
理论计算机科学
人工智能
算法
大地测量学
遗传学
生物
操作系统
地理
作者
Cheng-Hao Yang,Yinshui Xia,Zhufei Chu,Xiaojing Zha
出处
期刊:IEEE Transactions on Circuits and Systems Ii-express Briefs
[Institute of Electrical and Electronics Engineers]
日期:2022-04-19
卷期号:69 (8): 3600-3604
被引量:14
标识
DOI:10.1109/tcsii.2022.3168344
摘要
As a key step in the IC design flow, logic synthesis involves various logic optimization algorithms to be iteratively applied to the circuit. However, how these algorithms are used is usually determined by heuristics, and it does not always yield well optimizations on all circuits. To achieve well optimized results, engineers need to tune the sequence consisting of these logic optimization algorithms based on their knowledge. To overcome this limitation, in this brief, reinforcement learning (RL) proximal policy optimization (PPO) is proposed to train an agent to tune the optimization sequence. Specifically, graph isomorphic network with edge feature aggregation capability (GINE) is used to learn circuit representations and use circuit representations as state representations for the reinforcement learning agent. Furthermore, to enable the agent learning from historical operations, the Long Short-Term Memory (LSTM) is further embedded in reinforcement learning. The evaluation of the EPFL arithmetic benchmark shows that our model improves the area optimization under the delay constraint by 21.21% over existing methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI