强化学习
计算机科学
路径(计算)
参数统计
钢筋
箔法
运动规划
人工智能
材料科学
机器人
数学
计算机网络
统计
复合材料
作者
Z.J. Wang,Rongping Lin,Zhijie Zhao,Xi Chen,Pengming Guo,Ning Yang,Zhechao Wang,Dixia Fan
标识
DOI:10.1017/jfm.2023.1096
摘要
To optimize flapping foil performance, in the current study we apply deep reinforcement learning (DRL) to plan foil non-parametric motion, as the traditional control techniques and simplified motions cannot fully model nonlinear, unsteady and high-dimensional foil–vortex interactions. Therefore, a DRL training framework is proposed based on the proximal policy optimization algorithm and the transformer architecture, where the policy is initialized from the sinusoidal expert display. We first demonstrate the effectiveness of the proposed DRL-training framework, learning the coherent foil flapping motion to generate thrust. Furthermore, by adjusting reward functions and action thresholds, DRL-optimized foil trajectories can gain significant enhancement in both thrust and efficiency compared with the sinusoidal motion. Last, through visualization of wake morphology and instantaneous pressure distributions, it is found that DRL-optimized foil can adaptively adjust the phases between motion and shedding vortices to improve hydrodynamic performance. Our results give a hint of how to solve complex fluid manipulation problems using the DRL method.
科研通智能强力驱动
Strongly Powered by AbleSci AI