CN115618497A - An airfoil optimization design method based on deep reinforcement learning - Google Patents
An airfoil optimization design method based on deep reinforcement learning Download PDFInfo
- Publication number
- CN115618497A CN115618497A CN202211374735.1A CN202211374735A CN115618497A CN 115618497 A CN115618497 A CN 115618497A CN 202211374735 A CN202211374735 A CN 202211374735A CN 115618497 A CN115618497 A CN 115618497A
- Authority
- CN
- China
- Prior art keywords
- reward
- airfoil
- value
- model
- optimization
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/10—Geometric CAD
- G06F30/15—Vehicle, aircraft or watercraft design
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/28—Design optimisation, verification or simulation using fluid dynamics, e.g. using Navier-Stokes equations or computational fluid dynamics [CFD]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2111/00—Details relating to CAD techniques
- G06F2111/04—Constraint-based CAD
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2111/00—Details relating to CAD techniques
- G06F2111/10—Numerical modelling
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2113/00—Details relating to the application field
- G06F2113/08—Fluids
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2119/00—Details relating to the type or aim of the analysis or the optimisation
- G06F2119/14—Force analysis or force optimisation, e.g. static or dynamic forces
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T90/00—Enabling technologies or technologies with a potential or indirect contribution to GHG emissions mitigation
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Geometry (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Optimization (AREA)
- Computing Systems (AREA)
- Mathematical Analysis (AREA)
- Artificial Intelligence (AREA)
- Fluid Mechanics (AREA)
- Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Algebra (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Automation & Control Theory (AREA)
- Aviation & Aerospace Engineering (AREA)
- Computational Mathematics (AREA)
- Structures Of Non-Positive Displacement Pumps (AREA)
- Feedback Control In General (AREA)
Abstract
本发明提出了一种基于深度强化学习的翼型优化设计方法,该方法不同于监督性学习,而是有着自主学习策略,长期奖励最大的特性,是一种更接近智能化的优化方法,且具有可迁移性以及策略的普适性。若是在一定范围内改变设计条件,例如来流马赫数、雷诺数等,原来优化得到的策略依旧能够提供初始优化方向,优化目标在很短的步数内就可以有明显的提升。
The present invention proposes an airfoil optimization design method based on deep reinforcement learning. This method is different from supervised learning, but has the characteristics of independent learning strategy and the largest long-term reward. It is an optimization method closer to intelligence, and It has the portability and the universality of the strategy. If the design conditions are changed within a certain range, such as incoming flow Mach number, Reynolds number, etc., the strategy obtained from the original optimization can still provide the initial optimization direction, and the optimization goal can be significantly improved within a very short number of steps.
Description
技术领域technical field
本发明属于飞机设计领域,提出一种基于深度强化学习的翼型优化设计方法。The invention belongs to the field of aircraft design, and proposes an airfoil optimization design method based on deep reinforcement learning.
背景技术Background technique
翼型作为飞行器的主要组成部分,不仅提升飞行所需的升力,还保证飞行器的安定性和操作性。随着对翼型气动性能的了解与探索,多个翼型库也相应建立。目前尚不存在一种完全满足所有设计状态的超级翼型,需要根据设计用途和工况等因素,从翼型库中初步选择符合要求的基准翼型,基于基准翼型,结合设计要求不断改进基准翼型。对于改进基准翼型,最初的方法是使用风洞进行重复实验,这会消耗巨大的物力人力。As the main component of the aircraft, the airfoil not only improves the lift required for flight, but also ensures the stability and operability of the aircraft. With the understanding and exploration of the aerodynamic performance of the airfoil, multiple airfoil libraries have also been established accordingly. At present, there is no super airfoil that fully satisfies all design states. It is necessary to initially select a benchmark airfoil that meets the requirements from the airfoil library based on factors such as design use and working conditions. Based on the benchmark airfoil, it is continuously improved in combination with design requirements. base airfoil. For improving the reference airfoil, the initial method is to use the wind tunnel to conduct repeated experiments, which will consume huge material and manpower.
进入21世纪以来,科学技术的进步为计算流体力学(CFD)技术注入了新的活力,CFD技术很快成为了分析和解决流体问题的主要手段。使用计算流体力学技术可以大幅缩短优化周期,节省大量的人力物力,由此可以让设计师进行多次重复计算,直至达到获得预期的翼型,例如进化和遗传算法在气动优化问题的广泛应用。然而进化和遗传算法对优化过程中产生的大量计算数据的利用率较低。Since the beginning of the 21st century, the advancement of science and technology has injected new vitality into Computational Fluid Dynamics (CFD) technology, and CFD technology has quickly become the main means of analyzing and solving fluid problems. The use of computational fluid dynamics technology can greatly shorten the optimization cycle and save a lot of manpower and material resources, which allows designers to perform repeated calculations until the desired airfoil is achieved, such as the extensive application of evolutionary and genetic algorithms in aerodynamic optimization problems. However, the evolutionary and genetic algorithms have a low utilization rate for the large amount of computational data generated in the optimization process.
机器学习在流体力学中的应用在近几年内飞速发展,目前机器学习在优化设计中最为常见的是基于响应面的性能快速预测方法,利用已有数据构建输入输出之间的映射加速优化过程,部分取代CFD技术在优化过程中担任的“角色”,属于监督性学习。而与监督性学习不同,强化学习构建的是环境的状态参数到动作参数的映射,其目的是迭代更新与环境交互的策略以获得最大累积奖励。此时不同于监督学习中要求预测的真实输出,而是对应到某一动作的奖励。翼型设计问题并不是单独追求某个性能的最优,而是在反复权衡比较之后的整体性能最优,是一项关联复杂的系统工程,往往需要结合工程师自身的经验和理解进行手动修型。而强化学习的训练过程与工程师积累经验的过程十分相似。类比于传统的试错法,强化学习中负责做出动作的智能体通过训练过程中持续地采取不同的“动作”并观察设计结果在未来一段时间内(或采取一系列动作之后)的收益,更新采取动作的策略。随着收益的增加,可以认为该智能体在一定程度上获得与工程师相同的设计经验。在翼型优化方面,李润泽使用强化学习对翼型压力分布进行优化以减小跨声速翼型的阻力,Viquerat进行了带约束和不带约束的优化尝试,强化学习在翼型优化上仍处于起步阶段,存在着智能体找不到方向、陷入局部最优的可能性。总的来说,在飞行器气动优化设计中将强化学习应用到翼型优化中可以有效提高优化效率,具有广泛的应用场景。The application of machine learning in fluid mechanics has developed rapidly in recent years. At present, the most common method of machine learning in optimization design is the rapid performance prediction method based on response surface, which uses existing data to construct the mapping between input and output to speed up the optimization process. Partially replacing the "role" played by CFD technology in the optimization process belongs to supervised learning. Unlike supervised learning, reinforcement learning builds a mapping from the state parameters of the environment to action parameters, and its purpose is to iteratively update the strategy for interacting with the environment to obtain the maximum cumulative reward. At this time, it is different from the real output that is required to be predicted in supervised learning, but corresponds to the reward of a certain action. The problem of airfoil design is not to pursue the optimum of a certain performance alone, but to achieve the optimum of the overall performance after repeated weighing and comparison. It is a complex system engineering, which often needs to be manually modified by combining the experience and understanding of the engineer himself. . The training process of reinforcement learning is very similar to the process of engineers accumulating experience. Analogous to the traditional trial-and-error method, the agent responsible for making actions in reinforcement learning continuously takes different "actions" during the training process and observes the benefits of the design results for a period of time in the future (or after taking a series of actions), Update the policy to take action. As the payoff increases, the agent can be considered to gain the same design experience as the engineer to a certain extent. In terms of airfoil optimization, Li Runze used reinforcement learning to optimize the pressure distribution of the airfoil to reduce the resistance of the transonic airfoil. Viquerat conducted optimization attempts with constraints and without constraints. In the initial stage, there is a possibility that the agent cannot find its direction and falls into a local optimum. In general, applying reinforcement learning to airfoil optimization in aircraft aerodynamic optimization design can effectively improve optimization efficiency and has a wide range of application scenarios.
发明内容Contents of the invention
由于当前基于CFD的方法优化效率较低,为提升翼型优化效率,本发明提出了一种基于深度强化学习的翼型优化设计方法,该方法不同于监督性学习,而是有着自主学习策略,长期奖励最大的特性,是一种更接近智能化的优化方法,且具有可迁移性,或者说策略的普适性。Due to the low optimization efficiency of the current CFD-based method, in order to improve the efficiency of airfoil optimization, the present invention proposes an airfoil optimization design method based on deep reinforcement learning. This method is different from supervised learning, but has an independent learning strategy. The biggest feature of long-term rewards is a more intelligent optimization method with transferability, or the universality of strategies.
本发明的技术方案为:Technical scheme of the present invention is:
所述一种基于深度强化学习的翼型优化设计方法,包括以下步骤:Described a kind of airfoil optimization design method based on depth reinforcement learning, comprises the following steps:
步骤1:用自由曲面变形法进行翼型的几何参数化:在基准翼型周围建立自由曲面变形控制框,建立控制框与翼型的映射关系,通过更改控制框点的位置,得到新的翼型;Step 1: Use the free-form surface deformation method to parameterize the geometry of the airfoil: establish a free-form surface deformation control frame around the reference airfoil, establish the mapping relationship between the control frame and the airfoil, and obtain a new airfoil by changing the position of the control frame points type;
步骤2:建立优化设计模型,根据飞行要求,确认单设计目标和约束条件,其中设计目标和约束条件为翼型的各种气动参数,例如升力系数、阻力系数和翼型厚度等,将设计目标和约束条件用数学表达式表示。一般的单目标优化问题可以写成以下数学形式:Step 2: Establish an optimized design model, and confirm the single design goals and constraints according to the flight requirements, where the design goals and constraints are various aerodynamic parameters of the airfoil, such as lift coefficient, drag coefficient and airfoil thickness, etc., and the design goals and constraints expressed in mathematical expressions. A general single-objective optimization problem can be written in the following mathematical form:
Minimize:f(x)Minimize: f(x)
subject to:gw(x)≥0,w=1,2,···,Wsubject to:g w (x)≥0,w=1,2,···,W
hr(x)=0,r=1,2,···,Rh r (x)=0,r=1,2,...,R
其中x为设计变量,f(x)为目标函数,gw(x)为不等式约束,共W个;hr(x)为等式约束,共R个。Among them, x is the design variable, f(x) is the objective function, g w (x) is the inequality constraint, there are W in total; h r (x) is the equality constraint, there are R in total.
步骤3:根据优化目标和约束条件建立奖励函数,其中总奖励值是由各个气动参数奖励值线性和得到,其中达到目标增加奖励值,满足约束不增加奖励值,不满足约束减小奖励值,同时在目标奖励值和约束奖励值乘以不同的系数用以平衡目标和约束之间的量级差异。Step 3: Establish a reward function according to the optimization objective and constraint conditions, where the total reward value is obtained by the linear sum of the reward values of each aerodynamic parameter, where the reward value is increased when the target is reached, the reward value is not increased if the constraint is satisfied, and the reward value is reduced if the constraint is not satisfied. At the same time, the target reward value and constraint reward value are multiplied by different coefficients to balance the magnitude difference between the target and constraint.
步骤4:建立智能体,包含策略模型π和价值函数模型,策略函数模型可以输出动作策略,而价值函数模型可以输出优势估计和价值函数,策略模型和价值函数模型均使用含有两层隐藏层的人工神经网络模型,隐藏层节点数为64。初始化策略模型参数θ0和价值函数模型参数将翼型中有关设计目标和约束条件的气动参数作为状态,包括翼型的升力系数、阻力系数、最大厚度和力矩系数。将基准机翼的气动参数作为初始状态s0。Step 4: Establish an agent, including a policy model π and a value function model. The policy function model can output an action strategy, and the value function model can output an advantage estimate and a value function. Both the policy model and the value function model use a two-layer hidden layer Artificial neural network model, the number of hidden layer nodes is 64. Initialize policy model parameters θ 0 and value function model parameters The aerodynamic parameters related to design goals and constraints in the airfoil are taken as the state, including the lift coefficient, drag coefficient, maximum thickness and moment coefficient of the airfoil. The aerodynamic parameters of the reference wing are taken as the initial state s 0 .
步骤5:当前智能体由策略模型根据状态和奖励值给出动作a,也就是新的设计变量。Step 5: The current agent is given an action a by the policy model according to the state and reward value, which is the new design variable.
步骤6:对翼型实施动作得到新翼型。Step 6: Perform actions on the airfoil to get a new airfoil.
步骤7:对新翼型建立结构化网格,使用开源求解器CFL3D进行翼型绕流数值模拟,主控方程为雷诺平均N-S方程,湍流模型为k-ωSST模型,计算得到翼型的升力系数、阻力系数、最大厚度和力矩系数,确认新翼型的气动参数为新状态。Step 7: Establish a structured grid for the new airfoil, use the open source solver CFL3D to perform numerical simulation of the flow around the airfoil, the main control equation is the Reynolds average N-S equation, the turbulence model is the k-ωSST model, and the lift coefficient of the airfoil is calculated , drag coefficient, maximum thickness and moment coefficient, confirm that the aerodynamic parameters of the new airfoil are in the new state.
步骤8:由计算得到的气动参数根据步骤3中的奖励函数计算得到奖励值。Step 8: Calculate the reward value from the calculated aerodynamic parameters according to the reward function in
步骤9:由当前策略模型循环重复步骤5-8共e-1次,得到包含每次循环得到的状态和动作的轨迹和奖励值{re},轨迹τ={s0,a0,···,se-1,ae-1,se},其中s0和a0为初始状态和动作,se为e步状态,se-1为e-1步状态,ae-1为e-1步动作。Step 9: Repeat steps 5-8 for a total of e-1 times from the current policy model to obtain the trajectory and reward value {r e } including the state and action obtained in each cycle, trajectory τ={s 0 ,a 0 ,· ··,s e-1 ,a e-1 ,s e }, where s 0 and a 0 are the initial state and action, s e is the e-step state, s e-1 is the e-1 step state, a e- 1 is e-1 step action.
步骤10:基于当前策略模型重复步骤5-9共n-1次,得到n个的轨迹和奖励值。Step 10: Repeat steps 5-9 a total of n-1 times based on the current policy model to obtain n trajectories and reward values.
步骤11:根据得到的n条轨迹参数和奖励值,基于当前智能体的价值函数模型计算优势估计,也就是每一个动作a的期望奖励与该状态下所有可能的动作的期望奖励的均值的差值。Step 11: According to the obtained n trajectory parameters and reward values, calculate the advantage estimate based on the value function model of the current agent, that is, the difference between the expected reward of each action a and the mean value of the expected reward of all possible actions in this state value.
步骤12:根据优势估计、轨迹和奖励值构建损失函数,基于随机梯度下降算法优化策略模型参数θ和价值函数模型参数优化目标为损失函数最小,用优化后的参数更新策略模型和价值函数模型,得到新的策略模型和价值函数模型。Step 12: Construct a loss function based on the advantage estimate, trajectory and reward value, and optimize the policy model parameters θ and value function model parameters based on the stochastic gradient descent algorithm The optimization goal is to minimize the loss function, and the optimized parameters are used to update the strategy model and value function model to obtain a new strategy model and value function model.
步骤13:循环重复步骤5-12直至达到损失函数不再降低,完成训练。Step 13: Repeat steps 5-12 until the loss function is no longer reduced, and the training is completed.
有益效果Beneficial effect
1.本发明提出的基于深度强化学习的翼型优化设计方法,有自主学习策略和长期奖励最大的特性,相较遗传算法等算法仅仅将大量计算数据作为优化目标和约束的评价标准,深度强化学习尝试学习优化过程中得到的经验,可以提高数据的利用效率,减小计算量,提高优化效率。1. The airfoil optimization design method based on deep reinforcement learning proposed by the present invention has the characteristics of self-learning strategy and long-term reward maximum. Compared with genetic algorithm and other algorithms, only a large amount of calculation data is used as the evaluation standard of optimization goals and constraints, and the depth reinforcement Learning and trying to learn the experience gained in the optimization process can improve the efficiency of data utilization, reduce the amount of calculation, and improve the efficiency of optimization.
2.本发明提出的基于深度强化学习的翼型优化设计方法,深度强化学习优化得到的策略模型具有一定程度的普适性或者说可迁移性:若是在一定范围内改变设计条件,例如来流马赫数、雷诺数等。原来优化得到的策略依旧能够提供初始优化方向,优化目标在很短的步数内就可以有明显的提升,这是传统优化方法所不具备的特性和优势。2. The airfoil optimization design method based on deep reinforcement learning proposed by the present invention, the strategy model obtained by deep reinforcement learning optimization has a certain degree of universality or transferability: if the design conditions are changed within a certain range, such as incoming flow Mach number, Reynolds number, etc. The strategy obtained from the original optimization can still provide the initial optimization direction, and the optimization goal can be significantly improved within a short number of steps, which is a feature and advantage that traditional optimization methods do not have.
本发明的附加方面和优点将在下面的描述中部分给出,部分将从下面的描述中变得明显,或通过本发明的实践了解到。Additional aspects and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention.
附图说明Description of drawings
本发明的上述和/或附加的方面和优点从结合下面附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present invention will become apparent and comprehensible from the description of the embodiments in conjunction with the following drawings, wherein:
图1为本发明翼型优化方法的整体流程图。Fig. 1 is the overall flow chart of the airfoil optimization method of the present invention.
图2为自由曲面变形法控制框。Figure 2 is the control frame of the free-form surface deformation method.
图3为翼型设计空间示意图。Figure 3 is a schematic diagram of the airfoil design space.
图4为全连接网络示意图。Figure 4 is a schematic diagram of a fully connected network.
图5为强化学习优化流程。Figure 5 shows the reinforcement learning optimization process.
图6为翼型的网格全貌与局部图。Figure 6 shows the overall view and partial view of the mesh of the airfoil.
图7为本文近端优化策略(PPO)的翼型优化阻力收敛曲线。Fig. 7 is the airfoil optimization resistance convergence curve of the proximal optimization strategy (PPO) in this paper.
图8为对比策略非支配排序遗传算法(NSGA-II)的翼型优化阻力收敛曲线。Fig. 8 is the airfoil optimization drag convergence curve of the comparison strategy non-dominated sorting genetic algorithm (NSGA-II).
图9为采用预训练的PPO策略的翼型优化阻力收敛曲线。Figure 9 is the airfoil optimization drag convergence curve using the pre-trained PPO strategy.
具体实施方式detailed description
下面详细描述本发明的实施例,所述实施例是示例性的,旨在用于解释本发明,而不能理解为对本发明的限制。Embodiments of the present invention are described in detail below, and the embodiments are exemplary and intended to explain the present invention, but should not be construed as limiting the present invention.
步骤1:本实例使用RAE2822作为基准翼型,采用自由曲面变形法(FFD)对基准翼型参数化,自由曲面变形法控制框如图2所示,设计变量由控制上翼面和下翼面的14个点组成,每个设计点的y方向的扰动范围为±0.02,翼型的设计空间如图3所示。Step 1: In this example, RAE2822 is used as the reference airfoil, and the free-form surface deformation method (FFD) is used to parameterize the reference airfoil. The control frame of the free-form surface deformation method is shown in Figure 2. The design variables are controlled by the upper and lower airfoils It consists of 14 points, and the perturbation range of each design point in the y direction is ±0.02. The design space of the airfoil is shown in Figure 3.
步骤2:设计目标是使翼型在设计状态下阻力最小,同时满足三个约束条件:(1)升力系数不减,(2)翼型的力矩系数的绝对值不增,(3)翼型的最大厚度不减。优化设计问题的数学表达式如下:Step 2: The design goal is to minimize the resistance of the airfoil in the design state, while satisfying three constraints: (1) the lift coefficient does not decrease, (2) the absolute value of the moment coefficient of the airfoil does not increase, (3) the airfoil The maximum thickness does not decrease. The mathematical expression of the optimization design problem is as follows:
Minimize:Cd Minimize: C d
Subject to:Cl≥Cl0 Subject to: C l ≥ C l0
|Cm|≤|Cm0||C m |≤|C m0 |
t≥t0 t≥t 0
其中Cd为翼型阻力系数、Cl为翼型升力系数、Cl0为基准翼型的升力系数、Cm为翼型的力矩系数、Cm0为基准翼型的力矩系数、t为翼型的厚度、t0为基准翼型的厚度。Among them, C d is the drag coefficient of the airfoil, C l is the lift coefficient of the airfoil, C l0 is the lift coefficient of the reference airfoil, C m is the moment coefficient of the airfoil, C m0 is the moment coefficient of the reference airfoil, and t is the airfoil , t 0 is the thickness of the reference airfoil.
步骤3:根据步骤2中的设计目标和约束条件构建奖励函数如下:Step 3: Construct the reward function according to the design goals and constraints in
如果Cd-Cd0<0则rewardcd=100×(Cd-Cd0)。If C d -C d0 <0, reward cd =100×(C d -C d0 ).
如果Cl-Cl0<0则rewardcl=10×(Cl-Cl0),否则令rewardcl=0。If C l -C l0 <0 then reward cl =10×(C l -C l0 ), otherwise let reward cl =0.
如果Cm-Cm0<0则令rewardcm=100×(Cm-Cm0),否则令rewardcm=0。If C m -C m0 <0, set reward cm =100×(C m -C m0 ), otherwise set reward cm =0.
如果t-t0<0则令rewardt=50×(t-t0),否则令rewardt=0。If tt 0 <0, set reward t =50×(tt 0 ), otherwise set reward t =0.
最终奖励值reward=rewardcd+rewardcl+rewardcm+100rewardt。The final reward value reward = reward cd + reward cl + reward cm + 100 reward t .
其中rewardcd为阻力系数奖励值,rewardcl为升力系数奖励值,rewardcm为力矩系数奖励值,rewardt为翼型厚度奖励值,10、50和100用于平衡目标和约束之间的量级差距。where reward cd is the reward value of the drag coefficient, reward cl is the reward value of the lift coefficient, reward cm is the reward value of the moment coefficient, reward t is the reward value of the airfoil thickness, and 10, 50 and 100 are used to balance the magnitude between the target and the constraint gap.
步骤4:建立智能体,包含策略模型π和价值函数模型,策略函数模型输出动作策略,而价值函数模型输出优势估计和价值函数,模型均使用含有两层隐藏层的人工神经网络模型,基于PyTorch编写,隐藏层节点数为64。Step 4: Establish an agent, including a policy model π and a value function model. The policy function model outputs an action strategy, while the value function model outputs an advantage estimate and a value function. The models all use an artificial neural network model with two hidden layers, based on PyTorch Write, the number of hidden layer nodes is 64.
人工神经网络是一种前馈神经网络,将多个神经元节点连接在一起,每一层的输出作为下一层的输入,通过权值运算,组成有分组层次的计算结构,全连接网络示意图如图4所示。初始化策略模型参数θ0和价值函数模型参数将翼型的升力系数、阻力系数、最大厚度和力矩系数作为状态,这里将基准机翼的升力系数、阻力系数、最大厚度和力矩系数作为初始状态s0。强化学习优化流程中各部分关系如图5所示。Artificial neural network is a kind of feedforward neural network, which connects multiple neuron nodes together, and the output of each layer is used as the input of the next layer. Through weight calculation, it forms a computing structure with grouping level. Schematic diagram of fully connected network As shown in Figure 4. Initialize policy model parameters θ 0 and value function model parameters Take the lift coefficient, drag coefficient, maximum thickness and moment coefficient of the airfoil as the state, and here take the lift coefficient, drag coefficient, maximum thickness and moment coefficient of the reference wing as the initial state s 0 . The relationship between various parts in the reinforcement learning optimization process is shown in Figure 5.
步骤5:当前智能体由策略模型根据状态和奖励值给出动作,也就是新的翼型设计变量。Step 5: The current agent is given an action by the policy model according to the state and reward value, which is the new airfoil design variable.
步骤6:对翼型实施动作得到新的翼型。Step 6: Perform actions on the airfoil to get a new airfoil.
步骤7:对二维新翼型建立C型结构化网格,图6给出了翼型的网格全貌与局部图。计算工况为Ma=0.734,Re=6.5×106,α=2.79°,其中Ma为来流马赫数,Re为来流雷诺数,α为来流攻角。在不考虑彻体力和外加热的影响时,使用开源求解器CFL3D进行翼型绕流数值模拟计算,主控方程为雷诺平均N-S方程,湍流模型为k-ωSST湍流模型,计算得到新翼型的阻力系数cd和力矩系数cm,升力系数cl和翼型厚度t。确认新翼型的气动参数为新状态。Step 7: Establish a C-shaped structured grid for the two-dimensional new airfoil. Figure 6 shows the overall and partial grid of the airfoil. The calculation condition is Ma=0.734, Re=6.5×10 6 , α=2.79°, where Ma is the Mach number of the incoming flow, Re is the Reynolds number of the incoming flow, and α is the angle of attack of the incoming flow. When the influence of physical force and external heating is not considered, the open source solver CFL3D is used for numerical simulation calculation of the flow around the airfoil. The main control equation is the Reynolds average NS equation, and the turbulence model is the k-ωSST turbulence model. Drag coefficient c d and moment coefficient c m , lift coefficient c l and airfoil thickness t. Confirm that the aerodynamic parameters of the new airfoil are in the new state.
步骤8:由计算得到的cd、cm、cl和t根据步骤3中的奖励函数计算得到奖励值。Step 8: Calculate the reward value from the calculated c d , cm , c l and t according to the reward function in
步骤9:由当前策略模型循环重复步骤5-8共e-1次,得到包含每次循环得到的状态和动作的轨迹和奖励值{re},轨迹τ={s0,a0,···,se-1,ae-1,se},其中s0和a0为初始状态和动作,se为e步状态,se-1为e-1步状态,ae-1为e-1步动作。Step 9: Repeat steps 5-8 for a total of e-1 times from the current policy model to obtain the trajectory and reward value {r e } including the state and action obtained in each cycle, trajectory τ={s 0 ,a 0 ,· ··,s e-1 ,a e-1 ,s e }, where s 0 and a 0 are the initial state and action, s e is the e-step state, s e-1 is the e-1 step state, a e- 1 is e-1 step action.
步骤10:基于当前策略模型重复步骤5-9共得到n条轨迹和奖励值。Step 10: Repeat steps 5-9 based on the current policy model to obtain a total of n trajectories and reward values.
步骤11:根据得到的轨迹参数,基于当前价值函数模型计算优势估计,优势估计为每一个动作a的期望奖励与该状态下所有的动作的期望奖励的均值的插值,如果优势估计大于0则说明该动作a优于平均,否则次于平均。Step 11: According to the obtained trajectory parameters, calculate the advantage estimate based on the current value function model. The advantage estimate is the interpolation between the expected reward of each action a and the average value of the expected reward of all actions in this state. If the advantage estimate is greater than 0, it means The action a is better than average, otherwise inferior to average.
步骤12:根据优势估计、轨迹和奖励值构建损失函数,基于随机梯度下降算法优化策略模型参数θ和价值函数模型参数优化目标为损失函数最小,用优化后的参数更新策略模型和价值函数模型,得到新的策略模型和价值函数模型。Step 12: Construct a loss function based on the advantage estimate, trajectory and reward value, and optimize the policy model parameters θ and value function model parameters based on the stochastic gradient descent algorithm The optimization goal is to minimize the loss function, and the optimized parameters are used to update the strategy model and value function model to obtain a new strategy model and value function model.
步骤13:循环重复步骤5-12直至达到损失函数不再降低,完成训练。Step 13: Repeat steps 5-12 until the loss function is no longer reduced, and the training is completed.
近端策略优化(PPO)是一种深度强化学习算法。图7给出了由本文近端优化策略的优化阻力收敛曲线,从中可以看出翼型阻力有明显改善,相比初始翼型RAE2822,阻力减少50框左右,约有30%的提升。为校验PPO在单目标约束气动优化设计中的优化效果,此处采用和非支配排序遗传算法(NSGA-II)作为对比算法,其中NSGA-II的种群规模为100,进化代数100,优化收敛曲线如图8所示;PPO的优化效果与NSGA-II相比仍有差距,于是采用预训练的方法提升PPO的优化效果。预训练就是基于类似问题中已训练完成的模型,这样可以为优化提供一个更好的初始点,预训练可以避免陷入局部最优,得到更好的优化结果。Proximal Policy Optimization (PPO) is a deep reinforcement learning algorithm. Figure 7 shows the optimal resistance convergence curve of the proximal optimization strategy in this paper. It can be seen that the airfoil resistance has been significantly improved. Compared with the initial airfoil RAE2822, the resistance is reduced by about 50 frames, which is about 30% improved. In order to verify the optimization effect of PPO in the single-objective constrained aerodynamic optimization design, the non-dominated sorting genetic algorithm (NSGA-II) is used here as a comparison algorithm. The population size of NSGA-II is 100, the evolution algebra is 100, and the optimization convergence The curve is shown in Figure 8; there is still a gap between the optimization effect of PPO and NSGA-II, so the pre-training method is used to improve the optimization effect of PPO. Pre-training is based on models that have been trained in similar problems, which can provide a better initial point for optimization. Pre-training can avoid falling into local optimum and get better optimization results.
对网络进行预训练,经过预训练的PPO的优化收敛曲线如图9所示,两种方法优化得到的优化结果与所需的计算量对比如表1所示。可以观察到优化的收敛趋势对比遗传算法明显更加稳定,两者收敛趋势相同,在阻力优化效果相近的情况下,本文经过预训练PPO策略的计算量为遗传算法的0.05倍,计算量得到了明显的减小。如表2,当固定计算量为500时,本文的PPO策略和经预训练后的PPO策略的阻力优化效果也远远好于遗传算法。可以看出本发明基于深度强化学习的翼型优化设计方法,相较遗传算法可以显著减小计算量,同时经过预训练后也不容易陷入局部最优。The network is pre-trained, and the optimization convergence curve of the pre-trained PPO is shown in Figure 9. The optimization results obtained by the two optimization methods and the required calculation amount are compared in Table 1. It can be observed that the optimized convergence trend is significantly more stable than the genetic algorithm, and the convergence trend of the two is the same. In the case of similar resistance optimization effects, the calculation amount of the pre-trained PPO strategy in this paper is 0.05 times that of the genetic algorithm, and the calculation amount has been significantly reduced. decrease. As shown in Table 2, when the fixed calculation amount is 500, the resistance optimization effect of the PPO strategy and the pre-trained PPO strategy in this paper is also far better than the genetic algorithm. It can be seen that the airfoil optimization design method based on deep reinforcement learning of the present invention can significantly reduce the amount of calculation compared with the genetic algorithm, and at the same time, it is not easy to fall into local optimum after pre-training.
表1Table 1
表2Table 2
接下来改变设计条件以说明强化学习得到的策略具有一定的普适性:保持基础翼型不变,设计状态由Ma=0.734,Re=6.5×106,α=2.79°分别变为1#:Ma=0.72,Re=1×107,α=2.79°,2#:Ma=0.76,Re=8×106,α=2.79°和3#:Ma=0.6,Re=6.5×106,α=2.79°。为了验证策略的有效性,直接使用已有策略对三个状态优化5步、10步、15步和20步,观察阻力的变化。结果如表3所示,可以看出在三个设计状态中,1#经过20步优化,阻力下降7.4%;2#阻力下降3.7%;3#阻力下降6.8%,且三个设计状态的优化结果约束均符合要求。总而言之,强化学习学习到了具有一定普适性的策略,可以为优化提供方向和初步优化结果。Next, change the design conditions to illustrate that the strategy obtained by reinforcement learning has certain universality: keep the basic airfoil unchanged, and the design state changes from Ma=0.734, Re=6.5×10 6 , α=2.79° to 1# respectively: Ma=0.72, Re=1×10 7 , α=2.79°, 2#: Ma=0.76, Re=8×10 6 , α=2.79° and 3#: Ma=0.6, Re=6.5×10 6 , α = 2.79°. In order to verify the effectiveness of the strategy, the existing strategy is directly used to optimize the three states for 5 steps, 10 steps, 15 steps and 20 steps, and observe the change of resistance. The results are shown in Table 3. It can be seen that in the three design states, after 20 steps of optimization, the resistance of 1# decreased by 7.4%; the resistance of 2# decreased by 3.7%; the resistance of 3# decreased by 6.8%, and the optimization of the three design states The resulting constraints all meet the requirements. All in all, reinforcement learning has learned a certain universal strategy, which can provide direction and preliminary optimization results for optimization.
表3table 3
尽管上面已经示出和描述了本发明的实施例,可以理解的是,上述实施例是示例性的,不能理解为对本发明的限制,本领域的普通技术人员在不脱离本发明的原理和宗旨的情况下在本发明的范围内可以对上述实施例进行变化、修改、替换和变型。Although the embodiments of the present invention have been shown and described above, it can be understood that the above embodiments are exemplary and cannot be construed as limitations to the present invention. Variations, modifications, substitutions, and modifications to the above-described embodiments are possible within the scope of the present invention.
Claims (8)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202211374735.1A CN115618497B (en) | 2022-11-04 | 2022-11-04 | A method for airfoil optimization design based on deep reinforcement learning |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202211374735.1A CN115618497B (en) | 2022-11-04 | 2022-11-04 | A method for airfoil optimization design based on deep reinforcement learning |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN115618497A true CN115618497A (en) | 2023-01-17 |
| CN115618497B CN115618497B (en) | 2025-03-21 |
Family
ID=84876923
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202211374735.1A Active CN115618497B (en) | 2022-11-04 | 2022-11-04 | A method for airfoil optimization design based on deep reinforcement learning |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN115618497B (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117634334A (en) * | 2023-10-27 | 2024-03-01 | 西北工业大学 | Fighter aircraft wing profile optimization design method considering pneumatic/stealth and wide-speed-domain wing profile family |
| CN119249911A (en) * | 2024-12-03 | 2025-01-03 | 西北工业大学 | A design method for efficient flow control based on transfer learning |
| WO2025055320A1 (en) * | 2023-09-15 | 2025-03-20 | 上海师范大学 | Airfoil profile optimization design method based on deep learning and reinforcement learning |
| WO2025100881A1 (en) * | 2023-11-09 | 2025-05-15 | 포항공과대학교 산학협력단 | Cfd automation method for optimal airfoil flow analysis based on reinforcement training, cfd airfoil flow analysis method, and cfd airfoil flow analysis device |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109614631A (en) * | 2018-10-18 | 2019-04-12 | 清华大学 | Fully automatic aerodynamic optimization method for aircraft based on reinforcement learning and transfer learning |
| CN110488861A (en) * | 2019-07-30 | 2019-11-22 | 北京邮电大学 | Unmanned plane track optimizing method, device and unmanned plane based on deeply study |
| US20220004191A1 (en) * | 2020-07-01 | 2022-01-06 | Wuhan University Of Technology | Usv formation path-following method based on deep reinforcement learning |
| CN114626277A (en) * | 2022-04-02 | 2022-06-14 | 浙江大学 | Active flow control method based on reinforcement learning |
| CN114861368A (en) * | 2022-06-13 | 2022-08-05 | 中南大学 | A Construction Method of Learning Model for Railway Profile Design Based on Proximal Strategy |
-
2022
- 2022-11-04 CN CN202211374735.1A patent/CN115618497B/en active Active
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109614631A (en) * | 2018-10-18 | 2019-04-12 | 清华大学 | Fully automatic aerodynamic optimization method for aircraft based on reinforcement learning and transfer learning |
| CN110488861A (en) * | 2019-07-30 | 2019-11-22 | 北京邮电大学 | Unmanned plane track optimizing method, device and unmanned plane based on deeply study |
| US20220004191A1 (en) * | 2020-07-01 | 2022-01-06 | Wuhan University Of Technology | Usv formation path-following method based on deep reinforcement learning |
| CN114626277A (en) * | 2022-04-02 | 2022-06-14 | 浙江大学 | Active flow control method based on reinforcement learning |
| CN114861368A (en) * | 2022-06-13 | 2022-08-05 | 中南大学 | A Construction Method of Learning Model for Railway Profile Design Based on Proximal Strategy |
Non-Patent Citations (1)
| Title |
|---|
| 丁存伟;杨旭东: "一种旋翼翼型多点多约束气动优化设计策略", 航空计算技术, vol. 43, no. 1, 15 January 2013 (2013-01-15), pages 52 - 57 * |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2025055320A1 (en) * | 2023-09-15 | 2025-03-20 | 上海师范大学 | Airfoil profile optimization design method based on deep learning and reinforcement learning |
| CN117634334A (en) * | 2023-10-27 | 2024-03-01 | 西北工业大学 | Fighter aircraft wing profile optimization design method considering pneumatic/stealth and wide-speed-domain wing profile family |
| WO2025100881A1 (en) * | 2023-11-09 | 2025-05-15 | 포항공과대학교 산학협력단 | Cfd automation method for optimal airfoil flow analysis based on reinforcement training, cfd airfoil flow analysis method, and cfd airfoil flow analysis device |
| CN119249911A (en) * | 2024-12-03 | 2025-01-03 | 西北工业大学 | A design method for efficient flow control based on transfer learning |
| CN119249911B (en) * | 2024-12-03 | 2025-04-04 | 西北工业大学 | Flow active control synergy design method based on transfer learning |
Also Published As
| Publication number | Publication date |
|---|---|
| CN115618497B (en) | 2025-03-21 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN115618497A (en) | An airfoil optimization design method based on deep reinforcement learning | |
| CN108637020B (en) | Self-adaptive variation PSO-BP neural network strip steel convexity prediction method | |
| CN110238839A (en) | A control method for multi-axis hole assembly using environment prediction to optimize non-model robot | |
| CN112084727A (en) | A Transition Prediction Method Based on Neural Network | |
| CN114626277B (en) | An Active Flow Control Method Based on Reinforcement Learning | |
| CN110286586A (en) | A Hybrid Modeling Method for Magnetorheological Damper | |
| Dong et al. | Reinforcement learning-based wind farm control: Toward large farm applications via automatic grouping and transfer learning | |
| CN114676639A (en) | Aircraft aerodynamic shape optimization method, device and medium based on neural network | |
| CN115293052A (en) | On-line optimization control method, storage medium and device for active power flow in power system | |
| CN114818128B (en) | Modeling method and optimizing method for ship body local curved surface optimizing neural network | |
| CN115438584A (en) | A deep learning-based airfoil aerodynamic prediction method | |
| CN108764577A (en) | Online time series prediction technique based on dynamic fuzzy Cognitive Map | |
| CN116702292A (en) | Pneumatic optimization method for wind nozzle of flat steel box girder based on deep reinforcement learning | |
| CN114384931B (en) | Multi-target optimal control method and equipment for unmanned aerial vehicle based on strategy gradient | |
| CN117970782B (en) | Fuzzy PID control method based on fish scale evolution GSOM improvement | |
| CN120354752B (en) | Intelligent design simulation system for fluid mechanical through-flow component | |
| CN113300884B (en) | A step-by-step network traffic prediction method based on GWO-SVR | |
| Zeng et al. | DDPG-based continuous thickness and tension coupling control for the unsteady cold rolling process | |
| CN115167102A (en) | Reinforced learning self-adaptive PID control method based on parallel dominant motion evaluation | |
| CN118519342A (en) | Prediction model construction method and pump storage unit working condition dynamic adjustment method | |
| CN117057225A (en) | Self-adaptive learning gas valve high-speed high-frequency high-precision servo and performance reconstruction method | |
| CN116594288A (en) | Control method and system based on longhorn beetle whisker fuzzy PID | |
| CN116047904A (en) | Human-simulation-reality hybrid training method for learning robot operation skills | |
| CN118657072B (en) | Method, system and device for adaptive voltage regulator parameter optimization | |
| CN119518833A (en) | A method and system for coordinated control of power grid frequency |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |