首页 | 官方网站   微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 203 毫秒
1.
为了解决初始和终端确定的一类离散时间非线性系统有限时间优化控制,利用动态规划原理求解过程中遇到维数灾的问题,提出了基于神经网络的自适应动态规划近似优化控制.在分析动态规划求解遇到维数灾的基础上,进而给出了迭代ADP算法,并采用神经网络近似代价函数和控制律来实现迭代ADP算法,设计近似优化控制器.通过mat lab实验仿真结果表明,采用迭代ADP算法能够避免求解中遇到的维数灾,从而有效地实现了一类离散时间非线性系统的有限时间近似优化控制.  相似文献   

2.
设计了一种基于折扣广义值迭代的智能算法,用于解决一类复杂非线性系统的最优跟踪控制问题.通过选取合适的初始值,值迭代过程中的代价函数将以单调递减的形式收敛到最优代价函数.基于单调递减的值迭代算法,在不同折扣因子的作用下,讨论了迭代跟踪控制律的可容许性和误差系统的渐近稳定性.为了促进算法的实现,建立一个数据驱动的模型网络用...  相似文献   

3.
文章提出了一种新的模糊神经网络(FNN:FuzzyNeuralNetwork)控制的变步长盲均衡算法,利用模糊神经网络控制盲均衡算法的迭代步长,以得到更好的均衡性能。该文设计出模糊神经网络控制器的结构并给出状态方程,提出了新的代价函数,推导出控制器参数的迭代公式。计算机仿真表明,该算法与传统恒模(CMA:ConstantModulusAlgorithm)盲均衡算法相比,具有稳定性好的优点。  相似文献   

4.
分析了射频识别中码间干扰的产生原因,针对码间干扰的克服,提出了一种基于时变步长神经网络盲均衡算法.采用三层前馈神经网络结构,传递函数选用f(x)=x+0.005sin πX,代价函数选用传统恒模盲均衡算法的代价函数,步长控制因子为均方误差非线性函数,推导了算法的迭代公式,选择了待定参数,并进行了计算机仿真,验证了所提算法的有效性.  相似文献   

5.
基于频域共轭梯度法的交替迭代复原算法研究   总被引:1,自引:0,他引:1  
对迭代盲目去卷积复原方法进行了研究,提出了基于频域共轭梯度法的交替迭代优化复原算法,在频域上构造了关于目标图像和点扩展函数频谱的误差代价函数,将共轭梯度法引入到频谱误差代价函数的极小化过程中,并将空域非负性和频域带限等先验约束知识融合到对目标图像和点扩展函数的迭代交替优化估计过程中,取得了预期的复原效果,增强了算法的抗噪性和稳定性。在微机上进行了一系列的复原实验,实验结果表明算法复原效果好,抗噪能力强,速度较快,且能恢复具有复杂背景的目标图像。  相似文献   

6.
为了减轻D2D通信在资源复用模式下的互干扰问题,提升蜂窝网络均衡性能收益,提出一种基于纳什均衡的功率控制博弈算法。算法中将互干扰用户间的功率控制过程描述为静态博弈模型,用户之间根据最小化代价函数的博弈决策,通过多步迭代调节发射功率,使系统收敛至纳什均衡的优化状态。在用户代价函数设计中,综合考虑了能耗及传输速率影响,同时给出了博弈算法纳什均衡存在性以及收敛性的证明。仿真实验表明,在最优响应策略及能耗因子的有效约束下,互干扰用户更理智的选择发射功率,系统拥有较好均衡性收益的同时能耗进一步降低。  相似文献   

7.
基于数据自适应评判的离散2-D系统零和博弈最优控制   总被引:1,自引:1,他引:0  
提出了基于一种迭代自适应评判设计(ACD)算法解决一类离散时间Roesser型2-D系统的二人零和对策问题. 文章主要思想是采用自适应评判技术迭代的获得最优控制对使得性能指标函数达到零和对策的鞍点. 所提出的ACD可以通过输入输出数据进行实现而不需要系统的模型. 为了实现迭代ACD算法, 神经网络分别用来近似性能指标函数和计算最优控制率. 最后最优控制策略将应用到空气干燥过程控制中以证明其有效性.  相似文献   

8.
针对一类典型的带有控制约束的非线性离散时间系统,提出了一种基于自适应动态规划(adaptive dynamic programmmg,ADP)算法的多设定值跟踪控制方法,并对其收敛性和稳定性做了严格分析.在ADP迭代跟踪控制的基础上,根据多模型控制的思想,设置阶梯状的参考轨迹,使得系统状态逐渐地跟踪到最终设定值,保证了系统的稳定性,极大地减小超调量,加快了响应时间,改善控制品质;同时由于控制器约束的存在,引入非二次型的性能指标函数,使得控制量始终在有界的范围内变化.最后对仿真结果进行了分析,结果表明了此方法的可行性和有效性.  相似文献   

9.
在利用二阶统计量实现盲源分离问题中,混迭矩阵经过白化以后转变成了酉矩阵。针对酉矩阵各列之间相互正交的特性,提出一种关于酉矩阵某一列的最小二乘对称代价函数。通过基于梯度下降法的三迭代算法,交替估计三二次代价函数中的各组待定参数,搜索代价函数最小点,从而得到对应能量最大信号源的酉矩阵的一列。利用系统化的多步分解算法(MSA),依次估计酉矩阵的一列,最终得到整个酉矩阵的估计。仿真结果表明,与经典的通过连续Givens旋转求酉矩阵的SOBI算法相比,该算法全局拒噪水平至少改善了9 dB,而所需计算时间仅为SOBI的二分之一,更有效地解决了盲源分离问题。  相似文献   

10.
设计了一种基于事件的迭代自适应评判算法,用于解决一类非仿射系统的零和博弈最优跟踪控制问题.通过数值求解方法得到参考轨迹的稳定控制,进而将未知非线性系统的零和博弈最优跟踪控制问题转化为误差系统的最优调节问题.为了保证闭环系统在具有良好控制性能的基础上有效地提高资源利用率,引入一个合适的事件触发条件来获得阶段性更新的跟踪策略对.然后,根据设计的触发条件,采用Lyapunov方法证明误差系统的渐近稳定性.接着,通过构建四个神经网络,来促进所提算法的实现.为了提高目标轨迹对应稳定控制的精度,采用模型网络直接逼近未知系统函数而不是误差动态系统.构建评判网络、执行网络和扰动网络用于近似迭代代价函数和迭代跟踪策略对.最后,通过两个仿真实例,验证该控制方法的可行性和有效性.  相似文献   

11.
In this paper, a novel iterative adaptive dynamic programming (ADP) algorithm is developed to solve infinite horizon optimal control problems for discrete-time nonlinear systems. When the iterative control law and iterative performance index function in each iteration cannot be accurately obtained, it is shown that the iterative controls can make the performance index function converge to within a finite error bound of the optimal performance index function. Stability properties are presented to show that the system can be stabilized under the iterative control law which makes the present iterative ADP algorithm feasible for implementation both on-line and off-line. Neural networks are used to approximate the iterative performance index function and compute the iterative control policy, respectively, to implement the iterative ADP algorithm. Finally, two simulation examples are given to illustrate the performance of the present method.  相似文献   

12.
In this paper, we aim to solve the finite horizon optimal control problem for a class of discrete-time nonlinear systems with unfixed initial state using adaptive dynamic programming (ADP) approach. A new ε-optimal control algorithm based on the iterative ADP approach is proposed which makes the performance index function converge iteratively to the greatest lower bound of all performance indices within an error according to ε within finite time. The optimal number of control steps can also be obtained by the proposed ε-optimal control algorithm for the situation where the initial state of the system is unfixed. Neural networks are used to approximate the performance index function and compute the optimal control policy, respectively, for facilitating the implementation of the ε-optimal control algorithm. Finally, a simulation example is given to show the results of the proposed method.  相似文献   

13.
In this paper, a novel value iteration adaptive dynamic programming (ADP) algorithm, called “generalized value iteration ADP” algorithm, is developed to solve infinite horizon optimal tracking control problems for a class of discrete-time nonlinear systems. The developed generalized value iteration ADP algorithm permits an arbitrary positive semi-definite function to initialize it, which overcomes the disadvantage of traditional value iteration algorithms. Convergence property is developed to guarantee that the iterative performance index function will converge to the optimum. Neural networks are used to approximate the iterative performance index function and compute the iterative control policy, respectively, to implement the iterative ADP algorithm. Finally, a simulation example is given to illustrate the performance of the developed algorithm.  相似文献   

14.
In this paper, we aim to solve the finite-horizon optimal control problem for a class of non-linear discrete-time switched systems using adaptive dynamic programming(ADP) algorithm. A new ε-optimal control scheme based on the iterative ADP algorithm is presented which makes the value function converge iteratively to the greatest lower bound of all value function indices within an error according to ε within finite time. Two neural networks are used as parametric structures to implement the iterative ADP algorithm with ε-error bound, which aim at approximating the value function and the control policy, respectively. And then, the optimal control policy is obtained. Finally, a simulation example is included to illustrate the applicability of the proposed method.  相似文献   

15.
In this paper, a new dual iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for a class of nonlinear systems with time-delays in state and control variables. The idea is to use the dynamic programming theory to solve the expressions of the optimal performance index function and control. Then, the dual iterative ADP algorithm is introduced to obtain the optimal solutions iteratively, where in each iteration, the performance index function and the system states are both updated. Convergence analysis is presented to prove the performance index function to reach the optimum by the proposed method. Neural networks are used to approximate the performance index function and compute the optimal control policy, respectively, for facilitating the implementation of the dual iterative ADP algorithm. Simulation examples are given to demonstrate the validity of the proposed optimal control scheme.  相似文献   

16.
In this paper, a novel neural-network-based iterative adaptive dynamic programming (ADP) algorithm is proposed. It aims at solving the optimal control problem of a class of nonlinear discrete-time systems with control constraints. By introducing a generalized nonquadratic functional, the iterative ADP algorithm through globalized dual heuristic programming technique is developed to design optimal controller with convergence analysis. Three neural networks are constructed as parametric structures to facilitate the implementation of the iterative algorithm. They are used for approximating at each iteration the cost function, the optimal control law, and the controlled nonlinear discrete-time system, respectively. A simulation example is also provided to verify the effectiveness of the control scheme in solving the constrained optimal control problem.  相似文献   

17.
In this paper, we consider the feedback control on nonzero-sum linear quadratic (LQ) differential games in finite horizon for discrete-time stochastic systems with Markovian jump parameters and multiplicative noise. Four-coupled generalized difference Riccati equations (GDREs) are obtained, which are essential to find the optimal Nash equilibrium strategies and the optimal cost values of the LQ differential games. Furthermore, an iterative algorithm is given to solve the four-coupled GDREs. Finally, a suboptimal solution of the LQ differential games is proposed based on a convex optimization approach and a simplification of the suboptimal solution is given. Simulation examples are presented to illustrate the effectiveness of the iterative algorithm and the suboptimal solution.  相似文献   

18.
In this paper, a finite-horizon neuro-optimal tracking control strategy for a class of discrete-time nonlinear systems is proposed. Through system transformation, the optimal tracking problem is converted into designing a finite-horizon optimal regulator for the tracking error dynamics. Then, with convergence analysis in terms of cost function and control law, the iterative adaptive dynamic programming (ADP) algorithm via heuristic dynamic programming (HDP) technique is introduced to obtain the finite-horizon optimal tracking controller which makes the cost function close to its optimal value within an ?-error bound. Three neural networks are used as parametric structures to implement the algorithm, which aims at approximating the cost function, the control law, and the error dynamics, respectively. Two simulation examples are included to complement the theoretical discussions.  相似文献   

19.
Based on adaptive dynamic programming (ADP), the fixed-point tracking control problem is solved by a value iteration (Ⅵ) algorithm. First, a class of discrete-time (DT) nonlinear system with disturbance is considered. Second, the convergence of a Ⅵ algorithm is given. It is proven that the iterative cost function precisely converges to the optimal value, and the control input and disturbance input also converges to the optimal values. Third, a novel analysis pertaining to the range of the discount factor is presented, where the cost function serves as a Lyapunov function. Finally, neural networks (NNs) are employed to approximate the cost function, the control law, and the disturbance law. Simulation examples are given to illustrate the effective performance of the proposed method.   相似文献   

20.
This paper will present an approximate/adaptive dynamic programming(ADP) algorithm,that uses the idea of integral reinforcement learning(IRL),to determine online the Nash equilibrium solution for the two-player zerosum differential game with linear dynamics and infinite horizon quadratic cost.The algorithm is built around an iterative method that has been developed in the control engineering community for solving the continuous-time game algebraic Riccati equation(CT-GARE),which underlies the game problem.We here show how the ADP techniques will enhance the capabilities of the offline method allowing an online solution without the requirement of complete knowledge of the system dynamics.The feasibility of the ADP scheme is demonstrated in simulation for a power system control application.The adaptation goal is the best control policy that will face in an optimal manner the highest load disturbance.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号