首页 | 官方网站   微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 218 毫秒
1.
Glowworm swarm optimization (GSO) algorithm is the one of the newest nature inspired heuristics for optimization problems. In order to enhances accuracy and convergence rate of the GSO, two strategies about the movement phase of GSO are proposed. One is the greedy acceptance criteria for the glowworms update their position one-dimension by one-dimension. The other is the new movement formulas which are inspired by artificial bee colony algorithm (ABC) and particle swarm optimization (PSO). To compare and analyze the performance of our proposed improvement GSO, a number of experiments are carried out on a set of well-known benchmark global optimization problems. The effects of the parameters about the improvement algorithms are discussed by uniform design experiment. Numerical results reveal that the proposed algorithms can find better solutions when compared to classical GSO and other heuristic algorithms and are powerful search algorithms for various global optimization problems.  相似文献   

2.
具有混合群智能行为的萤火虫群优化算法研究   总被引:1,自引:1,他引:0  
吴斌  崔志勇  倪卫红 《计算机科学》2012,39(5):198-200,228
萤火虫群优化算法是一种新型的群智能优化算法,基本的萤火虫群优化算法存在收敛精度低等问题。为了提高算法的性能,借鉴蜂群和鸟群的群体智能行为,改进萤火虫群优化算法的移动策略。运用均匀设计调整改进算法的参数取值。若干经典测试问题的实验仿真结果表明,引入混合智能行为大幅提升了算法的优化性能。  相似文献   

3.
Whale Optimization Algorithm (WOA), as a new population-based optimization algorithm, performs well in solving optimization problems. However, when tackling high-dimensional global optimization problems, WOA tends to fall into local optimal solutions and has slow convergence rate and low solution accuracy. To address these problems, a whale optimization algorithm based on quadratic interpolation (QIWOA) is presented. On the one hand, a modified exploration process by introducing a new parameter is proposed to efficiently search the regions and deal with the premature convergence problem. On the other hand, quadratic interpolation around the best search agent helps QIWOA to improve the exploitation ability and the solution accuracy. Moreover, the algorithm tries to make a balance between exploitation and exploration. QIWOA is compared with several state-of-the-art algorithms on 30 high-dimensional benchmark functions with dimensions ranging from 100 to 2000. The experimental results show that QIWOA has faster convergence rate and higher solution accuracy than both WOA and other population-based algorithms. For functions with a flat or sharp bottom, QIWOA is difficult to find the global optimum, but it still performs best compared with other algorithms.  相似文献   

4.
In recent years, particle swarm optimization (PSO) has extensively applied in various optimization problems because of its simple structure. Although the PSO may find local optima or exhibit slow convergence speed when solving complex multimodal problems. Also, the algorithm requires setting several parameters, and tuning the parameters is a challenging for some optimization problems. To address these issues, an improved PSO scheme is proposed in this study. The algorithm, called non-parametric particle swarm optimization (NP-PSO) enhances the global exploration and the local exploitation in PSO without tuning any algorithmic parameter. NP-PSO combines local and global topologies with two quadratic interpolation operations to increase the search ability. Nineteen (19) unimodal and multimodal nonlinear benchmark functions are selected to compare the performance of NP-PSO with several well-known PSO algorithms. The experimental results showed that the proposed method considerably enhances the efficiency of PSO algorithm in terms of solution accuracy, convergence speed, global optimality, and algorithm reliability.  相似文献   

5.
针对标准群搜索优化算法在解决一些复杂优化问题时容易陷入局部最优且收敛速度较慢的问题,提出一种应用反向学习和差分进化的群搜索优化算法(Group Search Optimization with Opposition-based Learning and Diffe-rential Evolution,OBDGSO)。该算法利用一般动态反向学习机制产生反向种群,扩大算法的全局勘探范围;对种群中较优解个体实施差分进化的变异操作,实现在较优解附近的局部开采,以改善算法的求解精度和收敛速度。这两种策略在GSO算法中相互协同,以更好地平衡算法的全局搜索能力和局部开采能力。将OBDGSO算法和另外4种群智能算法在12个基准测试函数上进行实验,结果表明OBDGSO算法在求解精度和收敛速度上具有较显著的性能优势。  相似文献   

6.
Stochastic optimization algorithms like genetic algorithms (GAs) and particle swarm optimization (PSO) algorithms perform global optimization but waste computational effort by doing a random search. On the other hand deterministic algorithms like gradient descent converge rapidly but may get stuck in local minima of multimodal functions. Thus, an approach that combines the strengths of stochastic and deterministic optimization schemes but avoids their weaknesses is of interest. This paper presents a new hybrid optimization algorithm that combines the PSO algorithm and gradient-based local search algorithms to achieve faster convergence and better accuracy of final solution without getting trapped in local minima. In the new gradient-based PSO algorithm, referred to as the GPSO algorithm, the PSO algorithm is used for global exploration and a gradient based scheme is used for accurate local exploration. The global minimum is located by a process of finding progressively better local minima. The GPSO algorithm avoids the use of inertial weights and constriction coefficients which can cause the PSO algorithm to converge to a local minimum if improperly chosen. The De Jong test suite of benchmark optimization problems was used to test the new algorithm and facilitate comparison with the classical PSO algorithm. The GPSO algorithm is compared to four different refinements of the PSO algorithm from the literature and shown to converge faster to a significantly more accurate final solution for a variety of benchmark test functions.  相似文献   

7.
基于剪枝策略的骨干粒子群算法   总被引:1,自引:0,他引:1  

为了优化算法的全局探索能力和局部开发能力, 提出一种基于两方面改进的骨干粒子群算法. 提出一种进化方程, 通过即时搜索域的分析说明该方程可以改善粒子多样性. 提出粒子群“剪枝”策略: 每当粒子搜索到新的群体最优位置时, 剪去该粒子, 同时初始化一个新位置以安插该粒子. 理论分析指出, 在增强全局探索能力的同时, 合适的剪枝策略能增加局部开发能力. 实验结果表明, 所提出算法的性能较几种经典PSO 算法有显著的提升.

  相似文献   

8.
粒子群优化算法是一种随机优化算法,但它不依概率1收敛到全局最优解。因此提出一种新的依概率收敛的粒子群优化算法。在该算法中,首先引入了具有探索和开发能力的两个变异算子,并依一定概率对粒子当前最好位置应用这两个算子,然后证明了该算法是依概率1收敛到ε-最优解。最后,把该算法应用到13个典型的测试函数中,并与其他粒子群优化算法比较,数值结果表明所给出的算法能够提高求解精度和收敛速度。  相似文献   

9.
Particle swarm optimization (PSO) has shown its competitive performance for solving benchmark and real-world optimization problems. Nevertheless, it requires better control of exploration/exploitation searches to prevent the premature convergence of swarms. Thus, this paper proposes a new PSO variant called PSO with adaptive time-varying topology connectivity (PSO-ATVTC) that employs an ATVTC module and a new learning framework. The proposed ATVTC module specifically aims to balance the algorithm's exploration/exploitation searches by varying the particle's topology connectivity with time according to its searching performance. The proposed learning framework consists of a new velocity update mechanism and a new neighborhood search operator to improve the algorithm's performance. A comprehensive study was conducted on 24 benchmark functions and one real-world problem. Compared with nine well-established PSO variants and six other cutting-edge metaheuristic search algorithms, the searching performance of PSO-ATVTC was proven to be more prominent in majority of the tested problems.  相似文献   

10.
阴阳对优化算法是一种新颖的轻量级随机优化算法,利用两点(全局探索点P 2和局部开发点P 1)的迭代交换来实现优化搜索。用户定义参数直接影响该算法的全局探索和局部开发之间的平衡,并且对算法的性能有着重要的影响。为提高该算法的优化性能,首先分析了原算法的用户定义参数(缩放因子α)对于性能的影响,随后提出用户定义参数线性与非线性递减三种改进的阴阳对优化算法。采用2013年进化计算大会中单目标实参算法竞赛中使用的28个测试函数进行性能评估,结果表明相比于原算法,改进后的算法具有更高的计算精度和更快的收敛速度。最后通过一个工程优化任务来展示改进后算法的性能。  相似文献   

11.
粒子群优化算法已成为求解多目标优化问题的有效方法之一,而速度更新公式中的惯性、局部和全局3个速度项的系数的动态合理设置是算法优化效率的关键问题。为解决现有算法仅单独设置各速度项系数导致优化效率不高的问题,提出了一种均衡各速度项系数的多目标粒子群优化算法。该方法旨在通过粒子的局部最优和全局最优的信息来引导种群的进化方向,动态调整每一个粒子速度项系数来均衡惯性、局部和全局3个速度项在搜索中的作用,从而更为准确地刻画算法的搜索能力和搜索精度,更好地平衡算法的探究和探索能力,进一步提高粒子群优化算法解决复杂多目标优化问题的效率。在7个标准测试函数上进行实验,并与5种经典的进化算法进行对比,结果表明新算法在综合指标IGD以及多样性评估指标Δ评分上具有更好的收敛速度和分布性,验证了新算法的有效性。  相似文献   

12.
针对资产数目和投资资金比例受约束的投资组合选择这一NP难问题,基于混沌搜索、粒子群优化和引力搜索算法提出了一种新的混合元启发式搜索算法。该算法能很好地平衡开发能力和勘探能力,有效抑制了算法早熟收敛现象。标准测试函数的测试结果表明混合算法与标准的粒子群优化和引力搜索算法相比具有更好的寻优效率;实证分析进一步对混合算法与遗传算法及粒子群优化算法在求解这类投资组合选择问题的性能进行了比较。数值结果表明,混合算法在搜索具有高预期回报的非支配投资组合方面表现更好,取得了更为满意的结果。  相似文献   

13.
龙文  伍铁斌 《控制与决策》2017,32(10):1749-1757
提出一种协调探索和开发能力的灰狼优化算法.利用佳点集方法初始化灰狼个体的位置,为全局搜索多样性奠定基础;为协调算法的全局探索和局部开发能力,给出一种基于正切三角函数描述的非线性动态变化控制参数;为加快算法的收敛速度,受粒子群优化算法个体记忆功能的启发,设计一种新的个体位置更新公式.10个标准函数的测试结果表明,改进灰狼优化(IGWO)算法能够有效地协调其对问题搜索空间的探索和开发能力.  相似文献   

14.
This paper introduces a new algorithmic nature-inspired approach that uses particle swarm optimization (PSO) with different neighborhood topologies, for successfully solving one of the most computationally complex problems, the permutation flowshop scheduling problem (PFSP). The PFSP belongs to the class of combinatorial optimization problems characterized as NP-hard and, thus, heuristic and metaheuristic techniques have been used in order to find high quality solutions in reasonable computational time. The proposed algorithm for the solution of the PFSP, the PSO with expanding neighborhood topology, combines a PSO algorithm, the variable neighborhood search strategy and a path relinking strategy. As, in general, the structure of the social network affects strongly a PSO algorithm, the proposed method using an expanding neighborhood topology manages to increase the performance of the algorithm. As the algorithm starts from a small size neighborhood and by increasing (expanding) in each iteration the size of the neighborhood, it ends to a neighborhood that includes all the swarm, and it manages to take advantage of the exploration abilities of a global neighborhood structure and of the exploitation abilities of a local neighborhood structure. In order to test the effectiveness and the efficiency of the proposed method, we use a set of benchmark instances of different sizes and compare the proposed method with a number of other PSO algorithms and other algorithms from the literature.  相似文献   

15.
Most of the recent proposed particle swarm optimization (PSO) algorithms do not offer the alternative learning strategies when the particles fail to improve their fitness during the searching process. Motivated by this fact, we improve the cutting edge teaching–learning-based optimization (TLBO) algorithm and adapt the enhanced framework into the PSO, thereby develop a teaching and peer-learning PSO (TPLPSO) algorithm. To be specific, the TPLPSO adopts two learning phases, namely the teaching and peer-learning phases. The particle firstly enters into the teaching phase and updates its velocity based on its historical best and the global best information. Particle that fails to improve its fitness in the teaching phase then enters into the peer-learning phase, where an exemplar is selected as the guidance particle. Additionally, a stagnation prevention strategy (SPS) is employed to alleviate the premature convergence issue. The proposed TPLPSO is extensively evaluated on 20 benchmark problems with different features, as well as one real-world problem. Experimental results reveal that the TPLPSO exhibits competitive performances when compared with ten other PSO variants and seven state-of-the-art metaheuristic search algorithms.  相似文献   

16.
Particle swarm optimization (PSO) has received increasing interest from the optimization community due to its simplicity in implementation and its inexpensive computational overhead. However, PSO has premature convergence, especially in complex multimodal functions. Extremal optimization (EO) is a recently developed local-search heuristic method and has been successfully applied to a wide variety of hard optimization problems. To overcome the limitation of PSO, this paper proposes a novel hybrid algorithm, called hybrid PSO–EO algorithm, through introducing EO to PSO. The hybrid approach elegantly combines the exploration ability of PSO with the exploitation ability of EO. We testify the performance of the proposed approach on a suite of unimodal/multimodal benchmark functions and provide comparisons with other meta-heuristics. The proposed approach is shown to have superior performance and great capability of preventing premature convergence across it comparing favorably with the other algorithms.  相似文献   

17.
基于Pareto熵的多目标粒子群优化算法   总被引:4,自引:0,他引:4  
胡旺  Gary G. YEN  张鑫 《软件学报》2014,25(5):1025-1050
粒子群优化算法因形式简洁、收敛快速和参数调节机制灵活等优点,同时一次运行可得到多个解,且能逼近非凸或不连续的Pareto最优前端,因而被认为是求解多目标优化问题最具潜力的方法之一.但当粒子群优化算法从单目标问题扩展到多目标问题时,Pareto最优解集的存储与维护、全局和个体最优解的选择以及开发与开采的平衡等问题亦随之出现.通过目标空间变换方法,采用Pareto前端在被称为平行格坐标系统的新目标空间中的分布熵及差熵评估种群的多样性及进化状态,并以此为反馈信息来设计进化策略,使得算法能够兼顾近似Pareto前端的收敛性和多样性.同时,引入格占优和格距离密度的概念来评估Pareto最优解的个体环境适应度,以此建立外部档案更新方法和全局最优解选择机制,最终形成了基于Pareto熵的多目标粒子群优化算法.实验结果表明:在IGD性能指标上,与另外8种对等算法相比,该算法在由ZDT和DTLZ系列组成的12个多目标测试问题集中表现出了显著的性能优势.  相似文献   

18.
A new hybrid optimization algorithm is proposed for minimization of continuous multi-modal functions. The algorithm called Global Simplex Optimization (GSO) is a population set based Evolutionary Algorithm (EA) incorporating a special multi-stage, stochastic and weighted version of the reflection operator of the classical simplex method. An optional mutation operator has also been tested and then removed from the structure of the final algorithm in favor of simplicity and because of insignificant effect on performance. The promising performance achieved by GSO is demonstrated by comparisons made to some other state-of-the-art global optimization algorithms over a set of conventional benchmark problems.  相似文献   

19.
一种自适应柯西变异的反向学习粒子群优化算法   总被引:1,自引:0,他引:1  
针对传统粒子群优化算法易出现早熟的问题,提出了一种自适应变异的反向学习粒子群优化算法。该算法在一般性反向学习方法的基础上,提出了自适应柯西变异策略(ACM)。采用一般性反向学习策略生成反向解,可扩大搜索空间,增强算法的全局勘探能力。为避免粒子陷入局部最优解而导致搜索停滞现象的发生,采用ACM策略对当前最优粒子进行扰动,自适应地获取变异点,在有效提高算法局部开采能力的同时,使算法能更加平稳快速地收敛到全局最优解。为进一步平衡算法的全局搜索与局部探测能力,采用非线性的自适应惯性权值。将算法在14个测试函数上与多种基于反向学习策略的PSO算法进行对比,实验结果表明提出的算法在解的精度以及收敛速度上得到了大幅度的提高。  相似文献   

20.
In this paper, a modified particle swarm optimization (PSO) algorithm is developed for solving multimodal function optimization problems. The difference between the proposed method and the general PSO is to split up the original single population into several subpopulations according to the order of particles. The best particle within each subpopulation is recorded and then applied into the velocity updating formula to replace the original global best particle in the whole population. To update all particles in each subpopulation, the modified velocity formula is utilized. Based on the idea of multiple subpopulations, for the multimodal function optimization the several optima including the global and local solutions may probably be found by these best particles separately. To show the efficiency of the proposed method, two kinds of function optimizations are provided, including a single modal function optimization and a complex multimodal function optimization. Simulation results will demonstrate the convergence behavior of particles by the number of iterations, and the global and local system solutions are solved by these best particles of subpopulations.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号