scholarly journals A Hybrid Rao-NM Algorithm for Image Template Matching

Entropy ◽  
2021 ◽  
Vol 23 (6) ◽  
pp. 678
Author(s):  
Xinran Liu ◽  
Zhongju Wang ◽  
Long Wang ◽  
Chao Huang ◽  
Xiong Luo

This paper proposes a hybrid Rao-Nelder–Mead (Rao-NM) algorithm for image template matching is proposed. The developed algorithm incorporates the Rao-1 algorithm and NM algorithm serially. Thus, the powerful global search capability of the Rao-1 algorithm and local search capability of NM algorithm is fully exploited. It can quickly and accurately search for the high-quality optimal solution on the basis of ensuring global convergence. The computing time is highly reduced, while the matching accuracy is significantly improved. Four commonly applied optimization problems and three image datasets are employed to assess the performance of the proposed method. Meanwhile, three commonly used algorithms, including generic Rao-1 algorithm, particle swarm optimization (PSO), genetic algorithm (GA), are considered as benchmarking algorithms. The experiment results demonstrate that the proposed method is effective and efficient in solving image matching problems.

2019 ◽  
Vol 61 (4) ◽  
pp. 177-185
Author(s):  
Moritz Mühlenthaler ◽  
Alexander Raß

Abstract A discrete particle swarm optimization (PSO) algorithm is a randomized search heuristic for discrete optimization problems. A fundamental question about randomized search heuristics is how long it takes, in expectation, until an optimal solution is found. We give an overview of recent developments related to this question for discrete PSO algorithms. In particular, we give a comparison of known upper and lower bounds of expected runtimes and briefly discuss the techniques used to obtain these bounds.


2016 ◽  
Vol 11 (1) ◽  
pp. 58-67 ◽  
Author(s):  
S Sarathambekai ◽  
K Umamaheswari

Discrete particle swarm optimization is one of the most recently developed population-based meta-heuristic optimization algorithm in swarm intelligence that can be used in any discrete optimization problems. This article presents a discrete particle swarm optimization algorithm to efficiently schedule the tasks in the heterogeneous multiprocessor systems. All the optimization algorithms share a common algorithmic step, namely population initialization. It plays a significant role because it can affect the convergence speed and also the quality of the final solution. The random initialization is the most commonly used method in majority of the evolutionary algorithms to generate solutions in the initial population. The initial good quality solutions can facilitate the algorithm to locate the optimal solution or else it may prevent the algorithm from finding the optimal solution. Intelligence should be incorporated to generate the initial population in order to avoid the premature convergence. This article presents a discrete particle swarm optimization algorithm, which incorporates opposition-based technique to generate initial population and greedy algorithm to balance the load of the processors. Make span, flow time, and reliability cost are three different measures used to evaluate the efficiency of the proposed discrete particle swarm optimization algorithm for scheduling independent tasks in distributed systems. Computational simulations are done based on a set of benchmark instances to assess the performance of the proposed algorithm.


2011 ◽  
Vol 383-390 ◽  
pp. 7208-7213
Author(s):  
De Kun Tan

To overcome the shortage of standard Particle Swarm Optimization(SPSO) on premature convergence, Quantum-behaved Particle Swarm Optimization (QPSO) is presented to solve engineering constrained optimization problem. QPSO algorithm is a novel PSO algorithm model in terms of quantum mechanics. The model is based on Delta potential, and we think the particle has the behavior of quanta. Because the particle doesn’t have a certain trajectory, it has more randomicity than the particle which has fixed path in PSO, thus the QPSO more easily escapes from local optima, and has more capability to seek the global optimal solution. In the period of iterative optimization, outside point method is used to deal with those particles that violate the constraints. Furthermore, compared with other intelligent algorithms, the QPSO is verified by two instances of engineering constrained optimization, experimental results indicate that the algorithm performs better in terms of accuracy and robustness.


Symmetry ◽  
2020 ◽  
Vol 12 (6) ◽  
pp. 922 ◽  
Author(s):  
Yuji Du ◽  
Fanfan Xu

As a meta-heuristic algoriTthm, particle swarm optimization (PSO) has the advantages of having a simple principle, few required parameters, easy realization and strong adaptability. However, it is easy to fall into a local optimum in the early stage of iteration. Aiming at this shortcoming, this paper presents a hybrid multi-step probability selection particle swarm optimization with sine chaotic inertial weight and symmetric tangent chaotic acceleration coefficients (MPSPSO-ST), which can strengthen the overall performance of PSO to a large extent. Firstly, we propose a hybrid multi-step probability selection update mechanism (MPSPSO), which skillfully uses a multi-step process and roulette wheel selection to improve the performance. In order to achieve a good balance between global search capability and local search capability to further enhance the performance of the method, we also design sine chaotic inertial weight and symmetric tangent chaotic acceleration coefficients inspired by chaos mechanism and trigonometric functions, which are integrated into the MPSPSO-ST algorithm. This strategy enables the diversity of the swarm to be preserved to discourage premature convergence. To evaluate the effectiveness of the MPSPSO-ST algorithm, we conducted extensive experiments with 20 classic benchmark functions. The experimental results show that the MPSPSO-ST algorithm has faster convergence speed, higher optimization accuracy and better robustness, which is competitive in solving numerical optimization problems and outperforms a lot of classical PSO variants and well-known optimization algorithms.


Author(s):  
I. V. Kozin ◽  
S. E. Batovskiy

It is known that a large number of applied optimization problems can’t be exactly solved nowadays, because their computational complexity is related to the NP-hard class. In many cases metaheuristics of various types are used to search for approximate solutions, but the choice of the concrete metaheuristic has open question of the quality of the chosen method. There are several possible solutions to this problem, one of which is the verification of metaheuristic algorithms using examples from known test libraries with known records. Another approach to solving the problem of evaluating the quality of algorithms is to compare the "new" algorithm with other algorithms, the work of which has already been investigated. The construction a generator of random problems with a known optimal solution can solve the problem of obtaining "average" estimates of the accuracy for used algorithm in comparison with other methods. The article considers the construction of generators of random non-waste maps of rec-tangular cutting with restrictions on the rectangles of limited sizes. The existence of sets of such cards forms the basis of test problems for checking the quality of approximate algorithms for searching for optimal solution. Rectangular cutting, which is considered in the article, is also the basis for building cuts using more complex shapes. As the simplest method of generating random rectangular non-waste maps, considered a method that uses guillotine cutting. Also, a more complex algorithm for generating a random rectangular cut is given, whose job is to generate a random dot grid and remove some random points from this grid. Much attention is paid to the implementation of the above methods, since the main purpose of the article is to simplify using of generators in practice. All the above algorithms are already used in the software system for testing evolution-aryfragmentary algorithms for various classes of optimization problems on the graphs


Energies ◽  
2021 ◽  
Vol 14 (24) ◽  
pp. 8575
Author(s):  
Rehan Ali Khan ◽  
Shiyou Yang ◽  
Shafiullah Khan ◽  
Shah Fahad ◽  
Kalimullah

Particle Swarm Optimization (PSO) is a member of the swarm intelligence-based on a metaheuristic approach which is inspired by the natural deeds of bird flocking and fish schooling. In comparison to other traditional methods, the model of PSO is widely recognized as a simple algorithm and easy to implement. However, the traditional PSO’s have two primary issues: premature convergence and loss of diversity. These problems arise at the latter stages of the evolution process when dealing with high-dimensional, complex and electromagnetic inverse problems. To address these types of issues in the PSO approach, we proposed an Improved PSO (IPSO) which employs a dynamic control parameter as well as an adaptive mutation mechanism. The main proposal of the novel adaptive mutation operator is to prevent the diversity loss of the optimization process while the dynamic factor comprises the balance between exploration and exploitation in the search domain. The experimental outcomes achieved by solving complicated and extremely high-dimensional optimization problems were also validated on superconducting magnetic energy storage devices (SMES). According to numerical and experimental analysis, the IPSO delivers a better optimal solution than the other solutions described, particularly in the early computational evaluation of the generation.


2018 ◽  
Vol 232 ◽  
pp. 03015
Author(s):  
Changjun Wen ◽  
Changlian Liu ◽  
Heng Zhang ◽  
Hongliang Wang

The particle swarm optimization (PSO) is a widely used tool for solving optimization problems in the field of engineering technology. However, PSO is likely to fall into local optimum, which has the disadvantages of slow convergence speed and low convergence precision. In view of the above shortcomings, a particle swarm optimization with Gaussian disturbance is proposed. With introducing the Gaussian disturbance in the self-cognition part and social cognition part of the algorithm, this method can improve the convergence speed and precision of the algorithm, which can also improve the ability of the algorithm to escape the local optimal solution. The algorithm is simulated by Griewank function after the several evolutionary modes of GDPSO algorithm are analyzed. The experimental results show that the convergence speed and the optimization precision of the GDPSO is better than that of PSO.


Author(s):  
Shengyu Pei

How to solve constrained optimization problems constitutes an important part of the research on optimization problems. In this paper, a hybrid immune clonal particle swarm optimization multi-objective algorithm is proposed to solve constrained optimization problems. In the proposed algorithm, the population is first initialized with the theory of good point set. Then, differential evolution is adopted to improve the local optimal solution of each particle, with immune clonal strategy incorporated to improve each particle. As a final step, sub-swarm is used to enhance the position and velocity of individual particle. The new algorithm has been tested on 24 standard test functions and three engineering optimization problems, whose results show that the new algorithm has good performance in both robustness and convergence.


2018 ◽  
Vol 57 (05/06) ◽  
pp. 280-286 ◽  
Author(s):  
Silviu Bejinariu ◽  
Hariton Costin

Computational Intelligence Re-meets Medical Image Processing Analysis of Machine Learning Algorithms for Diagnosis of Diffuse Lung Diseases Background In the last decades, new optimization methods based on the nature's intelligence were developed. These metaheuristics can find a nearly optimal solution faster than other traditional algorithms even for high-dimensional optimization problems. All these algorithms have a similar structure, the difference being made by the strategies used during the evolutionary process. Objectives A set of three nature-inspired algorithms, including Cuckoo Search algorithm (CSA), Particle Swarm Optimization (PSO), and Multi-Swarm Optimization (MSO), are compared in terms of strategies used in the evolutionary process and also of the results obtained in case of particular optimization problems. Methods The three algorithms were applied for biomedical image registration (IR) and compared in terms of performances. The expected geometric transform has seven parameters and is composed of rotation against a point in the image, scaling on both axis with different factors, and translation. Results The evaluation consisted of 25 runs of each IR procedure and revealed that (1) PSO offers the most precise solutions; (2) CSA and MSO are more stable in the sense that their solutions are less scattered; and (3) MSO and PSO have a higher convergence speed. Conclusions The evaluation of PSO, MSO, and CSA was made for multimodal IR problems. It is possible that for other optimization problems and also for other settings of the optimization algorithms, the results can be different. Therefore, the nature-inspired algorithms demonstrated their efficacy for this class of optimization problems.


Symmetry ◽  
2020 ◽  
Vol 12 (4) ◽  
pp. 635
Author(s):  
Zong-Sheng Wang ◽  
Jung Lee ◽  
Chang Geun Song ◽  
Sun-Jeong Kim

The imperialist competitive algorithm combined with chaos theory (CICA) demonstrates excellent performance in global optimization problems. However, its computational complexity increases with the introduction of chaotic maps. To address this, we integrate CICA with a dropout strategy that randomly samples the dimensions of each solution at each iteration of the computation. We investigate the potential of the proposed algorithm with different chaotic maps through six symmetric and six asymmetric benchmark functions. We also apply the proposed algorithm to AUVs’ path planning application showing its performance and effectiveness in solving real problems. The simulation results show that the proposed algorithm not only has low computational complexity but also enhances local search capability near the globally optimal solution with an insignificant loss in the success rate.


Sign in / Sign up

Export Citation Format

Share Document