Optimization of Shell and Tube Heat Exchanger Using Alpha Tuning Elephant Herding Optimization (EHO) Technique

Author(s):  
Jiten Makadia ◽  
C.D. Sankhavara

Swarm Intelligence algorithms like PSO (Particle Swarm Optimization), ACO (Ant Colony Optimization), ABC (Artificial Bee Colony), Glow-worm swarm Optimization, etc. have been utilized by researchers for solving optimization problems. This work presents the application of a novel modified EHO (Elephant Herding Optimization) for cost optimization of shell and tube heat exchanger. A comparison of the results obtained by EHO in two benchmark problems shows that it is superior to those obtained with genetic algorithm and particle swarm optimization. The overall cost reduction is 13.3 % and 9.68% for both the benchmark problem compared to PSO. Results indicate that EHO can be effectively utilized for solving real-life optimization problems.

Author(s):  
T. O. Ting ◽  
H. C. Ting ◽  
T. S. Lee

In this work, a hybrid Taguchi-Particle Swarm Optimization (TPSO) is proposed to solve global numerical optimization problems with continuous and discrete variables. This hybrid algorithm combines the well-known Particle Swarm Optimization Algorithm with the established Taguchi method, which has been an important tool for robust design. This paper presents the improvements obtained despite the simplicity of the hybridization process. The Taguchi method is run only once in every PSO iteration and therefore does not give significant impact in terms of computational cost. The method creates a more diversified population, which also contributes to the success of avoiding premature convergence. The proposed method is effectively applied to solve 13 benchmark problems. This study’s results show drastic improvements in comparison with the standard PSO algorithm involving continuous and discrete variables on high dimensional benchmark functions.


2020 ◽  
Vol 2020 ◽  
pp. 1-26
Author(s):  
Wusi Yang ◽  
Li Chen ◽  
Yi Wang ◽  
Maosheng Zhang

The recently proposed multiobjective particle swarm optimization algorithm based on competition mechanism algorithm cannot effectively deal with many-objective optimization problems, which is characterized by relatively poor convergence and diversity, and long computing runtime. In this paper, a novel multi/many-objective particle swarm optimization algorithm based on competition mechanism is proposed, which maintains population diversity by the maximum and minimum angle between ordinary and extreme individuals. And the recently proposed θ-dominance is adopted to further enhance the performance of the algorithm. The proposed algorithm is evaluated on the standard benchmark problems DTLZ, WFG, and UF1-9 and compared with the four recently proposed multiobjective particle swarm optimization algorithms and four state-of-the-art many-objective evolutionary optimization algorithms. The experimental results indicate that the proposed algorithm has better convergence and diversity, and its performance is superior to other comparative algorithms on most test instances.


2015 ◽  
Vol 10 (2) ◽  
pp. 81-96 ◽  
Author(s):  
Sandip K. Lahiri ◽  
Nadeem Muhammed Khalfe

Abstract Owing to the wide utilization of shell and tube heat exchangers (STHEs) in industrial processes, their cost minimization is an important target for both designers and users. Traditional design approaches are based on iterative procedures which gradually change the design and geometric parameters until satisfying a given heat duty and set of geometric and operational constraints. Although well proven, this kind of approach is time-consuming and may not lead to cost-effective design. The present study explores the use of non-traditional optimization technique called hybrid particle swarm optimization (PSO) and ant colony optimization (ACO), for design optimization of STHEs from economic point of view. The PSO applies for global optimization and ant colony approach is employed to update positions of particles to attain rapidly the feasible solution space. ACO works as a local search, wherein ants apply pheromone-guided mechanism to update the positions found by the particles in the earlier stage. The optimization procedure involves the selection of the major geometric parameters such as tube diameters, tube length, baffle spacing, number of tube passes, tube layout, type of head, baffle cut, etc. and minimization of total annual cost is considered as design target. The methodology takes into account the geometric and operational constraints typically recommended by design codes. Three different case studies are presented to demonstrate the effectiveness and accuracy of proposed algorithm. The examples analyzed show that the hybrid PSO and ACO algorithm provides a valuable tool for optimal design of heat exchanger. The hybrid PSO and ACO approach is able to reduce the total cost of heat exchanger as compare to cost obtained by previously reported genetic algorithm (GA) approach. The result comparisons with particle swarm optimizer and other optimization algorithms (GA) demonstrate the effectiveness of the presented method.


2012 ◽  
Vol 239-240 ◽  
pp. 1027-1032 ◽  
Author(s):  
Qing Guo Wei ◽  
Yan Mei Wang ◽  
Zong Wu Lu

Applying many electrodes is undesirable for real-life brain-computer interface (BCI) application since the recording preparation can be troublesome and time-consuming. Multi-objective particle swarm optimization (MOPSO) has been widely utilized to solve multi-objective optimization problems and thus can be employed for channel selection. This paper presented a novel method named cultural-based MOPSO (CMOPSO) for channel selection in motor imagery based BCI. The CMOPSO method introduces a cultural framework to adapt the personalized flight parameters of the mutated particles. A comparison between the proposed algorithm and typical L1-norm algorithm was conducted, and the results showed that the proposed approach is more effective in selecting a smaller subset of channels while maintaining the classification accuracy unreduced.


2019 ◽  
Vol 8 (3) ◽  
pp. 8259-8265

Particle swarm optimization (PSO) is one of the most capable algorithms that reside to the swarm intelligence (SI) systems. Recently, it becomes very popular and renowned because of the easy implementation in complex/real life optimization problems. However, PSO has some observable drawbacks such as diversity maintenance, pre convergence and/or slow convergence speed. The ultimate success of PSO depends on the velocity update of the particles. Velocity has a significant dependence on its multiplied coefficient like inertia weight and acceleration factors. To increase the ability of PSO, this paper introduced an enriched PSO (namely ePSO), to solve hard optimization problems more precisely, efficiently and reliably. In ePSO novel gradually decreased inertia weight (as an alternative of a fixed constant value) and new gradually decreased and/or increased acceleration factors (meant for cognitive and social modules) is introduced. Proposed ePSO is used to solve four well known typical unconstrained benchmark functions and four complex unconstrained real life problems. The overall observation shows that proposed new algorithm ePSO is fitter than the compared algorithms significantly and statistically. Moreover, the convergence accuracy and speed of ePSO are also improved effectively


2014 ◽  
Vol 2014 ◽  
pp. 1-23 ◽  
Author(s):  
Martins Akugbe Arasomwan ◽  
Aderemi Oluyinka Adewumi

A new local search technique is proposed and used to improve the performance of particle swarm optimization algorithms by addressing the problem of premature convergence. In the proposed local search technique, a potential particle position in the solution search space is collectively constructed by a number of randomly selected particles in the swarm. The number of times the selection is made varies with the dimension of the optimization problem and each selected particle donates the value in the location of its randomly selected dimension from its personal best. After constructing the potential particle position, some local search is done around its neighbourhood in comparison with the current swarm global best position. It is then used to replace the global best particle position if it is found to be better; otherwise no replacement is made. Using some well-studied benchmark problems with low and high dimensions, numerical simulations were used to validate the performance of the improved algorithms. Comparisons were made with four different PSO variants, two of the variants implement different local search technique while the other two do not. Results show that the improved algorithms could obtain better quality solution while demonstrating better convergence velocity and precision, stability, robustness, and global-local search ability than the competing variants.


2014 ◽  
Vol 4 (3) ◽  
pp. 189-204 ◽  
Author(s):  
Simone A. Ludwig

Abstract Adaptive Particle Swarm Optimization (PSO) variants have become popular in recent years. The main idea of these adaptive PSO variants is that they adaptively change their search behavior during the optimization process based on information gathered during the run. Adaptive PSO variants have shown to be able to solve a wide range of difficult optimization problems efficiently and effectively. In this paper we propose a Repulsive Self-adaptive Acceleration PSO (RSAPSO) variant that adaptively optimizes the velocity weights of every particle at every iteration. The velocity weights include the acceleration constants as well as the inertia weight that are responsible for the balance between exploration and exploitation. Our proposed RSAPSO variant optimizes the velocity weights that are then used to search for the optimal solution of the problem (e.g., benchmark function). We compare RSAPSO to four known adaptive PSO variants (decreasing weight PSO, time-varying acceleration coefficients PSO, guaranteed convergence PSO, and attractive and repulsive PSO) on twenty benchmark problems. The results show that RSAPSO achives better results compared to the known PSO variants on difficult optimization problems that require large numbers of function evaluations.


Sign in / Sign up

Export Citation Format

Share Document