local optima
Recently Published Documents





Federico Antonello ◽  
Piero Baraldi ◽  
Enrico Zio ◽  
Luigi Serio

AbstractIn this work, a Multi-Objective Evolutionary Algorithm (MOEA) is developed to identify Functional Dependencies (FDEPs) in Complex Technical Infrastructures (CTIs) from alarm data. The objectives of the search are the maximization of a measure of novelty, which drives the exploration of the solution space avoiding to get trapped in local optima, and of a measure of dependency among alarms, which drives the uncovering of functional dependencies. The main contribution of the work is the direct identification of patterns of dependent alarms; this avoids going through the preliminary step of mining association rules, as typically done by state-of-the-art methods which, however, fail to identify rare functional dependencies due to the need of setting a balanced minimum occurrence threshold. The proposed framework for FDEPs identification is applied to a synthetic alarm database generated by a simulated CTI model and to a real large-scale database of alarms collected at the CTI of CERN (European Organization for Nuclear Research). The obtained results show that the framework enables the thorough exploration of the solution space and captures also rare functional dependencies.

2022 ◽  
Vol 2022 ◽  
pp. 1-13
Jhansi Rani Kaka ◽  
K. Satya Prasad

Early diagnosis of Alzheimer’s helps a doctor to decide the treatment for the patient based on the stages. The existing methods involve applying the deep learning methods for Alzheimer’s classification and have the limitations of overfitting problems. Some researchers were involved in applying the feature selection based on the optimization method, having limitations of easily trapping into local optima and poor convergence. In this research, Differential Evolution-Multiclass Support Vector Machine (DE-MSVM) is proposed to increase the performance of Alzheimer’s classification. The image normalization method is applied to enhance the quality of the image and represent the features effectively. The AlexNet model is applied to the normalized images to extract the features and also applied for feature selection. The Differential Evolution method applies Pareto Optimal Front for nondominated feature selection. This helps to select the feature that represents the characteristics of the input images. The selected features are applied in the MSVM method to represent in high dimension and classify Alzheimer’s. The DE-MSVM method has accuracy of 98.13% in the axial slice, and the existing whale optimization with MSVM has 95.23% accuracy.

Fukui Li ◽  
Jingyuan He ◽  
Mingliang Zhou ◽  
Bin Fang

Local search algorithms are widely applied in solving large-scale distributed constraint optimization problem (DCOP). Distributed stochastic algorithm (DSA) is a typical local search algorithm to solve DCOP. However, DSA has some drawbacks including easily falling into local optima and the unfairness of assignment choice. This paper presents a novel local search algorithm named VLSs to solve the issues. In VLSs, sampling according to the probability corresponding to assignment is introduced to enable each agent to choose other promising values. Besides, each agent alternately performs a greedy choice among multiple parallel solutions to reduce the chance of falling into local optima and a variance adjustment mechanism to guide the search into a relatively good initial solution in a periodic manner. We give the proof of variance adjustment mechanism rationality and theoretical explanation of impact of greed among multiple parallel solutions. The experimental results show the superiority of VLSs over state-of-the-art DCOP algorithms.

2022 ◽  
Francisco Daniel Filip Duarte

Abstract In optimization tasks, it is interesting to achieve a set of efficient solutions instead of one single output, in the case the best solution is not suitable. Many niching methods offer a diversified response, yet some important problems are common: (1) The most interesting solutions of each local optimum are not identified. Thus, the output is the overall population of solutions, which increases the work of the designer in verifying which solution is the most interesting. (2) Existing niching algorithms tend to distribute the solutions on the most promising regions, over-populating some local optima and sub-populating others, which leads to poor optimization.To solve these challenges, a novel niching method is presented, named local optimum ranking 2 (LOR2). This sorting methodology favors the exploration of a defined number of local optima and ranks each local population by objective value within each local optimum. Thus, is performed a multi-focus exploration, with an equalized number of solutions on each local optimum, while identifying which solutions are the local apices. To exemplify its application, the LOR2 algorithm is applied in the design optimization of a metallic cantilever beam. It achieves a set of efficient and diverse design configurations, offering both performance and diversity for structural design challenges.In addition, a second experiment describes how the algorithm can be applied to segment the domain of any function, into a mesh of similar sized or custom-sized elements. Thus, it can significantly simplify metamodels and reduce their computation time.

2022 ◽  
pp. 166-201
Asha Gowda Karegowda ◽  
Devika G.

Artificial neural networks (ANN) are often more suitable for classification problems. Even then, training of ANN is a surviving challenge task for large and high dimensional natured search space problems. These hitches are more for applications that involves process of fine tuning of ANN control parameters: weights and bias. There is no single search and optimization method that suits the weights and bias of ANN for all the problems. The traditional heuristic approach fails because of their poorer convergence speed and chances of ending up with local optima. In this connection, the meta-heuristic algorithms prove to provide consistent solution for optimizing ANN training parameters. This chapter will provide critics on both heuristics and meta-heuristic existing literature for training neural networks algorithms, applicability, and reliability on parameter optimization. In addition, the real-time applications of ANN will be presented. Finally, future directions to be explored in the field of ANN are presented which will of potential interest for upcoming researchers.

2022 ◽  
Vol 13 (1) ◽  
pp. 0-0

This paper reports the use of a nature-inspired metaheuristic algorithm known as ‘Whale Optimization Algorithm’ (WOA) for multimodal image registration. WOA is based on the hunting behaviour of Humpback whales and provides better exploration and exploitation of the search space with small possibility of trapping in local optima. Though WOA is used in various optimization problems, no detailed study is available for its use in image registration. For this study different sets of NIR and visible images are considered. The registration results are compared with the other state of the art image registration methods. The results show that WOA is a very competitive algorithm for NIR-visible image registration. With the advantages of better exploration of search space and local optima avoidance, the algorithm can be a suitable choice for multimodal image registration.

2022 ◽  
Vol 19 (1) ◽  
pp. 473-512
Rong Zheng ◽  
Heming Jia ◽  
Laith Abualigah ◽  
Qingxin Liu ◽  

<abstract> <p>Arithmetic optimization algorithm (AOA) is a newly proposed meta-heuristic method which is inspired by the arithmetic operators in mathematics. However, the AOA has the weaknesses of insufficient exploration capability and is likely to fall into local optima. To improve the searching quality of original AOA, this paper presents an improved AOA (IAOA) integrated with proposed forced switching mechanism (FSM). The enhanced algorithm uses the random math optimizer probability (<italic>RMOP</italic>) to increase the population diversity for better global search. And then the forced switching mechanism is introduced into the AOA to help the search agents jump out of the local optima. When the search agents cannot find better positions within a certain number of iterations, the proposed FSM will make them conduct the exploratory behavior. Thus the cases of being trapped into local optima can be avoided effectively. The proposed IAOA is extensively tested by twenty-three classical benchmark functions and ten CEC2020 test functions and compared with the AOA and other well-known optimization algorithms. The experimental results show that the proposed algorithm is superior to other comparative algorithms on most of the test functions. Furthermore, the test results of two training problems of multi-layer perceptron (MLP) and three classical engineering design problems also indicate that the proposed IAOA is highly effective when dealing with real-world problems.</p> </abstract>

2022 ◽  
Vol 13 (1) ◽  
pp. 0-0

Clustering is an unsupervised machine learning technique that optimally organizes the data objects in a group of clusters. In present work, a meta-heuristic algorithm based on cat intelligence is adopted for optimizing clustering problems. Further, to make the cat swarm algorithm (CSO) more robust for partitional clustering, some modifications are incorporated in it. These modifications include an improved solution search equation for balancing global and local searches, accelerated velocity equation for addressing diversity, especially in tracing mode. Furthermore, a neighborhood-based search strategy is introduced to handle the local optima and premature convergence problems. The performance of enhanced cat swarm optimization (ECSO) algorithm is tested on eight real-life datasets and compared with the well-known clustering algorithms. The simulation results confirm that the proposed algorithm attains the optimal results than other clustering algorithms.

Sign in / Sign up

Export Citation Format

Share Document