scholarly journals Energy Efficient Routing in Iot Based on Chronological Atom Search Optimization Algorithm

Internet of Things (IoT) is a developing technology used for enhancing the communication and providing flexibility to humans. The data packets are transmitted through better routing path using the IoT nodes in the network. This research presents an effective optimization algorithm for selecting the effective routing path for the transmission of the speech signal to the destination effectively. The optimization algorithm namely, chronological atom search optimization (chronological ASO) algorithm is developed for selecting the routing path. The chronological ASO algorithm is developed by integrating the chronological concept in ASO algorithm. After the initialization of the network, the effective routing path is selected using the fitness parameters, such as energy, delay and distance. Then, the speech signal is transmitted from the source node to the destination node by selecting the routing path having maximum fitness function. The performance analysis of the proposed chronological ASO algorithm is done based on the metrics, such as energy, throughput, and PDR. When compared to the existing methods, the proposed method provided a maximal energy of 0.2375, maximal PDR of 46.1538, and maximal throughput of 0.0792 for 50 nodes and obtained a maximal energy of 0.1381, maximal PDR of 32.0988, and maximal throughput of 0.0395, respectively.

2021 ◽  
Vol 11 (10) ◽  
pp. 4382
Author(s):  
Ali Sadeghi ◽  
Sajjad Amiri Doumari ◽  
Mohammad Dehghani ◽  
Zeinab Montazeri ◽  
Pavel Trojovský ◽  
...  

Optimization is the science that presents a solution among the available solutions considering an optimization problem’s limitations. Optimization algorithms have been introduced as efficient tools for solving optimization problems. These algorithms are designed based on various natural phenomena, behavior, the lifestyle of living beings, physical laws, rules of games, etc. In this paper, a new optimization algorithm called the good and bad groups-based optimizer (GBGBO) is introduced to solve various optimization problems. In GBGBO, population members update under the influence of two groups named the good group and the bad group. The good group consists of a certain number of the population members with better fitness function than other members and the bad group consists of a number of the population members with worse fitness function than other members of the population. GBGBO is mathematically modeled and its performance in solving optimization problems was tested on a set of twenty-three different objective functions. In addition, for further analysis, the results obtained from the proposed algorithm were compared with eight optimization algorithms: genetic algorithm (GA), particle swarm optimization (PSO), gravitational search algorithm (GSA), teaching–learning-based optimization (TLBO), gray wolf optimizer (GWO), and the whale optimization algorithm (WOA), tunicate swarm algorithm (TSA), and marine predators algorithm (MPA). The results show that the proposed GBGBO algorithm has a good ability to solve various optimization problems and is more competitive than other similar algorithms.


2021 ◽  
Vol 104 (2) ◽  
pp. 003685042110254
Author(s):  
Armaghan Mohsin ◽  
Yazan Alsmadi ◽  
Ali Arshad Uppal ◽  
Sardar Muhammad Gulfam

In this paper, a novel modified optimization algorithm is presented, which combines Nelder-Mead (NM) method with a gradient-based approach. The well-known Nelder Mead optimization technique is widely used but it suffers from convergence issues in higher dimensional complex problems. Unlike the NM, in this proposed technique we have focused on two issues of the NM approach, one is shape of the simplex which is reshaped at each iteration according to the objective function, so we used a fixed shape of the simplex and we regenerate the simplex at each iteration and the second issue is related to reflection and expansion steps of the NM technique in each iteration, NM used fixed value of [Formula: see text], that is, [Formula: see text]  = 1 for reflection and [Formula: see text]  = 2 for expansion and replace the worst point of the simplex with that new point in each iteration. In this way NM search the optimum point. In proposed algorithm the optimum value of the parameter [Formula: see text] is computed and then centroid of new simplex is originated at this optimum point and regenerate the simplex with this centroid in each iteration that optimum value of [Formula: see text] will ensure the fast convergence of the proposed technique. The proposed algorithm has been applied to the real time implementation of the transversal adaptive filter. The application used to demonstrate the performance of the proposed technique is a well-known convex optimization problem having quadratic cost function, and results show that the proposed technique shows fast convergence than the Nelder-Mead method for lower dimension problems and the proposed technique has also good convergence for higher dimensions, that is, for higher filter taps problem. The proposed technique has also been compared with stochastic techniques like LMS and NLMS (benchmark) techniques. The proposed technique shows good results against LMS. The comparison shows that the modified algorithm guarantees quite acceptable convergence with improved accuracy for higher dimensional identification problems.


2021 ◽  
Author(s):  
Rekha G ◽  
Krishna Reddy V ◽  
chandrashekar jatoth ◽  
Ugo Fiore

Abstract Class imbalance problems have attracted the research community but a few works have focused on feature selection with imbalanced datasets. To handle class imbalance problems, we developed a novel fitness function for feature selection using the chaotic salp swarm optimization algorithm, an efficient meta-heuristic optimization algorithm that has been successfully used in a wide range of optimization problems. This paper proposes an Adaboost algorithm with chaotic salp swarm optimization. The most discriminating features are selected using salp swarm optimization and Adaboost classifiers are thereafter trained on the features selected. Experiments show the ability of the proposed technique to find the optimal features with performance maximization of Adaboost.


Sign in / Sign up

Export Citation Format

Share Document