Modified Firefly Algorithm With Chaos Theory for Feature Selection

2019 ◽  
Vol 10 (2) ◽  
pp. 1-20 ◽  
Author(s):  
Sujata Dash ◽  
Ruppa Thulasiram ◽  
Parimala Thulasiraman

Conventional algorithms such as gradient-based optimization methods usually struggle to deal with high-dimensional non-linear problems and often land up with local minima. Recently developed nature-inspired optimization algorithms are the best approaches for finding global solutions for combinatorial optimization problems like microarray datasets. In this article, a novel hybrid swarm intelligence-based meta-search algorithm is proposed by combining a heuristic method called conditional mutual information maximization with chaos-based firefly algorithm. The combined algorithm is computed in an iterative manner to boost the sharing of information between fireflies, enhancing the search efficiency of chaos-based firefly algorithm and reduces the computational complexities of feature selection. The meta-search model is implemented using a well-established classifier, such as support vector machine as the modeler in a wrapper approach. The chaos-based firefly algorithm increases the global search mobility of fireflies. The efficiency of the model is studied over high-dimensional disease datasets and compared with standard firefly algorithm, particle swarm optimization, and genetic algorithm in the same experimental environment to establish its superiority of feature selection over selected counterparts.

2021 ◽  
Vol 12 (2) ◽  
pp. 1-15
Author(s):  
Khadoudja Ghanem ◽  
Abdesslem Layeb

Backtracking search optimization algorithm is a recent stochastic-based global search algorithm for solving real-valued numerical optimization problems. In this paper, a binary version of backtracking algorithm is proposed to deal with 0-1 optimization problems such as feature selection and knapsack problems. Feature selection is the process of selecting a subset of relevant features for use in model construction. Irrelevant features can negatively impact model performances. On the other hand, knapsack problem is a well-known optimization problem used to assess discrete algorithms. The objective of this research is to evaluate the discrete version of backtracking algorithm on the two mentioned problems and compare obtained results with other binary optimization algorithms using four usual classifiers: logistic regression, decision tree, random forest, and support vector machine. Empirical study on biological microarray data and experiments on 0-1 knapsack problems show the effectiveness of the binary algorithm and its ability to achieve good quality solutions for both problems.


2011 ◽  
Vol 14 (2) ◽  
pp. 22-28
Author(s):  
Hung Vo Duong

In this research, Tabu search algorithm, a heuristic method for solving combinatorial optimization problems, has been applied for type 2 problems of assembly line balancing. For type 2 problems, two methodologies are developed for problem solving. Method 1 is direct solving for type 2 problems, and method 2 gives solving through type 1 problems. As such, Tabu search algorithm for type 1 problem is employed for problem solving at second stage. The success of this research points out empty workstations (unnecessary) to reduce investment cost and operational costs. Moreover, the range of cycle time and number of workststions are provided for selection.


Author(s):  
VLADIMIR NIKULIN ◽  
TIAN-HSIANG HUANG ◽  
GEOFFREY J. MCLACHLAN

The method presented in this paper is novel as a natural combination of two mutually dependent steps. Feature selection is a key element (first step) in our classification system, which was employed during the 2010 International RSCTC data mining (bioinformatics) Challenge. The second step may be implemented using any suitable classifier such as linear regression, support vector machine or neural networks. We conducted leave-one-out (LOO) experiments with several feature selection techniques and classifiers. Based on the LOO evaluations, we decided to use feature selection with the separation type Wilcoxon-based criterion for all final submissions. The method presented in this paper was tested successfully during the RSCTC data mining Challenge, where we achieved the top score in the Basic track.


2020 ◽  
pp. 3397-3407
Author(s):  
Nur Syafiqah Mohd Nafis ◽  
Suryanti Awang

Text documents are unstructured and high dimensional. Effective feature selection is required to select the most important and significant feature from the sparse feature space. Thus, this paper proposed an embedded feature selection technique based on Term Frequency-Inverse Document Frequency (TF-IDF) and Support Vector Machine-Recursive Feature Elimination (SVM-RFE) for unstructured and high dimensional text classificationhis technique has the ability to measure the feature’s importance in a high-dimensional text document. In addition, it aims to increase the efficiency of the feature selection. Hence, obtaining a promising text classification accuracy. TF-IDF act as a filter approach which measures features importance of the text documents at the first stage. SVM-RFE utilized a backward feature elimination scheme to recursively remove insignificant features from the filtered feature subsets at the second stage. This research executes sets of experiments using a text document retrieved from a benchmark repository comprising a collection of Twitter posts. Pre-processing processes are applied to extract relevant features. After that, the pre-processed features are divided into training and testing datasets. Next, feature selection is implemented on the training dataset by calculating the TF-IDF score for each feature. SVM-RFE is applied for feature ranking as the next feature selection step. Only top-rank features will be selected for text classification using the SVM classifier. Based on the experiments, it shows that the proposed technique able to achieve 98% accuracy that outperformed other existing techniques. In conclusion, the proposed technique able to select the significant features in the unstructured and high dimensional text document.


2021 ◽  
Vol 12 (2) ◽  
pp. 85-99
Author(s):  
Nassima Dif ◽  
Zakaria Elberrichi

Hybrid metaheuristics has received a lot of attention lately to solve combinatorial optimization problems. The purpose of hybridization is to create a cooperation between metaheuristics for better solutions. Most proposed works were interested in static hybridization. The objective of this work is to propose a novel dynamic hybridization method (GPBD) that generates the most suitable sequential hybridization between GA, PSO, BAT, and DE metaheuristics, according to each problem. The authors choose to test this approach for solving the best feature selection problem in a wrapper tactic, performed on face image recognition datasets, with the k-nearest neighbor (KNN) learning algorithm. The comparative study of the metaheuristics and their hybridization GPBD shows that the proposed approach achieved the best results. It was definitely competitive with other filter approaches proposed in the literature. It achieved a perfect accuracy score of 100% for Orl10P, Pix10P, and PIE10P datasets.


2018 ◽  
Vol 2018 ◽  
pp. 1-21 ◽  
Author(s):  
Sizhou Sun ◽  
Jingqi Fu ◽  
Feng Zhu ◽  
Nan Xiong

The aims of this study contribute to a new hybrid model by combining ensemble empirical mode decomposition (EEMD) with multikernel function least square support vector machine (MKLSSVM) optimized by hybrid gravitation search algorithm (HGSA) for short-term wind speed prediction. In the forecasting process, EEMD is adopted to make the original wind speed data decomposed into intrinsic mode functions (IMFs) and one residual firstly. Then, partial autocorrelation function (PACF) is applied to identify the correlation between the corresponding decomposed components. Subsequently, the MKLSSVM using multikernel function of radial basis function (RBF) and polynomial (Poly) kernel function by weight coefficient is exploited as core forecasting engine to make the short-term wind speed prediction. To improve the regression performance, the binary-value GSA (BGSA) in HGSA is utilized as feature selection approach to remove the ineffective candidates and reconstruct the most relevant feature input-matrix for the forecasting engine, while real-value GSA (RGSA) makes the parameter combination optimization of MKLSSVM model. In the end, these respective decomposed subseries forecasting results are combined into the final forecasting values by aggregate calculation. Numerical results and comparable analysis illustrate the excellent performance of the EEMD-HGSA-MKLSSVM model when applied in the short-term wind speed forecasting.


2013 ◽  
Vol 411-414 ◽  
pp. 1904-1910
Author(s):  
Kai Zhong Jiang ◽  
Tian Bo Wang ◽  
Zhong Tuan Zheng ◽  
Yu Zhou

An algorithm based on free search is proposed for the combinatorial optimization problems. In this algorithm, a feasible solution is converted into a full permutation of all the elements and a transformation of one solution into another solution can be interpreted the transformation of one permutation into another permutation. Then, the algorithm is combined with intersection elimination. The discrete free search algorithm greatly improves the convergence rate of the search process and enhances the quality of the results. The experiment results on TSP standard data show that the performance of the proposed algorithm is increased by about 2.7% than that of the genetic algorithm.


Sign in / Sign up

Export Citation Format

Share Document