scholarly journals Novel clustering-based pruning algorithms

2020 ◽  
Vol 23 (3) ◽  
pp. 1049-1058
Author(s):  
Paweł Zyblewski ◽  
Michał Woźniak
Keyword(s):  
2017 ◽  
Vol 3 ◽  
pp. e137 ◽  
Author(s):  
Mona Alshahrani ◽  
Othman Soufan ◽  
Arturo Magana-Mora ◽  
Vladimir B. Bajic

Background Artificial neural networks (ANNs) are a robust class of machine learning models and are a frequent choice for solving classification problems. However, determining the structure of the ANNs is not trivial as a large number of weights (connection links) may lead to overfitting the training data. Although several ANN pruning algorithms have been proposed for the simplification of ANNs, these algorithms are not able to efficiently cope with intricate ANN structures required for complex classification problems. Methods We developed DANNP, a web-based tool, that implements parallelized versions of several ANN pruning algorithms. The DANNP tool uses a modified version of the Fast Compressed Neural Network software implemented in C++ to considerably enhance the running time of the ANN pruning algorithms we implemented. In addition to the performance evaluation of the pruned ANNs, we systematically compared the set of features that remained in the pruned ANN with those obtained by different state-of-the-art feature selection (FS) methods. Results Although the ANN pruning algorithms are not entirely parallelizable, DANNP was able to speed up the ANN pruning up to eight times on a 32-core machine, compared to the serial implementations. To assess the impact of the ANN pruning by DANNP tool, we used 16 datasets from different domains. In eight out of the 16 datasets, DANNP significantly reduced the number of weights by 70%–99%, while maintaining a competitive or better model performance compared to the unpruned ANN. Finally, we used a naïve Bayes classifier derived with the features selected as a byproduct of the ANN pruning and demonstrated that its accuracy is comparable to those obtained by the classifiers trained with the features selected by several state-of-the-art FS methods. The FS ranking methodology proposed in this study allows the users to identify the most discriminant features of the problem at hand. To the best of our knowledge, DANNP (publicly available at www.cbrc.kaust.edu.sa/dannp) is the only available and on-line accessible tool that provides multiple parallelized ANN pruning options. Datasets and DANNP code can be obtained at www.cbrc.kaust.edu.sa/dannp/data.php and https://doi.org/10.5281/zenodo.1001086.


2001 ◽  
Vol 14 ◽  
pp. 1-28 ◽  
Author(s):  
R. I. Brafman

In recent years, there is a growing awareness of the importance of reachability and relevance-based pruning techniques for planning, but little work specifically targets these techniques. In this paper, we compare the ability of two classes of algorithms to propagate and discover reachability and relevance constraints in classical planning problems. The first class of algorithms operates on SAT encoded planning problems obtained using the linear and Graphplan encoding schemes. It applies unit-propagation and more general resolution steps (involving larger clauses) to these plan encodings. The second class operates at the plan level and contains two families of pruning algorithms: Reachable-k and Relevant-k. Reachable-k provides a coherent description of a number of existing forward pruning techniques used in numerous algorithms, while Relevant-k captures different grades of backward pruning. Our results shed light on the ability of different plan-encoding schemes to propagate information forward and backward and on the relative merit of plan-level and SAT-level pruning methods.


2006 ◽  
Vol 16 (04) ◽  
pp. 283-293 ◽  
Author(s):  
PEI-YI HAO ◽  
JUNG-HSIEN CHIANG

This paper presents the pruning and model-selecting algorithms to the support vector learning for sample classification and function regression. When constructing RBF network by support vector learning we occasionally obtain redundant support vectors which do not significantly affect the final classification and function approximation results. The pruning algorithms primarily based on the sensitivity measure and the penalty term. The kernel function parameters and the position of each support vector are updated in order to have minimal increase in error, and this makes the structure of SVM network more flexible. We illustrate this approach with synthetic data simulation and face detection problem in order to demonstrate the pruning effectiveness.


ICGA Journal ◽  
1998 ◽  
Vol 21 (3) ◽  
pp. 193-193
Author(s):  
D. Carmel ◽  
S. Markovitch
Keyword(s):  

2015 ◽  
Vol 7 (2) ◽  
pp. 89-103 ◽  
Author(s):  
Jian Wang ◽  
Guoling Yang ◽  
Shan Liu ◽  
Jacek M. Zurada

Abstract Gradient descent method is one of the popular methods to train feedforward neural networks. Batch and incremental modes are the two most common methods to practically implement the gradient-based training for such networks. Furthermore, since generalization is an important property and quality criterion of a trained network, pruning algorithms with the addition of regularization terms have been widely used as an efficient way to achieve good generalization. In this paper, we review the convergence property and other performance aspects of recently researched training approaches based on different penalization terms. In addition, we show the smoothing approximation tricks when the penalty term is non-differentiable at origin.


2011 ◽  
Vol 23 (2) ◽  
pp. 271-280 ◽  
Author(s):  
Chyon Hae Kim ◽  
◽  
Hiroshi Tsujino ◽  
Shigeki Sugano ◽  

This paper addresses optimal motion for general machines. Approximation for optimal motion requires a global path planning algorithm that precisely calculates the whole dynamics of a machine in a brief calculation. We propose a path planning algorithm that consists of path searching and pruning algorithms. The pruning algorithmis based on our analysis of state resemblance in general phase space. To confirm precision, calculation cost, optimality and applicability of the proposed algorithm, we conducted several shortest time path planning experiments for the dynamic models of double inverted pendulums. Precision to reach the goal states of the pendulums was better than other algorithms. Calculation cost was 58 times faster at least. We could tune optimality of proposed algorithm via resolution parameters. A positive correlation between optimality and resolutions was confirmed. Applicability was confirmed in a torque based position and velocity feedback control simulation. As a result of this simulation, the double inverted pendulums tracked planned motion under noise while keeping within torque limitations.


2014 ◽  
Vol 2014 ◽  
pp. 1-8 ◽  
Author(s):  
Nuha A. S. Alwan

Multiobjective optimization methods for routing in static wireless mesh networks (WMNs), with more than one QoS measure to be optimized, are highly challenging. To optimize the performance for a given end-to-end route in a static network, the most common metrics that need to be optimized or bounded are the path capacity and the end-to-end delay. In this work, we focus on combining desirable properties of these two metrics by minimizing a weighted metrics sum via a Dijkstra-based algorithm. The approach is directed towards fast convergence rather than optimality. It is shown that the resulting algorithm provides more satisfactory results than simple Dijkstra-based pruning algorithms in terms of simultaneously achieving high capacity and small delay. The effect of changing the weighting factor on the proposed algorithm performance is investigated.


2014 ◽  
Vol 42 (3) ◽  
pp. 406-429 ◽  
Author(s):  
Qun Dai ◽  
Meiling Li

Sign in / Sign up

Export Citation Format

Share Document