stochastic local search
Recently Published Documents


TOTAL DOCUMENTS

195
(FIVE YEARS 41)

H-INDEX

19
(FIVE YEARS 2)

Author(s):  
Daniel Loscos ◽  
Narciso Martí-Oliet ◽  
Ismael Rodríguez

2021 ◽  
Vol 4 ◽  
Author(s):  
Sophie Burkhardt ◽  
Jannis Brugger ◽  
Nicolas Wagner ◽  
Zahra Ahmadi ◽  
Kristian Kersting ◽  
...  

Classification approaches that allow to extract logical rules such as decision trees are often considered to be more interpretable than neural networks. Also, logical rules are comparatively easy to verify with any possible input. This is an important part in systems that aim to ensure correct operation of a given model. However, for high-dimensional input data such as images, the individual symbols, i.e. pixels, are not easily interpretable. Therefore, rule-based approaches are not typically used for this kind of high-dimensional data. We introduce the concept of first-order convolutional rules, which are logical rules that can be extracted using a convolutional neural network (CNN), and whose complexity depends on the size of the convolutional filter and not on the dimensionality of the input. Our approach is based on rule extraction from binary neural networks with stochastic local search. We show how to extract rules that are not necessarily short, but characteristic of the input, and easy to visualize. Our experiments show that the proposed approach is able to model the functionality of the neural network while at the same time producing interpretable logical rules. Thus, we demonstrate the potential of rule-based approaches for images which allows to combine advantages of neural networks and rule learning.


Author(s):  
Ivan Sergienko ◽  
Vladimir Shylo ◽  
Valentyna Roshchyn ◽  
Petro Shylo

Introduction. Solving large-scale discrete optimization problems requires the processing of large-scale data in a reasonable time. Efficient solving is only possible by using multiprocessor computer systems. However, it is a daunting challenge to adapt existing optimization algorithms to get all the benefits of these parallel computing systems. The available computational resources are ineffective without efficient and scalable parallel methods. In this connection, the algorithm unions (portfolios and teams) play a crucial role in the parallel processing of discrete optimization problems. The purpose. The purpose of this paper is to research the efficiency of the algorithm portfolios by solving the weighted max-cut problem. The research is carried out in two stages using stochastic local search algorithms. Results. In this paper, we investigate homogeneous and non-homogeneous algorithm portfolios. We developed the homogeneous portfolios of two stochastic local optimization algorithms for the weighted max-cut problem, which has numerous applications. The results confirm the advantages of the proposed methods. Conclusions. Algorithm portfolios could be used to solve well-known discrete optimization problems of unprecedented scale and significantly improve their solving time. Further, we propose using communication between algorithms, namely teams and portfolios of algorithm teams. The algorithms in a team communicate with each other to boost overall performance. It is supposed that algorithm communication allows enhancing the best features of the developed algorithms and would improve the computational times and solution quality. The underlying algorithms should be able to utilize relevant data that is being communicated effectively to achieve any computational benefit from communication. Keywords: Discrete optimization, algorithm portfolios, computational experiment.


2021 ◽  
pp. 115206
Author(s):  
Leslie Pérez Cáceres ◽  
Ignacio Araya ◽  
Guillermo Cabrera-Guerrero

2021 ◽  
Vol 23 (04) ◽  
pp. 278-295
Author(s):  
Mohammad Zakarai ◽  
◽  
Hegazy Zaher ◽  
Naglaa Ragaa ◽  
◽  
...  

The assembly line balancing problems have great importance in research and industry fields. They allow minimizing the learning aspects and guaranteeing a fixed number of products per day. This paper introduces a new problem that combines the multi-manned concept with the U-shaped lines with time and space constraints under uncertainty. The processing time of the tasks is considered as random variables with known means and variances. Therefore, chance-constraints appear in the cycle time constraints. In addition, each task has an associated area, where the assigned tasks per station are restricted by a total area. The proposed algorithm for solving the problem is a stochastic local search algorithm. The parameter levels of the proposed algorithm are optimized by the Taguchi method to cover the small, medium, and large-sized problems. Well-known benchmark problems have been adapted to cover the new model. The computational results showed the importance of the new problem and the efficiency of the proposed algorithm.


Author(s):  
Henry Kautz ◽  
Ashish Sabharwal ◽  
Bart Selman

Research on incomplete algorithms for satisfiability testing lead to some of the first scalable SAT solvers in the early 1990’s. Unlike systematic solvers often based on an exhaustive branching and backtracking search, incomplete methods are generally based on stochastic local search. On problems from a variety of domains, such incomplete methods for SAT can significantly outperform DPLL-based methods. While the early greedy algorithms already showed promise, especially on random instances, the introduction of randomization and so-called uphill moves during the search significantly extended the reach of incomplete algorithms for SAT. This chapter discusses such algorithms, along with a few key techniques that helped boost their performance such as focusing on variables appearing in currently unsatisfied clauses, devising methods to efficiently pull the search out of local minima through clause re-weighting, and adaptive noise mechanisms. The chapter also briefly discusses a formal foundation for some of the techniques based on the discrete Lagrangian method.


2021 ◽  
Vol 17 (1) ◽  
pp. 1-10
Author(s):  
Hayder Al-Behadili

In today’s world, the data generated by many applications are increasing drastically, and finding an optimal subset of features from the data has become a crucial task. The main objective of this review is to analyze and comprehend different stochastic local search algorithms to find an optimal feature subset. Simulated annealing, tabu search, genetic programming, genetic algorithm, particle swarm optimization, artificial bee colony, grey wolf optimization, and bat algorithm, which have been used in feature selection, are discussed. This review also highlights the filter and wrapper approaches for feature selection. Furthermore, this review highlights the main components of stochastic local search algorithms, categorizes these algorithms in accordance with the type, and discusses the promising research directions for such algorithms in future research of feature selection.


Author(s):  
Muhamet Kastrati ◽  
Marenglen Biba

The main objective of this paper is to provide a state-of-the-art review, analyze and discuss stochastic local search techniques used for solving hard combinatorial problems. It begins with a short introduction, motivation and some basic notation on combinatorial problems, search paradigms and other relevant features of searching techniques as needed for background. In the following a brief overview of the stochastic local search methods along with an analysis of the state-of-the-art stochastic local search algorithms is given. Finally, the last part of the paper present and discuss some of the most latest trends in application of stochastic local search algorithms in machine learning, data mining and some other areas of science and engineering. We conclude with a discussion on capabilities and limitations of stochastic local search algorithms.


Sign in / Sign up

Export Citation Format

Share Document