Journal:Informatica
Volume 11, Issue 3 (2000), pp. 281–296
Abstract
In this paper we present an algorithm for generating quadratic assignment problem (QAP) instances with known provably optimal solution. The flow matrix of such instances is constructed from the matrices corresponding to special graphs whose size may reach the dimension of the problem. In this respect, the algorithm generalizes some existing algorithms based on the iterative selection of triangles only. The set of instances which can be produced by the algorithm is NP-hard. Using multi-start descent heuristic for the QAP, we compare experimentally such test cases against those created by several existing generators and against Nugent-type problems from the QAPLIB as well.
Journal:Informatica
Volume 8, Issue 4 (1997), pp. 465–476
Abstract
The problem of parameter clustering on the basis of their correlation matrix is considered. The convergence in probability of parameter clustering based on the simulated annealing is investigated theoretically.
Journal:Informatica
Volume 8, Issue 3 (1997), pp. 377–400
Abstract
In this paper we define a class of edge-weighted graphs having nonnegatively valued bisections. We show experimentally that complete such graphs with more than three vertices and also some special graphs with only positive edges can be applied to improve the existing lower bounds for a version of the quadratic assignment problem, namely with a matrix composed of rectilinear distances between points in the Euclidean space.
Journal:Informatica
Volume 5, Issues 1-2 (1994), pp. 241–255
Abstract
Neural networks are often characterized as highly nonlinear systems of fairly large amount of parameters (in order of 103 – 104). This fact makes the optimization of parameters to be a nontrivial problem. But the astonishing moment is that the local optimization technique is widely used and yields reliable convergence in many cases. Obviously, the optimization of neural networks is high-dimensional, multi-extremal problem, so, as usual, the global optimization methods would be applied in this case. On the basis of Perceptron-like unit (which is the building block for the most architectures of neural networks) we analyze why the local optimization technique is so successful in the field of neural networks. The result is that a linear approximation of the neural network can be sufficient to evaluate the start point for the local optimization procedure in the nonlinear regime. This result can help in developing faster and more robust algorithms for the optimization of neural network parameters.
Journal:Informatica
Volume 1, Issue 1 (1990), pp. 20–39
Abstract
In this paper we deal with the problem of extremal parameter grouping. The problem formulation, the algorithms of parameter grouping and the fields of implementation are presented. The deterministic algorithms of extremal parameter grouping often find the local maximum of the functional, characterizing the quality of a partition. The problem has been formulated as a problem of combinatorial optimization and attempted to be solved using the simulated annealing strategy. The algorithms, realizing such a strategy and devoted to the solving of the problem concerned, are proposed and investigated.