Результаты поиска по 'minimization problem':
Найдено статей: 67
  1. Shushko N.I., Barashov E.B., Krasotkin S.A., Lemtuzhnikova D.V.
    Solving traveling salesman problem via clustering and a new algorithm for merging tours
    Computer Research and Modeling, 2025, v. 17, no. 1, pp. 45-58

    Traditional methods for solving the traveling salesman problem are not effective for high-dimensional problems due to their high computational complexity. One of the most effective ways to solve this problem is the decomposition approach, which includes three main stages: clustering vertices, solving subproblems within each cluster and then merging the obtained solutions into a final solution. This article focuses on the third stage — merging cycles of solving subproblems — since this stage is not always given sufficient attention, which leads to less accurate final solutions of the problem. The paper proposes a new modified Sigal algorithm for merging cycles. To evaluate its effectiveness, it is compared with two algorithms for merging cycles — the method of connecting midpoints of edges and an algorithm based on closeness of cluster centroids. The dependence of quality of solving subproblems on algorithms used for merging cycles is investigated. Sigal’s modified algorithm performs pairwise clustering and minimizes total distance. The centroid method focuses on connecting clusters based on closeness of centroids, and an algorithm using mid-points estimates the distance between mid-points of edges. Two types of clustering — k-means and affinity propagation — were also considered. Numerical experiments were performed using the TSPLIB dataset with different numbers of cities and topologies to test effectiveness of proposed algorithm. The study analyzes errors caused by the order in which clusters were merged, the quality of solving subtasks and number of clusters. Experiments show that the modified Sigal algorithm has the smallest median final distance and the most stable results compared to other methods. Results indicate that the quality of the final solution obtained using the modified Sigal algorithm is more stable depending on the sequence of merging clusters. Improving the quality of solving subproblems usually results in linear improvement of the final solution, but the pooling algorithm rarely affects the degree of this improvement.

  2. Pavlov P.A.
    Mathematical models and methods for organizing calculations in SMP systems
    Computer Research and Modeling, 2025, v. 17, no. 3, pp. 423-436

    The paper proposes and investigates a mathematical model of a distributed computing system of parallel interacting processes competing for the use of a limited number of copies of a structured software resource. In cases of unlimited and limited parallelism by the number of processors of a multiprocessor system, the problems of determining operational and exact values of the execution time of heterogeneous and identically distributed competing processes in a synchronous mode are solved, which ensures a linear order of execution of blocks of a structured software resource within each of the processes without delays. The obtained results can be used in a comparative analysis of mathematical relationships for calculating the implementation time of a set of parallel distributed interacting competing processes, a mathematical study of the efficiency and optimality of the organization of distributed computing, solving problems of constructing an optimal layout of blocks of an identically distributed system, finding the optimal number of processors that provide the directive execution time of given volumes of computations. The proposed models and methods open up new prospects for solving problems of optimal distribution of limited computing resources, synchronization of a set of interacting competing processes, minimization of system costs when executing parallel distributed processes.

  3. Dudarov S.P., Diev A.N., Fedosova N.A., Koltsova E.M.
    Simulation of properties of composite materials reinforced by carbon nanotubes using perceptron complexes
    Computer Research and Modeling, 2015, v. 7, no. 2, pp. 253-262

    Use of algorithms based on neural networks can be inefficient for small amounts of experimental data. Authors consider a solution of this problem in the context of modelling of properties of ceramic composite materials reinforced with carbon nanotubes using perceptron complex. This approach allowed us to obtain a mathematical description of the object of study with a minimal amount of input data (the amount of necessary experimental samples decreased 2–3.3 times). Authors considered different versions of perceptron complex structures. They found that the most appropriate structure has perceptron complex with breakthrough of two input variables. The relative error was only 6%. The selected perceptron complex was shown to be effective for predicting the properties of ceramic composites. The relative errors for output components were 0.3%, 4.2%, 0.4%, 2.9%, and 11.8%.

    Views (last year): 2. Citations: 1 (RSCI).
  4. Khokhlov N.I., Stetsyuk V.O., Mitskovets I.A.
    Overset grids approach for topography modeling in elastic-wave modeling using the grid-characteristic method
    Computer Research and Modeling, 2019, v. 11, no. 6, pp. 1049-1059

    While modeling seismic wave propagation, it is important to take into account nontrivial topography, as this topography causes multiple complex phenomena, such as diffraction at rough surfaces, complex propagation of Rayleigh waves, and side effects caused by wave interference. The primary goal of this research is to construct a method that implements the free surface on topography, utilizing an overset curved grid for characterization, while keeping the main grid structured rectangular. For a combination of the regular and curve-linear grid, the workability of the grid characteristics method using overset grids (also known as the Chimera grid approach) is analyzed. One of the benefits of this approach is computational complexity reduction, caused by the fact that simulation in a regular, homogeneous physical area using a sparse regular rectangle grid is simpler. The simplification of the mesh building mechanism (one grid is regular, and the other can be automatically built using surface data) is a side effect. Despite its simplicity, the method we propose allows us to increase the digitalization of fractured regions and minimize the Courant number. This paper contains various comparisons of modeling results produced by the proposed method-based solver, and results produced by the well-known solver specfem2d, as well as previous modeling results for the same problems. The drawback of the method is that an interpolation error can worsen an overall model accuracy and reduce the computational schema order. Some countermeasures against it are described. For this paper, only two-dimensional models are analyzed. However, the method we propose can be applied to the three-dimensional problems with minimal adaptation required.

  5. Gladin E.L., Borodich E.D.
    Variance reduction for minimax problems with a small dimension of one of the variables
    Computer Research and Modeling, 2022, v. 14, no. 2, pp. 257-275

    The paper is devoted to convex-concave saddle point problems where the objective is a sum of a large number of functions. Such problems attract considerable attention of the mathematical community due to the variety of applications in machine learning, including adversarial learning, adversarial attacks and robust reinforcement learning, to name a few. The individual functions in the sum usually represent losses related to examples from a data set. Additionally, the formulation admits a possibly nonsmooth composite term. Such terms often reflect regularization in machine learning problems. We assume that the dimension of one of the variable groups is relatively small (about a hundred or less), and the other one is large. This case arises, for example, when one considers the dual formulation for a minimization problem with a moderate number of constraints. The proposed approach is based on using Vaidya’s cutting plane method to minimize with respect to the outer block of variables. This optimization algorithm is especially effective when the dimension of the problem is not very large. An inexact oracle for Vaidya’s method is calculated via an approximate solution of the inner maximization problem, which is solved by the accelerated variance reduced algorithm Katyusha. Thus, we leverage the structure of the problem to achieve fast convergence. Separate complexity bounds for gradients of different components with respect to different variables are obtained in the study. The proposed approach is imposing very mild assumptions about the objective. In particular, neither strong convexity nor smoothness is required with respect to the low-dimensional variable group. The number of steps of the proposed algorithm as well as the arithmetic complexity of each step explicitly depend on the dimensionality of the outer variable, hence the assumption that it is relatively small.

  6. Borisova O.V., Borisov I.I., Nuzhdin K.A., Ledykov A.M., Kolyubin S.A.
    Computational design of closed-chain linkages: synthesis of ergonomic spine support module of exosuit
    Computer Research and Modeling, 2022, v. 14, no. 6, pp. 1269-1280

    The article focuses on the problem of mechanisms’ co-design for robotic systems to perform adaptive physical interaction with an unstructured environment, including physical human robot interaction. The co-design means simultaneous optimization of mechanics and control system, ensuring optimal behavior and performance of the system. Mechanics optimization refers to the search for optimal structure, geometric parameters, mass distribution among the links and their compliance; control refers to the search for motion trajectories for mechanism’s joints. The paper presents a generalized method of structural-parametric synthesis of underactuated mechanisms with closed kinematics for robotic systems for various purposes, e. g., it was previously used for the co-design of fingers’ mechanisms for anthropomorphic gripper and legs’ mechanisms for galloping robots. The method implements the concept of morphological computation of control laws due to the features of mechanical design, minimizing the control effort from the algorithmic component of the control system, which reduces the requirements for the level of technical equipment and reduces energy consumption. In this paper, the proposed method is used to optimize the structure and geometric parameters of the passive mechanism of the back support module of an industrial exosuit. Human movements are diverse and non-deterministic when compared with the movements of autonomous robots, which complicates the design of wearable robotic devices. To reduce injuries, fatigue and increase the productivity of workers, the synthesized industrial exosuit should not only compensate for loads, but also not interfere with the natural human motions. To test the developed exosuit, kinematic datasets from motion capture of an entire human body during industrial operations were used. The proposed method of structural-parametric synthesis was used to improve the ergonomics of a wearable robotic device. Verification of the synthesized mechanism was carried out using simulation: the passive module of the back is attached to two geometric primitives that move the chest and pelvis of the exosuit operator in accordance with the motion capture data. The ergonomics of the back module is quantified by the distance between the joints connecting the upper and bottom parts of the exosuit; minimizing deviation from the average value corresponds to a lesser limitation of the operator’s movement, i. e. greater ergonomics. The article provides a detailed description of the method of structural-parametric synthesis, an example of synthesis of an exosuit module and the results of simulation.

  7. Grenkin G.V.
    On the uniqueness of identification of reaction rate parameters in a combustion model
    Computer Research and Modeling, 2023, v. 15, no. 6, pp. 1469-1476

    A model of combustion of premixed mixture of gases with one global chemical reaction is considered, the model includes equations of the second order for temperature of mixture and concentrations of fuel and oxidizer, and the right-hand sides of these equations contain the reaction rate function. This function depends on five unknown parameters of the global reaction and serves as approximation to multistep reaction mechanism. The model is reduced, after replacement of variables, to one equation of the second order for temperature of mixture that transforms to a first-order equation for temperature derivative depending on temperature that contains a parameter of flame propagation velocity. Thus, for computing the parameter of burning velocity, one has to solve Dirichlet problem for first-order equation, and after that a model dependence of burning velocity on mixture equivalence ratio at specified reaction rate parameters will be obtained. Given the experimental data of dependence of burning velocity on mixture equivalence ratio, the problem of optimal selection of reaction rate parameters is stated, based on minimization of the mean square deviation of model values of burning velocity on experimental ones. The aim of our study is analysis of uniqueness of this problem solution. To this end, we apply computational experiment during which the problem of global search of optima is solved using multistart of gradient descent. The computational experiment clarifies that the inverse problem in this statement is underdetermined, and every time, when running gradient descent from a selected starting point, it converges to a new limit point. The structure of the set of limit points in the five-dimensional space is analyzed, and it is shown that this set can be described with three linear equations. Therefore, it might be incorrect to tabulate all five parameters of reaction rate based on just one match criterion between model and experimental data of flame propagation velocity. The conclusion of our study is that in order to tabulate reaction rate parameters correctly, it is necessary to specify the values of two of them, based on additional optimality criteria.

  8. The work is devoted to the problem of creating a model with stationary parameters using historical data under conditions of unknown disturbances. The case is considered when a representative sample of object states can be formed using historical data accumulated only over a significant period of time. It is assumed that unknown disturbances can act in a wide frequency range and may have low-frequency and trend components. In such a situation, including data from different time periods in the sample can lead to inconsistencies and greatly reduce the accuracy of the model. The paper provides an overview of approaches and methods for data harmonization. In this case, the main attention is paid to data sampling. An assessment is made of the applicability of various data sampling options as a tool for reducing the level of uncertainty. We propose a method for identifying a self-leveling object model using data accumulated over a significant period of time under conditions of unknown disturbances with a wide frequency range. The method is focused on creating a model with stationary parameters that does not require periodic reconfiguration to new conditions. The method is based on the combined use of sampling and presentation of data from individual periods of time in the form of increments relative to the initial point in time for the period. This makes it possible to reduce the number of parameters that characterize unknown disturbances with a minimum of assumptions that limit the application of the method. As a result, the dimensionality of the search problem is reduced and the computational costs associated with setting up the model are minimized. It is possible to configure both linear and, in some cases, nonlinear models. The method was used to develop a model of closed cooling of steel on a unit for continuous hot-dip galvanizing of steel strip. The model can be used for predictive control of thermal processes and for selecting strip speed. It is shown that the method makes it possible to develop a model of thermal processes from a closed cooling section under conditions of unknown disturbances, including low-frequency components.

  9. Malovichko M.S., Petrov I.B.
    On numerical solution of joint inverse geophysical problems with structural constraints
    Computer Research and Modeling, 2020, v. 12, no. 2, pp. 329-343

    Inverse geophysical problems are difficult to solve due to their mathematically incorrect formulation and large computational complexity. Geophysical exploration in frontier areas is even more complicated due to the lack of reliable geological information. In this case, inversion methods that allow interpretation of several types of geophysical data together are recognized to be of major importance. This paper is dedicated to one of such inversion methods, which is based on minimization of the determinant of the Gram matrix for a set of model vectors. Within the framework of this approach, we minimize a nonlinear functional, which consists of squared norms of data residual of different types, the sum of stabilizing functionals and a term that measures the structural similarity between different model vectors. We apply this approach to seismic and electromagnetic synthetic data set. Specifically, we study joint inversion of acoustic pressure response together with controlled-source electrical field imposing structural constraints on resulting electrical conductivity and P-wave velocity distributions.

    We start off this note with the problem formulation and present the numerical method for inverse problem. We implemented the conjugate-gradient algorithm for non-linear optimization. The efficiency of our approach is demonstrated in numerical experiments, in which the true 3D electrical conductivity model was assumed to be known, but the velocity model was constructed during inversion of seismic data. The true velocity model was based on a simplified geology structure of a marine prospect. Synthetic seismic data was used as an input for our minimization algorithm. The resulting velocity model not only fit to the data but also has structural similarity with the given conductivity model. Our tests have shown that optimally chosen weight of the Gramian term may improve resolution of the final models considerably.

  10. Danilova M.Y., Malinovskiy G.S.
    Averaged heavy-ball method
    Computer Research and Modeling, 2022, v. 14, no. 2, pp. 277-308

    First-order optimization methods are workhorses in a wide range of modern applications in economics, physics, biology, machine learning, control, and other fields. Among other first-order methods accelerated and momentum ones obtain special attention because of their practical efficiency. The heavy-ball method (HB) is one of the first momentum methods. The method was proposed in 1964 and the first analysis was conducted for quadratic strongly convex functions. Since then a number of variations of HB have been proposed and analyzed. In particular, HB is known for its simplicity in implementation and its performance on nonconvex problems. However, as other momentum methods, it has nonmonotone behavior, and for optimal parameters, the method suffers from the so-called peak effect. To address this issue, in this paper, we consider an averaged version of the heavy-ball method (AHB). We show that for quadratic problems AHB has a smaller maximal deviation from the solution than HB. Moreover, for general convex and strongly convex functions, we prove non-accelerated rates of global convergence of AHB, its weighted version WAHB, and for AHB with restarts R-AHB. To the best of our knowledge, such guarantees for HB with averaging were not explicitly proven for strongly convex problems in the existing works. Finally, we conduct several numerical experiments on minimizing quadratic and nonquadratic functions to demonstrate the advantages of using averaging for HB. Moreover, we also tested one more modification of AHB called the tail-averaged heavy-ball method (TAHB). In the experiments, we observed that HB with a properly adjusted averaging scheme converges faster than HB without averaging and has smaller oscillations.

Pages: « first previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"