Результаты поиска по 'sets of results':
Найдено статей: 131
  1. Bozhko A.N., Livantsov V.E.
    Optimization of geometric analysis strategy in CAD-systems
    Computer Research and Modeling, 2024, v. 16, no. 4, pp. 825-840

    Computer-aided assembly planning for complex products is an important engineering and scientific problem. The assembly sequence and content of assembly operations largely depend on the mechanical structure and geometric properties of a product. An overview of geometric modeling methods that are used in modern computer-aided design systems is provided. Modeling geometric obstacles in assembly using collision detection, motion planning, and virtual reality is very computationally intensive. Combinatorial methods provide only weak necessary conditions for geometric reasoning. The important problem of minimizing the number of geometric tests during the synthesis of assembly operations and processes is considered. A formalization of this problem is based on a hypergraph model of the mechanical structure of the product. This model provides a correct mathematical description of coherent and sequential assembly operations. The key concept of the geometric situation is introduced. This is a configuration of product parts that requires analysis for freedom from obstacles and this analysis gives interpretable results. A mathematical description of geometric heredity during the assembly of complex products is proposed. Two axioms of heredity allow us to extend the results of testing one geometric situation to many other situations. The problem of minimizing the number of geometric tests is posed as a non-antagonistic game between decision maker and nature, in which it is required to color the vertices of an ordered set in two colors. The vertices represent geometric situations, and the color is a metaphor for the result of a collision-free test. The decision maker’s move is to select an uncolored vertex; nature’s answer is its color. The game requires you to color an ordered set in a minimum number of moves by decision maker. The project situation in which the decision maker makes a decision under risk conditions is discussed. A method for calculating the probabilities of coloring the vertices of an ordered set is proposed. The basic pure strategies of rational behavior in this game are described. An original synthetic criterion for making rational decisions under risk conditions has been developed. Two heuristics are proposed that can be used to color ordered sets of high cardinality and complex structure.

  2. Klimenko A.A., Ougolnitsky G.A.
    Subsystem “Developer” as a part of the Retail Payment System
    Computer Research and Modeling, 2013, v. 5, no. 1, pp. 25-36

    In this paper we consider one of the core subsystems of the retail payment system named “Developer”. The Queuing System for modeling this subsystem was developed and information about it is provided. The task for the assignment problem was set up and solved (the modification of the Hungarian algorithm was used). Information about Agent Based Model for subsystem “Developer” and the results of the simulation experiments are given.

  3. Kuznetsov M.B., Polezhaev A.A.
    The mechanism of formation of oscillons — localized oscillatory structures
    Computer Research and Modeling, 2015, v. 7, no. 6, pp. 1177-1184

    A formal model mechanism of oscillon formation is proposed. These structures were found in a variety of physical systems and a chemical Belousov–Jabotinsky reaction proceeding in an aerosol OT water-inoil microemulsion. Via the proposed mechanism oscillons occur as a result of interaction of two subsystems. In the first subsystem for a proper set of parameters solitary stationary structures may arise as a result of hard local excitation. These structures influence spatial distribution of the second subsystem parameter that leads to local oscillations in the subsystem.

    Views (last year): 6. Citations: 1 (RSCI).
  4. Fisher J.V., Schelyaev A.E.
    Verification of calculated characteristics of supersonic turbulent jets
    Computer Research and Modeling, 2017, v. 9, no. 1, pp. 21-35

    Verification results of supersonic turbulent jets computational characteristics are presented. Numerical simulation of axisymmetric nozzle operating is realized using FlowVision CFD. Open test cases for CFD are used. The test cases include Seiner tests with exit Mach number of 2.0 both fully-expanded and under-expanded $(P/P_0 = 1.47)$. Fully-expanded nozzle investigated with wide range of flow temperature (300…3000 K). The considered studies include simulation downstream from the nozzle exit diameter. Next numerical investigation is presented at an exit Mach number of 2.02 and a free-stream Mach number of 2.2. Geometric model of convergent- divergent nozzle rebuilt from original Putnam experiment. This study is set with nozzle pressure ratio of 8.12 and total temperature of 317 K.

    The paper provides a comparison of obtained FlowVision results with experimental data and another current CFD studies. A comparison of the calculated characteristics and experimental data indicates a good agreement. The best coincidence with Seiner's experimental velocity distribution (about 7 % at far field for the first case) obtained using two-equation $k–\varepsilon$ standard turbulence model with Wilcox compressibility correction. Predicted Mach number distribution at $Y/D = 1$ for Putnam nozzle presents accuracy of 3 %.

    General guidelines for simulation of supersonic turbulent jets in the FlowVision software are formulated in the given paper. Grid convergence determined the optimal cell rate. In order to calculate the design regime, it is recommended to build a grid, containing not less than 40 cells from the axis of symmetry to the nozzle wall. In order to calculate an off-design regime, it is necessary to resolve the shock waves. For this purpose, not less than 80 cells is required in the radial direction. Investigation of the influence of turbulence model on the flow characteristics has shown that the version of the SST $k–\omega$ turbulence model implemented in the FlowVision software essentially underpredicts the axial velocity. The standard $k–\varepsilon$ model without compressibility correction also underpredicts the axial velocity. These calculations agree well with calculations in other CFD codes using the standard $k–\varepsilon$ model. The in-home $k–\varepsilon$ turbulence model KEFV with compressibility correction a little bit overpredicts the axial velocity. Since, the best results are obtained using the standard $k–\varepsilon$ model combined with the Wilcox compressibility correction, this model is recommended for the problems discussed.

    The developed methodology can be regarded as a basis for numerical investigations of more complex nozzle flows.

    Views (last year): 43.
  5. Sviridenko A.B.
    Direct multiplicative methods for sparse matrices. Newton methods
    Computer Research and Modeling, 2017, v. 9, no. 5, pp. 679-703

    We consider a numerically stable direct multiplicative algorithm of solving linear equations systems, which takes into account the sparseness of matrices presented in a packed form. The advantage of the algorithm is the ability to minimize the filling of the main rows of multipliers without losing the accuracy of the results. Moreover, changes in the position of the next processed row of the matrix are not made, what allows using static data storage formats. Linear system solving by a direct multiplicative algorithm is, like the solving with $LU$-decomposition, just another scheme of the Gaussian elimination method implementation.

    In this paper, this algorithm is the basis for solving the following problems:

    Problem 1. Setting the descent direction in Newtonian methods of unconditional optimization by integrating one of the known techniques of constructing an essentially positive definite matrix. This approach allows us to weaken or remove additional specific difficulties caused by the need to solve large equation systems with sparse matrices presented in a packed form.

    Problem 2. Construction of a new mathematical formulation of the problem of quadratic programming and a new form of specifying necessary and sufficient optimality conditions. They are quite simple and can be used to construct mathematical programming methods, for example, to find the minimum of a quadratic function on a polyhedral set of constraints, based on solving linear equations systems, which dimension is not higher than the number of variables of the objective function.

    Problem 3. Construction of a continuous analogue of the problem of minimizing a real quadratic polynomial in Boolean variables and a new form of defining necessary and sufficient conditions of optimality for the development of methods for solving them in polynomial time. As a result, the original problem is reduced to the problem of finding the minimum distance between the origin and the angular point of a convex polyhedron, which is a perturbation of the $n$-dimensional cube and is described by a system of double linear inequalities with an upper triangular matrix of coefficients with units on the main diagonal. Only two faces are subject to investigation, one of which or both contains the vertices closest to the origin. To calculate them, it is sufficient to solve $4n – 4$ linear equations systems and choose among them all the nearest equidistant vertices in polynomial time. The problem of minimizing a quadratic polynomial is $NP$-hard, since an $NP$-hard problem about a vertex covering for an arbitrary graph comes down to it. It follows therefrom that $P = NP$, which is based on the development beyond the limits of integer optimization methods.

    Views (last year): 7. Citations: 1 (RSCI).
  6. Sviridenko A.B.
    Direct multiplicative methods for sparse matrices. Quadratic programming
    Computer Research and Modeling, 2018, v. 10, no. 4, pp. 407-420

    A numerically stable direct multiplicative method for solving systems of linear equations that takes into account the sparseness of matrices presented in a packed form is considered. The advantage of the method is the calculation of the Cholesky factors for a positive definite matrix of the system of equations and its solution within the framework of one procedure. And also in the possibility of minimizing the filling of the main rows of multipliers without losing the accuracy of the results, and no changes are made to the position of the next processed row of the matrix, which allows using static data storage formats. The solution of the system of linear equations by a direct multiplicative algorithm is, like the solution with LU-decomposition, just another scheme for implementing the Gaussian elimination method.

    The calculation of the Cholesky factors for a positive definite matrix of the system and its solution underlies the construction of a new mathematical formulation of the unconditional problem of quadratic programming and a new form of specifying necessary and sufficient conditions for optimality that are quite simple and are used in this paper to construct a new mathematical formulation for the problem of quadratic programming on a polyhedral set of constraints, which is the problem of finding the minimum distance between the origin ordinate and polyhedral boundary by means of a set of constraints and linear algebra dimensional geometry.

    To determine the distance, it is proposed to apply the known exact method based on solving systems of linear equations whose dimension is not higher than the number of variables of the objective function. The distances are determined by the construction of perpendiculars to the faces of a polyhedron of different dimensions. To reduce the number of faces examined, the proposed method involves a special order of sorting the faces. Only the faces containing the vertex closest to the point of the unconditional extremum and visible from this point are subject to investigation. In the case of the presence of several nearest equidistant vertices, we investigate a face containing all these vertices and faces of smaller dimension that have at least two common nearest vertices with the first face.

    Views (last year): 32.
  7. Basalaev A.V., Kloss Y.Y., Lubimov D.U., Knyazev A.N., Shuvalov P.V., Sherbakov D.V., Nahapetyan A.V.
    A problem-modeling environment for the numerical solution of the Boltzmann equation on a cluster architecture for analyzing gas-kinetic processes in the interelectrode gap of thermal emission converters
    Computer Research and Modeling, 2019, v. 11, no. 2, pp. 219-232

    This paper is devoted to the application of the method of numerical solution of the Boltzmann equation for the solution of the problem of modeling the behavior of radionuclides in the cavity of the interelectric gap of a multielement electrogenerating channel. The analysis of gas-kinetic processes of thermionic converters is important for proving the design of the power-generating channel. The paper reviews two constructive schemes of the channel: with one- and two-way withdrawal of gaseous fission products into a vacuum-cesium system. The analysis uses a two-dimensional transport equation of the second-order accuracy for the solution of the left-hand side and the projection method for solving the right-hand side — the collision integral. In the course of the work, a software package was implemented that makes it possible to calculate on the cluster architecture by using the algorithm of parallelizing the left-hand side of the equation; the paper contains the results of the analysis of the dependence of the calculation efficiency on the number of parallel nodes. The paper contains calculations of data on the distribution of pressures of gaseous fission products in the gap cavity, calculations use various sets of initial pressures and flows; the dependency of the radionuclide pressure in the collector region was determined as a function of cesium pressures at the ends of the gap. The tests in the loop channel of a nuclear reactor confirm the obtained results.

    Views (last year): 24.
  8. Antipova S.A., Vorobiev A.A.
    The purposeful transformation of mathematical models based on strategic reflection
    Computer Research and Modeling, 2019, v. 11, no. 5, pp. 815-831

    The study of complex processes in various spheres of human activity is traditionally based on the use of mathematical models. In modern conditions, the development and application of such models is greatly simplified by the presence of high-speed computer equipment and specialized tools that allow, in fact, designing models from pre-prepared modules. Despite this, the known problems associated with ensuring the adequacy of the model, the reliability of the original data, the implementation in practice of the simulation results, the excessively large dimension of the original data, the joint application of sufficiency heterogeneous mathematical models in terms of complexity and integration of the simulated processes are becoming increasingly important. The more critical may be the external constraints imposed on the value of the optimized functional, and often unattainable within the framework of the constructed model. It is logical to assume that in order to fulfill these restrictions, a purposeful transformation of the original model is necessary, that is, the transition to a mathematical model with a deliberately improved solution. The new model will obviously have a different internal structure (a set of parameters and their interrelations), as well as other formats (areas of definition) of the source data. The possibilities of purposeful change of the initial model investigated by the authors are based on the realization of the idea of strategic reflection. The most difficult in mathematical terms practical implementation of the author's idea is the use of simulation models, for which the algorithms for finding optimal solutions have known limitations, and the study of sensitivity in most cases is very difficult. On the example of consideration of rather standard discrete- event simulation model the article presents typical methodological techniques that allow ranking variable parameters by sensitivity and, in the future, to expand the scope of definition of variable parameter to which the simulation model is most sensitive. In the transition to the “improved” model, it is also possible to simultaneously exclude parameters from it, the influence of which on the optimized functional is insignificant, and vice versa — the introduction of new parameters corresponding to real processes into the model.

  9. Bulinskaya E.V.
    Isotropic Multidimensional Catalytic Branching Random Walk with Regularly Varying Tails
    Computer Research and Modeling, 2019, v. 11, no. 6, pp. 1033-1039

    The study completes a series of the author’s works devoted to the spread of particles population in supercritical catalytic branching random walk (CBRW) on a multidimensional lattice. The CBRW model describes the evolution of a system of particles combining their random movement with branching (reproduction and death) which only occurs at fixed points of the lattice. The set of such catalytic points is assumed to be finite and arbitrary. In the supercritical regime the size of population, initiated by a parent particle, increases exponentially with positive probability. The rate of the spread depends essentially on the distribution tails of the random walk jump. If the jump distribution has “light tails”, the “population front”, formed by the particles most distant from the origin, moves linearly in time and the limiting shape of the front is a convex surface. When the random walk jump has independent coordinates with a semiexponential distribution, the population spreads with a power rate in time and the limiting shape of the front is a star-shape nonconvex surface. So far, for regularly varying tails (“heavy” tails), we have considered the problem of scaled front propagation assuming independence of components of the random walk jump. Now, without this hypothesis, we examine an “isotropic” case, when the rate of decay of the jumps distribution in different directions is given by the same regularly varying function. We specify the probability that, for time going to infinity, the limiting random set formed by appropriately scaled positions of population particles belongs to a set $B$ containing the origin with its neighborhood, in $\mathbb{R}^d$. In contrast to the previous results, the random cloud of particles with normalized positions in the time limit will not concentrate on coordinate axes with probability one.

  10. Krechet V.G., Oshurko V.B., Kisser A.E.
    Cosmological models of the Universe without a Beginning and without a singularity
    Computer Research and Modeling, 2021, v. 13, no. 3, pp. 473-486

    A new type of cosmological models for the Universe that has no Beginning and evolves from the infinitely distant past is considered.

    These models are alternative to the cosmological models based on the Big Bang theory according to which the Universe has a finite age and was formed from an initial singularity.

    In our opinion, there are certain problems in the Big Bang theory that our cosmological models do not have.

    In our cosmological models, the Universe evolves by compression from the infinitely distant past tending a finite minimum of distances between objects of the order of the Compton wavelength $\lambda_C$ of hadrons and the maximum density of matter corresponding to the hadron era of the Universe. Then it expands progressing through all the stages of evolution established by astronomical observations up to the era of inflation.

    The material basis that sets the fundamental nature of the evolution of the Universe in the our cosmological models is a nonlinear Dirac spinor field $\psi(x^k)$ with nonlinearity in the Lagrangian of the field of type $\beta(\bar{\psi}\psi)^n$ ($\beta = const$, $n$ is a rational number), where $\psi(x^k)$ is the 4-component Dirac spinor, and $\psi$ is the conjugate spinor.

    In addition to the spinor field $\psi$ in cosmological models, we have other components of matter in the form of an ideal liquid with the equation of state $p = w\varepsilon$ $(w = const)$ at different values of the coefficient $w (−1 < w < 1)$. Additional components affect the evolution of the Universe and all stages of evolution occur in accordance with established observation data. Here $p$ is the pressure, $\varepsilon = \rho c^2$ is the energy density, $\rho$ is the mass density, and $c$ is the speed of light in a vacuum.

    We have shown that cosmological models with a nonlinear spinor field with a nonlinearity coefficient $n = 2$ are the closest to reality.

    In this case, the nonlinear spinor field is described by the Dirac equation with cubic nonlinearity.

    But this is the Ivanenko–Heisenberg nonlinear spinor equation which W.Heisenberg used to construct a unified spinor theory of matter.

    It is an amazing coincidence that the same nonlinear spinor equation can be the basis for constructing a theory of two different fundamental objects of nature — the evolving Universe and physical matter.

    The developments of the cosmological models are supplemented by their computer researches the results of which are presented graphically in the work.

Pages: previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"