All issues
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Parallel calculations in the Darwin PIC-model
Computer Research and Modeling, 2015, v. 7, no. 1, pp. 61-69Views (last year): 2.The approach to parallel implementation of low-frequency PIC-algorithms is proposed, taking into account peculiarity of the nonradiative (Darwin) field approximation. Its advantages and specifics of adaptation to the base computer types for high performance calculations are discussed.
-
Estimation of the probability of spontaneous synthesis of computational structures in relation to the implementation of parallel information processing
Computer Research and Modeling, 2021, v. 13, no. 4, pp. 677-696We consider a model of spontaneous formation of a computational structure in the human brain for solving a given class of tasks in the process of performing a series of similar tasks. The model is based on a special definition of a numerical measure of the complexity of the solution algorithm. This measure has an informational property: the complexity of a computational structure consisting of two independent structures is equal to the sum of the complexities of these structures. Then the probability of spontaneous occurrence of the structure depends exponentially on the complexity of the structure. The exponential coefficient requires experimental determination for each type of problem. It may depend on the form of presentation of the source data and the procedure for issuing the result. This estimation method was applied to the results of a series of experiments that determined the strategy for solving a series of similar problems with a growing number of initial data. These experiments were described in previously published papers. Two main strategies were considered: sequential execution of the computational algorithm, or the use of parallel computing in those tasks where it is effective. These strategies differ in how calculations are performed. Using an estimate of the complexity of schemes, you can use the empirical probability of one of the strategies to calculate the probability of the other. The calculations performed showed a good match between the calculated and empirical probabilities. This confirms the hypothesis about the spontaneous formation of structures that solve the problem during the initial training of a person. The paper contains a brief description of experiments, detailed computational schemes and a strict definition of the complexity measure of computational structures and the conclusion of the dependence of the probability of structure formation on its complexity.
-
Solution to a two-dimensional nonlinear heat equation using null field method
Computer Research and Modeling, 2023, v. 15, no. 6, pp. 1449-1467The paper deals with a heat wave motion problem for a degenerate second-order nonlinear parabolic equation with power nonlinearity. The considered boundary condition specifies in a plane the motion equation of the circular zero front of the heat wave. A new numerical-analytical algorithm for solving the problem is proposed. A solution is constructed stepby- step in time using difference time discretization. At each time step, a boundary value problem for the Poisson equation corresponding to the original equation at a fixed time is considered. This problem is, in fact, an inverse Cauchy problem in the domain whose initial boundary is free of boundary conditions and two boundary conditions (Neumann and Dirichlet) are specified on a current boundary (heat wave). A solution of this problem is constructed as the sum of a particular solution to the nonhomogeneous Poisson equation and a solution to the corresponding Laplace equation satisfying the boundary conditions. Since the inhomogeneity depends on the desired function and its derivatives, an iterative solution procedure is used. The particular solution is sought by the collocation method using inhomogeneity expansion in radial basis functions. The inverse Cauchy problem for the Laplace equation is solved by the null field method as applied to a circular domain with a circular hole. This method is used for the first time to solve such problem. The calculation algorithm is optimized by parallelizing the computations. The parallelization of the computations allows us to realize effectively the algorithm on high performance computing servers. The algorithm is implemented as a program, which is parallelized by using the OpenMP standard for the C++ language, suitable for calculations with parallel cycles. The effectiveness of the algorithm and the robustness of the program are tested by the comparison of the calculation results with the known exact solution as well as with the numerical solution obtained earlier by the authors with the use of the boundary element method. The implemented computational experiment shows good convergence of the iteration processes and higher calculation accuracy of the proposed new algorithm than of the previously developed one. The solution analysis allows us to select the radial basis functions which are most suitable for the proposed algorithm.
-
Mathematical simulation of vortex motion in the astrophysical objects on the basis of the gas-dynamic model
Computer Research and Modeling, 2018, v. 10, no. 5, pp. 631-643Views (last year): 27.The application of a conservative numerical method of fluxes is examined for studying the vortex structures in the massive, fast-turned compact astrophysical objects, which are in self-gravity conditions. The simulation is accomplished for the objects with different mass and rotational speed. The pictures of the vortex structure of objects are visualized. In the calculations the gas-dynamic model is used, in which gas is accepted perfected and nonviscous. Numerical procedure is based on the finite-difference approximation of the conservation laws of the additive characteristics of medium for the finite volume. The “upwind” approximations of the densities of distribution of mass, components of momentum and total energy are applied. For the simulation of the objects, which possess fast-spin motion, the control of conservation for the component of moment of momentun is carried out during calculation. Evolutionary calculation is carried out on the basis of the parallel algorithms, realized on the computer complex of cluster architecture. Algorithms are based on the standardized system of message transfer Message Passing Interface (MPI). The blocking procedures of exchange and non-blocking procedures of exchange with control of the completion of operation are used. The parallelization on the space in two or three directions is carried out depending on the size of integration area and parameters of computational grid. For each subarea the parallelization based on the physical factors is carried out also: the calculations of gas dynamics part and gravitational forces are realized on the different processors, that allows to raise the efficiency of algorithms. The real possibility of the direct calculation of gravitational forces by means of the summation of interaction between all finite volumes in the integration area is shown. For the finite volume methods this approach seems to more consecutive than the solution of Poisson’s equation for the gravitational potential. Numerical calculations were carried out on the computer complex of cluster architecture with the peak productivity 523 TFlops. In the calculations up to thousand processors was used.
-
Stiffness modeling for anthropomorphic robots
Computer Research and Modeling, 2019, v. 11, no. 4, pp. 631-651Views (last year): 3.In the work modeling method of anthropomorphic platforms is presented. An elastostatic stiffness model is used to determine positioning errors in the robot’s lower limbs. One of the main problems in achieving a fast and stable gait are deflections caused by the flexibility in the elements of the robot. This problem was solved using virtual joint modeling to predict stiffness and deformation caused by the robot weight and external forces.
To simulate a robot in the single-support phase, the robot is represented as a serial kinematic chain with a base at the supporting leg point of contact and an end effector in the swing leg foot. In the double support phase robot modeled as a parallel manipulator with an end effector in the pelvis. In this work, two cases of stiffness modeling are used: taking into account the compliance of the links and joints and taking into account only the compliance of joints. In the last case, joint compliances also include part of the link compliances. The joint stiffness parameters have been identified for two anthropomorphic robots: a small platform and a full-sized AR-601M.
Deflections maps were calculated using identified stiffness parameters and showing errors depending on the position of the robot end effector in the workspace. The errors in Z directions have maximum amplitude, due to the influence of the robot mass on its structure.
-
The computational algorithm for studying internal laminar flows of a multicomponent gas with different-scale chemical processes
Computer Research and Modeling, 2023, v. 15, no. 5, pp. 1169-1187The article presented the computational algorithm developed to study chemical processes in the internal flows of a multicomponent gas under the influence of laser radiation. The mathematical model is the gas dynamics’ equations with chemical reactions at low Mach numbers. It takes into account dissipative terms that describe the dynamics of a viscous heat-conducting medium with diffusion, chemical reactions and energy supply by laser radiation. This mathematical model is characterized by the presence of several very different time and spatial scales. The computational algorithm is based on a splitting scheme by physical processes. Each time integration step is divided into the following blocks: solving the equations of chemical kinetics, solving the equation for the radiation intensity, solving the convection-diffusion equations, calculating the dynamic component of pressure and calculating the correction of the velocity vector. The solution of a stiff system of chemical kinetics equations is carried out using a specialized explicit second-order accuracy scheme or a plug-in RADAU5 module. Numerical Rusanov flows and a WENO scheme of an increased order of approximation are used to find convective terms in the equations. The code based on the obtained algorithm has been developed using MPI parallel computing technology. The developed code is used to calculate the pyrolysis of ethane with radical reactions. The superequilibrium concentrations’ formation of radicals in the reactor volume is studied in detail. Numerical simulation of the reaction gas flow in a flat tube with laser radiation supply is carried out, which is in demand for the interpretation of experimental results. It is shown that laser radiation significantly increases the conversion of ethane and yields of target products at short lengths closer to the entrance to the reaction zone. Reducing the effective length of the reaction zone allows us to offer new solutions in the design of ethane conversion reactors into valuable hydrocarbons. The developed algorithm and program will find their application in the creation of new technologies of laser thermochemistry.
-
Building a high-performance computing system for simulation of gas dynamics
Computer Research and Modeling, 2010, v. 2, no. 3, pp. 309-317Views (last year): 5. Citations: 6 (RSCI).The aim of research is to develop software system for solving gas dynamic problem in multiply connected integration domains of regular shape by high-performance computing system. Comparison of the various technologies of parallel computing has been done. The program complex is implemented using multithreaded parallel systems to organize both multi-core and massively parallel calculation. The comparison of numerical results with known model problems solutions has been done. Research of performance of different computing platforms has been done.
-
High-speed penetration. Discrete-element simulation and experiments
Computer Research and Modeling, 2017, v. 9, no. 6, pp. 937-944Views (last year): 13. Citations: 4 (RSCI).The paper presents the results of numerical simulation and experimental data on the high-speed penetration of the impactor into the obstacle. In the calculations, a discrete-element model has been used, based on the representation of the impactor and the target by a set of close packed interconnected particles. This class of models finds an increasingly wide application in the problems of high-speed interaction of bodies. In the previous works of the authors, the questions of application of the discrete-element model to the problem of the penetration of spherical impactors into massive targets were considered. On the basis of a comparative analysis of the data of computational and physical experiments, it was found out that for a wide class of high-speed penetration problems, a high accuracy of discrete-element modeling can be achieved using the two-parameter Lennard–Jones potential. The binding energy was identified as a function of the dynamic hardness of materials. It was shown that the use of this approach makes it possible to describe accurately the penetration process in the range of impactor velocities 500–2500 m/c.
In this paper, we compare the results of discrete-element modeling with experimental data on penetration of high-strength targets of different thickness by steel impactors. The use of computational parallelization technologies on graphic processors in combination with 3D visualization and animation of the results makes it possible to obtain detailed spatio-temporal patterns of the penetration process and compare them with experimental data.
A comparative analysis of the experimental and calculated data has shown a sufficiently high accuracy of discrete-element modeling for a wide range of target thicknesses: for thin targets pierced with preservation of the integrity of the deformed impactor, for targets of medium thickness, pierced with practically complete fragmentation of the impactor at the exit from the target, and for thick impenetrable targets.
-
The Solver of Boltzmann equation on unstructured spatial grids
Computer Research and Modeling, 2019, v. 11, no. 3, pp. 427-447Views (last year): 13.The purpose of this work is to develop a universal computer program (solver) which solves kinetic Boltzmann equation for simulations of rarefied gas flows in complexly shaped devices. The structure of the solver is described in details. Its efficiency is demonstrated on an example of calculations of a modern many tubes Knudsen pump. The kinetic Boltzmann equation is solved by finite-difference method on discrete grid in spatial and velocity spaces. The differential advection operator is approximated by finite difference method. The calculation of the collision integral is based on the conservative projection method.
In the developed computational program the unstructured spatial mesh is generated using GMSH and may include prisms, tetrahedrons, hexahedrons and pyramids. The mesh is denser in areas of flow with large gradients of gas parameters. A three-dimensional velocity grid consists of cubic cells of equal volume.
A huge amount of calculations requires effective parallelization of the algorithm which is implemented in the program with the use of Message Passing Interface (MPI) technology. An information transfer from one node to another is implemented as a kind of boundary condition. As a result, every MPI node contains the information about only its part of the grid.
The main result of the work is presented in the graph of pressure difference in 2 reservoirs connected by a multitube Knudsen pump from Knudsen number. This characteristic of the Knudsen pump obtained by numerical methods shows the quality of the pump. Distributions of pressure, temperature and gas concentration in a steady state inside the pump and the reservoirs are presented as well.
The correctness of the solver is checked using two special test solutions of more simple boundary problems — test with temperature distribution between 2 planes with different temperatures and test with conservation of total gas mass.
The correctness of the obtained data for multitube Knudsen pump is checked using denser spatial and velocity grids, using more collisions in collision integral per time step.
-
A modified model of the effect of stress concentration near a broken fiber on the tensile strength of high-strength composites (MLLS-6)
Computer Research and Modeling, 2020, v. 12, no. 3, pp. 559-573The article proposes a model for assessing the potential strength of a composite material based on modern fibers with brittle fracture.
Materials consisting of parallel cylindrical fibers that are quasi-statically stretched in one direction are simulated. It is assumed that the sample is not less than 100 pieces, which corresponds to almost significant cases. It is known that the fibers have a distribution of ultimate deformation in the sample and are not destroyed at the same moment. Usually the distribution of their properties is described by the Weibull–Gnedenko statistical distribution. To simulate the strength of the composite, a model of fiber breaks accumulation is used. It is assumed that the fibers united by the polymer matrix are crushed to twice the inefficient length — the distance at which the stresses increase from the end of the broken fiber to the middle one. However, this model greatly overestimates the strength of composites with brittle fibers. For example, carbon and glass fibers are destroyed in this way.
In some cases, earlier attempts were made to take into account the stress concentration near the broken fiber (Hedgepest model, Ermolenko model, shear analysis), but such models either required a lot of initial data or did not coincide with the experiment. In addition, such models idealize the packing of fibers in the composite to the regular hexagonal packing.
The model combines the shear analysis approach to stress distribution near the destroyed fiber and the statistical approach of fiber strength based on the Weibull–Gnedenko distribution, while introducing a number of assumptions that simplify the calculation without loss of accuracy.
It is assumed that the stress concentration on the adjacent fiber increases the probability of its destruction in accordance with the Weibull distribution, and the number of such fibers with an increased probability of destruction is directly related to the number already destroyed before. All initial data can be obtained from simple experiments. It is shown that accounting for redistribution only for the nearest fibers gives an accurate forecast.
This allowed a complete calculation of the strength of the composite. The experimental data obtained by us on carbon fibers, glass fibers and model composites based on them (CFRP, GFRP), confirm some of the conclusions of the model.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"