Результаты поиска по 'modernization':
Найдено статей: 74
  1. Dzhinchvelashvili G.A., Dzerzhinsky R.I., Denisenkova N.N.
    Quantitative assessment of seismic risk and energy concepts of earthquake engineering
    Computer Research and Modeling, 2018, v. 10, no. 1, pp. 61-76

    Currently, earthquake-resistant design of buildings based on the power calculation and presentation of effect of the earthquake static equivalent forces, which are calculated using elastic response spectra (linear-spectral method) that connects the law of motion of the soil with the absolute acceleration of the model in a nonlinear oscillator.

    This approach does not directly take into account either the influence of the duration of strong motion or the plastic behavior of the structure. Frequency content and duration of ground vibrations directly affect the energy received by the building and causing damage to its elements. Unlike power or kinematic calculation of the seismic effect on the structure can be interpreted without considering separately the forces and displacements and to provide, as the product of both variables, i.e., the work or input energy (maximum energy that can be purchased building to the earthquake).

    With the energy approach of seismic design, it is necessary to evaluate the input seismic energy in the structure and its distribution among various structural components.

    The article provides substantiation of the energy approach in the design of earthquake-resistant buildings and structures instead of the currently used method based on the power calculation and presentation of effect of the earthquake static equivalent forces, which are calculated using spectra of the reaction.

    Noted that interest in the use of energy concepts in earthquake-resistant design began with the works of Housner, which provided the seismic force in the form of the input seismic energy, using the range of speeds, and suggested that the damage in elastic-plastic system and elastic system causes one and the same input seismic energy.

    The indices of the determination of the input energy of the earthquake, proposed by various authors, are given in this paper. It is shown that modern approaches to ensuring seismic stability of structures, based on the representation of the earthquake effect as a static equivalent force, do not adequately describe the behavior of the system during an earthquake.

    In this paper, based on quantitative estimates of seismic risk analyzes developed in the NRU MSUCE Standard Organization (STO) “Seismic resistance structures. The main design provisions”. In the developed document a step forward with respect to the optimal design of earthquake-resistant structures.

    The proposed concept of using the achievements of modern methods of calculation of buildings and structures on seismic effects, which are harmonized with the Eurocodes and are not contrary to the system of national regulations.

    Views (last year): 21.
  2. Krotov K.V., Skatkov A.V.
    Optimization of task package execution planning in multi-stage systems under restrictions and the formation of sets
    Computer Research and Modeling, 2021, v. 13, no. 5, pp. 917-946

    Modern methods of complex planning the execution of task packages in multistage systems are characterized by the presence of restrictions on the dimension of the problem being solved, the impossibility of guaranteed obtaining effective solutions for various values of its input parameters, as well as the impossibility of registration the conditions for the formation of sets from the result and the restriction on the interval duration of time of the system operating. The decomposition of the generalized function of the system into a set of hierarchically interconnected subfunctions is implemented to solve the problem of scheduling the execution of task packages with generating sets of results and the restriction on the interval duration of time for the functioning of the system. The use of decomposition made it possible to employ the hierarchical approach for planning the execution of task packages in multistage systems, which provides the determination of decisions by the composition of task groups at the first level of the hierarchy decisions by the composition of task packages groups executed during time intervals of limited duration at the second level and schedules for executing packages at the third level the hierarchy. In order to evaluate decisions on the composition of packages, the results of their execution, obtained during the specified time intervals, are distributed among the packages. The apparatus of the theory of hierarchical games is used to determine complex solutions. A model of a hierarchical game for making decisions by the compositions of packages, groups of packages and schedules of executing packages is built, which is a system of hierarchically interconnected criteria for optimizing decisions. The model registers the condition for the formation of sets from the results of the execution of task packages and restriction on duration of time intervals of its operating. The problem of determining the compositions of task packages and groups of task packages is NP-hard; therefore, its solution requires the use of approximate optimization methods. In order to optimize groups of task packages, the construction of a method for formulating initial solutions by their compositions has been implemented, which are further optimized. Moreover, a algorithm for distributing the results of executing task packages obtained during time intervals of limited duration by sets is formulated. The method of local solutions optimization by composition of packages groups, in accordance with which packages are excluded from groups, the results of which are not included in sets, and packages, that aren’t included in any group, is proposed. The software implementation of the considered method of complex optimization of the compositions of task packages, groups of task packages, and schedules for executing task packages from groups (including the implementation of the method for optimizing the compositions of groups of task packages) has been performed. With its use, studies of the features of the considered planning task are carried out. Conclusion are formulated concerning the dependence of the efficiency of scheduling the execution of task packages in multistage system under the introduced conditions from the input parameters of the problem. The use of the method of local optimization of the compositions of groups of task packages allows to increase the number of formed sets from the results of task execution in packages from groups by 60% in comparison with fixed groups (which do not imply optimization).

  3. Bykov N.V.
    A simulation model of connected automated vehicles platoon dynamics in a heterogeneous traffic flow
    Computer Research and Modeling, 2022, v. 14, no. 5, pp. 1041-1058

    The gradual incorporation of automated vehicles into the global transport networks leads to the need to develop tools to assess the impact of this process on various aspects of traffic. This implies a more organized movement of automated vehicles which can form uniformly moving platoons. The influence of the formation and movement of these platoons on the dynamics of traffic flow is of great interest. The currently most developed traffic flow models are based on the cellular automaton approach. They are mainly developed in the direction of increasing accuracy. This inevitably leads to the complication of models, which in their modern form have significantly moved away from the original philosophy of cellular automata, which implies simplicity and schematicity of models at the level of evolution rules, leading, however, to a complex organized behavior of the system. In the present paper, a simulation model of connected automated vehicles platoon dynamics in a heterogeneous transport system is proposed, consisting of two types of agents (vehicles): human-driven and automated. The description of the temporal evolution of the system is based on modified rules 184 and 240 for elementary cellular automata. Human-driven vehicles move according to rule 184 with the addition of accidental braking, the probability of which depends on the distance to the vehicle in front. For automated vehicles, a combination of rules is used depending on the type of nearest neighbors, regardless of the distance to them, which brings non-local interaction to the model. At the same time, it is considered that a group of sequentially moving connected automated vehicles can form an organized platoon. The influence of the ratio of types of vehicles in the system on the characteristics of the traffic flow during free movement on a circular one-lane and two-lane roads, as well as in the presence of a traffic light, is studied. The simulation results show that the effect of platoon formation is significant for a freeway traffic flow; the presence of a traffic light reduces the positive effect by about half. The movement of platoons of connected automated vehicles on two-lane roads with the possibility of lane changing was also studied. It is shown that considering the types of neighboring vehicles (automated or human-driven) when changing lanes for automated vehicles has a positive effect on the characteristics of the traffic flow.

  4. Kutalev A.A., Lapina A.A.
    Modern ways to overcome neural networks catastrophic forgetting and empirical investigations on their structural issues
    Computer Research and Modeling, 2023, v. 15, no. 1, pp. 45-56

    This paper presents the results of experimental validation of some structural issues concerning the practical use of methods to overcome catastrophic forgetting of neural networks. A comparison of current effective methods like EWC (Elastic Weight Consolidation) and WVA (Weight Velocity Attenuation) is made and their advantages and disadvantages are considered. It is shown that EWC is better for tasks where full retention of learned skills is required on all the tasks in the training queue, while WVA is more suitable for sequential tasks with very limited computational resources, or when reuse of representations and acceleration of learning from task to task is required rather than exact retention of the skills. The attenuation of the WVA method must be applied to the optimization step, i. e. to the increments of neural network weights, rather than to the loss function gradient itself, and this is true for any gradient optimization method except the simplest stochastic gradient descent (SGD). The choice of the optimal weights attenuation function between the hyperbolic function and the exponent is considered. It is shown that hyperbolic attenuation is preferable because, despite comparable quality at optimal values of the hyperparameter of the WVA method, it is more robust to hyperparameter deviations from the optimal value (this hyperparameter in the WVA method provides a balance between preservation of old skills and learning a new skill). Empirical observations are presented that support the hypothesis that the optimal value of this hyperparameter does not depend on the number of tasks in the sequential learning queue. And, consequently, this hyperparameter can be picked up on a small number of tasks and used on longer sequences.

  5. Bogdanov A.V., Degtyarev A.B., Khramushin V.N.
    High performance computations on hybrid systems: will "grand challenges" be solved?
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 429-437

    Based on CFD computations we provide the analysis of the possibilities for using modern hybrid distributed computational environments for large complex system simulation. We argue that only multilevel approach supported by new mathematical models of transport properties, dynamical representation of the problem with transport and internal processes separated, and modern paradigm of programming, taking into account specific properties of heterogeneous system, will make it possible to scale the problem effectively.

    Views (last year): 7. Citations: 8 (RSCI).
  6. Succi G., Ivanov V.V.
    Comparison of mobile operating systems based on models of growth reliability of the software
    Computer Research and Modeling, 2018, v. 10, no. 3, pp. 325-334

    Evaluation of software reliability is an important part of the process of developing modern software. Many studies are aimed at improving models for measuring and predicting the reliability of software products. However, little attention is paid to approaches to comparing existing systems in terms of software reliability. Despite the enormous importance for practice (and for managing software development), a complete and proven comparison methodology does not exist. In this article, we propose a software reliability comparison methodology in which software reliability growth models are widely used. The proposed methodology has the following features: it provides certain level of flexibility and abstraction while keeping objectivity, i.e. providing measurable comparison criteria. Also, given the comparison methodology with a set of SRGMs and evaluation criteria it becomes much easier to disseminate information about reliability of wide range of software systems. The methodology was evaluated on the example of three mobile operating systems with open source: Sailfish, Tizen, CyanogenMod.

    A byproduct of our study is a comparison of the three analyzed Open Source mobile operating systems. The goal of this research is to determine which OS is stronger in terms of reliability. To this end we have performed a GQM analysis and we have identified 3 questions and 8 metrics. Considering the comparison of metrics, it appears that Sailfish is in most case the best performing OS. However, it is also the OS that performs the worst in most cases. On the contrary, Tizen scores the best in 3 cases out of 8, but the worst only in one case out of 8.

    Views (last year): 29.
  7. Gasparyan M.M., Samonov A.S., Sazykina T.A., Ostapov E.L., Sakmarov A.V., Shahatarov O.K.
    The Solver of Boltzmann equation on unstructured spatial grids
    Computer Research and Modeling, 2019, v. 11, no. 3, pp. 427-447

    The purpose of this work is to develop a universal computer program (solver) which solves kinetic Boltzmann equation for simulations of rarefied gas flows in complexly shaped devices. The structure of the solver is described in details. Its efficiency is demonstrated on an example of calculations of a modern many tubes Knudsen pump. The kinetic Boltzmann equation is solved by finite-difference method on discrete grid in spatial and velocity spaces. The differential advection operator is approximated by finite difference method. The calculation of the collision integral is based on the conservative projection method.

    In the developed computational program the unstructured spatial mesh is generated using GMSH and may include prisms, tetrahedrons, hexahedrons and pyramids. The mesh is denser in areas of flow with large gradients of gas parameters. A three-dimensional velocity grid consists of cubic cells of equal volume.

    A huge amount of calculations requires effective parallelization of the algorithm which is implemented in the program with the use of Message Passing Interface (MPI) technology. An information transfer from one node to another is implemented as a kind of boundary condition. As a result, every MPI node contains the information about only its part of the grid.

    The main result of the work is presented in the graph of pressure difference in 2 reservoirs connected by a multitube Knudsen pump from Knudsen number. This characteristic of the Knudsen pump obtained by numerical methods shows the quality of the pump. Distributions of pressure, temperature and gas concentration in a steady state inside the pump and the reservoirs are presented as well.

    The correctness of the solver is checked using two special test solutions of more simple boundary problems — test with temperature distribution between 2 planes with different temperatures and test with conservation of total gas mass.

    The correctness of the obtained data for multitube Knudsen pump is checked using denser spatial and velocity grids, using more collisions in collision integral per time step.

    Views (last year): 13.
  8. Mikheyev P.V., Gorynin G.L., Borisova L.R.
    A modified model of the effect of stress concentration near a broken fiber on the tensile strength of high-strength composites (MLLS-6)
    Computer Research and Modeling, 2020, v. 12, no. 3, pp. 559-573

    The article proposes a model for assessing the potential strength of a composite material based on modern fibers with brittle fracture.

    Materials consisting of parallel cylindrical fibers that are quasi-statically stretched in one direction are simulated. It is assumed that the sample is not less than 100 pieces, which corresponds to almost significant cases. It is known that the fibers have a distribution of ultimate deformation in the sample and are not destroyed at the same moment. Usually the distribution of their properties is described by the Weibull–Gnedenko statistical distribution. To simulate the strength of the composite, a model of fiber breaks accumulation is used. It is assumed that the fibers united by the polymer matrix are crushed to twice the inefficient length — the distance at which the stresses increase from the end of the broken fiber to the middle one. However, this model greatly overestimates the strength of composites with brittle fibers. For example, carbon and glass fibers are destroyed in this way.

    In some cases, earlier attempts were made to take into account the stress concentration near the broken fiber (Hedgepest model, Ermolenko model, shear analysis), but such models either required a lot of initial data or did not coincide with the experiment. In addition, such models idealize the packing of fibers in the composite to the regular hexagonal packing.

    The model combines the shear analysis approach to stress distribution near the destroyed fiber and the statistical approach of fiber strength based on the Weibull–Gnedenko distribution, while introducing a number of assumptions that simplify the calculation without loss of accuracy.

    It is assumed that the stress concentration on the adjacent fiber increases the probability of its destruction in accordance with the Weibull distribution, and the number of such fibers with an increased probability of destruction is directly related to the number already destroyed before. All initial data can be obtained from simple experiments. It is shown that accounting for redistribution only for the nearest fibers gives an accurate forecast.

    This allowed a complete calculation of the strength of the composite. The experimental data obtained by us on carbon fibers, glass fibers and model composites based on them (CFRP, GFRP), confirm some of the conclusions of the model.

  9. Kotliarova E.V., Gasnikov A.V., Gasnikova E.V., Yarmoshik D.V.
    Finding equilibrium in two-stage traffic assignment model
    Computer Research and Modeling, 2021, v. 13, no. 2, pp. 365-379

    Authors describe a two-stage traffic assignment model. It contains of two blocks. The first block consists of a model for calculating a correspondence (demand) matrix, whereas the second block is a traffic assignment model. The first model calculates a matrix of correspondences using a matrix of transport costs (it characterizes the required volumes of movement from one area to another, it is time in this case). To solve this problem, authors propose to use one of the most popular methods of calculating the correspondence matrix in urban studies — the entropy model. The second model describes exactly how the needs for displacement specified by the correspondence matrix are distributed along the possible paths. Knowing the ways of the flows distribution along the paths, it is possible to calculate the cost matrix. Equilibrium in a two-stage model is a fixed point in the sequence of these two models. In practice the problem of finding a fixed point can be solved by the fixed-point iteration method. Unfortunately, at the moment the issue of convergence and estimations of the convergence rate for this method has not been studied quite thoroughly. In addition, the numerical implementation of the algorithm results in many problems. In particular, if the starting point is incorrect, situations may arise where the algorithm requires extremely large numbers to be computed and exceeds the available memory even on the most modern computers. Therefore the article proposes a method for reducing the problem of finding the equilibrium to the problem of the convex non-smooth optimization. Also a numerical method for solving the obtained optimization problem is proposed. Numerical experiments were carried out for both methods of solving the problem. The authors used data for Vladivostok (for this city information from various sources was processed and collected in a new dataset) and two smaller cities in the USA. It was not possible to achieve convergence by the method of fixed-point iteration, whereas the second model for the same dataset demonstrated convergence rate $k^{-1.67}$.

  10. Danilova M.Y., Malinovskiy G.S.
    Averaged heavy-ball method
    Computer Research and Modeling, 2022, v. 14, no. 2, pp. 277-308

    First-order optimization methods are workhorses in a wide range of modern applications in economics, physics, biology, machine learning, control, and other fields. Among other first-order methods accelerated and momentum ones obtain special attention because of their practical efficiency. The heavy-ball method (HB) is one of the first momentum methods. The method was proposed in 1964 and the first analysis was conducted for quadratic strongly convex functions. Since then a number of variations of HB have been proposed and analyzed. In particular, HB is known for its simplicity in implementation and its performance on nonconvex problems. However, as other momentum methods, it has nonmonotone behavior, and for optimal parameters, the method suffers from the so-called peak effect. To address this issue, in this paper, we consider an averaged version of the heavy-ball method (AHB). We show that for quadratic problems AHB has a smaller maximal deviation from the solution than HB. Moreover, for general convex and strongly convex functions, we prove non-accelerated rates of global convergence of AHB, its weighted version WAHB, and for AHB with restarts R-AHB. To the best of our knowledge, such guarantees for HB with averaging were not explicitly proven for strongly convex problems in the existing works. Finally, we conduct several numerical experiments on minimizing quadratic and nonquadratic functions to demonstrate the advantages of using averaging for HB. Moreover, we also tested one more modification of AHB called the tail-averaged heavy-ball method (TAHB). In the experiments, we observed that HB with a properly adjusted averaging scheme converges faster than HB without averaging and has smaller oscillations.

Pages: « first previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"