All issues
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Reduction of decision rule of multivariate interpolation and approximation method in the problem of data classification
Computer Research and Modeling, 2016, v. 8, no. 3, pp. 475-484Views (last year): 5.This article explores a method of machine learning based on the theory of random functions. One of the main problems of this method is that decision rule of a model becomes more complicated as the number of training dataset examples increases. The decision rule of the model is the most probable realization of a random function and it's represented as a polynomial with the number of terms equal to the number of training examples. In this article we will show the quick way of the number of training dataset examples reduction and, accordingly, the complexity of the decision rule. Reducing the number of examples of training dataset is due to the search and removal of weak elements that have little effect on the final form of the decision function, and noise sampling elements. For each $(x_i,y_i)$-th element sample was introduced the concept of value, which is expressed by the deviation of the estimated value of the decision function of the model at the point $x_i$, built without the $i$-th element, from the true value $y_i$. Also we show the possibility of indirect using weak elements in the process of training model without increasing the number of terms in the decision function. At the experimental part of the article, we show how changed amount of data affects to the ability of the method of generalizing in the classification task.
-
Four-factor computing experiment for the random walk on a two-dimensional square field
Computer Research and Modeling, 2017, v. 9, no. 6, pp. 905-918Views (last year): 21.Nowadays the random search became a widespread and effective tool for solving different complex optimization and adaptation problems. In this work, the problem of an average duration of a random search for one object by another is regarded, depending on various factors on a square field. The problem solution was carried out by holding total experiment with 4 factors and orthogonal plan with 54 lines. Within each line, the initial conditions and the cellular automaton transition rules were simulated and the duration of the search for one object by another was measured. As a result, the regression model of average duration of a random search for an object depending on the four factors considered, specifying the initial positions of two objects, the conditions of their movement and detection is constructed. The most significant factors among the factors considered in the work that determine the average search time are determined. An interpretation is carried out in the problem of random search for an object from the constructed model. The important result of the work is that the qualitative and quantitative influence of initial positions of objects, the size of the lattice and the transition rules on the average duration of search is revealed by means of model obtained. It is shown that the initial neighborhood of objects on the lattice does not guarantee a quick search, if each of them moves. In addition, it is quantitatively estimated how many times the average time of searching for an object can increase or decrease with increasing the speed of the searching object by 1 unit, and also with increasing the field size by 1 unit, with different initial positions of the two objects. The exponential nature of the growth in the number of steps for searching for an object with an increase in the lattice size for other fixed factors is revealed. The conditions for the greatest increase in the average search duration are found: the maximum distance of objects in combination with the immobility of one of them when the field size is changed by 1 unit. (that is, for example, with $4 \times 4$ at $5 \times 5$) can increase the average search duration in $e^{1.69} \approx 5.42$. The task presented in the work may be relevant from the point of view of application both in the landmark for ensuring the security of the state, and, for example, in the theory of mass service.
-
Development of anisotropic nonlinear noise-reduction algorithm for computed tomography data with context dynamic threshold
Computer Research and Modeling, 2019, v. 11, no. 2, pp. 233-248Views (last year): 21.The article deals with the development of the noise-reduction algorithm based on anisotropic nonlinear data filtering of computed tomography (CT). Analysis of domestic and foreign literature has shown that the most effective algorithms for noise reduction of CT data use complex methods for analyzing and processing data, such as bilateral, adaptive, three-dimensional and other types of filtrations. However, a combination of such techniques is rarely used in practice due to long processing time per slice. In this regard, it was decided to develop an efficient and fast algorithm for noise-reduction based on simplified bilateral filtration method with three-dimensional data accumulation. The algorithm was developed on C ++11 programming language in Microsoft Visual Studio 2015. The main difference of the developed noise reduction algorithm is the use an improved mathematical model of CT noise, based on the distribution of Poisson and Gauss from the logarithmic value, developed earlier by our team. This allows a more accurate determination of the noise level and, thus, the threshold of data processing. As the result of the noise reduction algorithm, processed CT data with lower noise level were obtained. Visual evaluation of the data showed the increased information content of the processed data, compared to original data, the clarity of the mapping of homogeneous regions, and a significant reduction in noise in processing areas. Assessing the numerical results of the algorithm showed a decrease in the standard deviation (SD) level by more than 6 times in the processed areas, and high rates of the determination coefficient showed that the data were not distorted and changed only due to the removal of noise. Usage of newly developed context dynamic threshold made it possible to decrease SD level on every area of data. The main difference of the developed threshold is its simplicity and speed, achieved by preliminary estimation of the data array and derivation of the threshold values that are put in correspondence with each pixel of the CT. The principle of its work is based on threshold criteria, which fits well both into the developed noise reduction algorithm based on anisotropic nonlinear filtration, and another algorithm of noise-reduction. The algorithm successfully functions as part of the MultiVox workstation and is being prepared for implementation in a single radiological network of the city of Moscow.
-
Difference scheme for solving problems of hydrodynamics for large grid Peclet numbers
Computer Research and Modeling, 2019, v. 11, no. 5, pp. 833-848The paper discusses the development and application of the accounting rectangular cell fullness method with material substance, in particular, a liquid, to increase the smoothness and accuracy of a finite-difference solution of hydrodynamic problems with a complex shape of the boundary surface. Two problems of computational hydrodynamics are considered to study the possibilities of the proposed difference schemes: the spatial-twodimensional flow of a viscous fluid between two coaxial semi-cylinders and the transfer of substances between coaxial semi-cylinders. Discretization of diffusion and convection operators was performed on the basis of the integro-interpolation method, taking into account taking into account the fullness of cells and without it. It is proposed to use a difference scheme, for solving the problem of diffusion – convection at large grid Peclet numbers, that takes into account the cell population function, and a scheme on the basis of linear combination of the Upwind and Standard Leapfrog difference schemes with weight coefficients obtained by minimizing the approximation error at small Courant numbers. As a reference, an analytical solution describing the Couette – Taylor flow is used to estimate the accuracy of the numerical solution. The relative error of calculations reaches 70% in the case of the direct use of rectangular grids (stepwise approximation of the boundaries), under the same conditions using the proposed method allows to reduce the error to 6%. It is shown that the fragmentation of a rectangular grid by 2–8 times in each of the spatial directions does not lead to the same increase in the accuracy that numerical solutions have, obtained taking into account the fullness of the cells. The proposed difference schemes on the basis of linear combination of the Upwind and Standard Leapfrog difference schemes with weighting factors of 2/3 and 1/3, respectively, obtained by minimizing the order of approximation error, for the diffusion – convection problem have a lower grid viscosity and, as a corollary, more precisely, describe the behavior of the solution in the case of large grid Peclet numbers.
-
Neural network analysis of transportation flows of urban aglomeration using the data from public video cameras
Computer Research and Modeling, 2021, v. 13, no. 2, pp. 305-318Correct modeling of complex dynamics of urban transportation flows requires the collection of large volumes of empirical data to specify types of the modes and their identification. At the same time, setting a large number of observation posts is expensive and technically not always feasible. All this results in insufficient factographic support for the traffic control systems as well as for urban planners with the obvious consequences for the quality of their decisions. As one of the means to provide large-scale data collection at least for the qualitative situation analysis, the wide-area video cameras are used in different situation centers. There they are analyzed by human operators who are responsible for observation and control. Some video cameras provided their videos for common access, which makes them a valuable resource for transportation studies. However, there are significant problems with getting qualitative data from such cameras, which relate to the theory and practice of image processing. This study is devoted to the practical application of certain mainstream neuro-networking technologies for the estimation of essential characteristics of actual transportation flows. The problems arising in processing these data are analyzed, and their solutions are suggested. The convolution neural networks are used for tracking, and the methods for obtaining basic parameters of transportation flows from these observations are studied. The simplified neural networks are used for the preparation of training sets for the deep learning neural network YOLOv4 which is later used for the estimation of speed and density of automobile flows.
-
Estimation of the probability of spontaneous synthesis of computational structures in relation to the implementation of parallel information processing
Computer Research and Modeling, 2021, v. 13, no. 4, pp. 677-696We consider a model of spontaneous formation of a computational structure in the human brain for solving a given class of tasks in the process of performing a series of similar tasks. The model is based on a special definition of a numerical measure of the complexity of the solution algorithm. This measure has an informational property: the complexity of a computational structure consisting of two independent structures is equal to the sum of the complexities of these structures. Then the probability of spontaneous occurrence of the structure depends exponentially on the complexity of the structure. The exponential coefficient requires experimental determination for each type of problem. It may depend on the form of presentation of the source data and the procedure for issuing the result. This estimation method was applied to the results of a series of experiments that determined the strategy for solving a series of similar problems with a growing number of initial data. These experiments were described in previously published papers. Two main strategies were considered: sequential execution of the computational algorithm, or the use of parallel computing in those tasks where it is effective. These strategies differ in how calculations are performed. Using an estimate of the complexity of schemes, you can use the empirical probability of one of the strategies to calculate the probability of the other. The calculations performed showed a good match between the calculated and empirical probabilities. This confirms the hypothesis about the spontaneous formation of structures that solve the problem during the initial training of a person. The paper contains a brief description of experiments, detailed computational schemes and a strict definition of the complexity measure of computational structures and the conclusion of the dependence of the probability of structure formation on its complexity.
-
Analysis of mechanical structures of complex technical systems
Computer Research and Modeling, 2021, v. 13, no. 5, pp. 903-916The work is devoted to the structural analysis of complex technical systems. Mechanical structures are considered, the properties of which affect the behavior of products during assembly, repair and operation. The main source of data on parts and mechanical connections between them is a hypergraph. This model formalizes the multidimensional basing relation. The hypergraph correctly describes the connectivity and mutual coordination of parts, which is achieved during the assembly of the product. When developing complex products in CAD systems, an engineer often makes serious design mistakes: overbasing of parts and non-sequential assembly operations. Effective ways of identifying these structural defects have been proposed. It is shown that the property of independent assembly can be represented as a closure operator whose domain is the boolean of the set of product parts. The images of this operator are connected and coordinated subsets of parts that can be assembled independently. A lattice model is described, which is the state space of the product during assembly, disassembly and decomposition into assembly units. The lattice model serves as a source of various structural information about the project. Numerical estimates of the cardinality of the set of admissible alternatives in the problems of choosing an assembly sequence and decomposition into assembly units are proposed. For many technical operations (for example, control, testing, etc.), it is necessary to mount all the operand parts in one assembly unit. A simple formalization of the technical conditions requiring the inclusion (exclusion) of parts in the assembly unit (from the assembly unit) has been developed. A theorem that gives an mathematical description of product decomposition into assembly units in exact lattice terms is given. A method for numerical evaluation of the robustness of the mechanical structure of a complex technical system is proposed.
-
Discrete simulation of the road restoration process
Computer Research and Modeling, 2022, v. 14, no. 6, pp. 1255-1268This work contains a description of the results of modeling the process of maintaining the readiness of a section of the road network under strikes of with specified parameters. A one-dimensional section of road up to 40 km long with a total number of strikes up to 100 during the work of the brigade is considered. A simulation model has been developed for carrying out work to maintain it in working condition by several groups (engineering teams) that are part of the engineering and road division. A multicopter-type unmanned aerial vehicle is used to search for the points of appearance of obstacles. Life cycle schemes of the main participants of the tactical scene have been developed and an event-driven model of the tactical scene has been built. The format of the event log generated as a result of simulation modeling of the process of maintaining a road section is proposed. To visualize the process of maintaining the readiness of a road section, it is proposed to use visualization in the cyclogram format.
An XSL style has been developed for building a cyclogram based on an event log. As an algorithm for making a decision on the assignment of barriers to brigades, the simplest algorithm has been adopted, prescribing choosing the nearest barrier. A criterion describing the effectiveness of maintenance work on the site based on the assessment of the average speed of vehicles on the road section is proposed. Graphs of the dependence of the criterion value and the root-meansquare error depending on the length of the maintained section are plotted and an estimate is obtained for the maximum length of the road section maintained in a state of readiness with specified values for the selected quality indicator with specified characteristics of striking and performance of repair crews. The expediency of carrying out work to maintain readiness by several brigades that are part of the engineering and road division operating autonomously is shown.
The influence of the speed of the unmanned aerial vehicle on the ability to maintain the readiness of the road section is analyzed. The speed range for from 10 to 70 km/h is considered, which corresponds to the technical capabilities of multicoptertype reconnaissance unmanned aerial vehicles. The simulation results can be used as part of a complex simulation model of an army offensive or defensive operation and for solving the problem of optimizing the assignment of tasks to maintain the readiness of road sections to engineering and road brigades. The proposed approach may be of interest for the development of military-oriented strategy games.
Keywords: simulation, optimal maintenance of the road. -
Multi-particle Brownian Dynamics software ProKSim for protein-protein interactions modeling
Computer Research and Modeling, 2013, v. 5, no. 1, pp. 47-64Views (last year): 4. Citations: 8 (RSCI).Protein-protein interactions are of central importance for virtually every process in living matter. Modeling the dynamics of protein association is crucial for understanding their functionality. This paper proposes novel simulation software ProKSim (Protein Kinetics Simulator) for modeling of protein interactions by means of the multi-particle Brownian Dynamics. Effect of long-range electrostatic interactions on the process of transient encounter complex formation is numerically estimated. Investigation of transient encounter complex formation was performed for three pairs of proteins: ferredoxin and ferredoxin:NADP+-redustase, plastocyanin and cytochrome f, barnase and barstar.
-
Analysis of the possibility of investigation of hydrodynamic responses and landing dynamics of space module impacting water with FlowVision CFD software
Computer Research and Modeling, 2017, v. 9, no. 1, pp. 47-55Views (last year): 10.The results of verification carried out for investigations of hydrodynamic effect on reentry conicalsegmental space vehicle are presented in the paper. The program complex Flow Vision is used for this analysis. The purpose of the study is verification of using Flow Vision program complex for problem solving mentioned above on the base of comparison between calculated and experimental data, obtained on the Apollo landing models and new development reentry spacecraft of manned transporting spaceship designed by RSC Energia. The comparison was carried out through the data of pressure values on spacecraft model surfaces during its water landing and inertia center motion parameters.
The results of study show good agreement between experimental and calculated data of force effects on vehicle construction during water landing and its motion parameters in the water medium. Computer simulation sufficiently well reproduces influence of initial velocities & water entry angles variations on water landing process.
Using of computer simulation provides simultaneous acquisition of all data information needed for investigation of water landing peculiarities during construction design, notably, hydrodynamic effects for structural strength calculations, parameters and dynamics of center mass motion and vehicle revolution around center mass for estimation water landing conditions, as well as vehicle stability after landing.
Obtained results confirm suitability of using Flow Vision program complex for water landing vehicle investigations and investigations of influence of different landing regimes through wide initial condition change range, that permits considerably decrease extent of expensive experimental tests and realize landing conditions which are sufficiently complicated for realizing in model physical experiments.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"