All issues
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Optimal threshold selection algorithms for multi-label classification: property study
Computer Research and Modeling, 2022, v. 14, no. 6, pp. 1221-1238Multi-label classification models arise in various areas of life, which is explained by an increasing amount of information that requires prompt analysis. One of the mathematical methods for solving this problem is a plug-in approach, at the first stage of which, for each class, a certain ranking function is built, ordering all objects in some way, and at the second stage, the optimal thresholds are selected, the objects on one side of which are assigned to the current class, and on the other — to the other. Thresholds are chosen to maximize the target quality measure. The algorithms which properties are investigated in this article are devoted to the second stage of the plug-in approach which is the choice of the optimal threshold vector. This step becomes non-trivial if the $F$-measure of average precision and recall is used as the target quality assessment since it does not allow independent threshold optimization in each class. In problems of extreme multi-label classification, the number of classes can reach hundreds of thousands, so the original optimization problem is reduced to the problem of searching a fixed point of a specially introduced transformation $\boldsymbol V$, defined on a unit square on the plane of average precision $P$ and recall $R$. Using this transformation, two algorithms are proposed for optimization: the $F$-measure linearization method and the method of $\boldsymbol V$ domain analysis. The properties of algorithms are studied when applied to multi-label classification data sets of various sizes and origin, in particular, the dependence of the error on the number of classes, on the $F$-measure parameter, and on the internal parameters of methods under study. The peculiarity of both algorithms work when used for problems with the domain of $\boldsymbol V$, containing large linear boundaries, was found. In case when the optimal point is located in the vicinity of these boundaries, the errors of both methods do not decrease with an increase in the number of classes. In this case, the linearization method quite accurately determines the argument of the optimal point, while the method of $\boldsymbol V$ domain analysis — the polar radius.
-
Subsystem “Developer” as a part of the Retail Payment System
Computer Research and Modeling, 2013, v. 5, no. 1, pp. 25-36In this paper we consider one of the core subsystems of the retail payment system named “Developer”. The Queuing System for modeling this subsystem was developed and information about it is provided. The task for the assignment problem was set up and solved (the modification of the Hungarian algorithm was used). Information about Agent Based Model for subsystem “Developer” and the results of the simulation experiments are given.
-
Discrete simulation of the road restoration process
Computer Research and Modeling, 2022, v. 14, no. 6, pp. 1255-1268This work contains a description of the results of modeling the process of maintaining the readiness of a section of the road network under strikes of with specified parameters. A one-dimensional section of road up to 40 km long with a total number of strikes up to 100 during the work of the brigade is considered. A simulation model has been developed for carrying out work to maintain it in working condition by several groups (engineering teams) that are part of the engineering and road division. A multicopter-type unmanned aerial vehicle is used to search for the points of appearance of obstacles. Life cycle schemes of the main participants of the tactical scene have been developed and an event-driven model of the tactical scene has been built. The format of the event log generated as a result of simulation modeling of the process of maintaining a road section is proposed. To visualize the process of maintaining the readiness of a road section, it is proposed to use visualization in the cyclogram format.
An XSL style has been developed for building a cyclogram based on an event log. As an algorithm for making a decision on the assignment of barriers to brigades, the simplest algorithm has been adopted, prescribing choosing the nearest barrier. A criterion describing the effectiveness of maintenance work on the site based on the assessment of the average speed of vehicles on the road section is proposed. Graphs of the dependence of the criterion value and the root-meansquare error depending on the length of the maintained section are plotted and an estimate is obtained for the maximum length of the road section maintained in a state of readiness with specified values for the selected quality indicator with specified characteristics of striking and performance of repair crews. The expediency of carrying out work to maintain readiness by several brigades that are part of the engineering and road division operating autonomously is shown.
The influence of the speed of the unmanned aerial vehicle on the ability to maintain the readiness of the road section is analyzed. The speed range for from 10 to 70 km/h is considered, which corresponds to the technical capabilities of multicoptertype reconnaissance unmanned aerial vehicles. The simulation results can be used as part of a complex simulation model of an army offensive or defensive operation and for solving the problem of optimizing the assignment of tasks to maintain the readiness of road sections to engineering and road brigades. The proposed approach may be of interest for the development of military-oriented strategy games.
Keywords: simulation, optimal maintenance of the road. -
Modifications of the Frank –Wolfe algorithm in the problem of finding the equilibrium distribution of traffic flows
Computer Research and Modeling, 2024, v. 16, no. 1, pp. 53-68The paper presents various modifications of the Frank–Wolfe algorithm in the equilibrium traffic assignment problem. The Beckman model is used as a model for experiments. In this article, first of all, attention is paid to the choice of the direction of the basic step of the Frank–Wolfe algorithm. Algorithms will be presented: Conjugate Frank–Wolfe (CFW), Bi-conjugate Frank–Wolfe (BFW), Fukushima Frank –Wolfe (FFW). Each modification corresponds to different approaches to the choice of this direction. Some of these modifications are described in previous works of the authors. In this article, following algorithms will be proposed: N-conjugate Frank–Wolfe (NFW), Weighted Fukushima Frank–Wolfe (WFFW). These algorithms are some ideological continuation of the BFW and FFW algorithms. Thus, if the first algorithm used at each iteration the last two directions of the previous iterations to select the next direction conjugate to them, then the proposed algorithm NFW is using more than $N$ previous directions. In the case of Fukushima Frank–Wolfe, the average of several previous directions is taken as the next direction. According to this algorithm, a modification WFFW is proposed, which uses a exponential smoothing from previous directions. For comparative analysis, experiments with various modifications were carried out on several data sets representing urban structures and taken from publicly available sources. The relative gap value was taken as the quality metric. The experimental results showed the advantage of algorithms using the previous directions for step selection over the classic Frank–Wolfe algorithm. In addition, an improvement in efficiency was revealed when using more than two conjugate directions. For example, on various datasets, the modification 3FW showed the best convergence. In addition, the proposed modification WFFW often overtook FFW and CFW, although performed worse than NFW.
-
Finding equilibrium in two-stage traffic assignment model
Computer Research and Modeling, 2021, v. 13, no. 2, pp. 365-379Authors describe a two-stage traffic assignment model. It contains of two blocks. The first block consists of a model for calculating a correspondence (demand) matrix, whereas the second block is a traffic assignment model. The first model calculates a matrix of correspondences using a matrix of transport costs (it characterizes the required volumes of movement from one area to another, it is time in this case). To solve this problem, authors propose to use one of the most popular methods of calculating the correspondence matrix in urban studies — the entropy model. The second model describes exactly how the needs for displacement specified by the correspondence matrix are distributed along the possible paths. Knowing the ways of the flows distribution along the paths, it is possible to calculate the cost matrix. Equilibrium in a two-stage model is a fixed point in the sequence of these two models. In practice the problem of finding a fixed point can be solved by the fixed-point iteration method. Unfortunately, at the moment the issue of convergence and estimations of the convergence rate for this method has not been studied quite thoroughly. In addition, the numerical implementation of the algorithm results in many problems. In particular, if the starting point is incorrect, situations may arise where the algorithm requires extremely large numbers to be computed and exceeds the available memory even on the most modern computers. Therefore the article proposes a method for reducing the problem of finding the equilibrium to the problem of the convex non-smooth optimization. Also a numerical method for solving the obtained optimization problem is proposed. Numerical experiments were carried out for both methods of solving the problem. The authors used data for Vladivostok (for this city information from various sources was processed and collected in a new dataset) and two smaller cities in the USA. It was not possible to achieve convergence by the method of fixed-point iteration, whereas the second model for the same dataset demonstrated convergence rate $k^{-1.67}$.
-
The two geometric parameters influence study on the hydrostatic problem solution accuracy by the SPH method
Computer Research and Modeling, 2021, v. 13, no. 5, pp. 979-992The two significant geometric parameters are proposed that affect the physical quantities interpolation in the smoothed particle hydrodynamics method (SPH). They are: the smoothing coefficient which the particle size and the smoothing radius are connecting and the volume coefficient which determine correctly the particle mass for a given particles distribution in the medium.
In paper proposes a technique for these parameters influence assessing on the SPH method interpolations accuracy when the hydrostatic problem solving. The analytical functions of the relative error for the density and pressure gradient in the medium are introduced for the accuracy estimate. The relative error functions are dependent on the smoothing factor and the volume factor. Designating a specific interpolation form in SPH method allows the differential form of the relative error functions to the algebraic polynomial form converting. The root of this polynomial gives the smoothing coefficient values that provide the minimum interpolation error for an assigned volume coefficient.
In this work, the derivation and analysis of density and pressure gradient relative errors functions on a sample of popular nuclei with different smoothing radius was carried out. There is no common the smoothing coefficient value for all the considered kernels that provides the minimum error for both SPH interpolations. The nuclei representatives with different smoothing radius are identified which make it possible the smallest errors of SPH interpolations to provide when the hydrostatic problem solving. As well, certain kernels with different smoothing radius was determined which correct interpolation do not allow provide when the hydrostatic problem solving by the SPH method.
-
Mathematical modeling of the human capital dynamic
Computer Research and Modeling, 2019, v. 11, no. 2, pp. 329-342Views (last year): 34.In the conditions of the development of modern economy, human capital is one of the main factors of economic growth. The formation of human capital begins with the birth of a person and continues throughout life, so the value of human capital is inseparable from its carriers, which in turn makes it difficult to account for this factor. This has led to the fact that currently there are no generally accepted methods of calculating the value of human capital. There are only a few approaches to the measurement of human capital: the cost approach (by income or investment) and the index approach, of which the most well-known approach developed under the auspices of the UN.
This paper presents the assigned task in conjunction with the task of demographic dynamics solved in the time-age plane, which allows to more fully take into account the temporary changes in the demographic structure on the dynamics of human capital.
The task of demographic dynamics is posed within the framework of the Mac-Kendrick – von Foerster model on the basis of the equation of age structure dynamics. The form of distribution functions for births, deaths and migration of the population is determined on the basis of the available statistical information. The numerical solution of the problem is given. The analysis and forecast of demographic indicators are presented. The economic and mathematical model of human capital dynamics is formulated on the basis of the demographic dynamics problem. The problem of modeling the human capital dynamics considers three components of capital: educational, health and cultural (spiritual). Description of the evolution of human capital components uses an equation of the transfer equation type. Investments in human capital components are determined on the basis of budget expenditures and private expenditures, taking into account the characteristic time life cycle of demographic elements. A one-dimensional kinetic equation is used to predict the dynamics of the total human capital. The method of calculating the dynamics of this factor is given as a time function. The calculated data on the human capital dynamics are presented for the Russian Federation. As studies have shown, the value of human capital increased rapidly until 2008, in the future there was a period of stabilization, but after 2014 there is a negative dynamics of this value.
-
Identification of the author of the text by segmentation method
Computer Research and Modeling, 2022, v. 14, no. 5, pp. 1199-1210The paper describes a method for recognizing authors of literary texts by the proximity of fragments into which a separate text is divided to the standard of the author. The standard is the empirical frequency distribution of letter combinations, built on a training sample, which included expertly selected reliably known works of this author. A set of standards of different authors forms a library, within which the problem of identifying the author of an unknown text is solved. The proximity between texts is understood in the sense of the norm in L1 for the frequency vector of letter combinations, which is constructed for each fragment and for the text as a whole. The author of an unknown text is assigned the one whose standard is most often chosen as the closest for the set of fragments into which the text is divided. The length of the fragment is optimized based on the principle of the maximum difference in distances from fragments to standards in the problem of recognition of «friend–foe». The method was tested on the corpus of domestic and foreign (translated) authors. 1783 texts of 100 authors with a total volume of about 700 million characters were collected. In order to exclude the bias in the selection of authors, authors whose surnames began with the same letter were considered. In particular, for the letter L, the identification error was 12%. Along with a fairly high accuracy, this method has another important property: it allows you to estimate the probability that the standard of the author of the text in question is missing in the library. This probability can be estimated based on the results of the statistics of the nearest standards for small fragments of text. The paper also examines statistical digital portraits of writers: these are joint empirical distributions of the probability that a certain proportion of the text is identified at a given level of trust. The practical importance of these statistics is that the carriers of the corresponding distributions practically do not overlap for their own and other people’s standards, which makes it possible to recognize the reference distribution of letter combinations at a high level of confidence.
-
Study and optimization of wireless sensor network based on ZigBee protocol
Computer Research and Modeling, 2012, v. 4, no. 4, pp. 855-869Views (last year): 5. Citations: 12 (RSCI).Algorithms of wireless sensor networks operation based on modified ZigBee/IEEE 802.15.4 protocol stack and problems of energy saving with simultaneous decrease of network latency are studied. Theoretical computations are given. Roles distribution and routers schedule assignment algorithms are described. Both results of experiments carried out with real devices and results of simulations with ns-2 (open-source network simulator) are given and analyzed.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"