All issues
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
The use of cluster analysis methods for the study of a set of feasible solutions of the phase problem in biological crystallography
Computer Research and Modeling, 2010, v. 2, no. 1, pp. 91-101Views (last year): 2.X-ray diffraction experiment allows determining of magnitudes of complex coefficients in the decomposition of the studied electron density distribution into Fourier series. The determination of the lost in the experiment phase values poses the central problem of the method, namely the phase problem. Some methods for solving of the phase problem result in a set of feasible solutions. Cluster analysis method may be used to investigate the composition of this set and to extract one or several typical solutions. An essential feature of the approach is the estimation of the closeness of two solutions by the map correlation between two aligned Fourier syntheses calculated with the use of phase sets under comparison. An interactive computer program ClanGR was designed to perform this analysis.
-
Estimation of maximal values of biomass growth yield based on the mass-energy balance of cell metabolism
Computer Research and Modeling, 2019, v. 11, no. 4, pp. 723-750Views (last year): 2.The biomass growth yield is the ratio of the newly synthesized substance of growing cells to the amount of the consumed substrate, the source of matter and energy for cell growth. The yield is a characteristic of the efficiency of substrate conversion to cell biomass. The conversion is carried out by the cell metabolism, which is a complete aggregate of biochemical reactions occurring in the cells.
This work newly considers the problem of maximal cell growth yield prediction basing on balances of the whole living cell metabolism and its fragments called as partial metabolisms (PM). The following PM’s are used for the present consideration. During growth on any substrate we consider i) the standard constructive metabolism (SCM) which consists of identical pathways during growth of various organisms on any substrate. SCM starts from several standard compounds (nodal metabolites): glucose, acetyl-CoA 2-oxoglutarate, erythrose-4-phosphate, oxaloacetate, ribose-5- phosphate, 3-phosphoglycerate, phosphoenolpyruvate, and pyruvate, and ii) the full forward metabolism (FM) — the remaining part of the whole metabolism. The first one consumes high-energy bonds (HEB) formed by the second one. In this work we examine a generalized variant of the FM, when the possible presence of extracellular products, as well as the possibilities of both aerobic and anaerobic growth are taken into account. Instead of separate balances of each nodal metabolite formation as it was made in our previous work, this work deals at once with the whole aggregate of these metabolites. This makes the problem solution more compact and requiring a smaller number of biochemical quantities and substantially less computational time. An equation expressing the maximal biomass yield via specific amounts of HEB formed and consumed by the partial metabolisms has been derived. It includes the specific HEB consumption by SCM which is a universal biochemical parameter applicable to the wide range of organisms and growth substrates. To correctly determine this parameter, the full constructive metabolism and its forward part are considered for the growth of cells on glucose as the mostly studied substrate. We used here the found earlier properties of the elemental composition of lipid and lipid-free fractions of cell biomass. Numerical study of the effect of various interrelations between flows via different nodal metabolites has been made. It showed that the requirements of the SCM in high-energy bonds and NAD(P)H are practically constants. The found HEB-to-formed-biomass coefficient is an efficient tool for finding estimates of maximal biomass yield from substrates for which the primary metabolism is known. Calculation of ATP-to-substrate ratio necessary for the yield estimation has been made using the special computer program package, GenMetPath.
-
Views (last year): 8. Citations: 2 (RSCI).
A mathematical model that reflects the main features of the protests is constructed in this paper. An analytical solution was found with assuming that only excited part of the population involved in protests. The numerical value of the model coefficients was estimated from the real data for the cascade of protests that took place in Leipzig in 1989. The changes of the participants number in the protest action with influence the model coefficients was analysed.
-
Estimation of models parameters for time series with Markov switching regimes
Computer Research and Modeling, 2018, v. 10, no. 6, pp. 903-918Views (last year): 36.The paper considers the problem of estimating the parameters of time series described by regression models with Markov switching of two regimes at random instants of time with independent Gaussian noise. For the solution, we propose a variant of the EM algorithm based on the iterative procedure, during which an estimation of the regression parameters is performed for a given sequence of regime switching and an evaluation of the switching sequence for the given parameters of the regression models. In contrast to the well-known methods of estimating regression parameters in the models with Markov switching, which are based on the calculation of a posteriori probabilities of discrete states of the switching sequence, in the paper the estimates are calculated of the switching sequence, which are optimal by the criterion of the maximum of a posteriori probability. As a result, the proposed algorithm turns out to be simpler and requires less calculations. Computer modeling allows to reveal the factors influencing accuracy of estimation. Such factors include the number of observations, the number of unknown regression parameters, the degree of their difference in different modes of operation, and the signal-to-noise ratio which is associated with the coefficient of determination in regression models. The proposed algorithm is applied to the problem of estimating parameters in regression models for the rate of daily return of the RTS index, depending on the returns of the S&P 500 index and Gazprom shares for the period from 2013 to 2018. Comparison of the estimates of the parameters found using the proposed algorithm is carried out with the estimates that are formed using the EViews econometric package and with estimates of the ordinary least squares method without taking into account regimes switching. The account of regimes switching allows to receive more exact representation about structure of a statistical dependence of investigated variables. In switching models, the increase in the signal-to-noise ratio leads to the fact that the differences in the estimates produced by the proposed algorithm and using the EViews program are reduced.
-
Dynamic regimes of the stochastic “prey – predatory” model with competition and saturation
Computer Research and Modeling, 2019, v. 11, no. 3, pp. 515-531Views (last year): 28.We consider “predator – prey” model taking into account the competition of prey, predator for different from the prey resources, and their interaction described by the second type Holling trophic function. An analysis of the attractors is carried out depending on the coefficient of competition of predators. In the deterministic case, this model demonstrates the complex behavior associated with the local (Andronov –Hopf and saddlenode) and global (birth of a cycle from a separatrix loop) bifurcations. An important feature of this model is the disappearance of a stable cycle due to a saddle-node bifurcation. As a result of the presence of competition in both populations, parametric zones of mono- and bistability are observed. In parametric zones of bistability the system has either coexisting two equilibria or a cycle and equilibrium. Here, we investigate the geometrical arrangement of attractors and separatrices, which is the boundary of basins of attraction. Such a study is an important component in understanding of stochastic phenomena. In this model, the combination of the nonlinearity and random perturbations leads to the appearance of new phenomena with no analogues in the deterministic case, such as noise-induced transitions through the separatrix, stochastic excitability, and generation of mixed-mode oscillations. For the parametric study of these phenomena, we use the stochastic sensitivity function technique and the confidence domain method. In the bistability zones, we study the deformations of the equilibrium or oscillation regimes under stochastic perturbation. The geometric criterion for the occurrence of such qualitative changes is the intersection of confidence domains and the separatrix of the deterministic model. In the zone of monostability, we evolve the phenomena of explosive change in the size of population as well as extinction of one or both populations with minor changes in external conditions. With the help of the confidence domains method, we solve the problem of estimating the proximity of a stochastic population to dangerous boundaries, upon reaching which the coexistence of populations is destroyed and their extinction is observed.
-
Valuation of machines at the random process of their degradation and premature sales
Computer Research and Modeling, 2024, v. 16, no. 3, pp. 797-815The model of the process of using machinery and equipment is considered, which takes into account the probabilistic nature of the process of their operation and sale. It takes into account the possibility of random hidden failures, after which the condition of the machine deteriorates abruptly, as well as the randomly arising need for premature (before the end of its service life) sale of the machine, which requires, generally speaking, random time. The model is focused on assessing the market value and service life of machines in accordance with International Valuation Standards. Strictly speaking, the market value of a used machine depends on its technical condition, but in practice, appraisers only take into account its age, since generally accepted measures of the technical condition of machines do not yet exist. As a result, the market value of a used machine is assumed to be equal to the average market value of similar machines of the corresponding age. For these purposes, appraisers use coefficients that reflect the influence of the age of machines on their market value. Such coefficients are not always justified and do not take into account either the degradation of the machine or the probabilistic nature of the process of its use. The proposed model is based on the anticipation of benefits principle. In it, we characterize the state of the machine by the intensity of the benefits it brings. The machine is subjected to a complex Poisson failure process, and after failure its condition abruptly worsens and may even reach its limit. Situations also arise that preclude further use of the machine by its owner. In such situations, the owner puts the machine up for sale before the end of its service life (prematurely), and the sale requires a random timing. The model allows us to take into account the influence of such situations and construct an analytical relationship linking the market value of a machine with its condition, and calculate the average coefficients of change in the market value of machines with age. At the same time, it is also possible to take into account the influence of inflation and the scrap cost of the machine. We have found that the rate of prematurely sales has a significant impact on the cost of new and used machines. The model also allows us to take into account the influence of inflation and the scrap value of the machine. We have found that the rate of premature sales has a significant impact on the service life and market value of new and used machines. At the same time, the dependence of the market value of machines on age is largely determined by the coefficient of variation of the service life of the machines. The results obtained allow us to obtain more reasonable estimates of the market value of machines, including for the purposes of the system of national accounts.
-
Fuzzy modeling of human susceptibility to panic situations
Computer Research and Modeling, 2021, v. 13, no. 1, pp. 203-218The study of the mechanism for the development of mass panic in view of its extreme importance and social danger is an important scientific task. Available information about the mechanism of her development is based mainly on the work of psychologists and belongs to the category of inaccurate. Therefore, the theory of fuzzy sets has been chosen as a tool for developing a mathematical model of a person's susceptibility to panic situations. As a result of the study, an fuzzy model was developed, consisting of blocks: “Fuzzyfication”, where the degree of belonging of the values of the input parameters to fuzzy sets is calculated; “Inference” where, based on the degree of belonging of the input parameters, the resulting function of belonging of the output value to an odd model is calculated; “Defuzzyfication”, where using the center of gravity method, the only quantitative value of the output variable characterizing a person's susceptibility to panic situations is determined Since the real quantitative values for linguistic variables mental properties of a person are unknown, then to assess the quality of the developed model, without endangering people, it is not possible. Therefore, the quality of the results of fuzzy modeling was estimated by the calculated value of the determination coefficient R2, which showed that the developed fuzzy model belongs to the category of good quality models $(R^2 = 0.93)$, which confirms the legitimacy of the assumptions made during her development. In accordance with to the results of the simulation, human susceptibility to panic situations for sanguinics and cholerics can be attributed to “increased” (0.88), and for phlegmatics and melancholics — to “moderate” (0.38). This means that cholerics and sanguinics can become epicenters of panic and the initiators of stampede, and phlegmatics and melancholics — obstacles to evacuation routes. What should be taken into account when developing effective evacuation measures, the main task of which is to quickly and safely evacuate people from adverse conditions. In the approved methods, the calculation of normative values of safety parameters is based on simplified analytical models of human flow movement, because a large number of factors have to be taken into account, some of which are quantitatively uncertain. The obtained result in the form of quantitative estimates of a person's susceptibility to panic situations will increase the accuracy of calculations.
-
Efficient and error-free information hiding in the hybrid domain of digital images using metaheuristic optimization
Computer Research and Modeling, 2023, v. 15, no. 1, pp. 197-210Data hiding in digital images is a promising direction of cybersecurity. Digital steganography methods provide imperceptible transmission of secret data over an open communication channel. The information embedding efficiency depends on the embedding imperceptibility, capacity, and robustness. These quality criteria are mutually inverse, and the improvement of one indicator usually leads to the deterioration of the others. A balance between them can be achieved using metaheuristic optimization. Metaheuristics are a class of optimization algorithms that find an optimal, or close to an optimal solution for a variety of problems, including those that are difficult to formalize, by simulating various natural processes, for example, the evolution of species or the behavior of animals. In this study, we propose an approach to data hiding in the hybrid spatial-frequency domain of digital images based on metaheuristic optimization. Changing a block of image pixels according to some change matrix is considered as an embedding operation. We select the change matrix adaptively for each block using metaheuristic optimization algorithms. In this study, we compare the performance of three metaheuristics such as genetic algorithm, particle swarm optimization, and differential evolution to find the best change matrix. Experimental results showed that the proposed approach provides high imperceptibility of embedding, high capacity, and error-free extraction of embedded information. At the same time, storage of change matrices for each block is not required for further data extraction. This improves user experience and reduces the chance of an attacker discovering the steganographic attachment. Metaheuristics provided an increase in imperceptibility indicator, estimated by the PSNR metric, and the capacity of the previous algorithm for embedding information into the coefficients of the discrete cosine transform using the QIM method [Evsutin, Melman, Meshcheryakov, 2021] by 26.02% and 30.18%, respectively, for the genetic algorithm, 26.01% and 19.39% for particle swarm optimization, 27.30% and 28.73% for differential evolution.
-
Performance prediction for chosen types of loops over one-dimensional arrays with embedding-driven intermediate representations analysis
Computer Research and Modeling, 2023, v. 15, no. 1, pp. 211-224The method for mapping of intermediate representations (IR) set of C, C++ programs to vector embedding space is considered to create an empirical estimation framework for static performance prediction using LLVM compiler infrastructure. The usage of embeddings makes programs easier to compare due to avoiding Control Flow Graphs (CFG) and Data Flow Graphs (DFG) direct comparison. This method is based on transformation series of the initial IR such as: instrumentation — injection of artificial instructions in an instrumentation compiler’s pass depending on load offset delta in the current instruction compared to the previous one, mapping of instrumented IR into multidimensional vector with IR2Vec and dimension reduction with t-SNE (t-distributed stochastic neighbor embedding) method. The D1 cache miss ratio measured with perf stat tool is considered as performance metric. A heuristic criterion of programs having more or less cache miss ratio is given. This criterion is based on embeddings of programs in 2D-space. The instrumentation compiler’s pass developed in this work is described: how it generates and injects artificial instructions into IR within the used memory model. The software pipeline that implements the performance estimation based on LLVM compiler infrastructure is given. Computational experiments are performed on synthetic tests which are the sets of programs with the same CFGs but with different sequences of offsets used when accessing the one-dimensional array of a given size. The correlation coefficient between performance metric and distance to the worst program’s embedding is measured and proved to be negative regardless of t-SNE initialization. This fact proves the heuristic criterion to be true. The process of such synthetic tests generation is also considered. Moreover, the variety of performance metric in programs set in such a test is proposed as a metric to be improved with exploration of more tests generators.
-
Modelling of carbon dioxide net ecosystem exchange of hayfield on drained peat soil: land use scenario analysis
Computer Research and Modeling, 2020, v. 12, no. 6, pp. 1427-1449The data of episodic field measurements of carbon dioxide balance components (soil respiration — Rsoil, ecosystem respiration — Reco, net ecosystem exchange — NEE) of hayfields under use and abandoned one are interpreted by modelling. The field measurements were carried within five field campaigns in 2018 and 2019 on the drained part of the Dubna Peatland in Taldom District, Moscow Oblast, Russia. The territory is within humid continental climate zone. Peatland drainage was done out for milled peat extraction. After extraction was stopped, the residual peat deposit (1–1.5 m) was ploughed and grassed (Poa pratensis L.) for hay production. The current ground water level (GWL) varies from 0.3–0.5 m below the surface during wet and up to 1.0 m during dry periods. Daily dynamics of CO2 fluxes was measured using dynamic chamber method in 2018 (August) and 2019 (May, June, August) for abandoned ditch spacing only with sanitary mowing once in 5 years and the ditch spacing with annual mowing. NEE and Reco were measured on the sites with original vegetation, and Rsoil — after vegetation removal. To model a seasonal dynamics of NEE, the dependence of its components (Reco, Rsoil, and Gross ecosystematmosphere exchange of carbon dioxide — GEE) from soil and air temperature, GWL, photosynthetically active radiation, underground and aboveground plant biomass were used. The parametrization of the models has been carried out considering the stability of coefficients estimated by the bootstrap method. R2 (α = 0.05) between simulated and measured Reco was 0.44 (p < 0.0003) on abandoned and 0.59 (p < 0.04) on under use hayfield, and GEE was 0.57 (p < 0.0002) and 0.77 (p < 0.00001), respectively. Numerical experiments were carried out to assess the influence of different haymaking regime on NEE. It was found that NEE for the season (May 15 – September 30) did not differ much between the hayfield without mowing (4.5±1.0 tC·ha–1·season–1) and the abandoned one (6.2±1.4). Single mowing during the season leads to increase of NEE up to 6.5±0.9, and double mowing — up to 7.5±1.4 tC·ha–1·season–1. This means increase of carbon losses and CO2 emission into the atmosphere. Carbon loss on hayfield for both single and double mowing scenario was comparable with abandoned hayfield. The value of removed phytomass for single and double mowing was 0.8±0.1 tC·ha–1·season–1 and 1.4±0.1 (45% carbon content in dry phytomass) or 3.0 and 4.4 t·ha–1·season–1 of hay (17% moisture content). In comparison with the fallow, the removal of biomass of 0.8±0.1 at single and 1.4±0.1 tC·ha–1·season–1 double mowing is accompanied by an increase in carbon loss due to CO2 emissions, i.e., the growth of NEE by 0.3±0.1 and 1.3±0.6 tC·ha–1·season–1, respectively. This corresponds to the growth of NEE for each ton of withdrawn phytomass per hectare of 0.4±0.2 tС·ha–1·season–1 at single mowing, and 0.9±0.7 tС·ha–1·season–1 at double mowing. Therefore, single mowing is more justified in terms of carbon loss than double mowing. Extensive mowing does not increase CO2 emissions into the atmosphere and allows, in addition, to “replace” part of the carbon loss by agricultural production.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"