All issues
- 2025 Vol. 17
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Influence of random malignant cell motility on growing tumor front stability
Computer Research and Modeling, 2009, v. 1, no. 2, pp. 225-232Views (last year): 5. Citations: 7 (RSCI).Chemotaxis plays an important role in morphogenesis and processes of structure formation in nature. Both unicellular organisms and single cells in tissue demonstrate this property. In vitro experiments show that many types of transformed cell, especially metastatic competent, are capable for directed motion in response usually to chemical signal. There is a number of theoretical papers on mathematical modeling of tumour growth and invasion using Keller-Segel model for the chemotactic motility of cancer cells. One of the crucial questions for using the chemotactic term in modelling of tumour growth is a lack of reliable quantitative estimation of its parameters. The 2-D mathematical model of tumour growth and invasion, which takes into account only random cell motility and convective fluxes in compact tissue, has showed that due to competitive mechanism tumour can grow toward sources of nutrients in absence of chemotactic cell motility.
-
Modeling of behavior of panicked crowd in multi-floor branched space
Computer Research and Modeling, 2013, v. 5, no. 3, pp. 491-508Views (last year): 7. Citations: 10 (RSCI).The collective behavior of crowd leaving a room is modeled. The model is based on molecular dynamics approach with a mixture of socio-psychological and physical forces. The new algorithm for complicatedly branched space is proposed. It suggests that each individual develops its own plan of escape, which is stochastically transformed during the evolution. The algorithm includes also the separation of original space into rooms with possible exits selected by individuals according to their probability distribution. The model is calibrated on the base of empirical data provided by fire case in the nightclub “Lame Horse” (Perm, 2009). The algorithm is realized as an end-user Java software. It is assumed that this tool could help to test the buildings for their safety for humans.
-
A framework for medical image segmentation based on measuring diversity of pixel’s intensity utilizing interval approach
Computer Research and Modeling, 2021, v. 13, no. 5, pp. 1059-1066Segmentation of medical image is one of the most challenging tasks in analysis of medical image. It classifies the organs pixels or lesions from medical images background like MRI or CT scans, that is to provide critical information about the human organ’s volumes and shapes. In scientific imaging field, medical imaging is considered one of the most important topics due to the rapid and continuing progress in computerized medical image visualization, advances in analysis approaches and computer-aided diagnosis. Digital image processing becomes more important in healthcare field due to the growing use of direct digital imaging systems for medical diagnostics. Due to medical imaging techniques, approaches of image processing are now applicable in medicine. Generally, various transformations will be needed to extract image data. Also, a digital image can be considered an approximation of a real situation includes some uncertainty derived from the constraints on the process of vision. Since information on the level of uncertainty will influence an expert’s attitude. To address this challenge, we propose novel framework involving interval concept that consider a good tool for dealing with the uncertainty, In the proposed approach, the medical images are transformed into interval valued representation approach and entropies are defined for an image object and background. Then we determine a threshold for lower-bound image and for upper-bound image, and then calculate the mean value for the final output results. To demonstrate the effectiveness of the proposed framework, we evaluate it by using synthetic image and its ground truth. Experimental results showed how performance of the segmentation-based entropy threshold can be enhanced using proposed approach to overcome ambiguity.
-
Random forest of risk factors as a predictive tool for adverse events in clinical medicine
Computer Research and Modeling, 2025, v. 17, no. 5, pp. 987-1004The aim of study was to develop an ensemble machine learning method for constructing interpretable predictive models and to validate it using the example of predicting in-hospital mortality (IHM) in patients with ST-segment elevation myocardial infarction (STEMI).
A retrospective cohort study was conducted using data from 5446 electronic medical records of STEMI patients who underwent percutaneous coronary intervention (PCI). Patients were divided into two groups: 335 (6.2%) patients who died during hospitalization and 5111 (93.8%) patients with a favourable in-hospital outcome. A pool of potential predictors was formed using statistical methods. Through multimetric categorization (minimizing p-values, maximizing the area under the ROC curve (AUC), and SHAP value analysis), decision trees, and multivariable logistic regression (MLR), predictors were transformed into risk factors for IHM. Predictive models for IHM were developed using MLR, Random Forest Risk Factors (RandFRF), Stochastic Gradient Boosting (XGboost), Random Forest (RF), Adaptive boosting, Gradient Boosting, Light Gradient-Boosting Machine, Categorical Boosting (CatBoost), Explainable Boosting Machine and Stacking methods.
Authors developed the RandFRF method, which integrates the predictive outcomes of modified decision trees, identifies risk factors and ranks them based on their contribution to the risk of adverse outcomes. RandFRF enables the development of predictive models with high discriminative performance (AUC 0.908), comparable to models based on CatBoost and Stacking (AUC 0.904 and 0.908, respectively). In turn, risk factors provide clinicians with information on the patient’s risk group classification and the extent of their impact on the probability of IHM. The risk factors identified by RandFRF can serve not only as rationale for the prediction results but also as a basis for developing more accurate models.
-
Iterative decomposition methods in modelling the development of oligopolistic markets
Computer Research and Modeling, 2025, v. 17, no. 6, pp. 1237-1256One of the principles of forming a competitive market environment is to create conditions for economic agents to implement Nash – Cournot optimal strategies. With the standard approach to determining Nash – Cournot optimal market strategies, economic agents must have complete information about the indicators and dynamic characteristics of all market participants. Which is not true.
In this regard, to find Nash – Cournot optimal solutions in dynamic models, it is necessary to have a coordinator who has complete information about the participants. However, in the case of a large number of game participants, even if the coordinator has the necessary information, computational difficulties arise associated with the need to solve a large number of coupled equations (in the case of linear dynamic games — Riccati matrix equations).
In this regard, there is a need to decompose the general problem of determining optimal strategies for market participants into private (local) problems. Approaches based on the iterative decomposition of coupled matrix Riccati equations and the solution of local Riccati equations were studied for linear dynamic games with a quadratic criterion. This article considers a simpler approach to the iterative determination of the Nash – Cournot equilibrium in an oligopoly, by decomposition using operational calculus (operator method).
The proposed approach is based on the following procedure. A virtual coordinator, which has information about the parameters of the inverse demand function, forms prices for the prospective period. Oligopolists, given fixed price dynamics, determine their strategies in accordance with a slightly modified optimality criterion. The optimal volumes of production of the oligopolists are sent to the coordinator, who, based on the iterative algorithm, adjusts the price dynamics at the previous step.
The proposed procedure is illustrated by the example of a static and dynamic model of rational behavior of oligopoly participants who maximize the net present value (NPV). Using the methods of operational calculus (and in particular, the inverse Z-transformation), conditions are found under which the iterative procedure leads to equilibrium levels of price and production volumes in the case of linear dynamic games with both quadratic and nonlinear (concave) optimization criteria.
The approach considered is used in relation to examples of duopoly, triopoly, duopoly on the market with a differentiated product, duopoly with interacting oligopolists with a linear inverse demand function. Comparison of the results of calculating the dynamics of price and production volumes of oligopolists for the considered examples based on coupled equations of the matrix Riccati equations in Matlab (in the table — Riccati), as well as in accordance with the proposed iterative method in the widely available Excel system shows their practical identity.
In addition, the application of the proposed iterative procedure is illustrated by the example of a duopoly with a nonlinear demand function.
-
Control theory methods for creating market structures
Computer Research and Modeling, 2014, v. 6, no. 5, pp. 839-859Views (last year): 4. Citations: 4 (RSCI).Control theory methods for creating market structures are discussed for two cases: when market participants are pursuing aims 1) of maximal growth and 2) of maximum economic efficiency of their firms. For the first case method based on variable structure systems principles is developed. For the second case dynamic game approach is proposed based on computation of Nash–Cournot and Stackelberg strategies with the help of Z-transform.
-
Numerical simulation of sportsman's external flow
Computer Research and Modeling, 2017, v. 9, no. 2, pp. 331-344Views (last year): 29.Numerical simulation of moving sportsman external flow is presented. The unique method is developed for obtaining integral aerodynamic characteristics, which were the function of the flow regime (i.e. angle of attack, flow speed) and body position. Individual anthropometric characteristics and moving boundaries of sportsman (or sports equipment) during the race are taken into consideration.
Numerical simulation is realized using FlowVision CFD. The software is based on the finite volume method, high-performance numerical methods and reliable mathematical models of physical processes. A Cartesian computational grid is used by FlowVision, the grid generation is a completely automated process. Local grid adaptation is used for solving high-pressure gradient and object complex shape. Flow simulation process performed by solutions systems of equations describing movement of fluid and/or gas in the computational domain, including: mass, moment and energy conservation equations; state equations; turbulence model equations. FlowVision permits flow simulation near moving bodies by means of computational domain transformation according to the athlete shape changes in the motion. Ski jumper aerodynamic characteristics are studied during all phases: take-off performance in motion, in-run and flight. Projected investigation defined simulation method, which includes: inverted statement of sportsman external flow development (velocity of the motion is equal to air flow velocity, object is immobile); changes boundary of the body technology defining; multiple calculations with the national team member data projecting. The research results are identification of the main factors affected to jumping performance: aerodynamic forces, rotating moments etc. Developed method was tested with active sportsmen. Ski jumpers used this method during preparations for Sochi Olympic Games 2014. A comparison of the predicted characteristics and experimental data shows a good agreement. Method versatility is underlined by performing swimmer and skater flow simulation. Designed technology is applicable for sorts of natural and technical objects.
-
An effective segmentation approach for liver computed tomography scans using fuzzy exponential entropy
Computer Research and Modeling, 2021, v. 13, no. 1, pp. 195-202Accurate segmentation of liver plays important in contouring during diagnosis and the planning of treatment. Imaging technology analysis and processing are wide usage in medical diagnostics, and therapeutic applications. Liver segmentation referring to the process of automatic or semi-automatic detection of liver image boundaries. A major difficulty in segmentation of liver image is the high variability as; the human anatomy itself shows major variation modes. In this paper, a proposed approach for computed tomography (CT) liver segmentation is presented by combining exponential entropy and fuzzy c-partition. Entropy concept has been utilized in various applications in imaging computing domain. Threshold techniques based on entropy have attracted a considerable attention over the last years in image analysis and processing literatures and it is among the most powerful techniques in image segmentation. In the proposed approach, the computed tomography (CT) of liver is transformed into fuzzy domain and fuzzy entropies are defined for liver image object and background. In threshold selection procedure, the proposed approach considers not only the information of liver image background and object, but also interactions between them as the selection of threshold is done by find a proper parameter combination of membership function such that the total fuzzy exponential entropy is maximized. Differential Evolution (DE) algorithm is utilizing to optimize the exponential entropy measure to obtain image thresholds. Experimental results in different CT livers scan are done and the results demonstrate the efficient of the proposed approach. Based on the visual clarity of segmented images with varied threshold values using the proposed approach, it was observed that liver segmented image visual quality is better with the results higher level of threshold.
-
Mathematical methods for stabilizing the structure of social systems under external disturbances
Computer Research and Modeling, 2021, v. 13, no. 4, pp. 845-857The article considers a bilinear model of the influence of external disturbances on the stability of the structure of social systems. Approaches to the third-party stabilization of the initial system consisting of two groups are investigated — by reducing the initial system to a linear system with uncertain parameters and using the results of the theory of linear dynamic games with a quadratic criterion. The influence of the coefficients of the proposed model of the social system and the control parameters on the quality of the system stabilization is analyzed with the help of computer experiments. It is shown that the use of a minimax strategy by a third party in the form of feedback control leads to a relatively close convergence of the population of the second group (excited by external influences) to an acceptable level, even with unfavorable periodic dynamic perturbations.
The influence of one of the key coefficients in the criterion $(\varepsilon)$ used to compensate for the effects of external disturbances (the latter are present in the linear model in the form of uncertainty) on the quality of system stabilization is investigated. Using Z-transform, it is shown that a decrease in the coefficient $\varepsilon$ should lead to an increase in the values of the sum of the squares of the control. The computer calculations carried out in the article also show that the improvement of the convergence of the system structure to the equilibrium level with a decrease in this coefficient is achieved due to sharp changes in control in the initial period, which may induce the transition of some members of the quiet group to the second, excited group.
The article also examines the influence of the values of the model coefficients that characterize the level of social tension on the quality of management. Calculations show that an increase in the level of social tension (all other things being equal) leads to the need for a significant increase in the third party's stabilizing efforts, as well as the value of control at the transition period.
The results of the statistical modeling carried out in the article show that the calculated feedback controls successfully compensate for random disturbances on the social system (both in the form of «white» noise, and of autocorrelated disturbances).
-
A survey on the application of large language models in software engineering
Computer Research and Modeling, 2024, v. 16, no. 7, pp. 1715-1726Large Language Models (LLMs) are transforming software engineering by bridging the gap between natural language and programming languages. These models have revolutionized communication within development teams and the Software Development Life Cycle (SDLC) by enabling developers to interact with code using natural language, thereby improving workflow efficiency. This survey examines the impact of LLMs across various stages of the SDLC, including requirement gathering, system design, coding, debugging, testing, and documentation. LLMs have proven to be particularly useful in automating repetitive tasks such as code generation, refactoring, and bug detection, thus reducing manual effort and accelerating the development process. The integration of LLMs into the development process offers several advantages, including the automation of error correction, enhanced collaboration, and the ability to generate high-quality, functional code based on natural language input. Additionally, LLMs assist developers in understanding and implementing complex software requirements and design patterns. This paper also discusses the evolution of LLMs from simple code completion tools to sophisticated models capable of performing high-level software engineering tasks. However, despite their benefits, there are challenges associated with LLM adoption, such as issues related to model accuracy, interpretability, and potential biases. These limitations must be addressed to ensure the reliable deployment of LLMs in production environments. The paper concludes by identifying key areas for future research, including improving the adaptability of LLMs to specific software domains, enhancing their contextual understanding, and refining their capabilities to generate semantically accurate and efficient code. This survey provides valuable insights into the evolving role of LLMs in software engineering, offering a foundation for further exploration and practical implementation.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"




