All issues
- 2025 Vol. 17
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Multifractal and entropy statistics of seismic noise in Kamchatka in connection with the strongest earthquakes
Computer Research and Modeling, 2023, v. 15, no. 6, pp. 1507-1521The study of the properties of seismic noise in Kamchatka is based on the idea that noise is an important source of information about the processes preceding strong earthquakes. The hypothesis is considered that an increase in seismic hazard is accompanied by a simplification of the statistical structure of seismic noise and an increase in spatial correlations of its properties. The entropy of the distribution of squared wavelet coefficients, the width of the carrier of the multifractal singularity spectrum, and the Donoho – Johnstone index were used as statistics characterizing noise. The values of these parameters reflect the complexity: if a random signal is close in its properties to white noise, then the entropy is maximum, and the other two parameters are minimum. The statistics used are calculated for 6 station clusters. For each station cluster, daily median noise properties are calculated in successive 1-day time windows, resulting in an 18-dimensional (3 properties and 6 station clusters) time series of properties. To highlight the general properties of changes in noise parameters, a principal component method is used, which is applied for each cluster of stations, as a result of which the information is compressed into a 6-dimensional daily time series of principal components. Spatial noise coherences are estimated as a set of maximum pairwise quadratic coherence spectra between the principal components of station clusters in a sliding time window of 365 days. By calculating histograms of the distribution of cluster numbers in which the minimum and maximum values of noise statistics are achieved in a sliding time window of 365 days in length, the migration of seismic hazard areas was assessed in comparison with strong earthquakes with a magnitude of at least 7.
-
Detecting large fractures in geological media using convolutional neural networks
Computer Research and Modeling, 2025, v. 17, no. 5, pp. 889-901This paper considers the inverse problem of seismic exploration — determining the structure of the media based on the recorded wave response from it. Large cracks are considered as target objects, whose size and position are to be determined.
he direct problem is solved using the grid-characteristic method. The method allows using physically based algorithms for calculating outer boundaries of the region and contact boundaries inside the region. The crack is assumed to be thin, a special condition on the crack borders is used to describe the crack.
The inverse problem is solved using convolutional neural networks. The input data of the neural network are seismograms interpreted as images. The output data are masks describing the medium on a structured grid. Each element of such a grid belongs to one of two classes — either an element of a continuous geological massif, or an element through which a crack passes. This approach allows us to consider a medium with an unknown number of cracks.
The neural network is trained using only samples with one crack. The final testing of the trained network is performed using additional samples with several cracks. These samples are not involved in the training process. The purpose of testing under such conditions is to verify that the trained network has sufficient generality, recognizes signs of a crack in the signal, and does not suffer from overtraining on samples with a single crack in the media.
The paper shows that a convolutional network trained on samples with a single crack can be used to process data with multiple cracks. The networks detects fairly small cracks at great depths if they are sufficiently spatially separated from each other. In this case their wave responses are clearly distinguishable on the seismogram and can be interpreted by the neural network. If the cracks are close to each other, artifacts and interpretation errors may occur. This is due to the fact that on the seismogram the wave responses of close cracks merge. This cause the network to interpret several cracks located nearby as one. It should be noted that a similar error would most likely be made by a human during manual interpretation of the data. The paper provides examples of some such artifacts, distortions and recognition errors.
-
On one particular model of a mixture of the probability distributions in the radio measurements
Computer Research and Modeling, 2012, v. 4, no. 3, pp. 563-568Views (last year): 3. Citations: 7 (RSCI).This paper presents a model mixture of probability distributions of signal and noise. Typically, when analyzing the data under conditions of uncertainty it is necessary to use nonparametric tests. However, such an analysis of nonstationary data in the presence of uncertainty on the mean of the distribution and its parameters may be ineffective. The model involves the implementation of a case of a priori non-parametric uncertainty in the processing of the signal at a time when the separation of signal and noise are related to different general population, is feasible.
-
Model for building of the radio environment map for cognitive communication system based on LTE
Computer Research and Modeling, 2022, v. 14, no. 1, pp. 127-146The paper is devoted to the secondary use of spectrum in telecommunication networks. It is emphasized that one of the solutions to this problem is the use of cognitive radio technologies and dynamic spectrum access for the successful functioning of which a large amount of information is required, including the parameters of base stations and network subscribers. Storage and processing of information should be carried out using a radio environment map, which is a spatio-temporal database of all activity in the network and allows you to determine the frequencies available for use at a given time. The paper presents a two-level model for forming a map of the radio environment of a cellular communication system LTE, in which the local and global levels are highlighted, which is described by the following parameters: a set of frequencies, signal attenuation, signal propagation map, grid step, current time count. The key objects of the model are the base station and the subscriber unit. The main parameters of the base station include: name, identifier, cell coordinates, range number, radiation power, numbers of connected subscriber devices, dedicated resource blocks. For subscriber devices, the following parameters are used: name, identifier, location, current coordinates of the device cell, base station identifier, frequency range, numbers of resource blocks for communication with the station, radiation power, data transmission status, list of numbers of the nearest stations, schedules movement and communication sessions of devices. An algorithm for the implementation of the model is presented, taking into account the scenarios of movement and communication sessions of subscriber devices. A method for calculating a map of the radio environment at a point on a coordinate grid, taking into account losses during the propagation of radio signals from emitting devices, is presented. The software implementation of the model is performed using the MatLab package. The approaches are described that allow to increase the speed of its work. In the simulation, the choice of parameters was carried out taking into account the data of the existing communication systems and the economy of computing resources. The experimental results of the algorithm for the formation of a radio environment map are demonstrated, confirming the correctness of the developed model.
-
Signal and noise calculation at Rician data analysis by means of combining maximum likelihood technique and method of moments
Computer Research and Modeling, 2018, v. 10, no. 4, pp. 511-523Views (last year): 11.The paper develops a new mathematical method of the joint signal and noise calculation at the Rice statistical distribution based on combing the maximum likelihood method and the method of moments. The calculation of the sough-for values of signal and noise is implemented by processing the sampled measurements of the analyzed Rician signal’s amplitude. The explicit equations’ system has been obtained for required signal and noise parameters and the results of its numerical solution are provided confirming the efficiency of the proposed technique. It has been shown that solving the two-parameter task by means of the proposed technique does not lead to the increase of the volume of demanded calculative resources if compared with solving the task in one-parameter approximation. An analytical solution of the task has been obtained for the particular case of small value of the signal-to-noise ratio. The paper presents the investigation of the dependence of the sought for parameters estimation accuracy and dispersion on the quantity of measurements in experimental sample. According to the results of numerical experiments, the dispersion values of the estimated sought-for signal and noise parameters calculated by means of the proposed technique change in inverse proportion to the quantity of measurements in a sample. There has been implemented a comparison of the accuracy of the soughtfor Rician parameters’ estimation by means of the proposed technique and by earlier developed version of the method of moments. The problem having been considered in the paper is meaningful for the purposes of Rician data processing, in particular, at the systems of magnetic-resonance visualization, in devices of ultrasonic visualization, at optical signals’ analysis in range-measuring systems, at radar signals’ analysis, as well as at solving many other scientific and applied tasks that are adequately described by the Rice statistical model.
-
Mathematical modeling of carcinoma growth with a dynamic change in the phenotype of cells
Computer Research and Modeling, 2018, v. 10, no. 6, pp. 879-902Views (last year): 46.In this paper, we proposed a two-dimensional chemo-mechanical model of the growth of invasive carcinoma in epithelial tissue. Each cell is modeled by an elastic polygon, changing its shape and size under the influence of pressure forces acting from the tissue. The average size and shape of the cells have been calibrated on the basis of experimental data. The model allows to describe the dynamic deformations in epithelial tissue as a collective evolution of cells interacting through the exchange of mechanical and chemical signals. The general direction of tumor growth is controlled by a pre-established linear gradient of nutrient concentration. Growth and deformation of the tissue occurs due to the mechanisms of cell division and intercalation. We assume that carcinoma has a heterogeneous structure made up of cells of different phenotypes that perform various functions in the tumor. The main parameter that determines the phenotype of a cell is the degree of its adhesion to the adjacent cells. Three main phenotypes of cancer cells are distinguished: the epithelial (E) phenotype is represented by internal tumor cells, the mesenchymal (M) phenotype is represented by single cells and the intermediate phenotype is represented by the frontal tumor cells. We assume also that the phenotype of each cell under certain conditions can change dynamically due to epithelial-mesenchymal (EM) and inverse (ME) transitions. As for normal cells, we define the main E-phenotype, which is represented by ordinary cells with strong adhesion to each other. In addition, the normal cells that are adjacent to the tumor undergo a forced EM-transition and form an M-phenotype of healthy cells. Numerical simulations have shown that, depending on the values of the control parameters as well as a combination of possible phenotypes of healthy and cancer cells, the evolution of the tumor can result in a variety of cancer structures reflecting the self-organization of tumor cells of different phenotypes. We compare the structures obtained numerically with the morphological structures revealed in clinical studies of breast carcinoma: trabecular, solid, tubular, alveolar and discrete tumor structures with ameboid migration. The possible scenario of morphogenesis for each structure is discussed. We describe also the metastatic process during which a single cancer cell of ameboid phenotype moves due to intercalation in healthy epithelial tissue, then divides and undergoes a ME transition with the appearance of a secondary tumor.
-
Current issues in computational modeling of thrombosis, fibrinolysis, and thrombolysis
Computer Research and Modeling, 2024, v. 16, no. 4, pp. 975-995Hemostasis system is one of the key body’s defense systems, which is presented in all the liquid tissues and especially important in blood. Hemostatic response is triggered as a result of the vessel injury. The interaction between specialized cells and humoral systems leads to the formation of the initial hemostatic clot, which stops bleeding. After that the slow process of clot dissolution occurs. The formation of hemostatic plug is a unique physiological process, because during several minutes the hemostatic system generates complex structures on a scale ranging from microns for microvessel injury or damaged endothelial cell-cell contacts, to centimeters for damaged systemic arteries. Hemostatic response depends on the numerous coordinated processes, which include platelet adhesion and aggregation, granule secretion, platelet shape change, modification of the chemical composition of the lipid bilayer, clot contraction, and formation of the fibrin mesh due to activation of blood coagulation cascade. Computer modeling is a powerful tool, which is used to study this complex system at different levels of organization. This includes study of intracellular signaling in platelets, modelling humoral systems of blood coagulation and fibrinolysis, and development of the multiscale models of thrombus growth. There are two key issues of the computer modeling in biology: absence of the adequate physico-mathematical description of the existing experimental data due to the complexity of the biological processes, and high computational complexity of the models, which doesn’t allow to use them to test physiologically relevant scenarios. Here we discuss some key unresolved problems in the field, as well as the current progress in experimental research of hemostasis and thrombosis. New findings lead to reevaluation of the existing concepts and development of the novel computer models. We focus on the arterial thrombosis, venous thrombosis, thrombosis in microcirculation and the problems of fibrinolysis and thrombolysis. We also briefly discuss basic types of the existing mathematical models, their computational complexity, and principal issues in simulation of thrombus growth in arteries.
-
Optimization of the brain command dictionary based on the statistical proximity criterion in silent speech recognition task
Computer Research and Modeling, 2023, v. 15, no. 3, pp. 675-690In our research, we focus on the problem of classification for silent speech recognition to develop a brain– computer interface (BCI) based on electroencephalographic (EEG) data, which will be capable of assisting people with mental and physical disabilities and expanding human capabilities in everyday life. Our previous research has shown that the silent pronouncing of some words results in almost identical distributions of electroencephalographic signal data. Such a phenomenon has a suppressive impact on the quality of neural network model behavior. This paper proposes a data processing technique that distinguishes between statistically remote and inseparable classes in the dataset. Applying the proposed approach helps us reach the goal of maximizing the semantic load of the dictionary used in BCI.
Furthermore, we propose the existence of a statistical predictive criterion for the accuracy of binary classification of the words in a dictionary. Such a criterion aims to estimate the lower and the upper bounds of classifiers’ behavior only by measuring quantitative statistical properties of the data (in particular, using the Kolmogorov – Smirnov method). We show that higher levels of classification accuracy can be achieved by means of applying the proposed predictive criterion, making it possible to form an optimized dictionary in terms of semantic load for the EEG-based BCIs. Furthermore, using such a dictionary as a training dataset for classification problems grants the statistical remoteness of the classes by taking into account the semantic and phonetic properties of the corresponding words and improves the classification behavior of silent speech recognition models.
-
Reinforcement learning-based adaptive traffic signal control invariant to traffic signal configuration
Computer Research and Modeling, 2024, v. 16, no. 5, pp. 1253-1269In this paper, we propose an adaptive traffic signal control method invariant to the configuration of the traffic signal. The proposed method uses one neural network model to control traffic signals of various configurations, differing both in the number of controlled lanes and in the used traffic light control cycle (set of phases). To describe the state space, both dynamic information about the current state of the traffic flow and static data about the configuration of a controlled intersection are used. To increase the speed of model training and reduce the required amount of data required for model convergence, it is proposed to use an “expert” who provides additional data for model training. As an expert, we propose to use an adaptive control method based on maximizing the weighted flow of vehicles through an intersection. Experimental studies of the effectiveness of the developed method were carried out in a microscopic simulation software package. The obtained results confirmed the effectiveness of the proposed method in different simulation scenarios. The possibility of using the developed method in a simulation scenario that is not used in the training process was shown. We provide a comparison of the proposed method with other baseline solutions, including the method used as an “expert”. In most scenarios, the developed method showed the best results by average travel time and average waiting time criteria. The advantage over the method used as an expert, depending on the scenario under study, ranged from 2% to 12% according to the criterion of average vehicle waiting time and from 1% to 7% according to the criterion of average travel time.
-
Influence of random malignant cell motility on growing tumor front stability
Computer Research and Modeling, 2009, v. 1, no. 2, pp. 225-232Views (last year): 5. Citations: 7 (RSCI).Chemotaxis plays an important role in morphogenesis and processes of structure formation in nature. Both unicellular organisms and single cells in tissue demonstrate this property. In vitro experiments show that many types of transformed cell, especially metastatic competent, are capable for directed motion in response usually to chemical signal. There is a number of theoretical papers on mathematical modeling of tumour growth and invasion using Keller-Segel model for the chemotactic motility of cancer cells. One of the crucial questions for using the chemotactic term in modelling of tumour growth is a lack of reliable quantitative estimation of its parameters. The 2-D mathematical model of tumour growth and invasion, which takes into account only random cell motility and convective fluxes in compact tissue, has showed that due to competitive mechanism tumour can grow toward sources of nutrients in absence of chemotactic cell motility.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"




