All issues
- 2025 Vol. 17
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Investigation of the relationships of the size and production characteristics of phyto- and zooplankton in the Vistula and Curonian lagoons of the Baltic Sea. Part 1. The statistical analysis of long-term observation data and development of the structure for the mathematical model of the plankton food chain
Computer Research and Modeling, 2017, v. 9, no. 2, pp. 211-246In the paper the statistical relationships between the size and production characteristics of phytoplankton and zooplankton of the Vistula and Curonian lagoons, the Baltic Sea, were investigated. Research phytoplankton and zooplankton within the Russian part of the area of the Vistula and the Curonian lagoon was carried out on the monthly basis (from April to November) within the framework of long-term monitoring program on evaluating of ecological status of the lagoons. The size structure of plankton is the basis for understanding of the development of production processes, mechanisms of formation of the plankton species diversity and functioning of the lagoon ecosystems. As results of the work it was found that the maximum rate of photosynthesis and the integral value of the primary production with a change in cell volume of phytoplankton are changed according to a power law. The result shows that the smaller the size of algal cells in phytoplankton communities the more actively occur metabolism and the more effective they assimilate the solar energy. It is shown that the formation of plankton species diversity in ecosystems of lagoons is closely linked with the size structure of plankton communities and with features of development of the production processes. It is proposed the structure of a spatially homogenous mathematical model of the plankton food chain for the lagoon ecosystems taking into account the size spectrum and the characteristics of phytoplankton and zooplankton. The model parameters are the sizedependent indicators allometrically linked with average volumes of cells and organisms in different ranges of their sizes. In the model the algorithm for changes over time the coefficients of food preferences in the diet of zooplankton was proposed. Developed the size-dependent mathematical model of aquatic ecosystems allows to consider the impact of turbulent exchange on the size structure and temporal dynamics of the plankton food chain of the Vistula and Curonian lagoons. The model can be used to study the different regimes of dynamic behavior of plankton systems depending on the changes in the values of its parameters and external influences, as well as to quantify the redistribution of matter flows in ecosystems of the lagoons.
Keywords: ecosystem, nutrients, phytoplankton, zooplankton, plankton detritus, size structure, the maximum rate of photosynthesis, integrated primary production, zooplankton production, allometric scaling, Shannon index of species diversity, mathematical modeling, ecological simulation model, turbulent exchange.Views (last year): 9. -
A modified model of the effect of stress concentration near a broken fiber on the tensile strength of high-strength composites (MLLS-6)
Computer Research and Modeling, 2020, v. 12, no. 3, pp. 559-573The article proposes a model for assessing the potential strength of a composite material based on modern fibers with brittle fracture.
Materials consisting of parallel cylindrical fibers that are quasi-statically stretched in one direction are simulated. It is assumed that the sample is not less than 100 pieces, which corresponds to almost significant cases. It is known that the fibers have a distribution of ultimate deformation in the sample and are not destroyed at the same moment. Usually the distribution of their properties is described by the Weibull–Gnedenko statistical distribution. To simulate the strength of the composite, a model of fiber breaks accumulation is used. It is assumed that the fibers united by the polymer matrix are crushed to twice the inefficient length — the distance at which the stresses increase from the end of the broken fiber to the middle one. However, this model greatly overestimates the strength of composites with brittle fibers. For example, carbon and glass fibers are destroyed in this way.
In some cases, earlier attempts were made to take into account the stress concentration near the broken fiber (Hedgepest model, Ermolenko model, shear analysis), but such models either required a lot of initial data or did not coincide with the experiment. In addition, such models idealize the packing of fibers in the composite to the regular hexagonal packing.
The model combines the shear analysis approach to stress distribution near the destroyed fiber and the statistical approach of fiber strength based on the Weibull–Gnedenko distribution, while introducing a number of assumptions that simplify the calculation without loss of accuracy.
It is assumed that the stress concentration on the adjacent fiber increases the probability of its destruction in accordance with the Weibull distribution, and the number of such fibers with an increased probability of destruction is directly related to the number already destroyed before. All initial data can be obtained from simple experiments. It is shown that accounting for redistribution only for the nearest fibers gives an accurate forecast.
This allowed a complete calculation of the strength of the composite. The experimental data obtained by us on carbon fibers, glass fibers and model composites based on them (CFRP, GFRP), confirm some of the conclusions of the model.
-
Extraction of characters and events from narratives
Computer Research and Modeling, 2024, v. 16, no. 7, pp. 1593-1600Events and character extraction from narratives is a fundamental task in text analysis. The application of event extraction techniques ranges from the summarization of different documents to the analysis of medical notes. We identify events based on a framework named “four W” (Who, What, When, Where) to capture all the essential components like the actors, actions, time, and places. In this paper, we explore two prominent techniques for event extraction: statistical parsing of syntactic trees and semantic role labeling. While these techniques were investigated by different researchers in isolation, we directly compare the performance of the two approaches on our custom dataset, which we have annotated.
Our analysis shows that statistical parsing of syntactic trees outperforms semantic role labeling in event and character extraction, especially in identifying specific details. Nevertheless, semantic role labeling demonstrate good performance in correct actor identification. We evaluate the effectiveness of both approaches by comparing different metrics like precision, recall, and F1-scores, thus, demonstrating their respective advantages and limitations.
Moreover, as a part of our work, we propose different future applications of event extraction techniques that we plan to investigate. The areas where we want to apply these techniques include code analysis and source code authorship attribution. We consider using event extraction to retrieve key code elements as variable assignments and function calls, which can further help us to analyze the behavior of programs and identify the project’s contributors. Our work provides novel understandings of the performance and efficiency of statistical parsing and semantic role labeling techniques, offering researchers new directions for the application of these techniques.
-
Searching for connections between biological and physico-chemical characteristics of Rybinsk reservoir ecosystem. Part 1. Criteria of connection nonrandomness
Computer Research and Modeling, 2013, v. 5, no. 1, pp. 83-105Views (last year): 3. Citations: 6 (RSCI).Based on contents of phytoplankton pigments, fluorescence samples and some physico-chemical characteristics of the Rybinsk reservoir waters, searching for connections between biological and physicalchemical characteristics is working out. The standard methods of statistical analysis (correlation, regression), methods of description of connection between qualitative classes of characteristics, based on deviation of the studied characteristics distribution from independent distribution, are studied. A method of searching for boundaries of quality classes by criterion of maximum connection coefficient is offered.
-
Topological microstructure analysis of the TIP4P-EW water model
Computer Research and Modeling, 2014, v. 6, no. 3, pp. 415-426Views (last year): 1. Citations: 1 (RSCI).Molecular dynamics (MD) simulations of rigid water model TIP4P-EW at ambient conditions were carried out. Delaunay’s simplexes were considered as structural elements of liquid water. Topological criterion which allows to identify the water microstructure in snapshot of MD cell was used to allocate its dense part. Geometrical analysis of water Delaunay’s simplexes indicates their strong flatness in comparison with a regular tetrahedron that is fundamentally different from the results for dense part of simple liquids. The statistics of TIP4P-EW water clusters was investigated depending on their cardinality and connectivity. It is similar to the statistics for simple liquids and the structure of this dense part is also a fractal surface consisting of the free edges of the Delaunay’s simplexes.
-
The use of GIS INTEGRO in searching tasks for oil and gas deposits
Computer Research and Modeling, 2015, v. 7, no. 3, pp. 439-444Views (last year): 4.GIS INTEGRO is the geo-information software system forming the basis for the integrated interpretation of geophysical data in researching a deep structure of Earth. GIS INTEGRO combines a variety of computational and analytical applications for the solution of geological and geophysical problems. It includes various interfaces that allow you to change the form of representation of data (raster, vector, regular and irregular network of observations), the conversion unit of map projections, application blocks, including block integrated data analysis and decision prognostic and diagnostic tasks.
The methodological approach is based on integration and integrated analysis of geophysical data on regional profiles, geophysical potential fields and additional geological information on the study area. Analytical support includes packages transformations, filtering, statistical processing, calculation, finding of lineaments, solving direct and inverse tasks, integration of geographic information.
Technology and software and analytical support was tested in solving problems tectonic zoning in scale 1:200000, 1:1000000 in Yakutia, Kazakhstan, Rostov region, studying the deep structure of regional profiles 1:S, 1-SC, 2-SAT, 3-SAT and 2-DV, oil and gas forecast in the regions of Eastern Siberia, Brazil.
The article describes two possible approaches of parallel calculations for data processing 2D or 3D nets in the field of geophysical research. As an example presented realization in the environment of GRID of the application software ZondGeoStat (statistical sensing), which create 3D net model on the basis of data 2d net. The experience has demonstrated the high efficiency of the use of environment of GRID during realization of calculations in field of geophysical researches.
-
Game-theoretic model of coordinations of interests at innovative development of corporations
Computer Research and Modeling, 2016, v. 8, no. 4, pp. 673-684Views (last year): 9. Citations: 6 (RSCI).Dynamic game theoretic models of the corporative innovative development are investigated. The proposed models are based on concordance of private and public interests of agents. It is supposed that the structure of interests of each agent includes both private (personal interests) and public (interests of the whole company connected with its innovative development first) components. The agents allocate their personal resources between these two directions. The system dynamics is described by a difference (not differential) equation. The proposed model of innovative development is studied by simulation and the method of enumeration of the domains of feasible controls with a constant step. The main contribution of the paper consists in comparative analysis of efficiency of the methods of hierarchical control (compulsion or impulsion) for information structures of Stackelberg or Germeier (four structures) by means of the indices of system compatibility. The proposed model is a universal one and can be used for a scientifically grounded support of the programs of innovative development of any economic firm. The features of a specific company are considered in the process of model identification (a determination of the specific classes of model functions and numerical values of its parameters) which forms a separate complex problem and requires an analysis of the statistical data and expert estimations. The following assumptions about information rules of the hierarchical game are accepted: all players use open-loop strategies; the leader chooses and reports to the followers some values of administrative (compulsion) or economic (impulsion) control variables which can be only functions of time (Stackelberg games) or depend also on the followers’ controls (Germeier games); given the leader’s strategies all followers simultaneously and independently choose their strategies that gives a Nash equilibrium in the followers’ game. For a finite number of iterations the proposed algorithm of simulation modeling allows to build an approximate solution of the model or to conclude that it doesn’t exist. A reliability and efficiency of the proposed algorithm follow from the properties of the scenario method and the method of a direct ordered enumeration with a constant step. Some comprehensive conclusions about the comparative efficiency of methods of hierarchical control of innovations are received.
-
A hybrid multi-objective carpool route optimization technique using genetic algorithm and A* algorithm
Computer Research and Modeling, 2021, v. 13, no. 1, pp. 67-85Carpooling has gained considerable importance as an effective solution for reducing pollution, mitigation of traffic and congestion on the roads, reduced demand for parking facilities, lesser energy and fuel consumption and most importantly, reduction in carbon emission, thus improving the quality of life in cities. This work presents a hybrid GA-A* algorithm to obtain optimal routes for the carpooling problem in the domain of multiobjective optimization having multiple conflicting objectives. Though the Genetic Algorithm provides optimal solutions, the A* algorithm because of its efficiency in providing the shortest route between any two points based on heuristics, enhances the optimal routes obtained using the Genetic algorithm. The refined routes obtained using the GA-A* algorithm, are further subjected to dominance test to obtain non-dominating solutions based on Pareto-Optimality. The routes obtained maximize the profit of the service provider by minimizing the travel and detour distance as well as pick-up/drop costs while maximizing the utilization of the car. The proposed algorithm has been implemented over the Salt Lake area of Kolkata. Route distance and detour distance for the optimal routes obtained using the proposed algorithm are consistently lesser for the same number of passengers when compared to the corresponding results obtained from an existing algorithm. Various statistical analysis like boxplots have also confirmed that the proposed algorithm regularly performed better than the existing algorithm using only Genetic Algorithm.
-
Methodology and program for the storage and statistical analysis of the results of computer experiment
Computer Research and Modeling, 2013, v. 5, no. 4, pp. 589-595Views (last year): 1. Citations: 5 (RSCI).The problem of accumulation and the statistical analysis of computer experiment results are solved. The main experiment program is considered as the data source. The results of main experiment are collected on specially prepared sheet Excel with pre-organized structure for the accumulation, statistical processing and visualization of the data. The created method and the program are used at efficiency research of the scientific researches which are carried out by authors.
-
Modeling of axisymmetric deformation processes with taking into account the metal microstructure
Computer Research and Modeling, 2015, v. 7, no. 4, pp. 897-908Views (last year): 9. Citations: 1 (RSCI).The article describes the state of the art computer simulation in the field of metal forming processes, the main problem points of traditional methods were identified. The method, that allows to predict the deformation distribution in the volume of deformable metal with taking into account of microstructure behavioral characteristics in deformation load conditions, was described. The method for optimizing computational resources of multiscale models by using statistical similar representative volume elements (SSRVE) was presented. The modeling methods were tested on the process of single pass drawing of round rod from steel grade 20. In a comparative analysis of macro and micro levels models differences in quantitative terms of the stress-strain state and their local distribution have been identified. Microlevel model also allowed to detect the compressive stresses and strains, which were absent at the macro level model. Applying the SSRVE concept repeatedly lowered the calculation time of the model while maintaining the overall accuracy.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"




